tensorflow
- North America > Canada > Ontario > Toronto (0.14)
- Europe > Austria > Tyrol > Innsbruck (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- Europe > Germany > Baden-Württemberg > Karlsruhe Region > Karlsruhe (0.04)
- Europe > France (0.04)
Supplementary Material for Paper " Terra: Imperative-Symbolic Co-Execution of Imperative Deep Learning Programs " A Criteria for Node Equality When Merging Traces
TraceGraph, it compares the type, attributes, and the executed location of each operation. For example, the MatMul operation of TensorFlow has ' MatMul ' as GraphGenerator fails to match because of the different attributes. The pushed call id is popped when the function is returned. As same as the call id stack, Terra manages the loop id stack for the entire program execution. Current implementation of Terra does not consider multi-threading yet.
- Oceania > Australia > New South Wales > Sydney (0.04)
- North America > Mexico > Gulf of Mexico (0.04)
- Asia > South Korea > Seoul > Seoul (0.05)
- North America > United States > California > Alameda County > Berkeley (0.04)
DISCO: A Browser-Based Privacy-Preserving Framework for Distributed Collaborative Learning
Vignoud, Julien T. T., Rousset, Valérian, Guedj, Hugo El, Aleman, Ignacio, Bennaceur, Walid, Derinbay, Batuhan Faik, Ďurech, Eduard, Gengler, Damien, Giordano, Lucas, Grimberg, Felix, Lippoldt, Franziska, Kopidaki, Christina, Liu, Jiafan, Lopata, Lauris, Maire, Nathan, Mansat, Paul, Milenkoski, Martin, Omont, Emmanuel, Özgün, Güneş, Petrović, Mina, Posa, Francesco, Ridel, Morgan, Savini, Giorgio, Torne, Marcel, Trognon, Lucas, Unell, Alyssa, Zavertiaieva, Olena, Karimireddy, Sai Praneeth, Rabbani, Tahseen, Hartley, Mary-Anne, Jaggi, Martin
Data is often impractical to share for a range of well considered reasons, such as concerns over privacy, intellectual property, and legal constraints. This not only fragments the statistical power of predictive models, but creates an accessibility bias, where accuracy becomes inequitably distributed to those who have the resources to overcome these concerns. We present DISCO: an open-source DIStributed COllaborative learning platform accessible to non-technical users, offering a means to collaboratively build machine learning models without sharing any original data or requiring any programming knowledge. DISCO's web application trains models locally directly in the browser, making our tool cross-platform out-of-the-box, including smartphones. The modular design of \disco offers choices between federated and decentralized paradigms, various levels of privacy guarantees and several approaches to weight aggregation strategies that allow for model personalization and bias resilience in the collaborative training. Code repository is available at https://github.com/epfml/disco and a showcase web interface at https://discolab.ai
- Europe > Switzerland > Vaud > Lausanne (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine (1.00)
- Law (0.87)
Scalable Planning with Tensorflow for Hybrid Nonlinear Domains
Given recent deep learning results that demonstrate the ability to effectively optimize high-dimensional non-convex functions with gradient descent optimization on GPUs, we ask in this paper whether symbolic gradient optimization tools such as Tensorflow can be effective for planning in hybrid (mixed discrete and continuous) nonlinear domains with high dimensional state and action spaces? To this end, we demonstrate that hybrid planning with Tensorflow and RMSProp gradient descent is competitive with mixed integer linear program (MILP) based optimization on piecewise linear planning domains (where we can compute optimal solutions) and substantially outperforms state-of-the-art interior point methods for nonlinear planning domains. Furthermore, we remark that Tensorflow is highly scalable, converging to a strong plan on a large-scale concurrent domain with a total of 576,000 continuous action parameters distributed over a horizon of 96 time steps and 100 parallel instances in only 4 minutes. We provide a number of insights that clarify such strong performance including observations that despite long horizons, RMSProp avoids both the vanishing and exploding gradient problems. Together these results suggest a new frontier for highly scalable planning in nonlinear hybrid domains by leveraging GPUs and the power of recent advances in gradient descent with highly optimized toolkits like Tensorflow.
- North America > Canada > Ontario > Toronto (0.14)
- South America > Brazil > São Paulo (0.04)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.04)
- (6 more...)
- Energy (0.46)
- Information Technology (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Planning & Scheduling (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.99)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Gradient Descent (0.32)
- North America > Canada > Quebec > Montreal (0.04)
- Europe > Germany > Baden-Württemberg > Karlsruhe Region > Karlsruhe (0.04)
- Europe > France (0.04)
- Europe > Switzerland > Zürich > Zürich (0.14)
- North America > United States > Virginia (0.04)
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- (5 more...)
Supplementary Materials Outline of Supplementary Materials
Proofs of results stated in the main text are provided in Appendix A. Additional experimental results, including coverage plots, are provided in Appendix B. Therefore, the proof is completed by continuous mapping. Applying Taylor's theorem using the Lagrange form of the remainder, we have that, for some random From the proof of Eq. (7), we know that, with probability tending to 1, Figure 1 shows the fitted mean and covariance on a single draw of the quadratic dataset. In this section we provide details for the experimental setup used in the paper. The visualized covariance matrices were projecting to ensure positive semi-definiteness. In this section we present the multivariate algorithm for finite-difference IDM (FDIDM).