Skip to yearly menu bar Skip to main content


( events)   Timezone:  
Workshop
Sat Dec 09 08:00 AM -- 06:30 PM (PST) @ 104 C
The future of gradient-based machine learning software & techniques
Alex Wiltschko · Bart van MerriĆ«nboer · Pascal Lamblin





Workshop Home Page

Many algorithms in machine learning, computer vision, physical simulation, and other fields require the calculation of gradients and other derivatives. Manual derivation of gradients can be time consuming and error-prone. Automatic differentiation comprises a set of techniques to calculate the derivative of a numerical computation expressed as a computer program. These techniques are commonly used in atmospheric sciences and computational fluid dynamics, and have more recently also been adopted by machine learning researchers.

Practitioners across many fields have built a wide set of automatic differentiation tools, using different programming languages, computational primitives and intermediate compiler representations. Each of these choices comes with positive and negative trade-offs, in terms of their usability, flexibility and performance in specific domains.

This workshop will bring together researchers in the fields of automatic differentiation and machine learning to discuss ways in which advanced automatic differentiation frameworks and techniques can enable more advanced machine learning models, run large-scale machine learning on accelerators with better performance, and increase the usability of machine learning frameworks for practitioners. Topics for discussion will include:

* What abstractions (languages, kernels, interfaces, instruction sets) do we need to develop advanced automatic differentiation frameworks for the machine learning ecosystem?
* What different use cases exist in machine learning, from large-scale performance-critical models to small prototypes, and how should our toolsets reflect these needs?
* What advanced techniques from the automatic differentiation literature, such as checkpointing, differentiating through iterative processes or chaotic systems, cross-country elimination, etc., could be adopted by the ML community to enable research on new models?
* How can we foster greater collaboration between the fields of machine learning and automatic differentiation?

Introduction and opening remarks (Talk)
Beyond backprop: automatic differentiation in machine learning (Talk)
Automatic differentiation in PyTorch (Talk)
Morning Coffee Break (Break)
Optimal Smoothing for Pathwise Adjoints (Talk)
Poster session (Poster Session)
Algorithmic differentiation techniques in the deep learning context (Talk)
Some highlights on Source-to-Source Adjoint AD (Talk)
Afternoon Coffee Break (Break)
Divide-and-Conquer Checkpointing for Arbitrary Programs with No User Annotation (Talk)
Automatic Differentiation of Parallelised Convolutional Neural Networks - Lessons from Adjoint PDE Solvers (Talk)
Panel discussion (Discussion Panel)