Relay: A New IR for Machine Learning Frameworks
Machine learning powers diverse services in industry including search,
translation, recommendation systems, security, and more. The scale and
importance of these models require that they be efficient, expressive, and
portable across an array of heterogeneous hardware devices. These constraints are often at odds; in order to simultaneously accommodate them we propose a new high-level intermediate representation called Relay. Relay is a purely
functional, statically typed IR designed to balance efficient compilation,
expressiveness, and portability. We present a prototype of Relay and
highlight its important design decisions. Our prototype is part of the open
source NNVM compiler framework, which powers Amazon’s deep learning framework MxNet.
Mon 18 JunDisplayed time zone: Eastern Time (US & Canada) change
14:00 - 15:30 | |||
14:00 30mTalk | Relay: A New IR for Machine Learning Frameworks MAPL Jared Roesch University of Washington, USA, Steven Lyubomirsky University of Washington, USA, Logan Weber University of Washington, Josh Pollock University of Washington, Marisa Kirisame , Tianqi Chen , Zachary Tatlock University of Washington, Seattle | ||
14:30 30mTalk | Diesel - DSL for Linear Algebra and Neural Net Computations on GPUs MAPL Venmugil Elango NVIDIA, Norm Rubin NVIDIA, Mahesh Ravishankar , Hari Sandanagobalane NVIDIA, Vinod Grover | ||
15:00 30mTalk | Gen: probabilistic programming with fast custom inference via code generation MAPL File Attached |