From a04ea628c25ce48d38620b87ca76de88c1953478 Mon Sep 17 00:00:00 2001 From: Dan Fu Date: Sat, 2 Dec 2023 16:57:04 -0800 Subject: [PATCH] update links --- _data/talks_229s.yml | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/_data/talks_229s.yml b/_data/talks_229s.yml index d7d2c9e..e35c77c 100644 --- a/_data/talks_229s.yml +++ b/_data/talks_229s.yml @@ -36,5 +36,10 @@ title: Bringing Foundational Models to Consumer Devices via ML Compilation abstract: "Deploying deep learning models on various devices has become an important topic. Machine learning compilation is an emerging field that leverages compiler and automatic search techniques to accelerate AI models. ML compilation brings a unique set of challenges: emerging machine learning models; increasing hardware specialization brings a diverse set of acceleration primitives; growing tension between flexibility and performance. In this talk. I then discuss our experience in bringing foundational models to a variety of devices and hardware environments through machine learning compilation." bio: "Tianqi Chen is currently an Assistant Professor at the Machine Learning Department and Computer Science Department of Carnegie Mellon University. He is also the Chief Technologist of OctoML. He received his PhD. from the Paul G. Allen School of Computer Science & Engineering at the University of Washington. He has created many major learning systems that are widely adopted: XGBoost, TVM, and MLC-LLM." + livestream: https://www.youtube.com/watch?v=InoNMvjs_vo - speaker: Dan Fu (Stanford, Together) - date: M 12/04/23 \ No newline at end of file + date: M 12/04/23 + title: "Monarch Mixer: Making Foundation Models More Efficient" + abstract: "Machine learning models are increasingly being scaled in both sequence length and model dimension to reach longer contexts and better performance. However, existing architectures like Transformers scale quadratically along both these axes. In this talk I'll discuss Monarch Mixer (M2), a new architecture that uses the same sub-quadratic primitive along both sequence length and model dimension. M2 mixes information along the sequence and model dimensions using Monarch matrices, a simple class of expressive structured matrices that captures many linear transforms, achieves high hardware efficiency on GPUs, and scales sub-quadratically." + bio: "Dan Fu is a PhD student in the Computer Science Department at Stanford University, where he is co-advised by Christopher RĂ© and Kayvon Fatahalian. His research is at the intersection of systems and machine learning and focuses on developing algorithms and architectures to make machine learning more efficient." + livestream: https://www.youtube.com/watch?v=IS59IwGLvVs \ No newline at end of file