Technical Talks

View All

HuggingFace + Ray AIR Integration: A Python Developer’s Guide to Scaling Transformers

Jules Damji Jules Damji | Lead Developer Advocate | Anyscale
Antoni Baum Antoni Baum | Software Engineer | Anyscale Inc

Hugging Face Transformers is a popular open-source project with cutting-edge Machine Learning (ML). Still, meeting the computational requirements for advanced models it provides often requires scaling beyond a single machine. This session explores the integration between Hugging Face and Ray AI Runtime (AIR), allowing users to scale their model training and data loading seamlessly. We will dive deep into the implementation and API and explore how we can use Ray AIR to create an end-to-end Hugging Face workflow, from data ingest through fine-tuning and HPO to inference and serving.
 
 The computational and memory requirements for fine-tuning and training these models can be significant. To deal with this issue, the Ray team has developed a Hugging Face integration for Ray AI Runtime (AIR), allowing Transformers model training to be easily parallelized across multiple CPUs or GPUs in a Ray Cluster, saving time and money, all the while allowing to take advantage of the rich Ray ML ecosystem thanks to standard and common API.
 
In this session, we explore the integration between Hugging Face and Ray AIR, allowing users to scale their model training and data loading seamlessly. We will dive deep into the implementation and API and explore how we can use Ray AIR to create an end-to-end Hugging Face workflow, from data ingest through fine-tuning and HPO to inference and serving. 

 Key Takeaways: 

  •   Python developers and machine learning engineers can use * 
  •  Transformers and scale their language models.
  •  Get exposed to Ray AIR’s Python APIs for end-to-end Hugging Face and ML workflow.
  •  Understand how Ray AIR, built atop Ray, can scale your Python-based ML workloads.

Jules Damji
Jules Damji
Lead Developer Advocate | Anyscale

Jules S. Damji is a lead developer advocate at Anyscale Inc, an MLflow contributor, and co-author of Learning Spark, 2nd Edition. He is a hands-on developer with over 25 years of experience and has worked at leading companies, such as Sun Microsystems, Netscape, @Home, Opsware/LoudCloud, VeriSign, ProQuest, Hortonworks, and Databricks, building large-scale distributed systems. He holds a B.Sc and M.Sc in computer science (from Oregon State University and Cal State, Chico respectively), and an MA in political advocacy and communication (from Johns Hopkins University).

Antoni Baum
Antoni Baum
Software Engineer | Anyscale Inc

Antoni Baum is a software engineer at Anyscale, working on Ray Tune, XGBoost-Ray, Ray AIR, and other ML libraries. In his spare time, he contributes to various open source projects, trying to make machine learning more accessible and approachable.

FEATURED MEETINGS