Skip to content

The Race to Open-Source ChatGPT

Published: at 06:00 PM
Lewis Tunstall

Just one year ago, chatbots were out of fashion and most people hadn’t heard about large language models and techniques to align them with human preferences. Then, OpenAI broke the internet with ChatGPT and Meta followed suit by releasing the Llama series of language models which enabled the open-source community to build their very own capable chatbots. This has led to a rich ecosystem of open access datasets and models, with many of them hosted on the Hugging Face Hub.

In this talk, Lewis Tunstall will give an overview of the community’s efforts to create an open-source ChatGPT, and discuss some of the challenges involved in training and evaluating these models effectively.


Lewis is a Machine Learning Engineer in the research team at Hugging Face, and is the co-author of the bestseller “NLP with Transformers” book. He has previously built machine learning-powered applications for start-ups and enterprises in the domains of natural language processing, topological data analysis, and time series. He holds a PhD in Theoretical Physics, was a 2010 Fulbright Scholar and has held research positions in Australia, the USA, and Switzerland. His current work focuses on building tools and recipes to align language models with human and AI preferences through techniques like reinforcement learning.