Conversations with Hugging Face CTO
Contents
The following are the learnings from Hugging Face Interview in Oct 2019
- GPT2 from Open AI is impressive - Packaged in to Demo Application
- Conversational AI + Open Source package(Transformers)
- Half a million monthly active users
- Hard to good Deep Conversational AI
- Self starter - Was working in 2008 on ML and then moved on to do some software jobs
- I was curious to see what the number of downloads for various pre-trained models were. So, wrote a small Python program to get the downloads
|
|
Here are the top 25 models as of
|
|
- Amazed to see that there were 44 million downloads of distilbert and bert models. What am I doing ? Why am I not trying out something that seems to have revolutionized NLP ? 1.5 Million Downloads in one day
- Look at all the kind of models that are being used
- Distilbert
- BERT
- GPT2
- ROBERTA
- XLNET
- Different language specific models based on BERT
- AI and Conversational AI are going to be transformative
- 2005 - Did his engineering in Paris
- Most of the research professors were interested in theoretical stuff
- 2016 was the year - Use of Deep learning for applications in conversational AI
- Founders for a couple of companies
- 1600 repositories in github
- Raised 5 Million for NLP work - They are not making any money yet
- 10-15 member startup are doing amazing work
- Large Transformers are important in NLP
- Took a week off and write pytorch version of models
- 20,000 models every day a year back
- Super easy to download models and work on the various models
- It makes using pre-trained models democratic
- Conversational AI - Entered in to a competition - NuerIPS - Conversational AI
- Huggingface beat everyone
- Generative capacity
- Talk to Transformer project - Huggingface team loved it
- GPT and BERT are the most popular downloads
- XLNet(Google) and Roberta(Facebook) are also popular on Huggingface
- DistilBert cuts down 50% of weights from BERT and gives a similar performance
- Swift repository - Can ship the models to edge devices
- Apple uses Hugging face transformers on iPhone
- Ported GPT2 to iPhone but needs a large memory on iPhone
- Personal Dataset to train the Conversational AI
- Conditional text generation
- Code generation
- DistilBert - 2 months to build and develop
- Recommended courses
- Stanford Class - Put a small study group and work through the videos - Newer version of course is fantastic
- Fast AI for NLP
- Learn by doing - Make an open source contribution