Benedict Neo - Roadmap To Learn AI in 2024
Benedict Neo - Roadmap To Learn AI in 2024
Listen Share
S o, you want to learn AI? But you don’t know how or where to get started?
I wrote the Top 20 free Data Science, ML, and AI MOOCs on the Internet back in
2020. But I’ve realized that doing many courses isn’t the way.
To escape tutorial hell and really learn, you have to get hands-on, write algorithms
from scratch, implement papers, and do fun side projects using AI to solve
problems.
This article attempts to craft a free curriculum that follows that philosophy. I’m
working on some of these courses, so reach out on Twitter or Linkedin if you want
to learn together!
But first, a few notes on the curriculum and some advice on learning.
Top-down approach
This curriculum follows a top-down approach — code first, theory later.
I like to learn out of necessity. So, if I have to figure out something, a problem to
solve, or a prototype to make, I will reach far and wide for the information I need,
study, make sense of it, and then act on it.
Before I dump a bunch of links on you, I wish somebody had told me two important
things before I started learning anything.
Learn in Public
There’s a lot to learn, and you will never be done learning, especially with AI, when
new revolutionary papers and ideas are released weekly.
The biggest mistake you can make is to learn in private. You don’t create any
opportunities for yourself if you do that. You don’t have anything to show for it
besides being able to say you completed something. What matters more is what you
made of the information, how you turned it into knowledge to be shared with the
public, and what novel ideas and solutions came from that information.
Use Twitter
If you follow the right people and use it right, Twitter is the highest-value social
platform anyone can be on today.
DM people on Twitter. Be sincere, keep it short, and have a specific ask. This guide
on How to write a cold email by Sriram Krishnan can also apply to DMs.
How to tweet? Read Anatomy of a Tweet by Jason, creator of Instructor, who grew
from 0 → 14k followers in months.
DM me about what you’re up to! I’m always up for collaborating on cool projects.
Tools
∘ Python
∘ PyTorch
Machine Learning
∘ Write from Scratch
∘ Compete
∘ Do side projects
∘ Deploy them
∘ Supplementary
Deep Learning
∘ Fast.ai
∘ Do more competitions
∘ Implement papers
∘ Computer Vision
∘ NLP
Mathematics
Search
DALL·E
This is a great series on Math for ML from a programmer’s perspective: Math for
Machine Learning by Weights & Biases (code)
Read Introduction to Linear Algebra for Applied Machine Learning with Python
alongside the course.
If you want something more traditional, look at Imperial College London lectures —
Linear Algebra & Multivariate Calculus.
Supplementary
Tools
DALL·E
Python
Beginners start here: Practical Python Programming.
Supplementary
PyTorch
Videos
PyTorch Examples
Practice
Supplementary
Machine Learning
DALL·E
JeremyNixon/oracle
trekhleb/homemade-machine-learning
If you want a challenge, write PyTorch from scratch by following this course.
Compete
Apply what you learn in competitions.
Join ML competitions on platforms like bitgrit and Kaggle; find more in this
article.
Do side projects
Read Getting machine learning to production by Vicki Boykis
She also wrote about what she learned building Viberary, a semantic search for
books.
Get a dataset and build a model (i.e., use earthaccess to get NASA Earth data).
Deploy them
Get the models in production. Track your experiments. Learn how to monitor
models. Experience data and model drift firsthand.
Made With ML
chiphuyen/machine-learning-systems-design
Supplementary
Machine Learning with PyTorch and Scikit-Learn (code)
Deep Learning
If you want a more comprehensive, traditional course, check out UNIGE 14x050 —
Deep Learning by François Fleuret.
If you need to reach for theory at some point, these are great books.
Dive into Deep Learning (has code examples in PyTorch, NumPy/MXNet, JAX,
and TensorFlow)
Read The Little Book of Deep Learning on your phone instead of scrolling Twitter.
Do more competitions
PlantTraits2024 — FGVC11 | Kaggle (computer vision)
Implement papers
Check out labml.ai Annotated PyTorch Paper Implementations
Papers with Code is a great resource; here’s BERT explained on their website.
Below are some resources for the specializations within Deep Learning
Computer Vision
A lot of people recommend CS231n: Deep Learning for Computer Vision. It’s
challenging but worth it if you get through it.
Reinforcement Learning
For RL, these two are great:
NLP
Another great Stanford course, CS 224N | Natural Language Processing with Deep
Learning
The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning
Supplementary
Then Large Language Models in Five Formulas, by Alexander Rush — Cornell Tech
You can also look at GPT in 60 Lines of NumPy | Jay Mody while you’re at it.
It teaches prompt engineering, LLMOps, UX for LLMs, and how to launch an LLM
app in an hour.
As well as Patterns for Building LLM-based Systems & Products by Eugene Yan
Participate in hackathons
lablab.ai has new AI hackathons every week. Let me know if you want to team up!
If you want to go deeper into the theory and understand how everything works:
Read papers
A great article by Sebastian Raschka on Understanding Large Language Models,
where he lists some papers you should read.
He also recently published another article with papers you should read in January
2024, covering mistral models.
Paper
Blogs
Videos
You can code transformers from scratch now. But there’s still more.
Beyond Self-Attention: How a Small Language Model Predicts the Next Token
Llama from scratch (or how to implement a paper without crying) | Brian Kitano
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation
(DoRA) from Scratch
Some more links related to LLMs that are not exhaustive. Look at LLM Syllabus for a
more comprehensive syllabus for LLMs.
Prompt Engineering
Read Prompt Engineering | Lil’Log
ChatGPT Prompt Engineering for Developers by Ise Fulford (OpenAI) and Andrew
Ng
DeepLearning.ai also has other short courses you can enroll in for free.
Fine-tuning LLMs
Read the Hugging Face fine-tuning guide.
RAG
A great article by Anyscale: Building RAG-based LLM Applications for Production
For podcasts, the best I’ve found is Latent Space by Swyx & Alessio
They also have this newsletter, Smol Talk, which summarizes all big AI discords.
openai/syllabus.md
I’ve spent enough time writing and organizing this that it’s diminishing returns. It’s
time to learn and build.
If you’ve read this far, don’t forget to reach out or leave a comment :)
Want to discuss the latest developments in Data Science and AI with other data
scientists? Join our discord server!
Follow