Audience: Software engineer with a computer science background
Goal: Learn modern LLMs on a technical and foundational level with practical implementation
Time Commitment: 1β2 hours/day
Approach: Mixed theory and coding, compact but deep plan
Focus: Free resources prioritized, including YouTube, articles, and open-access papers
| Week | Focus Area |
|---|---|
| 1 | Fundamentals of ML & Deep Learning |
| 2 | NLP Basics & Word Embeddings |
| 3 | Transformers & Self-Attention |
| 4 | Language Modeling & Pretraining Objectives |
| 5 | Implementing Mini Language Models |
| 6 | Training & Fine-tuning Transformers |
| 7 | Serving, Prompting, and Ecosystem |
| 8 | Advanced Architectures & Research Papers |
| 9 | Capstone Implementation & Research Readings |
- Article: Machine Learning Crash Course β Google
- YouTube: 3Blue1Brown - Neural Networks
- Book (Optional): Deep Learning Book β Goodfellow β Chapter 6β7
- Colab Notebook: Implement a simple neural net with PyTorch Quickstart
- Kaggle: Intro to Deep Learning
- Article: CS224n Notes: Word Vectors
- YouTube: Lilian Weng - NLP from Scratch
- Implement: Word2Vec skip-gram (simplified)
- Code: Use
gensimto load embeddings and do similarity comparisons - Colab: Word2Vec from Scratch
- Article: The Illustrated Transformer
- Paper: Attention Is All You Need
- YouTube: Attention Explained β Jay Alammar
- Code: Build a transformer encoder layer using PyTorch
- Colab: Transformer from Scratch β Harvard NLP
- Article: Lil'Log on Language Modeling
- Paper: GPT-1: Improving Language Understanding
- Implement: A tiny character-level language model (use Andrej Karpathyβs nanoGPT)
- Experiment: Train it on small text datasets (e.g. Shakespeare, Python code)
- YouTube Series: Andrej Karpathyβs GPT From Scratch
- Repo: minGPT
- Follow Karpathyβs notebook and implement a transformer step by step
- Train it on simple text corpus and experiment with generation
- Colab: Fine-tune GPT-2 on your own dataset using HuggingFace
- Tool: Try
transformers.Traineranddatasets
- Article: Prompt Engineering Guide
- Article: LangChain Overview
- Try: Prompting OpenAI API (or HuggingFace models)
- Experiment: Build a chatbot or document Q&A with LangChain or LlamaIndex
-
Papers:
-
YouTube: Yannic Kilcherβs summaries of major papers
- Read and summarize papers
- Compare architectures using visual diagrams (e.g. llama vs. gpt vs. bert)
- Implement a mini LLM pipeline from scratch (train β inference)
- Pick a paper to replicate
- Optionally, write a blog post or dev diary of your journey
- Build a domain-specific chatbot
- Try model distillation or quantization
- Submit to HuggingFace Spaces
| Type | Resource |
|---|---|
| Framework | PyTorch |
| NLP Library | HuggingFace Transformers |
| Notebook | Google Colab |
| Datasets | HuggingFace Datasets |
| Books | Deep Learning by Goodfellow |
- Use Obsidian or Notion to journal your learning.
- Take notes from every video/article.
- Push implementation code to GitHub.
- Join communities like r/LocalLLaMA or HuggingFace forums.
Happy building! π»π§