Skip to content

Instantly share code, notes, and snippets.

@vritant24
Created June 12, 2025 22:23
Show Gist options
  • Select an option

  • Save vritant24/3c48dff782a5ebcb05c387e0559adfd0 to your computer and use it in GitHub Desktop.

Select an option

Save vritant24/3c48dff782a5ebcb05c387e0559adfd0 to your computer and use it in GitHub Desktop.

Learning Plan: Understanding and Implementing LLMs (Large Language Models)

Audience: Software engineer with a computer science background
Goal: Learn modern LLMs on a technical and foundational level with practical implementation
Time Commitment: 1–2 hours/day
Approach: Mixed theory and coding, compact but deep plan
Focus: Free resources prioritized, including YouTube, articles, and open-access papers


πŸ“… Weekly Timeline Overview

Week Focus Area
1 Fundamentals of ML & Deep Learning
2 NLP Basics & Word Embeddings
3 Transformers & Self-Attention
4 Language Modeling & Pretraining Objectives
5 Implementing Mini Language Models
6 Training & Fine-tuning Transformers
7 Serving, Prompting, and Ecosystem
8 Advanced Architectures & Research Papers
9 Capstone Implementation & Research Readings

πŸ“˜ Week 1: Machine Learning & Deep Learning Basics

Theory

Practice


πŸ“˜ Week 2: NLP Fundamentals & Embeddings

Theory

Practice

  • Implement: Word2Vec skip-gram (simplified)
  • Code: Use gensim to load embeddings and do similarity comparisons
  • Colab: Word2Vec from Scratch

πŸ“˜ Week 3: Transformers & Self-Attention

Theory

Practice


πŸ“˜ Week 4: Language Modeling & Pretraining

Theory

Practice

  • Implement: A tiny character-level language model (use Andrej Karpathy’s nanoGPT)
  • Experiment: Train it on small text datasets (e.g. Shakespeare, Python code)

πŸ“˜ Week 5: Implementing a Mini Language Model

Theory

Practice

  • Follow Karpathy’s notebook and implement a transformer step by step
  • Train it on simple text corpus and experiment with generation

πŸ“˜ Week 6: Training and Fine-tuning

Theory

Practice

  • Colab: Fine-tune GPT-2 on your own dataset using HuggingFace
  • Tool: Try transformers.Trainer and datasets

πŸ“˜ Week 7: Prompting, Serving, and Tools

Theory

Practice

  • Try: Prompting OpenAI API (or HuggingFace models)
  • Experiment: Build a chatbot or document Q&A with LangChain or LlamaIndex

πŸ“˜ Week 8: Advanced LLM Architectures & Research

Theory

Practice

  • Read and summarize papers
  • Compare architectures using visual diagrams (e.g. llama vs. gpt vs. bert)

πŸ“˜ Week 9: Capstone Project & Research

Goal

  • Implement a mini LLM pipeline from scratch (train β†’ inference)
  • Pick a paper to replicate
  • Optionally, write a blog post or dev diary of your journey

Suggestions

  • Build a domain-specific chatbot
  • Try model distillation or quantization
  • Submit to HuggingFace Spaces

βœ… Resources & Tools Summary

Type Resource
Framework PyTorch
NLP Library HuggingFace Transformers
Notebook Google Colab
Datasets HuggingFace Datasets
Books Deep Learning by Goodfellow

🧠 Tips for Success

  • Use Obsidian or Notion to journal your learning.
  • Take notes from every video/article.
  • Push implementation code to GitHub.
  • Join communities like r/LocalLLaMA or HuggingFace forums.

Happy building! πŸ’»πŸ§ 

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment