When I joined a research lab in May, I quickly realized there's a big gap between using OpenAI's API and actually training models. This guide aims to bridge that gap with carefully selected resources and a structured learning path.

Visual Learning Resources

These visualizations helped me understand the core concepts:

Essential Reading

Two books stand out for different learning styles:

Getting Started with Code

Hardware Options

  • Free: Google Colab - Great for learning
  • Local: M-series Macbooks or any modern laptop for small models

Software Stack

Ideas for First Projects

  1. Try to run MiniCPM or YOLOv8 locally.
  2. Finetune a model on a dataset from Roboflow
  3. Join a competition on Kaggle

Learning Resources

  • Take DeepLearning.AI courses
  • Join university ML clubs or classes (surrounding yourself with ML enthusiasts helps!)
  • Stay updated via tldr.tech/ai newsletter and follow key ML researchers ( scroll through my follows for reference)

Study Method

I use this prompt template with LLMs to learn new concepts:

Core Understanding
• Category (ML/math/startups/investments): [WHY]
• ELI5: [simple analogy with everyday objects]
• Technical: [detailed explanation]

Connections & Context
• Prerequisites → [core concepts needed]
• Builds into → [advanced applications]
• Related to → [parallel concepts]

Implementation
• Industry use: [practical applications]
• Common pitfalls: [key challenges]
• Your context: [relevance to ML/startups/grad school]

Flash Card
[in the format: Question == Answer] (to be added to remnote)

TL;DR
[One-paragraph essence + key insight]

Time Context: Today's date

Explain

Remember: Start small, focus on understanding fundamentals, and learn by implementing. While the gap between using APIs and training models may seem daunting at first, this structured approach and these resources can help make the transition more manageable.