Roadmap
Your journey to a first break in AI
Your learning path — one journey, step by step. The cohort runs 8 March 2026 — 7 June 2026 (3 months). Use your AI-based IDE and the community to complete each step. This roadmap is in progress; new steps get added as the cohort grows.
Step 1
First use of AI for coding
Set up a Quarto blog and host it on GitHub with an about-me page, blog posts, “Today I learned,” and other pages.
- Set up the project locally, link it to a GitHub repo, and configure GitHub Pages for deployment.
- Use your AI-based IDE to complete this setup.
You will learn:
- GitHub basics refresher
- Setting up a personal and blogging website
- How coding tools and SWE / AI agents work
Step 2
Run a model locally
Run a model locally using a basic inference setup (e.g. llama.cpp or another minimal setup).
You will learn:
- Basics of inference: decoding, KV cache
- Chat templates and system prompts
- Prompting basics, tags, and special tags
- Tokenization
Step 3
Inference deep dive
Go beyond running a model — understand how inference works under the hood and how to serve models.
You will learn:
- Inference engines and runtimes (vLLM, TGI, llama.cpp server)
- Batching, continuous batching, and throughput vs. latency
- Quantization (GGUF, GPTQ, AWQ) and when to use each
- Structured output, function calling, and tool use
- Serving and API design for inference endpoints
Step 4 — coming soon
Training fundamentals
Build the foundations to train and fine-tune models from scratch.
You will learn:
- PyTorch fundamentals: tensors, autograd, training loops
- Modelling: architectures (transformers, attention, MLP), building blocks
- Data pipelines: datasets, dataloaders, preprocessing
- Fine-tuning: LoRA, QLoRA, full fine-tune, adapters
- Distributed training: DDP, FSDP, multi-GPU and multi-node setups
- Experiment tracking and evaluation
Step 5 — coming soon
Build an AI product
Ship an AI-powered product end to end.
You will learn:
- Product thinking: problem → solution → users
- Building with APIs, RAG, agents, and tool use
- Frontend/backend integration for AI features
- Deployment, monitoring, and iteration
Step 6 — coming soon
Capstone project or open-source contribution
Prove what you’ve learned. Pick one: build a capstone project or make a meaningful contribution to an open-source AI project.
Options:
- Capstone: End-to-end project combining inference, training, or product skills — deployed, documented, and added to your public portfolio
- Open-source contribution: Submit a PR to an AI repo (model, library, dataset, docs) — get reviewed, merged, and credited
- Present your work to the cohort; get peer feedback
Why it matters: A shipped project or merged PR is the strongest signal on your profile when applying for your first AI role.
Step 1: First use of AI for coding — Quarto blog with GitHub
Goal: Create a Quarto blog and host it on GitHub with an about-me page, blog posts, “Today I learned,” and other pages.
- Set up the project locally, link it to a GitHub repository, and configure GitHub Pages for deployment.
- Use your AI-based IDE to complete this setup.
Learning objectives: GitHub basics refresher · Setting up a personal and blogging website · Understanding how coding tools or SWE / AI agents work
Step 2: Run a model locally — Basic inference setup
Goal: Run a model locally using a basic inference setup (e.g. llama.cpp or another minimal setup).
Topics: Basics of inference (decoding, KV cache) · Chat templates and system prompts · Prompting basics, tags, and special tags · Tokenization
Step 3: Inference deep dive
Goal: Go beyond running a model — understand how inference works under the hood and how to serve models.
Topics: Inference engines and runtimes (vLLM, TGI, llama.cpp server) · Batching and continuous batching · Quantization (GGUF, GPTQ, AWQ) · Structured output, function calling, and tool use · Serving and API design for inference endpoints
Step 4: Training fundamentals (coming soon)
Goal: Build the foundations to train and fine-tune models from scratch.
Topics: PyTorch fundamentals (tensors, autograd, training loops) · Modelling (transformers, attention, MLP) · Data pipelines · Fine-tuning (LoRA, QLoRA, full fine-tune, adapters) · Distributed training (DDP, FSDP, multi-GPU) · Experiment tracking and evaluation
Step 5: Build an AI product (coming soon)
Goal: Ship an AI-powered product end to end.
Topics: Product thinking · Building with APIs, RAG, agents, and tool use · Frontend/backend integration · Deployment, monitoring, and iteration
Step 6: Capstone project or open-source contribution (coming soon)
Goal: Prove what you’ve learned. Pick one: build a capstone project or make a meaningful contribution to an open-source AI project.
Options: Capstone (end-to-end project, deployed and documented, added to public portfolio) · Open-source contribution (PR to an AI repo — model, library, dataset, or docs) · Present your work to the cohort for peer feedback
More steps can be added as the roadmap grows. Suggest new modules via CONTRIBUTING.md or a pull request.