Skip to content

AI Courses

Curated university courses on LLMs, Transformers, Graph Models, Reasoning, and AI Safety — sorted by recency and quality.


Tier 1: Premier Research-Oriented Courses (2025–2026) Premier

Stanford CS336: Language Modeling from Scratch (Spring 2025)

Instructors: Percy Liang & Tatsunori Hashimoto Course site · GitHub

A deep, implementation-heavy graduate course on building LLMs entirely from the ground up. Covers data collection/cleaning, tokenizer and Transformer architecture implementation, model training at scale, and evaluation/deployment. Advanced topics: custom FlashAttention kernels, multi-GPU distribution, scaling laws, and alignment (SFT + RLHF).

Aspect Rating Notes
Depth A Low-level LLM internals; few courses match this
Relevance A State-of-the-art training techniques
Accessibility B Materials public; requires advanced skills + GPUs
Career Value A Ideal for AI researchers and expert engineers

Harvard CS 2881: AI Safety (Fall 2025)

Instructor: Boaz Barak Course site · YouTube playlist

Graduate-level course on AI alignment and safety. Covers generalization, optimization, robustness, interpretability, and alignment. Seminar-style with readings, midterm, and NeurIPS-style research final project. Student projects produce original research.

Aspect Rating Notes
Depth A Conceptually deep, tackles unsolved problems
Relevance A Critical for future AI development
Accessibility A Public lectures and materials
Career Value A For researchers pursuing AI safety

CMU 11-667: Large Language Models: Methods and Applications (Fall 2025)

Instructors: Jaromir Savelka & Yubin Kim Course site

Holistic graduate survey of LLM techniques. Part 1: architectures, pre-training, inference, evaluation, alignment, emergent capabilities. Part 2: scaling, efficiency, deployment (latency, cost, safety). Students implement and fine-tune models from scratch in PyTorch.

Aspect Rating Notes
Depth A- Thorough on core concepts and applications
Relevance A Covers everything needed for LLM work in 2025
Accessibility B- Syllabus public; lectures not openly shared
Career Value A Excellent for NLP/LLM-focused roles

CMU 11-868: Large Language Model Systems (Spring 2025)

Instructor: Lei Li Course site

Advanced systems engineering for LLMs. Covers efficient training algorithms, distributed training, memory optimization, inference serving, caching, model sharding, RLHF at scale, and continuous learning. Intersection of ML and systems research.

Aspect Rating Notes
Depth A Deep dive into LLM engineering
Relevance A Critical as LLMs move to production
Accessibility C+ Specialized content not openly provided
Career Value A Gold for AI infrastructure engineers

Stanford CS372: AI for Reasoning, Planning, and Decision Making (Spring 2025)

Instructor: Edward Y. Chang Course site · Stanford Online

Cutting-edge course on reasoning frameworks atop LLMs, targeting AGI-like abilities. Topics: multi-agent LLM systems, planning algorithms, temporal reasoning, validated decision-making, long-term memory, ethical alignment. Research-focused seminar on frontier AI.

Aspect Rating Notes
Depth A Forward-looking, tackles unsolved problems
Relevance A- Highly relevant for future AI, less for current
Accessibility C Limited; some lectures shared informally
Career Value A Goldmine for next-gen AI systems

Stanford CS224W: Machine Learning with Graphs (Fall 2025)

Instructor: Jure Leskovec Course site · Videos (2021)

Comprehensive deep dive into graph ML and GNNs. Covers PageRank, community detection, representation learning, GNN architectures, Graph Transformers, knowledge graph reasoning. Applications: social networks, recommender systems, biology.

Aspect Rating Notes
Depth A One of the most in-depth graph ML courses
Relevance A- Crucial for relational data domains
Accessibility B+ Materials free; older recordings available
Career Value A Foundational for graph ML research

CMU Advanced NLP (Fall 2025)

Course site · YouTube playlist

Graduate-level, practice-oriented coverage of modern NLP: transformers and LLMs, prompting and agents, alignment (RL/GRPO), evaluation, safety, and applied projects.


Tier 2: Excellent Foundational Courses (2025) Excellent

Stanford CME 295: Transformers & Large Language Models (Autumn 2025)

Instructors: Afshine Amidi & Shervine Amidi Course site · Cheatsheet · Syllabus + Videos

Comprehensive intro to Transformers and LLMs. Covers evolution of NLP, Transformer architecture (tokenization, embeddings, attention), LLM foundations (MoEs, decoding), training/tuning (SFT, RLHF, LoRA), RAG, and agentic workflows.

Aspect Rating Notes
Depth B+ Great breadth, solid depth for overview
Relevance A Ubiquitous transformer paradigm
Accessibility A Free lectures, slides, illustrated cheatsheet
Career Value A- Strong for practitioners and NLP foundations

Stanford CS 120: Introduction to AI Safety

Instructor: Max Lamparth Course site

Foundational AI safety course covering interpretability, robustness, RL safety, NLP issues. Topics span reinforcement learning, computer vision, and NLP. No strict prerequisites.

Aspect Rating Notes
Depth B+ Solid foundations
Relevance A Critical safety topics
Accessibility B+ Open to Stanford community
Career Value B+ Good entry point for AI safety

Stanford CS329H: Machine Learning from Human Preferences

Course site · YouTube playlist

Advanced topics: preference modeling, RLHF/GRPO, reward modeling, dataset construction, safety and alignment, evaluation beyond accuracy, applications to LLMs and agents.


MIT MAS.S60: How to AI (Almost) Anything (Spring 2025)

Instructor: Paul Liang (MIT Media Lab) Course site · Videos

Broad modern AI survey emphasizing foundation models and multimodal applications. Covers LLMs plus vision, audio, sensor data, music, art. Topics: multimodal fusion, cross-modal mapping, grounding AI in physical world.

Aspect Rating Notes
Depth B+ Wide range at solid graduate level
Relevance A Foundation models + multimodal AI
Accessibility A Free via MIT OCW
Career Value A Panoramic view of modern AI

Davidson CSC 381: Deep Learning (Fall 2022)

Instructor: Prof. Bryce Wiedenbeck YouTube playlist

"Hands down, the clearest and most effective explanation of Transformers on YouTube." — Highly praised lecture series covering neural networks, deep learning fundamentals, and an exceptionally clear treatment of Transformers and self-attention.

Aspect Rating Notes
Depth B+ Excellent pedagogical clarity
Relevance A Timeless Transformer foundations
Accessibility A Free on YouTube
Career Value A- Best starting point for understanding Transformers

Tier 3: Specialized & Domain Courses Specialized

CS-E4740: Federated Learning (Aalto University, Spring 2025)

Instructor: Alexander Jung Course site · GitHub · YouTube playlist

FL networks, design principles, core algorithms, variants, and trustworthy FL. Weekly lectures and Python/Jupyter exercises.


HAII-2024: Human-AI Interaction (Politecnico di Torino)

Instructors: Luigi De Russis, Alberto Monge Roffarello YouTube playlist

Ph.D. course on methods for designing, building, and evaluating human–AI interactions. Covers prototyping, UX for AI, evaluation methods, and case studies.


HAI Fall Conference 2022: AI in the Loop (Stanford HAI)

YouTube playlist

Human-in-the-loop AI decision-making processes. Explores keeping humans at the center of AI technologies.


Professional Programs & Certificates Professional

CMU Graduate Certificate: Generative AI & Large Language Models (Online, 2025)

Program site · Curriculum

12-month online program (3 courses). Build LLMs from scratch, fine-tune models, distributed training (SLURM), RLHF, compression, multimodal AI. Live evening classes for working professionals.

Aspect Rating Notes
Depth A Compressed master's-level curriculum
Relevance A Tuned to 2025 industry needs
Accessibility B Online but paid (~tuition-based)
Career Value A CMU credential + immediately applicable skills

MIT Professional: AI System Architecture & LLM Applications (July 2026)

Instructor: David Martinez (MIT Lincoln Laboratory) Program site

5-day intensive on designing, building, and deploying LLM-powered systems end-to-end. Covers architecture design, deployment at scale, testing, evaluation, responsible AI. Culminates in group project building an LLM application.

Aspect Rating Notes
Depth B Thorough overview for practitioners
Relevance A LLM deployment — critical skill
Accessibility D Paid (~$4,500), in-person Cambridge
Career Value A- Fast-track for technical leads/managers

Oxford (LMH): Advanced AI & ML: LLMs and NLP (Summer 2026)

Program site · NLP Course

3-week intensive (in-person or online). Covers Transformers, attention, LLM fine-tuning (LoRA, QLoRA), multimodal AI (vision-language), AI ethics. Hands-on labs.

Aspect Rating Notes
Depth B+ Rigorous for summer program
Relevance A- LLMs, NLP, vision-language
Accessibility B Online option; paid program
Career Value B+ Oxford credential + networking

Quick Reference by Topic

Topic Top Recommendations
Transformers (from scratch) Davidson CSC 381, Stanford CME 295
LLM Training & Systems Stanford CS336, CMU 11-868
LLM Applications CMU 11-667, MIT MAS.S60
AI Safety & Alignment Harvard CS 2881, Stanford CS 120, Stanford CS329H
Reasoning & AGI Stanford CS372
Graph ML & GNNs Stanford CS224W
Multimodal AI MIT MAS.S60, Oxford Summer
Professional Upskilling CMU Certificate, MIT Professional

Learning Path Recommendations

For beginners wanting to understand Transformers: 1. Davidson CSC 381 (clearest explanations) 2. Stanford CME 295 (comprehensive + cheatsheet)

For researchers building LLMs: 1. Stanford CS336 (build from scratch) 2. CMU 11-868 (systems engineering) 3. CMU 11-667 (methods + applications)

For AI safety focus: 1. Stanford CS 120 (foundations) 2. Harvard CS 2881 (research-level) 3. Stanford CS329H (human preferences)

For industry practitioners: 1. Stanford CME 295 (fast ramp-up) 2. CMU Certificate (structured + credential) 3. MIT Professional (intensive deployment)


Last updated: January 2026