Course curriculum

  • 1

    Module 1: Introduction to Language Models

    • What Are Language Models (and Why Should You Care)?

    • The Rise of Small Language Models

    • SLMs vs. LLMs: Speed, Scale, and Smarts Compared

    • Where SLMs Shine: Real-world Use Cases (Edge AI, Privacy-first Apps, Offline Assistants)

  • 2

    Module 2: Core Concepts in NLP and Model Architecture

    • Tokenisation and Vocabulary Design

    • Transformer Architecture: The Essentials

    • Attention Mechanisms Explained Simply

    • How Size Affects Performance, Speed, and Cost

  • 3

    Module 3: Training Small Language Models

    • Dataset Preparation and Cleaning

    • Choosing Model Size and Architecture (parameter budgets)

    • Training Objectives: Masked, Causal, and Instruction-based

    • Techniques for Efficient Training (Quantisation, Pruning, LoRA)

  • 4

    Module 4: Evaluation and Benchmarking

    • Measuring Performance: Perplexity, BLEU, and Beyond

    • Evaluating SLMs for Task-Specific Performance

    • Common Pitfalls in Small Model Evaluation

    • Ethical and Bias Considerations in Compact Models

  • 5

    Module 5: Deployment and Optimisation

    • On-device Inference (Edge, Mobile, and IoT)

    • Model Compression and Distillation

    • Latency, Memory, and Power Optimisation

    • Serving SLMs in Production (APIs, Containers, and Pipelines)

  • 6

    Module 6: Usecases of SLMS

    • Chatbots and Personal Assistants

    • Domain-Specific SLMs (e.g. Medical, Legal, Finance)

    • Multilingual and Low-resource Adaptations

    • Hybrid Systems: Combining SLMs with Retrieval or Rules

  • 7

    Module 7: Advanced Topics

    • Fine-tuning and Continual Learning

    • Instruction-tuning and Prompt Engineering for SLMs

    • Open-weight Models (Phi, TinyLlama, Mistral)

    • Future Directions in Compact AI

  • 8

    Study Material

    • Study Material