What's next? - Generative AI with SnapGPT

Objective

After completing this lesson, you will be able to understand statistical language models and next-token predictions that are the basis for modern Large Language Models.

Content

Video summary:

The video explains a simple generative-AI principle using n‑grams and shows how the same next-token prediction idea can generate text, music, and drawings. It walks through splitting corpora into n‑grams, building a statistical model of n‑grams, repeatedly predicting the next token to grow a sequence, and demos generating fairy tales, children’s-music improvisations, and doodle-style sketches.

Key points:

  • Domino/n‑gram view: break a corpus into overlapping n‑word (or n‑token) tiles so predicting the next token is like finding the next matching tile.
  • Build a model by computing 1…N grams (monograms, bigrams, trigrams, etc.), then filter matching n‑grams to pick plausible continuations (next-token prediction).
  • The same pipeline works for other sequential data: musical notes (pitch+duration) and drawings generatin the next token from directions and distances to generate new sequences.
  • Repeatedly append predicted tokens to the sequence to generate content; diversity of training data matters, and outputs are statistical/stylistic rather than semantically understanding.