Saikat's notes on AI
πŸ πŸ¦πŸ’ΌπŸ§‘β€πŸ’»
  • Hello world!
  • πŸš€LLM
    • The Evolution of Language Models: From Word2Vec to GPT-4
      • [1] Word2Vec - Efficient Estimation of Word Representations in Vector Space
      • [2] Seq2Seq - Sequence to Sequence Learning with Neural Networks
      • [3] Attention Mechanism - Neural Machine Translation by Jointly Learning to Align and Translate
      • [4] Transformers - Attention Is All You Need
      • [5] GPT - Improving Language Understanding by Generative Pre-Training
      • [6] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
      • [7] T5 - Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
      • [8] GPT2 - Language Models are Unsupervised Multitask Learners
  • Best LLM Resources on the internet
  • MPT-7B: A Revolutionary Leap in Language Models
  • From Rules to Vectors: How NLP Changed Over Time
Powered by GitBook
On this page
  • Talks
  • Lecture Videos
  • Courses
  • Blogs

Was this helpful?

Best LLM Resources on the internet

Talks

  • The Amazing AI Super Tutor for Students and Teachers | Sal Khan | TED

Lecture Videos

  • Let's build GPT: from scratch, in code, spelled out by Andrej Karpathy

  • GPT-4 - How does it work, and how do I build apps with it? - CS50 Tech Talk

  • Large Language Models from scratch: part 1 + part 2

Courses

  • CS 324 - Large Language Models

  • NLP Course by Huggingface : a really great course for understanding transformers, which is the backbone for LLMs.

  • COS 597G: Understanding Large Language Models

  • CSCI 601.771: Self-supervised Statistical Models

Blogs

  • https://jalammar.github.io => intuitive posts on language models & transformers. Great visuals.

Previous[8] GPT2 - Language Models are Unsupervised Multitask LearnersNextMPT-7B: A Revolutionary Leap in Language Models

Last updated 2 years ago

Was this helpful?