Notebooks — Overview
(1) Artificial Neural Networks (Basic Architecture) |
(2) Attention & Multi-Head Attention |
(3) Backpropagation |
(4) Backpropagation (Generalization) |
(5) Bias & Variance (Machine Learning) |
(6) Bias-Variance Decomposition |
(7) Building a GPT-Style LLM from Scratch |
(8) Building a Word Tokenizer from Scratch |
(9) Byte-Pair Encoding Tokenization |
(10) Curse of Dimensionality |
(11) Data Batching for Training LLMs |
(12) Data Normalization — Motivation & Overview |
(13) Data Preparation for Training LLMs — An Overview |
(14) Decision Trees |
(15) Decision Trees — CART (Classification and Regression Trees) |
(16) Dropout |
(17) Gradient Descent — The (Very) Basics |
(18) Handwritten Digit Recognition with Artificial Neural Networks (ANNs) |
(19) Implementing an ANN from Scratch (NumPy only) |
(20) Language Models |
(21) Linear Regression |
(22) Linear Regression — Assumptions & Caveats |
(23) LoRA Fine-Tuning — A Basic Example |
(24) Logistic Regression — Basics |
(25) Logistic Regression: The Math |
(26) Logit Distillation |
(27) Machine Translation with Transformers |
(28) Masking in Sequence Models |
(29) Mixture of Experts (MoE) |
(30) Model Fine-Tuning for LLMs — An Overview |
(31) Multinomial Naive Bayes (Basics) |
(32) NumPy — Basic Tutorial |
(33) Part-of-Speech (POS) Tagging (Basics) |
(34) Porter Stemmer |
(35) Positional Encodings — Overview |
(36) RNN-based Language Models |
(37) Recurrent Neural Networks — An Introduction |
(38) Resource-Efficient LLMs — An Overview |
(39) Retrieval-Augmented Generation (RAG) — A (Very) Basic Example |
(40) Retrieval-Augmented Generation (RAG) — Basics |
(41) Rotary Position Embeddings (RoPE) |
(42) Sinusoidal Positional Encodings (Original Transformer) |
(43) Stemming & Lemmatization |
(44) Subword Tokenization (WordPiece) |
(45) Text Classification with Recurrent Neural Networks (RNNs) |
(46) Text Normalization |
(47) Text Tokenization |
(48) The Linear Layer |
(49) The Math Behind Linear Regression |
(50) The Softmax Function |
(51) Token Indexing with Vocabularies |
(52) Training Word2Vec from Scratch |
(53) Transformers — Basic Architecture |
(54) Using Pretrained LLMs Locally — A Starter Guide |
(55) Vector Space Model |
(56) Word & Text Embeddings — An Overview |
(57) Working with Batches for Sequence Tasks |
(58) Working with the OpenAI API — An Introduction |
(this list of notebooks is auto-generated)