Covers neural networks, model architectures, training dynamics, inference, and AI safety/policy.


Foundations

  • neural-network — Overview: nodes, edges, weights, activations; links to specific architectures
  • multilayer-perceptron — Fully-connected feed-forward network (MLP); the simplest architecture
  • backpropagation — Algorithm that computes the cost-function gradient, , via the chain rule
  • computation-graph — DAG of mathematical operations recorded during the forward pass; the structure autograd traverses to compute gradients
  • gradient-descent — Iterative optimisation: step in the negative-gradient direction , , to minimise cost
  • cost-function — Scalar measure of how badly the network performs (e.g. mean squared error)
  • activation-function — Nonlinear function (sigmoid, ReLU) on layer’s weighted sum, giving a NN its expressive power

Architectures

Embedding & I/O

  • one-hot-encoding — Binary vector representation of categorical indices; prevents false ordinal relationships and acts as a row-select on the weight matrix
  • tokenization — Splitting text into subword tokens before embedding
  • word-embedding — Learned vectors per token; directions encode meaning; dot product measures alignment
  • unembedding — Final projection from residual stream to per-token scores
  • logits — Unnormalised pre-softmax scores over the vocabulary
  • softmax — Turns logits into a probability distribution; temperature controls sharpness

Training & Optimisation

  • pretraining — Self-supervised next-token prediction over massive text corpora
  • rlhf — Post-pretraining fine-tuning using human preferences to bend models toward assistant behaviour

Inference & Deployment

Language Models

  • large-language-model — Top-level concept: what an LLM is, how sampling works, why scale matters
  • gpt-3 — OpenAI’s 175B-parameter transformer; running example for all the parameter counts here

Interpretability

  • superposition — Why features are spread across many neurons rather than one-per-neuron
  • johnson-lindenstrauss-lemma — The math result explaining why high-dim spaces can pack many near-orthogonal directions

AI Safety & Alignment

Key Papers & Source Summaries

Neural Networks series (3b1b):

LLMs series (3b1b):

Queries

  • backprop-graph-terminology — Why “children”, “upstream”, and “downstream” mean what they do in a backprop computation graph

Key Figures

  • 3blue1brown — Grant Sanderson’s math/ML YouTube channel; source of the Neural Networks and LLMs series