Changelog
Append-only log of wiki ingests.
2026-04-20
- Initialized vault structure. Created raw/ and wiki/ folders, index.md, and CHANGELOG.md.
- Ingested week 1. Created week page
weeks/week-01.mdand 5 concept pages:perceptron.md,dot-product.md,decision-boundary.md,loss-function.md,maximum-likelihood-estimation.md. Updatedindex.md.
2026-04-24
- Ingested week 2. Created week page
weeks/week-02.mdand 5 concept pages:gradient-descent.md,learning-rate.md,sigmoid-function.md,binary-cross-entropy.md,gradient-descent-variants.md. Updatedindex.md. - Ingested week 3. Created week page
weeks/week-03.mdand 6 concept pages:multi-layer-perceptron.md,computation-graph.md,backpropagation.md,softmax.md,overfitting.md,regularization.md. Extendedbinary-cross-entropy.mdwith a multiclass section linking to softmax. Updatedindex.md.
2026-04-27 (week 9)
- Ingested week 9. Created week page
weeks/week-09.mdand 7 concept pages:language-model.md(the framing — , autoregressive vs masked, applications),n-gram-language-model.md(chain rule + Markov assumption + MLE counting; size scaling ; zeros problem; the two structural limitations — no long-distance dependencies, no synonymy generalisation — that motivate neural LMs),perplexity.md(per-word geometric-mean inverse probability; intrinsic vs extrinsic evaluation; intuition as average branching factor),decoding-strategies.md(PDF→CDF sampling primitive; greedy, beam, ancestral sampling, temperature, top-K, top-P),word-embedding.md(word2vec, GloVe, emergent linear analogy geometry, static vs contextualized; the recipe also generalises to images/multimodal — connects to representation learning),recurrent-neural-network.md(weight sharing across time, vanishing/exploding gradients, BPTT, sequential processing as the GPU bottleneck),lstm.md(cell state with additive forget/input/output gates as the residual-connection-of-recurrence; BiLSTM for ELMo). Wove the week-09 problem set into the week page (Q1 unigram drill confirming and ; Q2 bigram drill with greedy search demonstrating the cycle failure mode —<SOS> the cat sat on the cat sat on the— illustrating why greedy decoding gets stuck in local loops). Survey-level treatment of Transformer architectures (encoder-only BERT family, decoder-only GPT family, encoder-decoder T5/Whisper), instruction tuning, RLHF, and CLIP — these point forward to week 11’s deep dive into attention rather than getting their own concept atoms (week 11 will likely create transformer/attention/BERT/GPT atoms in depth). The week-09 l21 transcript is essentially empty (national holiday); slides + l19/l20 transcripts + Jurafsky/Martin lineage carried the lecture content. Updatedindex.md. Note: the cross-link fromrepresentation-learning.mdtoword-embedding.mdwould strengthen the cross-week story but was not added in this pass — could be done in a future cross-link pass.
2026-04-27 (week 8)
- Ingested week 8. Created week page
weeks/week-08.mdand 2 concept pages:diffusion-model.md(the comprehensive DDPM atom) andlatent-diffusion-model.md(Stable Diffusion as a distinct architecture). The DDPM page covers forward/reverse processes, the closed-form forward shortcut via , the noise-prediction parameterisation (predict , recover analytically), training Algorithm 1 and sampling Algorithm 2 from the DDPM paper, U-Net + timestep embedding, the quality/diversity/speed trilemma, and extensions (cascaded, super-resolution, image-to-image, segmentation). The LDM page covers the encoder/decoder around the diffusion process, the speed-up math, cross-attention conditioning for text-to-image, and disambiguates “latent” between AE / GAN-prior / LDM-grid uses (cross-linked tolatent-representation.md). Wove the week-08 problem set into the week page (Q1 6-statement T/F on DDPM determinism / U-Net role / latent-code confusion; Q2 mismatched-eye-colour receptive-field bug). Conditional diffusion is not a new concept atom — it’s “diffusion with added as U-Net input”; reuses the existingconditional-generative-model.mdfor the posterior-estimation framing. Updatedindex.md. Note: no Notion notes for week 8.
2026-04-26 (week 7)
- Wove 3Blue1Brown’s Bayes lesson into
bayes-theorem.md. Added a leading “Steve the librarian” intuition pump (Kahneman & Tversky) before the formula — concrete-counts framing (210 people: 10 librarians, 200 farmers; 40% vs 10% match rates → posterior) plus a> [!warning] COMMON MISCONCEPTIONcallout for base-rate neglect. Added a “geometry: restricted possibility space” section explaining Bayes as area-ratio in a square, with Grant’s verbatim “the actual math of probability is really just the math of proportions” quote in a tip box and his “rationality is not about knowing facts” line in a quote box. Added a recall question on base-rate neglect. Kept the penguin example in place (it’s the bridge to conditional generative models). The Steve example is now the intuition pump; penguin is the module application.
2026-04-26 (week 7)
- Ingested week 7. Created week page
weeks/week-07.mdand 4 concept pages:generative-model.md,generative-adversarial-network.md,bayes-theorem.md,conditional-generative-model.md. The GAN page consolidates discriminator/generator losses, theoretical-vs-practical generator loss with the saturating-gradient explanation, the min-max algorithm, the Jensen-Shannon proof sketch, and DCGAN/BigGAN/StyleGAN as variants. The Bayes page is foundational (will be reused for week 8 diffusion); covers prior/likelihood/posterior with the penguin walkthrough plus the MMSE-regression-converges-to-conditional-mean digression. The conditional-GM page builds on Bayes to motivate cGAN/pix2pix and explains the L1+GAN hybrid loss. Wove the week-07 problem set lessons (Q1 MSE-regression-on-digits → conditional mean is a blurry compromise, Q2 cGAN tensor sizes, Q3 GAN true/false statements) into the week page. Cross-linked -disambiguation (GAN’s is a prior sample, not an encoded latent) tolatent-representation.md. Updatedindex.md. Note: no Notion notes for week 7, so the wiki draws from slides + transcripts + problem set only.
2026-04-26 (continued)
- Ingested week 6. Created week page
weeks/week-06.mdand 7 concept pages:autoencoder.md,latent-representation.md,representation-learning.md,self-supervised-learning.md,contrastive-learning.md,pretext-task.md,clever-hans-effect.md. Thelatent-representation.mdpage explicitly defines “latent” and disambiguates the AE-vs-SimCLR naming clash (in AE: is kept; in SimCLR: is kept, is discarded). Cross-linked fromautoencoder.md,contrastive-learning.md,representation-learning.md, and the week page. - Wove in remaining Notion supplementary material into week 6. Added the translation/re-translation pedagogical framing to
autoencoder.md(motivates why the decoder exists as a grader, not the goal). Added the detective-vs-artist intuition tocontrastive-learning.md(why contrast beats reconstruction for representation quality). Added a t-SimCNE section tocontrastive-learning.mdcovering 2-d unsupervised dataset visualization with contrastive learning (slides 72–74). Added a word-embeddings aside torepresentation-learning.mdto anchor the cross-domain generality of the recipe. Copied 5 figures from the Notion export intoimages/week-06/(autoencoder architecture+loss, identity trap, latent walk on digit-7, MNIST clustering, compression pipeline) and embedded them across the week page andautoencoder.md. Wove the week-06 problem-set Q1 (skip-connections defeat AEs / lower loss ≠ better representation) and Q2 (combining hospital datasets → Clever Hans) intoautoencoder.mdandclever-hans-effect.md. Updatedindex.md. Note: week-06 ships only slides + problem set (no lecture transcripts), so the wiki draws primarily from slides and Notion notes.
2026-04-26
- Ingested week 4. Created week page
weeks/week-04.mdand 6 concept pages:image-representation.md,activation-functions.md,convolution.md,pooling.md,convolutional-neural-network.md,shift-invariance-equivariance.md. Updatedindex.md. - Wove in Notion supplementary material into week 4. Added geometric/decision-boundary intuition to
activation-functions.md; added “Weight sharing” section, end-to-end framing, “Three hyperparameters” section, and “What is a kernel (and a filter)?” definition toconvolution.md. Embedded 13 week-04 figures acrossconvolution.md,pooling.md, andconvolutional-neural-network.md. - Ingested week 5. Created week page
weeks/week-05.mdand 8 concept pages:weight-initialization.md,normalization.md,data-augmentation.md,dropout.md,residual-connection.md,transfer-learning.md,upsampling.md,u-net.md. Extendedlearning-rate.mdwith a learning-rate schedule section. Cross-linkedoverfitting.md,regularization.md, andconvolutional-neural-network.mdto the new concepts; added a forward-pointing ResNet subsection toconvolutional-neural-network.md. Copied 11 figures from the Notion export intoimages/week-05/with descriptive names; embedded across the new concept pages. Updatedindex.md. - Naming violations flagged (not auto-renamed):
raw/week-05/w5-notion.zipshould bew05-notion.zip(zero-padded). The Notion export folder name (raw/week-05/ExportBlock-…) is auto-generated; not in vault convention. Neither was renamed because the .zip is referenced as a source citation.