Policy Optimization for RL and LLM Alignment — Theory

The optimization story behind aligning language models with human preferences

February 17, 2026 · 41 min · Ji Hun Wang

Variational Autoencoders

On my favorite generative model of all time!

July 19, 2025 · 16 min · Ji Hun Wang