Text Diffusion: The Idea Hiding Inside BERT All Along
Auto regressive generation is sequential and diffusion uses much fewer passes in text generation.
Auto regressive generation is sequential and diffusion uses much fewer passes in text generation.
Cross-entropy loss isn’t a heuristic, it is maximum likelihood estimation with a sign flip. It also shows how the same math powers GPT training.
Understanding the basics of RLHF vs RLAIF vs RLVR for AI feedback comparison
Learning to rank with lambdarank multi objective pairwise ranking models using lightgbm