Cross Entropy Loss Connection to GPT Models
Cross-entropy loss isn’t a heuristic, it is maximum likelihood estimation with a sign flip. It also shows how the same math powers GPT training.
Cross-entropy loss isn’t a heuristic, it is maximum likelihood estimation with a sign flip. It also shows how the same math powers GPT training.
Understanding the basics of RLHF vs RLAIF vs RLVR for AI feedback comparison
Master the complete pipeline for fine-tuning Large Language Models using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) with Axolotl f...
Understanding RoPE Scaling and how it enables LLMs to handle longer contexts
Basics of gradient accumulation and gradient checkpointing to train LLMs
Understanding basics of flops and how they influence gpu computation
Lessons learned from deploying large language models in productions using vLLM
Tutorial to work on large datasets in memory Polars DataFrames in Python
Understanding asyncio using simple examples in python 3.7
Learning to use FTS5 with sqlite for full text search applications
Use pex python library to package a python project for deployment.
this line will show up as preview on the posts page
Create tfidf matrix in R just like using scikit-learn
Use pex python library to package a python project for deployment.
this line will show up as preview on the posts page
Use pex python library to package a python project for deployment.
this line will show up as preview on the posts page
Master the complete pipeline for fine-tuning Large Language Models using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) with Axolotl f...
Basics of gradient accumulation and gradient checkpointing to train LLMs
Create tfidf matrix in R just like using scikit-learn
Create tfidf matrix in R just like using scikit-learn
Learning to use FTS5 with sqlite for full text search applications
Learning to use FTS5 with sqlite for full text search applications
Understanding asyncio using simple examples in python 3.7
Understanding asyncio using simple examples in python 3.7
Lessons learned from deploying large language models in productions using vLLM
Understanding basics of flops and how they influence gpu computation
Basics of gradient accumulation and gradient checkpointing to train LLMs
Understanding RoPE Scaling and how it enables LLMs to handle longer contexts
Understanding RoPE Scaling and how it enables LLMs to handle longer contexts
Learning to rank with lambdarank multi objective pairwise ranking models using lightgbm
Learning to rank with lambdarank multi objective pairwise ranking models using lightgbm
Master the complete pipeline for fine-tuning Large Language Models using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) with Axolotl f...
Cross-entropy loss isn’t a heuristic, it is maximum likelihood estimation with a sign flip. It also shows how the same math powers GPT training.