Compare Masked Language Modeling and Next-Token Prediction for LLM pretraining. Learn which objective delivers better performance for understanding vs. generation tasks, and explore hybrid strategies.