N-Gram House

Tag: top-k

Controlling Length and Structure in LLM Outputs: Practical Decoding Parameters

Controlling Length and Structure in LLM Outputs: Practical Decoding Parameters

Learn how to control LLM output length and structure using decoding parameters like temperature, top-k, top-p, and repetition penalties. Practical settings for real-world use cases.

Categories

  • Machine Learning (66)
  • History (50)
  • Software Development (7)
  • Business AI Strategy (6)
  • AI Security (4)

Recent Posts

Time Savings from Generative AI: How Much Time Do Teams Really Get Back? Mar, 17 2026
Time Savings from Generative AI: How Much Time Do Teams Really Get Back?
Synthetic Data Generation with Multimodal Generative AI: Augmenting Datasets Jan, 11 2026
Synthetic Data Generation with Multimodal Generative AI: Augmenting Datasets
Encoder-Decoder vs Decoder-Only Transformers: What You Need to Know About Large Language Models Mar, 10 2026
Encoder-Decoder vs Decoder-Only Transformers: What You Need to Know About Large Language Models
When to Transition from Vibe-Coded MVPs to Production Engineering Oct, 15 2025
When to Transition from Vibe-Coded MVPs to Production Engineering
Debugging Large Language Models: Diagnosing Errors and Hallucinations Mar, 6 2026
Debugging Large Language Models: Diagnosing Errors and Hallucinations

Menu

  • About
  • Terms of Service
  • Privacy Policy
  • CCPA
  • Contact

© 2026. All rights reserved.