N-Gram House

Tag: top-k

Controlling Length and Structure in LLM Outputs: Practical Decoding Parameters

Controlling Length and Structure in LLM Outputs: Practical Decoding Parameters

Learn how to control LLM output length and structure using decoding parameters like temperature, top-k, top-p, and repetition penalties. Practical settings for real-world use cases.

Categories

  • History (50)
  • Machine Learning (9)

Recent Posts

How Layer Dropping and Early Exit Make Large Language Models Faster Feb, 4 2026
How Layer Dropping and Early Exit Make Large Language Models Faster
Latency Management for RAG Pipelines in Production LLM Systems Dec, 19 2025
Latency Management for RAG Pipelines in Production LLM Systems
Toolformer-Style Self-Supervision: How LLMs Learn to Use Tools on Their Own Nov, 17 2025
Toolformer-Style Self-Supervision: How LLMs Learn to Use Tools on Their Own
Autonomous Agents in Generative AI for Business Processes: From Plans to Actions Jun, 25 2025
Autonomous Agents in Generative AI for Business Processes: From Plans to Actions
Hybrid Search for RAG: Boost LLM Accuracy with Semantic and Keyword Retrieval Dec, 7 2025
Hybrid Search for RAG: Boost LLM Accuracy with Semantic and Keyword Retrieval

Menu

  • About
  • Terms of Service
  • Privacy Policy
  • CCPA
  • Contact

© 2026. All rights reserved.