Understanding LLM Temperature and Other Parameters
New · Open Source
Kerb - LLM Development Toolkit
Python toolkit for building production-ready LLM applications. Modular utilities for prompts, RAG, agents, structured outputs, and multi-provider support.
Was this section helpful?
The Curious Case of Neural Text Degeneration, Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, Yejin Choi, 2020International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1904.09751 - Introduces nucleus sampling (top-p) and analyzes its effectiveness compared to other decoding strategies for neural text generation, including temperature and top-k.
API Reference: Chat Completions, OpenAI, 2024 (OpenAI) - Official documentation detailing the generation parameters available in OpenAI's API, including temperature, top-p, max_tokens, and various penalties.
Speech and Language Processing (3rd ed. draft), Daniel Jurafsky and James H. Martin, 2025 - A standard reference for natural language processing, providing comprehensive background on language models, text generation, and the mechanics of token probability and sampling.
Generation Strategies for Language Models, Hugging Face, 2024 - Practical guide to various text generation strategies implemented in the Transformers library, offering clear explanations and examples for parameters like temperature, top-k, and top-p.