Christian Ullrich June 2025

LLM parameters for a ChatGPT-like experience


GPT4All

Parameter Phi-4-14B
Context Length 4096
Max Length 4096
Prompt Batch Size 128
Temperature 0,7
Top-P 0,9
Top-K 40
Min-P 0
Repeat Penalty Tokens 256
GPU Layers 40
Repeat Penalty 1,1

Text generation web UI

Kategorie Einstellung Phi-4-14B Qwen3-32B
Curve shape temperature 0,7 0,7
Curve cutoff min_p 0 0
top_n_sigma 1 1
top_p 0,9 0,9
top_k 40 40
typical_p 1 1
xtc_threshold 0 0
xtc_probability 0 0
Repetition suppression dry_multiplier 1 1
dry_allowed_length 1 1
dry_base 1 1
repetition_penalty 1,1 1,2
frequency_penalty 0 0
presence_penalty 0 0
repetition_penalty_range 256 256
Alternative sampling mirostat_mode 0 0
mirostat_tau 0 0
mirostat_eta 0 0
Other options max_new_tokens 512 1024
Maximum tokens/second 0 0
Maximum UI updates/second 5 5
Other settings 1 dynamic temperature ❌ (deactivated) ❌ (deactivated)
temperature_last ❌ (deactivated) ❌ (deactivated)
auto_max_new_tokens ❌ (deactivated) ❌ (deactivated)
Ban the eos_token ❌ (deactivated) ❌ (deactivated)
Add the bos_token to the beginning of prompts ✅ (activated) ✅ (activated)
Enable thinking ❌ (deactivated) ❌ (deactivated)
Activate text streaming ✅ (activated) ✅ (activated)
Other settings 2 Truncate the prompt up to this length 4096 8192
Seed (-1 for random) -1 -1

Homepage | Terms of service • Privacy policy • Legal notice