Christian Ullrich
June 2025
LLM parameters for a ChatGPT-like experience
GPT4All
Parameter |
Phi-4-14B |
Context Length |
4096 |
Max Length |
4096 |
Prompt Batch Size |
128 |
Temperature |
0,7 |
Top-P |
0,9 |
Top-K |
40 |
Min-P |
0 |
Repeat Penalty Tokens |
256 |
GPU Layers |
40 |
Repeat Penalty |
1,1 |
Text generation web UI
Kategorie |
Einstellung |
Phi-4-14B |
Qwen3-32B |
Curve shape |
temperature |
0,7 |
0,7 |
Curve cutoff |
min_p |
0 |
0 |
|
top_n_sigma |
1 |
1 |
|
top_p |
0,9 |
0,9 |
|
top_k |
40 |
40 |
|
typical_p |
1 |
1 |
|
xtc_threshold |
0 |
0 |
|
xtc_probability |
0 |
0 |
Repetition suppression |
dry_multiplier |
1 |
1 |
|
dry_allowed_length |
1 |
1 |
|
dry_base |
1 |
1 |
|
repetition_penalty |
1,1 |
1,2 |
|
frequency_penalty |
0 |
0 |
|
presence_penalty |
0 |
0 |
|
repetition_penalty_range |
256 |
256 |
Alternative sampling |
mirostat_mode |
0 |
0 |
|
mirostat_tau |
0 |
0 |
|
mirostat_eta |
0 |
0 |
Other options |
max_new_tokens |
512 |
1024 |
|
Maximum tokens/second |
0 |
0 |
|
Maximum UI updates/second |
5 |
5 |
Other settings 1 |
dynamic temperature |
❌ (deactivated) |
❌ (deactivated) |
|
temperature_last |
❌ (deactivated) |
❌ (deactivated) |
|
auto_max_new_tokens |
❌ (deactivated) |
❌ (deactivated) |
|
Ban the eos_token |
❌ (deactivated) |
❌ (deactivated) |
|
Add the bos_token to the beginning of prompts |
✅ (activated) |
✅ (activated) |
|
Enable thinking |
❌ (deactivated) |
❌ (deactivated) |
|
Activate text streaming |
✅ (activated) |
✅ (activated) |
Other settings 2 |
Truncate the prompt up to this length |
4096 |
8192 |
|
Seed (-1 for random) |
-1 |
-1 |
Homepage | Terms of service • Privacy policy • Legal notice