---
name: llm-tuning-patterns
description: LLM Tuning Patterns
user-invocable: false
---
# LLM Tuning Patterns
Evidence-based patterns for configuring LLM parameters, based on APOLLO and Godel-Prover research.
## Pattern
Different tasks require different LLM configurations. Use these evidence-based settings.
## Theorem Proving / Formal Reasoning
Based on APOLLO parity analysis:
| Parameter | Value | Rationale |
|-----------|-------|-----------|
| max_tokens | 4096 | Proofs need space for chain-of-thought |
| temperature | 0.6 | Higher creativity for tactic exploration |
| top_p | 0.95 | Allow diverse proof paths |
### Proof Plan Prompt
Always request a proof plan before tactics:
```
Given the theorem to prove:
[theorem statement]
First, write a high-level proof plan explaining your approach.
Then, suggest Lean 4 tactics to implement each step.
```
The proof plan (chain-of-thought) significantly improves tactic quality.
### Parallel Sampling
For hard proofs, use parallel sampling:
- Generate N=8-32 candidate proof attempts
- Use best-of-N selection
- Each sample at temperature 0.6-0.8
## Code Generation
| Parameter | Value | Rationale |
|-----------|-------|-----------|
| max_tokens | 2048 | Sufficient for most functions |
| temperature | 0.2-0.4 | Prefer deterministic output |
## Creative / Exploration Tasks
| Parameter | Value | Rationale |
|-----------|-------|-----------|
| max_tokens | 4096 | Space for exploration |
| temperature | 0.8-1.0 | Maximum creativity |
## Anti-Patterns
- **Too low tokens for proofs**: 512 tokens truncates chain-of-thought
- **Too low temperature for proofs**: 0.2 misses creative tactic paths
- **No proof plan**: Jumping to tactics without planning reduces success rate
## Source Sessions
- This session: APOLLO parity - increased max_tokens 512->4096, temp 0.2->0.6
- This session: Added proof plan prompt for chain-of-thought before tactics