Skip to content

Commit 0cb0f28

Browse files
committed
fix: increase model output tokens
1 parent 9ab2e88 commit 0cb0f28

1 file changed

Lines changed: 2 additions & 2 deletions

File tree

autoresearch/algo.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@
3131
MAX_ITERATIONS = 200
3232
TRAINING_TIMEOUT = 600 # 10 minutes
3333
MAX_MODEL_LEN = 65536 # larger context — dedicated GPU has plenty of room
34-
MAX_OUTPUT_TOKENS = 4096 # train.py is ~2K tokens; 4K is plenty and keeps generation fast
34+
MAX_OUTPUT_TOKENS = 10000 # train.py is ~8K tokens; need enough room for the full file
3535
TEMPERATURE = 0.7
3636
STAGNATION_THRESHOLD = 3 # consecutive non-improvements before nudge
3737
MAX_HISTORY_IN_PROMPT = 20 # only show last N iterations in prompt
@@ -432,7 +432,7 @@ def main():
432432
max_model_len=MAX_MODEL_LEN,
433433
dtype="auto",
434434
trust_remote_code=True,
435-
enforce_eager=True, # avoid DeltaNet compilation issues with Qwen3.5
435+
enforce_eager=True, # required: CUDA graphs fail for Qwen3.5 DeltaNet on this vLLM version
436436
)
437437
sampling_params = SamplingParams(
438438
max_tokens=MAX_OUTPUT_TOKENS,

0 commit comments

Comments
 (0)