We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 09aa5c8 commit 0e9160fCopy full SHA for 0e9160f
1 file changed
autoresearch/algo.py
@@ -22,7 +22,7 @@
22
GPU_MEMORY_UTILIZATION = 0.25 # ~35GB for LLM weights+KV cache, rest for training
23
MAX_ITERATIONS = 200
24
TRAINING_TIMEOUT = 600 # 10 minutes
25
-MAX_MODEL_LEN = 40960
+MAX_MODEL_LEN = 32000
26
MAX_OUTPUT_TOKENS = 16384 # max tokens for LLM output (enough for full train.py)
27
TEMPERATURE = 0.7
28
STAGNATION_THRESHOLD = 3 # consecutive non-improvements before nudge
@@ -419,6 +419,7 @@ def main():
419
max_model_len=MAX_MODEL_LEN,
420
dtype="auto",
421
trust_remote_code=True,
422
+ enforce_eager=True, # required: torch.compile incompatibility in this vLLM version
423
)
424
sampling_params = SamplingParams(
425
max_tokens=MAX_OUTPUT_TOKENS,
0 commit comments