You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Fix Greptile issues on top of llama_params_fit integration
Four follow-ups to the llama_params_fit adoption from PR #444:
1. CPU-only build no longer silently drops user gpu_layers override.
Previously, the CPU-only preprocessor block ran *after* the user
override and unconditionally set n_gpu_layers=0, discarding the
config value with no log. Now the override is scoped to the
GPU-build branch; the CPU-only branch logs a warning when a non-zero
user value is being ignored because no GPU backend is available.
2. n_ctx_min is now read from config["n_ctx_min"] (the inline comment
"Configurable parameter" previously didn't match behavior).
3. When llama_params_fit returns FAILURE or ERROR and the user's
requested context_size is larger than 2048, we now log that it's
being capped, instead of silently clamping.
4. Restore trailing newline at EOF (was stripped by the PR).
0 commit comments