Usage FAQ
What models are supported?
Model support
Note that SWE-agent is currently unlikely to perform well with small or local models.
Models are configured in models.py
(we're working on giving a complete list of model settings).
Here are some few examples:
gpt4
gpt4o
gpt4-turbo
claude-2
claude-opus
claude-sonnet
claude-haiku
claude-sonnet-3.5
deepseek-coder
azure:gpt4
Ollama support
Models served with an ollama server can be used by specifying --model
with ollama:model_name
and --host_url
to point to the url used to serve ollama (http://localhost:11434
by default). See more details about using ollama here.
python run.py --model_name ollama:deepseek-coder:6.7b-instruct \
--host_url http://localhost:11434 \
--data_path https://github.com/pvlib/pvlib-python/issues/1603 \
--config_file config/default_from_url.yaml
Models for testing
We also provide models for testing SWE-agent without spending any credits
HumanModel
andHumandThoughtModel
will prompt for input from the user that stands in for the output of the LM. This can be used to create new demonstrations.ReplayModel
takes a trajectory as input and "replays it"InstantEmptySubmitTestModel
will create an emptyreproduce.py
and then submit
Debugging
- If you get
Error code: 404
, please check your configured keys, in particular whether you setOPENAI_API_BASE_URL
correctly (if you're not using it, the line should be deleted or commented out). Also see this issue for reference.
-
Something broken? Report bug
-
Something unclear? Ask question