====== OpenDevin ======
OpenDevin is a platform for autonomous software engineers, powered by AI and LLMs. OpenDevin agents collaborate with human developers to write code, fix bugs, and ship features.
https://github.com/OpenDevin/OpenDevin
https://docs.all-hands.dev/modules/usage/intro
https://docs.all-hands.dev/modules/usage/llms/localLLMs
{{:ai:pasted:20240706-113044.png}}
Docker compose file for local setup with separate ollama docker instance. In the current OpenDevin version (0.7), the ollama model needs to be specifically entered into the compose file and/or the interface.
mkdir /opt/opendevin/base
mkdir /opt/opendevin/logs
name: opendevin
services:
opendevin:
image: ghcr.io/opendevin/opendevin:0.7
#image: ghcr.io/opendevin/opendevin:latest
container_name: opendevin
deploy:
resources:
reservations:
devices:
- driver: nvidia
device_ids: ['0']
capabilities: [gpu]
ports:
- 3300:3000
volumes:
- /opt/opendevin/base:/opt/workspace_base
- /opt/opendevin/logs:/app/logs
- /var/run/docker.sock:/var/run/docker.sock
restart: unless-stopped
environment:
- OPENAI_API_KEY=anything
- CUSTOM_LLM_PROVIDER=ollama
#- LLM_MODEL=deepseek-coder-v2:latest
- LLM_MODEL=codestral:latest
- LLM_EMBEDDING_MODEL="local"
- LLM_API_KEY=ollama
- LLM_BASE_URL=http://host.docker.internal:11434
- LLM_OLLAMA_BASE_URL=http://host.docker.internal:11434
- TZ=Europe/London
- SANDBOX_USER_ID=0
- WORKSPACE_MOUNT_PATH=/opt/opendevin/base/workspace
extra_hosts:
host.docker.internal: host-gateway