OpenDevin is a platform for autonomous software engineers, powered by AI and LLMs. OpenDevin agents collaborate with human developers to write code, fix bugs, and ship features.
https://github.com/OpenDevin/OpenDevin
https://docs.all-hands.dev/modules/usage/intro
https://docs.all-hands.dev/modules/usage/llms/localLLMs
Docker compose file for local setup with separate ollama docker instance. In the current OpenDevin version (0.7), the ollama model needs to be specifically entered into the compose file and/or the interface.
mkdir /opt/opendevin/base mkdir /opt/opendevin/logs
name: opendevin services: opendevin: image: ghcr.io/opendevin/opendevin:0.7 #image: ghcr.io/opendevin/opendevin:latest container_name: opendevin deploy: resources: reservations: devices: - driver: nvidia device_ids: ['0'] capabilities: [gpu] ports: - 3300:3000 volumes: - /opt/opendevin/base:/opt/workspace_base - /opt/opendevin/logs:/app/logs - /var/run/docker.sock:/var/run/docker.sock restart: unless-stopped environment: - OPENAI_API_KEY=anything - CUSTOM_LLM_PROVIDER=ollama #- LLM_MODEL=deepseek-coder-v2:latest - LLM_MODEL=codestral:latest - LLM_EMBEDDING_MODEL="local" - LLM_API_KEY=ollama - LLM_BASE_URL=http://host.docker.internal:11434 - LLM_OLLAMA_BASE_URL=http://host.docker.internal:11434 - TZ=Europe/London - SANDBOX_USER_ID=0 - WORKSPACE_MOUNT_PATH=/opt/opendevin/base/workspace extra_hosts: host.docker.internal: host-gateway