====== RagFlow ====== notes: add to /etc/sysctl.conf: vm.max_map_count=262144 sudo sysctl -w vm.max_map_count=262144 git clone https://github.com/infiniflow/ragflow.git cd ragflow/docker chmod +x ./entrypoint.sh adjust docker-compose.yml, add deploy: resources: reservations: devices: - driver: nvidia device_ids: ['0'] capabilities: [gpu] for change ports docker compose up -d Docker ENV Variables...: user_default_llm Newly signed-up users use LLM configured by this part. Otherwise, user need to configure his own LLM in setting. factory The LLM suppliers. "OpenAI", "Tongyi-Qianwen", "ZHIPU-AI", "Moonshot", "DeepSeek", "Baichuan", and "VolcEngine" are supported. api_key The corresponding API key of your assigned LLM vendor. ollama docker compose: docker compose -f docker-ollama.yml up -d run models (more on https://ollama.com/library ): preferred ones: mistral-openorca , llama3-chatqa? llama3, qwen2, phi3, mistral, mixtral, codegemma, codellama, dolphin-mixtral, deepseek-coder, mistral-openorca , codestral, codeqwen, orca2, llama3-gradient, llama3-chatqa (nvidia for RAG) and others... download/update model: docker exec -it ollama ollama pull codestral list available models: docker exec -it ollama ollama list run model locally: docker exec -it ollama ollama run llama3