User Tools

Site Tools


ai:ragflow

RagFlow

notes:

add to /etc/sysctl.conf:

vm.max_map_count=262144
sudo sysctl -w vm.max_map_count=262144
git clone https://github.com/infiniflow/ragflow.git

cd ragflow/docker
chmod +x ./entrypoint.sh

adjust docker-compose.yml, add

    deploy:
      resources:
        reservations:
          devices:
          - driver: nvidia
            device_ids: ['0']
            capabilities: [gpu]

for
change ports

docker compose up -d

Docker ENV Variables…:
user_default_llm

Newly signed-up users use LLM configured by this part. Otherwise, user need to configure his own LLM in setting.
factory

The LLM suppliers. “OpenAI”, “Tongyi-Qianwen”, “ZHIPU-AI”, “Moonshot”, “DeepSeek”, “Baichuan”, and “VolcEngine” are supported.
api_key
The corresponding API key of your assigned LLM vendor.

ollama docker compose:

docker compose -f docker-ollama.yml up -d

run models (more on https://ollama.com/library ):
preferred ones: mistral-openorca , llama3-chatqa?

llama3, qwen2, phi3, mistral, mixtral, codegemma, codellama, dolphin-mixtral, deepseek-coder, mistral-openorca , codestral, codeqwen, orca2, llama3-gradient,  llama3-chatqa (nvidia for RAG) and others...

download/update model:

docker exec -it ollama ollama pull codestral

list available models:

docker exec -it ollama ollama list

run model locally:

docker exec -it ollama ollama run llama3
ai/ragflow.txt · Last modified: 2024/07/11 17:51 by Wulf Rajek