mirror of
https://github.com/n8n-io/self-hosted-ai-starter-kit.git
synced 2026-03-15 16:58:07 +01:00
39 lines
1.5 KiB
Markdown
39 lines
1.5 KiB
Markdown
# n8n Demo setup
|
|
|
|
This repo helps quickly bootstrap an n8n demo environment using docker-compose.
|
|
|
|
### Requirements
|
|
- [Docker compose](https://docs.docker.com/compose/)
|
|
- **Optionally** an Nvidia GPU for faster inference on Ollama
|
|
|
|
### Setup
|
|
- Clone this repo
|
|
- **Optionally** edit the credentials in the `.env` file
|
|
- Start the containers:
|
|
- If you have an Nvidia GPU, run `docker compose --profile gpu-nvidia up`
|
|
- Otherwise to run inference services on your CPU, run `docker compose --profile cpu up`
|
|
- Wait a couple of minutes for all the containers to become healthy
|
|
- Open http://localhost:5678 in your browser and fill in the details
|
|
- Open the included workflow: http://localhost:5678/workflow/srOnR8PAY3u4RSwb
|
|
- Wait until Ollama has downloaded the `llama3.1` model (you can check the
|
|
docker console)
|
|
|
|
### Included service endpoints
|
|
- [n8n](http://localhost:5678/)
|
|
- [Ollama](http://localhost:11434/)
|
|
- [Qdrant](http://localhost:6333/dashboard)
|
|
|
|
### Local files
|
|
|
|
When running the demo for the first time, Docker will create a folder `shared`
|
|
next to the `docker-compose.yml` file. You can add files to that, that will be
|
|
visible on the `/data/shared` folder inside the n8n container, and you can use
|
|
that, for example, with the Local File Trigger node.
|
|
|
|
### Updating
|
|
- Run `docker compose pull` to fetch all the latest images
|
|
- If you use Ollama, use either `docker compose --profile cpu pull` or
|
|
`docker compose --profile gpu pull` to pull the correct ollama images
|
|
- Run `docker compose create && docker compose up -d` to update and restart
|
|
all the containers
|