mirror of
https://github.com/Skyvern-AI/skyvern.git
synced 2026-04-26 10:41:14 +00:00
Include .env in Docker build for easy llm setup (#2484)
Co-authored-by: Shuchang Zheng <shu@skyvern.com>
This commit is contained in:
parent
ed9dda6d8f
commit
829e46c726
4 changed files with 11 additions and 6 deletions
|
|
@ -10,6 +10,7 @@
|
||||||
**/.venv
|
**/.venv
|
||||||
**/.vscode
|
**/.vscode
|
||||||
*.env*
|
*.env*
|
||||||
|
!/.env
|
||||||
# Streamlit ignores
|
# Streamlit ignores
|
||||||
**/secrets*.toml
|
**/secrets*.toml
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,4 +1,5 @@
|
||||||
FROM python:3.11 AS requirements-stage
|
FROM python:3.11 as requirements-stage
|
||||||
|
# Run `skyvern init llm` before building to generate the .env file
|
||||||
|
|
||||||
WORKDIR /tmp
|
WORKDIR /tmp
|
||||||
RUN pip install poetry
|
RUN pip install poetry
|
||||||
|
|
|
||||||
|
|
@ -193,6 +193,7 @@ skyvern status
|
||||||
1. Make sure you have [Docker Desktop](https://www.docker.com/products/docker-desktop/) installed and running on your machine
|
1. Make sure you have [Docker Desktop](https://www.docker.com/products/docker-desktop/) installed and running on your machine
|
||||||
1. Make sure you don't have postgres running locally (Run `docker ps` to check)
|
1. Make sure you don't have postgres running locally (Run `docker ps` to check)
|
||||||
1. Clone the repository and navigate to the root directory
|
1. Clone the repository and navigate to the root directory
|
||||||
|
1. Run `skyvern init llm` to generate a `.env` file. This will be copied into the Docker image.
|
||||||
1. Fill in the LLM provider key on the [docker-compose.yml](./docker-compose.yml). *If you want to run Skyvern on a remote server, make sure you set the correct server ip for the UI container in [docker-compose.yml](./docker-compose.yml).*
|
1. Fill in the LLM provider key on the [docker-compose.yml](./docker-compose.yml). *If you want to run Skyvern on a remote server, make sure you set the correct server ip for the UI container in [docker-compose.yml](./docker-compose.yml).*
|
||||||
2. Run the following command via the commandline:
|
2. Run the following command via the commandline:
|
||||||
```bash
|
```bash
|
||||||
|
|
|
||||||
|
|
@ -21,6 +21,8 @@ services:
|
||||||
skyvern:
|
skyvern:
|
||||||
image: public.ecr.aws/skyvern/skyvern:latest
|
image: public.ecr.aws/skyvern/skyvern:latest
|
||||||
restart: on-failure
|
restart: on-failure
|
||||||
|
env_file:
|
||||||
|
- .env
|
||||||
# comment out if you want to externally call skyvern API
|
# comment out if you want to externally call skyvern API
|
||||||
ports:
|
ports:
|
||||||
- 8000:8000
|
- 8000:8000
|
||||||
|
|
@ -45,7 +47,7 @@ services:
|
||||||
# - BROWSER_REMOTE_DEBUGGING_URL=http://host.docker.internal:9222/
|
# - BROWSER_REMOTE_DEBUGGING_URL=http://host.docker.internal:9222/
|
||||||
|
|
||||||
# =========================
|
# =========================
|
||||||
# LLM Settings
|
# LLM Settings - Recommended to use skyvern CLI, `skyvern init llm` to setup your LLM's
|
||||||
# =========================
|
# =========================
|
||||||
# OpenAI Support:
|
# OpenAI Support:
|
||||||
# If you want to use OpenAI as your LLM provider, uncomment the following lines and fill in your OpenAI API key.
|
# If you want to use OpenAI as your LLM provider, uncomment the following lines and fill in your OpenAI API key.
|
||||||
|
|
@ -54,10 +56,10 @@ services:
|
||||||
# - OPENAI_API_KEY=<your_openai_key>
|
# - OPENAI_API_KEY=<your_openai_key>
|
||||||
# Gemini Support:
|
# Gemini Support:
|
||||||
# Gemini is a new LLM provider that is currently in beta. You can use it by uncommenting the following lines and filling in your Gemini API key.
|
# Gemini is a new LLM provider that is currently in beta. You can use it by uncommenting the following lines and filling in your Gemini API key.
|
||||||
- LLM_KEY=GEMINI
|
# - LLM_KEY=GEMINI
|
||||||
- ENABLE_GEMINI=true
|
# - ENABLE_GEMINI=true
|
||||||
- GEMINI_API_KEY=YOUR_GEMINI_KEY
|
# - GEMINI_API_KEY=YOUR_GEMINI_KEY
|
||||||
- LLM_KEY=GEMINI_2.5_PRO_PREVIEW_03_25
|
# - LLM_KEY=GEMINI_2.5_PRO_PREVIEW_03_25
|
||||||
# If you want to use other LLM provider, like azure and anthropic:
|
# If you want to use other LLM provider, like azure and anthropic:
|
||||||
# - ENABLE_ANTHROPIC=true
|
# - ENABLE_ANTHROPIC=true
|
||||||
# - LLM_KEY=ANTHROPIC_CLAUDE3.5_SONNET
|
# - LLM_KEY=ANTHROPIC_CLAUDE3.5_SONNET
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue