mirror of
https://github.com/onestardao/WFGY.git
synced 2026-04-28 03:29:51 +00:00
44 lines
1.4 KiB
Docker
44 lines
1.4 KiB
Docker
|
|
#######################################################################
|
|
# Dockerfile — WFGY SDK 1-click inference container (CPU by default) #
|
|
# • Build on Hugging-Face Spaces, Colab, or local Docker Desktop. #
|
|
# • Size ≈ 250 MB compressed (Python 3.10-slim + torch-cpu). #
|
|
# • Optional GPU: uncomment the four CUDA lines below. #
|
|
#######################################################################
|
|
|
|
# ---------- CPU base (unchanged) ----------
|
|
FROM python:3.10-slim
|
|
|
|
# ---------- OPTIONAL GPU base (comment out to stay CPU-only) ----------
|
|
# FROM nvidia/cuda:12.4.0-runtime-ubuntu22.04
|
|
|
|
WORKDIR /app
|
|
COPY . /app
|
|
|
|
# ----------- deps ------------------------------------------------------
|
|
# torch==2.1.2+cpu → 60 MB wheel
|
|
# if you built with CUDA base, replace with cuXX wheel
|
|
RUN pip install --no-cache-dir -e . \
|
|
gradio==4.28.1 \
|
|
transformers==4.38.2 \
|
|
torch==2.1.2+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html
|
|
|
|
EXPOSE 7860
|
|
CMD ["python", "gradio_app.py"]
|
|
```
|
|
|
|
### How to build & run locally
|
|
|
|
```bash
|
|
docker build -t wfgy-demo .
|
|
docker run -p 7860:7860 wfgy-demo
|
|
open http://localhost:7860
|
|
```
|
|
|
|
The image is still **slim, fast, and Space-compatible**.
|
|
If a user wants CUDA, they only need to switch the `FROM` line and the
|
|
wheel URL—no other changes.
|
|
|
|
---
|
|
|
|
|