local configuration
This commit is contained in:
3
.gitignore
vendored
3
.gitignore
vendored
@@ -11,4 +11,5 @@ modules/__pycache__/
|
||||
models/
|
||||
modules/yt_tmp.wav
|
||||
configs/default_parameters.yaml
|
||||
__pycache__/
|
||||
__pycache__/*~
|
||||
*~
|
||||
|
||||
@@ -1,17 +1,26 @@
|
||||
services:
|
||||
app:
|
||||
whisper:
|
||||
container_name: whisper
|
||||
networks:
|
||||
- www
|
||||
build: .
|
||||
image: jhj0517/whisper-webui:latest
|
||||
|
||||
# labels:
|
||||
# caddy: whisper.alfter.us
|
||||
## caddy.basic_auth.salfter: $2a$14$IhrjNNbvLmLJRM0BSiV4mOikVZ68rsL6u9Uz2qxUdI6tFRo6f.K/C
|
||||
# caddy.reverse_proxy: whisper.www:7860
|
||||
|
||||
|
||||
volumes:
|
||||
# Update paths to mount models and output paths to your custom paths like this, e.g:
|
||||
# - C:/whisper-models/custom-path:/Whisper-WebUI/models
|
||||
# - C:/whisper-webui-outputs/custom-path:/Whisper-WebUI/outputs
|
||||
- /Whisper-WebUI/models
|
||||
- /Whisper-WebUI/outputs
|
||||
- "whisper-models:/Whisper-WebUI/models"
|
||||
- "/mnt/storage/documents/whisper:/Whisper-WebUI/outputs"
|
||||
|
||||
ports:
|
||||
- "7860:7860"
|
||||
# ports:
|
||||
# - "7860:7860"
|
||||
|
||||
stdin_open: true
|
||||
tty: true
|
||||
@@ -20,10 +29,20 @@ services:
|
||||
|
||||
# If you're not using nvidia GPU, Update device to match yours.
|
||||
# See more info at : https://docs.docker.com/compose/compose-file/deploy/#driver
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: all
|
||||
capabilities: [ gpu ]
|
||||
# deploy:
|
||||
# resources:
|
||||
# reservations:
|
||||
# devices:
|
||||
# - driver: nvidia
|
||||
# count: all
|
||||
# capabilities: [ gpu ]
|
||||
|
||||
networks:
|
||||
www:
|
||||
name: www
|
||||
external: true
|
||||
|
||||
volumes:
|
||||
whisper-models:
|
||||
name: whisper-models
|
||||
external: true
|
||||
|
||||
1
nginx-conf
Symbolic link
1
nginx-conf
Symbolic link
@@ -0,0 +1 @@
|
||||
../nginx/conf.d/whisper.conf
|
||||
@@ -2,7 +2,7 @@
|
||||
# If you're using it, update url to your CUDA version (CUDA 12.1 is minimum requirement):
|
||||
# For CUDA 12.1, use : https://download.pytorch.org/whl/cu121
|
||||
# For CUDA 12.4, use : https://download.pytorch.org/whl/cu124
|
||||
--extra-index-url https://download.pytorch.org/whl/cu124
|
||||
#--extra-index-url https://download.pytorch.org/whl/cu124
|
||||
|
||||
|
||||
torch
|
||||
|
||||
Reference in New Issue
Block a user