# fly.toml app configuration file generated for medical-transcription-cpu on 2024-01-22T15:15:29-06:00 | |
# | |
# See https://fly.io/docs/reference/configuration/ for information about how to use this file. | |
# | |
app = "medical-transcription-cpu" | |
primary_region = "ord" | |
kill_signal = "SIGTERM" | |
# TODO: Enable the following line and install the necessary CUDA packages in the Dockerfile for GPU usage. | |
# The volumes will need to be recreated as well to match the GPU kind. | |
# See https://fly.io/docs/gpus/gpu-quickstart/ | |
# and https://fly.io/docs/gpus/getting-started-gpus/ | |
# vm.size = "a100-40gb" | |
[build] | |
[deploy] | |
release_command = "/app/bin/migrate" | |
[[mounts]] | |
source = "data" | |
destination = "/data" | |
initial_size = "40gb" | |
[env] | |
ELIXIR_ERL_OPTIONS = "-proto_dist inet6_tcp +sssdio 128" | |
AUDIO_TAGGER_CACHE_DIR="/data/cache/audio_tagger" | |
BUMBLEBEE_CACHE_DIR="/data/cache/bumblebee" | |
XLA_CACHE_DIR="/data/cache/xla" | |
# XLA_TARGET="cuda120" | |
PHX_HOST = "medical-transcription-cpu.fly.dev" | |
PORT = "8080" | |
[http_service] | |
internal_port = 8080 | |
force_https = true | |
auto_stop_machines = false | |
auto_start_machines = false | |
min_machines_running = 0 | |
processes = ["app"] | |
[http_service.concurrency] | |
type = "connections" | |
hard_limit = 1000 | |
soft_limit = 1000 | |
[[vm]] | |
cpu_kind = "shared" | |
cpus = 1 | |
memory_mb = 2048 | |
# To generate the vector embeddings for the ICD-9 codelist, we need more RAM available. | |
# fly scale vm shared-cpu-8x --memory 16384 | |