--- title: Seamless Streaming emoji: 📞 colorFrom: blue colorTo: yellow sdk: docker pinned: false suggested_hardware: t4-medium --- # Seamless Streaming demo ## Running on HF spaces You can simply duplicate the space to run it. ## Running locally ### Install backend seamless_server dependencies Please note: we *strongly* recommend running the model on GPU. If running for the first time, create conda environment and install the desired torch version. The example below is for PyTorch 2.1.1 and variant cu118. Check [here](https://pytorch.org/get-started/locally/) to find the command for your variant. Then install the rest of the requirements: ``` cd seamless_server conda create --name smlss_server python=3.8 conda activate smlss_server conda install pytorch torchvision torchaudio pytorch-cuda=11.8 -c pytorch -c nvidia pip install -r requirements.txt conda install -c conda-forge libsndfile==1.0.31 ``` ### Install frontend streaming-react-app dependencies ``` conda install -c conda-forge nodejs cd streaming-react-app npm install npm run build # this will create the dist/ folder ``` ### Running the server The server can be run locally with uvicorn below. Run the server in dev mode: ``` cd seamless_server uvicorn app_pubsub:app --reload --host localhost ``` Run the server in prod mode: ``` cd seamless_server uvicorn app_pubsub:app --host 0.0.0.0 ``` To enable additional logging from uvicorn pass `--log-level debug` or `--log-level trace`. ### Debuging If you enable "Server Debug Flag" when starting streaming from the client, this enables extensive debug logging and it saves audio files in /debug folder.