utkarsh-dixit commited on
Commit
526db9a
β€’
1 Parent(s): 8b378ca

feat: add dockerfile

Browse files
Files changed (1) hide show
  1. README.md +30 -10
README.md CHANGED
@@ -1,11 +1,31 @@
1
-
2
- title: WhisperFusion
3
- emoji: πŸŒ–
4
- colorFrom: pink
5
- colorTo: green
6
- sdk: docker
 
 
 
 
 
 
 
 
 
 
7
  pinned: false
8
- description: WhisperFusion is an AI chatbot that provides ultra-low latency conversations. It integrates Mistral, a Large Language Model (LLM), on top of the real-time speech-to-text pipeline. It utilizes OpenAI WhisperLive to convert spoken language into text in real-time and is optimized to run as TensorRT engines, ensuring high-performance and low-latency processing.
9
- installation: Install [TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM/blob/main/docs/source/installation.md) to build Whisper and Mistral TensorRT engines. Refer to the README and the [Dockerfile.multi](https://github.com/NVIDIA/TensorRT-LLM/blob/main/docker/Dockerfile.multi) to install the required packages in the base pytorch docker image.
10
- usage: Run the main.py script with the appropriate arguments to start the chatbot.
11
- source: This information is provided by Hugging Face.
 
 
 
 
 
 
 
 
 
 
 
1
+ title: "WhisperFusion"
2
+ emoji: "πŸŒ–"
3
+ colorFrom: "pink"
4
+ colorTo: "green"
5
+ sdk: "docker"
6
+ python_version: "3.10"
7
+ sdk_version: "latest"
8
+ suggested_hardware: "t4-small"
9
+ suggested_storage: "medium"
10
+ app_file: "examples/chatbot/html/main.py"
11
+ app_port: 7860
12
+ base_path: "/"
13
+ fullWidth: false
14
+ models: ["teknium/OpenHermes-2.5-Mistral-7B"]
15
+ datasets: []
16
+ tags: ["AI", "chatbot", "speech-to-text", "real-time", "TensorRT", "LLM"]
17
  pinned: false
18
+ hf_oauth: false
19
+ hf_oauth_scopes: []
20
+ disable_embedding: false
21
+ startup_duration_timeout: "30m"
22
+ custom_headers:
23
+ cross-origin-embedder-policy: "require-corp"
24
+ cross-origin-opener-policy: "same-origin"
25
+ cross-origin-resource-policy: "cross-origin"
26
+ preload_from_hub:
27
+ - "NVIDIA/TensorRT-LLM examples/whisper/whisper_small_en,examples/phi/phi_engine,examples/phi/phi-2"
28
+ description: "WhisperFusion is an AI chatbot that provides ultra-low latency conversations. It integrates Mistral, a Large Language Model (LLM), on top of the real-time speech-to-text pipeline. It utilizes OpenAI WhisperLive to convert spoken language into text in real-time and is optimized to run as TensorRT engines, ensuring high-performance and low-latency processing."
29
+ installation: "Install [TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM/blob/main/docs/source/installation.md) to build Whisper and Mistral TensorRT engines. Refer to the README and the [Dockerfile.multi](https://github.com/NVIDIA/TensorRT-LLM/blob/main/docker/Dockerfile.multi) to install the required packages in the base pytorch docker image."
30
+ usage: "Run the main.py script with the appropriate arguments to start the chatbot."
31
+ source: "This information is provided by Hugging Face."