Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -41,7 +41,7 @@ language_mapping = {
|
|
41 |
'Arabic': ('ar', 'ar-SA-HamedNeural'),
|
42 |
'Chinese (Simplified)': ('zh-CN', 'zh-CN-YunxiNeural'),
|
43 |
'Japanese': ('ja', 'ja-JP-KeitaNeural'),
|
44 |
-
'
|
45 |
'Hindi': ('hi', 'hi-IN-MadhurNeural'),
|
46 |
'Swedish': ('sv', 'sv-SE-MattiasNeural'),
|
47 |
'Danish': ('da', 'da-DK-JeppeNeural'),
|
@@ -60,7 +60,7 @@ def cleanup_files(*files):
|
|
60 |
os.remove(file)
|
61 |
print(f"Removed file: {file}")
|
62 |
|
63 |
-
@spaces.GPU(duration=
|
64 |
def transcribe_audio(file_path):
|
65 |
print(f"Starting transcription of file: {file_path}")
|
66 |
temp_audio = None
|
@@ -185,7 +185,7 @@ with gr.Blocks(theme=gr.themes.Soft()) as demo:
|
|
185 |
target_language = gr.Dropdown(
|
186 |
choices=list(language_mapping.keys()),
|
187 |
label="Target Language for Dubbing",
|
188 |
-
value="
|
189 |
)
|
190 |
use_wav2lip = gr.Checkbox(
|
191 |
label="Use Wav2Lip for lip sync",
|
|
|
41 |
'Arabic': ('ar', 'ar-SA-HamedNeural'),
|
42 |
'Chinese (Simplified)': ('zh-CN', 'zh-CN-YunxiNeural'),
|
43 |
'Japanese': ('ja', 'ja-JP-KeitaNeural'),
|
44 |
+
'Khmer': ('kh', 'kh-Kh-InJoonNeural'),
|
45 |
'Hindi': ('hi', 'hi-IN-MadhurNeural'),
|
46 |
'Swedish': ('sv', 'sv-SE-MattiasNeural'),
|
47 |
'Danish': ('da', 'da-DK-JeppeNeural'),
|
|
|
60 |
os.remove(file)
|
61 |
print(f"Removed file: {file}")
|
62 |
|
63 |
+
@spaces.GPU(duration=200)
|
64 |
def transcribe_audio(file_path):
|
65 |
print(f"Starting transcription of file: {file_path}")
|
66 |
temp_audio = None
|
|
|
185 |
target_language = gr.Dropdown(
|
186 |
choices=list(language_mapping.keys()),
|
187 |
label="Target Language for Dubbing",
|
188 |
+
value="English"
|
189 |
)
|
190 |
use_wav2lip = gr.Checkbox(
|
191 |
label="Use Wav2Lip for lip sync",
|