import openai import os import azure.cognitiveservices.speech as speechsdk LANGUAGE = "nl" AUDIO_FILE_NAME = "audio_response.wav" openai.api_key = os.environ['OPEN_AI_KEY'] speech_config = speechsdk.SpeechConfig(subscription=os.environ['AZURE_SPEECH_KEY'], region="westeurope") speech_config.speech_synthesis_voice_name = "nl-NL-ColetteNeural" speech_synthesizer = speechsdk.SpeechSynthesizer(speech_config=speech_config, audio_config=None) def transcribe(model: str, audio: str): audio_file = open(audio, "rb") transcript = openai.Audio.transcribe(model, audio_file, language=LANGUAGE) return transcript def gen_response(model: str, context: list): response = openai.ChatCompletion.create(model=model, messages=context) return response["choices"][0]["message"] def gen_voice(response, response_filename): reponse_audio = speech_synthesizer.speak_text_async(response['content']).get() stream = speechsdk.AudioDataStream(reponse_audio) stream.save_to_wav_file(response_filename) def respond(audio:str, context: list): transcript = transcribe("whisper-1", audio) context.append({"role": "user", "content": transcript['text']}) response = gen_response("gpt-3.5-turbo", context) context.append(response) gen_voice(response, AUDIO_FILE_NAME) return AUDIO_FILE_NAME def transcript(context: list): transcript = "" for m in context: if m["role"] != "system": transcript += m["role"] + " : " + m["content"] + "\n\n"