Spaces:
Sleeping
Sleeping
Commit
·
87ae702
1
Parent(s):
c57cd9a
Added audio I/O
Browse files
app.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
import gradio as gr
|
| 2 |
|
| 3 |
-
from llm import end_interview, get_problem, send_request
|
| 4 |
from options import languages_list, models, topics_list
|
| 5 |
|
| 6 |
|
|
@@ -8,14 +8,43 @@ def hide_settings():
|
|
| 8 |
init_acc = gr.Accordion("Settings", open=False)
|
| 9 |
start_btn = gr.Button("Generate a problem", interactive=False)
|
| 10 |
solution_acc = gr.Accordion("Solution", open=True)
|
| 11 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
|
| 14 |
def hide_solution():
|
| 15 |
solution_acc = gr.Accordion("Solution", open=False)
|
| 16 |
end_btn = gr.Button("Finish the interview", interactive=False)
|
| 17 |
problem_acc = gr.Accordion("Problem statement", open=False)
|
| 18 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
|
| 21 |
with gr.Blocks() as demo:
|
|
@@ -57,12 +86,22 @@ with gr.Blocks() as demo:
|
|
| 57 |
language_select = gr.Dropdown(
|
| 58 |
label="Select language", choices=languages_list, value="python", container=False, interactive=True
|
| 59 |
)
|
| 60 |
-
code = gr.Code(label="Solution", language=language_select.value, lines=
|
| 61 |
-
message = gr.Textbox(label="Message", lines=1)
|
| 62 |
-
# TODO: add voice input and output
|
| 63 |
with gr.Column(scale=1):
|
|
|
|
| 64 |
chat = gr.Chatbot(label="Chat history")
|
| 65 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 66 |
with gr.Accordion("Feedback", open=True) as feedback_acc:
|
| 67 |
feedback = gr.Markdown()
|
| 68 |
|
|
@@ -71,16 +110,28 @@ with gr.Blocks() as demo:
|
|
| 71 |
inputs=[requirements, difficulty_select, topic_select, model_select],
|
| 72 |
outputs=[description, chat_history],
|
| 73 |
scroll_to_output=True,
|
| 74 |
-
).then(fn=hide_settings, inputs=None, outputs=[init_acc, start_btn, solution_acc])
|
| 75 |
|
| 76 |
-
|
| 77 |
fn=send_request,
|
| 78 |
inputs=[code, previous_code, message, chat_history, chat, model_select],
|
| 79 |
outputs=[chat_history, chat, message, previous_code],
|
| 80 |
)
|
| 81 |
|
| 82 |
end_btn.click(fn=end_interview, inputs=[chat_history, model_select], outputs=feedback).then(
|
| 83 |
-
fn=hide_solution, inputs=None, outputs=[solution_acc, end_btn, problem_acc]
|
| 84 |
)
|
| 85 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 86 |
demo.launch()
|
|
|
|
| 1 |
import gradio as gr
|
| 2 |
|
| 3 |
+
from llm import end_interview, get_problem, read_last_message, send_request, transcribe_audio
|
| 4 |
from options import languages_list, models, topics_list
|
| 5 |
|
| 6 |
|
|
|
|
| 8 |
init_acc = gr.Accordion("Settings", open=False)
|
| 9 |
start_btn = gr.Button("Generate a problem", interactive=False)
|
| 10 |
solution_acc = gr.Accordion("Solution", open=True)
|
| 11 |
+
end_btn = gr.Button("Finish the interview", interactive=True)
|
| 12 |
+
send_btn = gr.Button("Send", interactive=True)
|
| 13 |
+
audio_input = gr.Audio(
|
| 14 |
+
label="Record audio",
|
| 15 |
+
sources=["microphone"],
|
| 16 |
+
type="filepath",
|
| 17 |
+
waveform_options={"show_controls": False},
|
| 18 |
+
interactive=True,
|
| 19 |
+
editable=False,
|
| 20 |
+
)
|
| 21 |
+
chat = [
|
| 22 |
+
(
|
| 23 |
+
None,
|
| 24 |
+
"Welcome to the interview! Please take a moment to read the problem statement. Then you can share you initial thoughts and ask any questions you may have. Good luck!",
|
| 25 |
+
)
|
| 26 |
+
]
|
| 27 |
+
return init_acc, start_btn, solution_acc, end_btn, send_btn, audio_input, chat
|
| 28 |
|
| 29 |
|
| 30 |
def hide_solution():
|
| 31 |
solution_acc = gr.Accordion("Solution", open=False)
|
| 32 |
end_btn = gr.Button("Finish the interview", interactive=False)
|
| 33 |
problem_acc = gr.Accordion("Problem statement", open=False)
|
| 34 |
+
send_btn = gr.Button("Send", interactive=False)
|
| 35 |
+
audio_input = gr.Audio(
|
| 36 |
+
label="Record audio",
|
| 37 |
+
sources=["microphone"],
|
| 38 |
+
type="filepath",
|
| 39 |
+
waveform_options={"show_controls": False},
|
| 40 |
+
interactive=False,
|
| 41 |
+
editable=False,
|
| 42 |
+
)
|
| 43 |
+
return solution_acc, end_btn, problem_acc, send_btn, audio_input
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
def return_none():
|
| 47 |
+
return None
|
| 48 |
|
| 49 |
|
| 50 |
with gr.Blocks() as demo:
|
|
|
|
| 86 |
language_select = gr.Dropdown(
|
| 87 |
label="Select language", choices=languages_list, value="python", container=False, interactive=True
|
| 88 |
)
|
| 89 |
+
code = gr.Code(label="Solution", language=language_select.value, lines=35)
|
|
|
|
|
|
|
| 90 |
with gr.Column(scale=1):
|
| 91 |
+
end_btn = gr.Button("Finish the interview", interactive=False)
|
| 92 |
chat = gr.Chatbot(label="Chat history")
|
| 93 |
+
audio_input = gr.Audio(
|
| 94 |
+
label="Record audio",
|
| 95 |
+
sources=["microphone"],
|
| 96 |
+
type="filepath",
|
| 97 |
+
waveform_options={"show_controls": False},
|
| 98 |
+
interactive=False,
|
| 99 |
+
editable=False,
|
| 100 |
+
)
|
| 101 |
+
audio_output = gr.Audio(label="Play audio", autoplay=True, visible=False)
|
| 102 |
+
message = gr.Textbox(label="Message", lines=3)
|
| 103 |
+
send_btn = gr.Button("Send", interactive=False)
|
| 104 |
+
|
| 105 |
with gr.Accordion("Feedback", open=True) as feedback_acc:
|
| 106 |
feedback = gr.Markdown()
|
| 107 |
|
|
|
|
| 110 |
inputs=[requirements, difficulty_select, topic_select, model_select],
|
| 111 |
outputs=[description, chat_history],
|
| 112 |
scroll_to_output=True,
|
| 113 |
+
).then(fn=hide_settings, inputs=None, outputs=[init_acc, start_btn, solution_acc, end_btn, send_btn, audio_input, chat])
|
| 114 |
|
| 115 |
+
send_btn.click(
|
| 116 |
fn=send_request,
|
| 117 |
inputs=[code, previous_code, message, chat_history, chat, model_select],
|
| 118 |
outputs=[chat_history, chat, message, previous_code],
|
| 119 |
)
|
| 120 |
|
| 121 |
end_btn.click(fn=end_interview, inputs=[chat_history, model_select], outputs=feedback).then(
|
| 122 |
+
fn=hide_solution, inputs=None, outputs=[solution_acc, end_btn, problem_acc, send_btn, audio_input]
|
| 123 |
)
|
| 124 |
|
| 125 |
+
audio_input.stop_recording(fn=transcribe_audio, inputs=[audio_input], outputs=[message]).then(
|
| 126 |
+
fn=return_none, inputs=None, outputs=[audio_input]
|
| 127 |
+
).then(
|
| 128 |
+
fn=send_request,
|
| 129 |
+
inputs=[code, previous_code, message, chat_history, chat, model_select],
|
| 130 |
+
outputs=[chat_history, chat, message, previous_code],
|
| 131 |
+
)
|
| 132 |
+
|
| 133 |
+
chat.change(fn=read_last_message, inputs=[chat], outputs=[audio_output])
|
| 134 |
+
|
| 135 |
+
audio_output.stop(fn=return_none, inputs=None, outputs=[audio_output])
|
| 136 |
+
|
| 137 |
demo.launch()
|
llm.py
CHANGED
|
@@ -68,7 +68,25 @@ def send_request(code, previous_code, message, chat_history, chat_display, model
|
|
| 68 |
reply = "There was an error processing your request."
|
| 69 |
|
| 70 |
chat_history.append({"role": "assistant", "content": json_reply})
|
| 71 |
-
|
| 72 |
chat_display.append([message, str(reply)])
|
| 73 |
|
| 74 |
return chat_history, chat_display, "", code
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 68 |
reply = "There was an error processing your request."
|
| 69 |
|
| 70 |
chat_history.append({"role": "assistant", "content": json_reply})
|
|
|
|
| 71 |
chat_display.append([message, str(reply)])
|
| 72 |
|
| 73 |
return chat_history, chat_display, "", code
|
| 74 |
+
|
| 75 |
+
|
| 76 |
+
def transcribe_audio(filename, client=client):
|
| 77 |
+
with open(filename, "rb") as audio_file:
|
| 78 |
+
transcription = client.audio.transcriptions.create(model="whisper-1", file=audio_file, response_format="text")
|
| 79 |
+
|
| 80 |
+
return transcription
|
| 81 |
+
|
| 82 |
+
|
| 83 |
+
def text_to_speech(text, client=client):
|
| 84 |
+
response = client.audio.speech.create(model="tts-1", voice="alloy", input=text)
|
| 85 |
+
return response.content
|
| 86 |
+
|
| 87 |
+
|
| 88 |
+
def read_last_message(chat_display):
|
| 89 |
+
last_message = chat_display[-1][1]
|
| 90 |
+
|
| 91 |
+
audio = text_to_speech(last_message)
|
| 92 |
+
return audio
|