minor refactoring

This commit is contained in:
SevaSk 2023-05-13 10:46:05 -04:00
parent 654464c8d9
commit 86b97d0039
2 changed files with 8 additions and 8 deletions

View File

@ -16,24 +16,24 @@ PHRASE_TIMEOUT = 3.01
MAX_PHRASES = 10 MAX_PHRASES = 10
class AudioTranscriber: class AudioTranscriber:
def __init__(self, default_mic : DefaultMicRecorder, default_speaker : DefaultSpeakerRecorder): def __init__(self, mic_source, speaker_source):
self.transcript_data = {"You": [], "Speaker": []} self.transcript_data = {"You": [], "Speaker": []}
self.transcript_changed_event = threading.Event() self.transcript_changed_event = threading.Event()
self.audio_model = whisper.load_model(os.path.join(os.getcwd(), 'tiny.en.pt')) self.audio_model = whisper.load_model(os.path.join(os.getcwd(), 'tiny.en.pt'))
self.audio_sources = { self.audio_sources = {
"You": { "You": {
"sample_rate": default_mic.source.SAMPLE_RATE, "sample_rate": mic_source.SAMPLE_RATE,
"sample_width": default_mic.source.SAMPLE_WIDTH, "sample_width": mic_source.SAMPLE_WIDTH,
"channels": default_mic.source.channels, "channels": mic_source.channels,
"last_sample": bytes(), "last_sample": bytes(),
"last_spoken": None, "last_spoken": None,
"new_phrase": True, "new_phrase": True,
"process_data_func": self.process_mic_data "process_data_func": self.process_mic_data
}, },
"Speaker": { "Speaker": {
"sample_rate": default_speaker.source.SAMPLE_RATE, "sample_rate": speaker_source.SAMPLE_RATE,
"sample_width": default_speaker.source.SAMPLE_WIDTH, "sample_width": speaker_source.SAMPLE_WIDTH,
"channels": default_speaker.source.channels, "channels": speaker_source.channels,
"last_sample": bytes(), "last_sample": bytes(),
"last_spoken": None, "last_spoken": None,
"new_phrase": True, "new_phrase": True,

View File

@ -72,7 +72,7 @@ def main():
speaker_audio_recorder = AudioRecorder.DefaultSpeakerRecorder() speaker_audio_recorder = AudioRecorder.DefaultSpeakerRecorder()
speaker_audio_recorder.record_into_queue(audio_queue) speaker_audio_recorder.record_into_queue(audio_queue)
global_transcriber = AudioTranscriber(user_audio_recorder, speaker_audio_recorder) global_transcriber = AudioTranscriber(user_audio_recorder.source, speaker_audio_recorder.source)
transcribe = threading.Thread(target=global_transcriber.transcribe_audio_queue, args=(audio_queue,)) transcribe = threading.Thread(target=global_transcriber.transcribe_audio_queue, args=(audio_queue,))
transcribe.start() transcribe.start()