-
Notifications
You must be signed in to change notification settings - Fork 508
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Speaker Identification #672
base: main
Are you sure you want to change the base?
Conversation
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
@EzraEllette did you test on windows btw? would be good that someone test on windows |
.join("pyannote") | ||
.join("segmentation-3.0.onnx"); | ||
|
||
let embedding_extractor = EmbeddingExtractor::new( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
any downside to load / unload the model at every chunk?
|
|
unrelated but fun:
whisper detect my voice in french (i spoke english) |
there are a few Thank you (something with VAD i suppose) but maybe not more than on main |
Okay there are some bug fixes to make. |
description
This PR adds speaker identification to screenpipe. Audio is segmented by speaker then transcribed. transcriptions now have a
speaker_id
column. new tablespeakers
was added withname
andmetadata
columns.speaker_embeddings
table was created with a one-to-many relationship for speaker and embeddings.related issue: #
type of change
how to test
Run the speaker_identification test. run
screenpipe-server/src/db.rs
tests.Use screenpipe.