-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: add LLM-based diarization optimization postprocessing #96
Comments
The more I look at Input: i.e. there is no change at all. [EDIT] I had better results when running the q8 ( Output (q8): However, I am able to get similar results with llama 3.2, qwen 2.5, granite or mistral, with a proper prompt, ex.
So, I'm wondering if we could not simply integrate langchain (ex. the Alternatively, we could use the https://huggingface.co/google/DiarizationLM-8b-Fisher-v2/blob/main/README.md Was there a specific reason you wanted to use the DiarizationLM project directly? |
See my comments in #95 about the code of
Great!
I have no experience with LangChain. What would that give us in this context? Let's externalise the model. Do you have any preference re: the client library? I recently found out about https://ai.pydantic.dev. It's still 'early beta' but I don't think the API for basic stuff is going to change much. And even if, it's going to be well documented.
No, it was just where I found out about the possibility of improving transcript diarization with LLMs. |
diarizationlm
and its pretrained models
https://ai.pydantic.dev/ is a good choice - I will personally be connecting it to ollama, but I think it's also possible to hook it to vLLM through the openai model. |
So, I played around with |
I have had much greater success with LLM transcription improvements using BAML especially with ollama and smaller models. For example, I used lm-structured-output-benchmark and added in @linozen I didn't see the pydantic-ai code in the repo/PRs, if you have that I could prob switch it to BAML easily for testing or I can share my benchmark test suite above for you to look at? |
Great, let's try it.
Here is the code in its current state. I have forgotten to put it on its proper branch. Now, it should be cleaned up. Feel free to play around with the code and potentially add BAML (Thanks for the tip!). Otherwise, I will take a crack sometime next week. |
Ah and @mrmichaeladavis, I’d live to see your benchmark code. |
This aims to add a script to improve the quality of diarized transcrips using the pre-trained models from the
diarizationlm
project.implementation via
llama-cpp-python
and the utility prompt building functions provided bydiarizationlm
The text was updated successfully, but these errors were encountered: