Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Find a solution to avoid CUDA OOM during diarization #186

Merged
merged 18 commits into from
Aug 7, 2023

Conversation

chainyo
Copy link
Contributor

@chainyo chainyo commented Aug 4, 2023

By implementing the SpeakerClustering model directly in the project, I'm looking to solve CUDA OOM problems during diarization.

This PR:

  • Split diarization sub-modules into multiple files under wordcab_transcribe/services/diarization/*.
  • Re-implement the code proposed in the NVIDIA NeMo package for Multiscale Auto-tuning Spectral Clustering with only the first segmentation and clustering steps (not an entire MSDD pipeline, beneficial?).
  • Makes the diarization process more modular and comprehensive.
  • Adds a dynamic multiscale strategy for long audio files to avoid CUDA OOM.
  • Make us close to avoid installing the nemo package (which adds a lot of constraints on Python and PyTorch versions...).

@chainyo chainyo added bug Something isn't working diarization Everything related to the diarization part labels Aug 4, 2023
@chainyo chainyo self-assigned this Aug 4, 2023
@chainyo chainyo linked an issue Aug 4, 2023 that may be closed by this pull request
@chainyo chainyo marked this pull request as ready for review August 7, 2023 09:55
@chainyo chainyo merged commit 69cf1c1 into main Aug 7, 2023
3 checks passed
@chainyo chainyo deleted the 129-error-in-diarization-cuda-out-of-memory branch August 7, 2023 10:14
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working diarization Everything related to the diarization part
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Error in diarization: CUDA out of memory
2 participants