Skip to content

Handling of large datasets and the cache parameter - strategies #5125

Answered by glenn-jocher
Nisse123 asked this question in Q&A
Discussion options

You must be logged in to vote

@Nisse123 you can try caching to disk to not use RAM:

python train.py --cache disk

You can experiment with additional workers but 60 seems to be far too high to be useful. If you are using a DGX2 you should also be training Multi-GPU naturally for fastest results. See Multi-GPU Training tutorial:

YOLOv5 Tutorials

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@Nisse123
Comment options

@glenn-jocher
Comment options

Answer selected by Nisse123
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants