-
Notifications
You must be signed in to change notification settings - Fork 402
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Source for v2 (mobile inference engine) #194
Comments
please release PowerInfer-2 so that it can be tested on low resource PCs (like llama.cpp) for a comparison. |
PowerInfer-2 will be open-sourced in the future. We’re refining it to untangle from our testing platform and making it accessible on PCs for the community. |
Can't wait to test your amazing work! |
same here! I wish to test it on low resource pc with no gpu or an old and small one. |
This is fantastic!, on my old smartphone with 6 Gb of memory the Meta-Llama-3-8B-Instruct-Q4_K_M.gguf model ran, I hope for v2 in the near future. |
when can i use it on anroid phone ? |
Is it possible you could release the testing platform and the code entangled with the testing platform, so that the reported results can be reproduced? |
Hello there!
I came across the v2 paper yesterday, and saw the updates on the project readme.
I am interested in porting the v2 framework to iOS. The goal is to complete a naive port at first, and then include metal shaders.
Any plans on releasing the source and instructions for running v2 on Android?
The text was updated successfully, but these errors were encountered: