Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature request : Support for PHI3 mini #210

Open
3 tasks
raymond-infinitecode opened this issue Jul 14, 2024 · 0 comments
Open
3 tasks

Feature request : Support for PHI3 mini #210

raymond-infinitecode opened this issue Jul 14, 2024 · 0 comments
Labels
enhancement New feature or request

Comments

@raymond-infinitecode
Copy link

Prerequisites

Before submitting your issue, please ensure the following:

  • I am running the latest version of PowerInfer. Development is rapid, and as of now, there are no tagged versions.
  • I have carefully read and followed the instructions in the README.md.
  • I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).

Feature Description

PHI3 mini is currently the most powerful SLM yet, but can we relu it to make it fast so a single Xeon server can serve hundreds of concurrent users with relu implementation ?

Motivation

Please provide a detailed written description of reasons why this feature is necessary and how it is useful to PowerInfer users.

Possible Implementation

Convert the Phi3 model to relu model

@raymond-infinitecode raymond-infinitecode added the enhancement New feature or request label Jul 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant