-
Notifications
You must be signed in to change notification settings - Fork 183
Issues: mit-han-lab/llm-awq
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Feature 'ldmatrix' requires target sm_75 or higher when building awq_inference_engine on Tesla V100
#223
opened Oct 5, 2024 by
ShobhaRajanna
AttributeError: 'LlamaConfig' object has no attribute 'rope_theta'
#222
opened Sep 30, 2024 by
lvtao65535
Unsupported NVHPC compiler found. nvc++ is the only NVHPC compiler
#220
opened Sep 17, 2024 by
SimWangArizona
"Expected all tensors to be on the same device" when running "Perform AWQ search" on Llama3
#219
opened Sep 10, 2024 by
charlesyju
Batch Processing not implemented for LlavaStreamGenerator
#216
opened Aug 12, 2024 by
rahulthakur319
NotImplementedError: <class 'transformers_modules.modeling_chatglm.ChatGLMForConditionalGeneration'>
#214
opened Aug 8, 2024 by
lihaofd
Add support for GPUs with compute capability lower than 8.0 for awq/kernels installation
#204
opened Jul 3, 2024 by
rahulthakur319
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.