Skip to content

Issues: SJTU-IPADS/PowerInfer

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

请问能和vllm共同使用吗 question Further information is requested
#202 opened Jun 26, 2024 by yadandan
How to convert ProSparse-LLaMA-2-13B model to .gguf? question Further information is requested
#201 opened Jun 23, 2024 by Graysonicc
3 tasks done
代码中使用的llama.cpp版本 question Further information is requested
#200 opened Jun 21, 2024 by weizhenhuan
windows下cmake编译失败
#199 opened Jun 19, 2024 by codetown
支持的量化类型 question Further information is requested
#196 opened Jun 14, 2024 by deleteeeee
Source for v2 (mobile inference engine) question Further information is requested
#194 opened Jun 12, 2024 by peeteeman
Need quite a long time to load the model question Further information is requested
#188 opened May 21, 2024 by meicale
Will this work with Falcon 2? question Further information is requested
#186 opened May 14, 2024 by aaronrmm
关于在A100显卡上测得的效果异常的疑问 question Further information is requested
#184 opened May 4, 2024 by bulaikexiansheng
请问大神有支持LLama 3 70B 的计划吗? enhancement New feature or request
#183 opened May 1, 2024 by xiasw81
在A100-80G上无法找到cuda的情况 question Further information is requested
#182 opened Apr 24, 2024 by bulaikexiansheng
two questions that i want to solve question Further information is requested
#167 opened Mar 18, 2024 by yeptttt
Will we have instruct fine-tuned model support in the future? question Further information is requested
#164 opened Mar 13, 2024 by ZeonfaiHo
3 tasks done
[Question]: High PPL on wikitext2 of ReLU-LLAMA-7B for language modeling tasks question Further information is requested
#162 opened Mar 11, 2024 by llCurious
3 tasks done
[ROCm] Is AMD ROCm support available in near future? enhancement New feature or request
#158 opened Feb 29, 2024 by Orion-zhen
3 tasks done
关于LLaMA-70B-PowerInfer-GGUF的chat版本 question Further information is requested
#143 opened Feb 6, 2024 by NerounCstate
possible to do one that can fit into 7GB vram? question Further information is requested
#141 opened Feb 4, 2024 by sprappcom
Will using only CPU be faster than llama.cpp? question Further information is requested
#140 opened Feb 2, 2024 by liutt1312
ProTip! Type g p on any issue or pull request to go back to the pull request listing page.