Repository Details
Shared by
HelloGitHub Rating
10.0
1 ratings
Microsoft's Open-Source 1-bit Large Model Inference Framework
Past 6 days Received 85 stars ✨
Free•MIT
Claim
Discuss
Collect
Share
38.9k
Stars
No
Chinese
Python
Language
Yes
Active
21
Contributors
315
Issues
Yes
Organization
None
Latest
4k
Forks
MIT
License
More

This is an inference framework designed by Microsoft specifically for CPU-based local inference and extreme compression (low-bit) of large models. It supports efficient and low-power inference for 1-bit/1.58-bit quantized models, compatible with models such as BitNet, Llama3-8B-1.58, and Falcon3. It is suitable for running large model inference tasks locally or on edge devices without the need for GPUs.
Comments
Rating:
No comments yet