Skip to content

Support for BitNet (Microsoft's native 1 bit LLM) #73

@DrMurx

Description

@DrMurx

How to use GitHub

  • Please use the 👍 reaction to show that you are interested into the same feature.
  • Please don't comment if you have no relevant information to add. It's just extra noise for everyone subscribed to this issue.
  • Subscribe to receive notifications on status change and new comments.

Feature request

Microsoft has released Bitnet B1.58-2B-4T under the MIT license. This is a 1.58 bit model with 2 billion parameters, natively trained for that precision (instead of being created by quantization of larger models).

The implication is that BitNet is fairly small, yet powerful, and can run solely on a CPU with acceptable inference speed.

Furthermore, Microsoft provides an own inference framework named BitNet.cpp, based on llama.cpp. My understanding is that this would allow a rather straightforward integration in existing codebase.

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions