| | --- |
| | license: mit |
| | library_name: transformers |
| | tags: |
| | - mlx |
| | - open4bits |
| | base_model: deepseek-ai/DeepSeek-R1 |
| | pipeline_tag: text-generation |
| | --- |
| | |
| | # Open4bits / DeepSeek-R1-MLX-2Bit |
| |
|
| | This repository provides the **DeepSeek-R1 model quantized to 2-bit in MLX format**, published by Open4bits to enable highly efficient local inference with minimal memory usage and broad hardware compatibility. |
| |
|
| | The underlying DeepSeek-R1 model and architecture are **developed and owned by DeepSeek AI**. This repository contains only a 2-bit quantized MLX conversion of the original model weights. |
| |
|
| | The model is designed for lightweight, high-performance text generation and instruction-following tasks, making it well suited for resource-constrained and local deployments. |
| |
|
| | --- |
| |
|
| | ## Model Overview |
| |
|
| | DeepSeek-R1 is a transformer-based large language model developed for strong general language understanding and generation. |
| | This release provides a **2-bit quantized checkpoint in MLX format**, enabling efficient inference on CPUs and supported accelerators with reduced memory footprint. |
| |
|
| | Open4bits has started supporting **MLX models** to broaden compatibility with emerging quantization formats and efficient runtimes. |
| |
|
| | --- |
| |
|
| | ## Model Details |
| |
|
| | * **Base Model:** DeepSeek-R1 |
| | * **Quantization:** 2-bit |
| | * **Format:** MLX |
| | * **Task:** Text generation, instruction following |
| | * **Weight tying:** Preserved |
| | * **Compatibility:** MLX-enabled inference engines and efficient runtimes |
| |
|
| | This quantized release is designed to balance strong generation performance with low resource requirements. |
| |
|
| | --- |
| |
|
| | ## Intended Use |
| |
|
| | This model is intended for: |
| |
|
| | * Local text generation and conversational applications |
| | * CPU-based or low-resource deployments |
| | * Research, prototyping, and experimentation |
| | * Self-hosted or offline AI systems |
| |
|
| | --- |
| |
|
| | ## Limitations |
| |
|
| | * Reduced performance compared to full-precision variants |
| | * Output quality depends on prompt design and inference settings |
| | * Not specifically tuned for highly specialized or domain-specific tasks |
| |
|
| | --- |
| |
|
| | ## License |
| |
|
| | This model follows the **MIT** as defined by the base model creators. |
| | Users must comply with the licensing conditions of the base DeepSeek-R1 model. |
| |
|
| | --- |
| |
|
| | ## Support |
| |
|
| | If you find this model useful, please consider supporting the project. |
| | Your support helps Open4bits continue releasing and maintaining high-quality, efficient open models for the community. |
| |
|
| |
|