Skip to content
#

int4-quantization

Here are 2 public repositories matching this topic...

Language: All
Filter by language

Implemented post-training quantisation (PTQ) on transformer-based reasoning models using 8-bit and 4-bit weight quantisation (INT8, INT4) with frameworks like PyTorch and Hugging Face Transformers. Leveraged libraries such as bitsandbytes to reduce model size and accelerate inference, while evaluating performance degradation on reasoning tasks. Com

  • Updated Apr 21, 2026
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the int4-quantization topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the int4-quantization topic, visit your repo's landing page and select "manage topics."

Learn more