Post-Training quantization perfomed on the model trained with CLIC dataset.
-
Updated
Sep 1, 2025 - Jupyter Notebook
8000
Post-Training quantization perfomed on the model trained with CLIC dataset.
[Best Paper Award IEEE EDGE 2024] Research experiments archive for post-training quantization with TensorRT.
Political Leaning Detection in the News Articles
Creating Snapdragon neural processing engine environment to convert protobuf(pb) files to dlc(deep learning container) files as well as use snpe for Quantizing neural networks. Mobile app development link https://github.com/anshumax/mobilenn
A set of notebooks analyzing neural network quantization, comparing symmetric and asymmetric schemes, calibration methods, and PTQ vs. QAT
TQ-DiT: Efficient Time-Aware Quantization for Diffusion Transformers
QuantLab-8bit is a reproducible benchmark of 8-bit quantization on compact vision backbones. It includes FP32 baselines, PTQ (dynamic & static), QAT, ONNX exports, parity checks, ORT CPU latency, and visual diagnostics.
A Quantization method for PyTorch framework. Implementing lower than 8 bits quantization-aware and post training quantization methods
Real-time video recognition with your Raspberry Pi!
In this repo you will understand .The process of reducing the precision of a model’s parameters and/or activations (e.g., from 32-bit floating point to 8-bit integers) to make neural networks smaller, faster, and more energy-efficient with minimal accuracy loss.
Hardware-verified SNN-equivalent intrusion detection system (IDS) on STM32N6 Neural-ART NPU. INT8 quantized MLP achieving 0.4561ms inference at 800MHz. First publicly documented IDS on a commodity MCU NPU with T=1 SNN-ANN equivalence validation.
EfficientNetV2 (Efficientnetv2-b2) and quantization int8 and fp32 (QAT and PTQ) on CK+ dataset . fine-tuning, augmentation, solving imbalanced dataset, etc.
This repo aims at replicating: "AdpQ: A Zero-shot Calibration Free Adaptive Post Training Quantization Method for LLMs" https://arxiv.org/abs/2405.13358
An example to quantize pretrained models from pytorch-image-models with PyTorch FX graph mode quantization
LLM Post-Training, RLHF, PPO, DPO, etc
A framework to train a ResUNet architecture, quantize, compile and execute it on an FPGA.
An Empirical Study of World Model Quantization
GPTQ post-training quantization from scratch — GPT-2, OPT, LLaMA support
Simple pipeline to efficiently train, evaluate and serve customised LLMs
Add a description, image, and links to the post-training-quantization topic page so that developers can more easily learn about it.
To associate your repository with the post-training-quantization topic, visit your repo's landing page and select "manage topics."