GreenWaves-Technologies / bfloat16Links
bfloat16 dtype for numpy
☆19Updated last year
Alternatives and similar repositories for bfloat16
Users that are interested in bfloat16 are comparing it to the libraries listed below
Sorting:
- PyTorch extension for emulating FP8 data formats on standard FP32 Xeon/GPU hardware.☆110Updated 7 months ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆149Updated 3 weeks ago
- ☆157Updated last year
- A tiny FP8 multiplication unit written in Verilog. TinyTapeout 2 submission.☆14Updated 2 years ago
- Framework to reduce autotune overhead to zero for well known deployments.☆79Updated last week
- ☆71Updated 8 months ago
- ☆152Updated 2 years ago
- Converting a deep neural network to integer-only inference in native C via uniform quantization and the fixed-point representation.☆25Updated 3 years ago
- Customized matrix multiplication kernels☆56Updated 3 years ago
- An experimental CPU backend for Triton (https//github.com/openai/triton)☆43Updated 4 months ago
- The official, proof-of-concept C++ implementation of PocketNN.☆34Updated last year
- ☆37Updated last year
- Trying to find out what is the minimal model that can achieve 99% accuracy on MNIST dataset☆25Updated 6 years ago
- GPTQ inference TVM kernel☆40Updated last year
- Explore training for quantized models☆20Updated this week
- ☆32Updated last year
- [TCAD 2021] Block Convolution: Towards Memory-Efficient Inference of Large-Scale CNNs on FPGA☆17Updated 3 years ago
- Prototype routines for GPU quantization written using PyTorch.