int8-inference
Here are 13 public repositories matching this topic...
Language:All
Sort:Most stars
Real-time portrait segmentation for mobile devices
- Updated
Jan 17, 2021 - Jupyter Notebook
Generate a quantization parameter file for ncnn framework int8 inference
- Updated
Jul 29, 2020 - Python
BEVFormer inference on TensorRT, including INT8 Quantization and Custom TensorRT Plugins (float/half/half2/int8).
- Updated
Nov 20, 2023 - Python
Quantization Aware Training
- Updated
Jan 13, 2024 - Python
将端上模型部署过程中,常见的问题以及解决办法记录并汇总,希望能给其他人带来一点帮助。
- Updated
Aug 17, 2022 - Python
TensorRT Int8 Python version sample. TensorRT Int8 Python 实现例子。TensorRT Int8 Pythonの例です
- Updated
Jan 28, 2019 - Python
GPT-J 6B inference on TensorRT with INT-8 precision
- Updated
Apr 5, 2023 - Python
VB.NET api wrapper for llm-inference chatllm.cpp
- Updated
Nov 26, 2024 - Visual Basic .NET
Generating tensorrt model using onnx
- Updated
Jun 22, 2023 - C++
C# api wrapper for llm-inference chatllm.cpp
- Updated
Nov 20, 2024 - C#
Compressed CNNs for airplane classification in satellite images (APoZ-based parameter pruning, INT8 weight quantization)
- Updated
Jun 10, 2020 - Python
it has support for openvino converted model of yolov7-int.xml ,yolov7x,
- Updated
Mar 6, 2023 - Python
Improve this page
Add a description, image, and links to theint8-inference topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with theint8-inference topic, visit your repo's landing page and select "manage topics."