TRTorch v0.0.2
Pre-release
Pre-release
TRTorch v0.0.2
Python API & PyTorch 1.5.0 Support
- This is a second alpha release of TRTorch. It bumps support for PyTorch to 1.5.0 and introduces a Python distribution for TRTorch.
- Also now includes full documentation https://nvidia.github.io/TRTorch
- Adds support for Post Training Quantization in C++
Dependencies
- Libtorch 1.5.0
- CUDA 10.2
- cuDNN 7.6.5
- TensorRT 7.0.0
Changelog
Bug Fixes
- //core/conversion: Check for calibrator before setting int8 mode (3afd209)
- //core/conversion/conversionctx: Check both tensor and eval maps (2d65ece)
- //core/conversion/converters/impl/element_wise: Fix broadcast (a9f33e4)
- //cpp: Remove deprecated script namespace (d70760f)
- //cpp/api: Better inital condition for the dataloader iterator to (8d22bdd)
- //cpp/api: Remove unecessary destructor in ptq class (fc70267)
- //cpp/api: set a default for calibrator (825be69)
- //cpp/ptq: remove some logging from ptq app (b989c7f)
- Address issues in PR (cd24f26)
- //cpp/ptq: Tracing model in eval mode wrecks accuracy in Libtorch (54a24b3)
- //docs: add nojekyll file (2a02cd5)
- //docs: fix version links (11555f7)
- //py: Build system issues (c1de126)
- //py: Ignore generated version file (9e37dc1)
- bypass jeykll, also add PR template (a41c400)
Features
- //core/conversion/conversionctx: Make op precision available at (78a1c61)
- //core/conversion/converters/impl/shuffle: Implement aten::resize (353f2d2)
- //core/execution: Type checking for the executor, now is the (2dd1ba3)
- //core/lowering: New freeze model pass and new exception (4acc3fd)
- //core/quantization: skeleton of INT8 PTQ calibrator (dd443a6)
- //core/util: New logging level for Graph Dumping (90c44b9)
- //cpp/api: Adding max batch size setting (1b25542)
- //cpp/api: Functional Dataloader based PTQ (f022dfe)
- //cpp/api: Remove the extra includes in the API header (2f86f84)
- //cpp/ptq: Add a feature to the dataset to use less than the full (5f36f47)
- //cpp/ptq/training: Training recipe for VGG16 Classifier on (676bf56)
- //lowering: centralize lowering and try to use PyTorch Conv2DBN folding (fad4a10)
- //py: API now produces valid engines that are consumable by (72bc1f7)
- //py: Inital introduction of the Python API (7088245)
- //py: Manylinux container and build system for multiple python (639c2a3)
- //py: Working portable package (482ef2c)
- //tests: New optional accuracy tests to check INT8 and FP16 (df74136)
- //cpp/api: Working INT8 Calibrator, also resolves #41 (5c0d737)
- aten::flatten: Adds a converter for aten flatten since MM is the (d945eb9)
- aten::matmul|aten::addmm: Adds support for aten::matmul and (c5b6202)
- Support non cxx11-abi builds for use in python api (83e0ed6)
- aten::size [static]: Implement a aten::size converter for static input size (0548540)
- conv2d_to_convolution: A pass to map aten::conv2d to _convolution (2c5c0d5)