New
v1.3.0
Added
- Support on-device training for MLP, CNN(lenet, resnet50, mobilnetv1), transformer/bert(text to speech)
- Support change model input and output names in X2bolt
- Support more graph optimizations : Transpose+Convolution, Swish, Quantization, Power+Scale
- Support dynamic output related operators : Shape, ConstantOfShape, GenerateProposals, NonZero, NonMaxSuppression, Reshape, etc
- Support more operators : GridSample, CumSum, OneHot, Round, Floor, Ceil
- Support more networks on CPU : yolov2, yolov3, yolov4, yolov5, faster-rcnn, mask-rcnn, retinanet, dfsmn, frill, conformer, unet, etc
- Support Armv8 int8 to accelerate NLP network
- Improve inference performance on avx2 CPU
- Support netron to visualize bolt model
- Support not to bind CPU core
- Add C API MemoryCheck to check bolt memory leak
Changed
- X2bolt add -I and -O options to change model input and output names.
- X2bolt add -t option to convert model for on-device training.
- C API CreateModel and AllocAllResultHandle return value is set to NULL when unsuccessful.
- install.sh add --neon option to close arm neon acceleration on old platform.