Accelerated NLP pipelines for fast inference on CPU and GPU. Built with Transformers, Optimum and ONNX Runtime
Accelerated NLP pipelines for fast inference on CPU and GPU. Built with
Transformers, Optimum and ONNX runtime.
Installation:
With PyPI:
pip install optimum-transformers
Or directly from GitHub:
pip install git+https://github.com/AlekseyKorshuk/optimum-transformers
Usage:
The pipeline API is similar to transformers pipeline
with just a few differences which are explained below.
Just provide the path/url to the model, and itβll download the model if needed from
the hub and automatically create onnx graph and run inference.