Accelerated NLP pipelines for fast inference on CPU and GPU. Built with Transformers, Optimum and ONNX Runtime
Accelerated NLP pipelines for fast inference đ on CPU and GPU. Built with đ¤Transformers, Optimum and ONNX runtime. Installation: With PyPI: pip install optimum-transformers Or directly from GitHub: pip install git+https://github.com/AlekseyKorshuk/optimum-transformers Usage: The pipeline API is similar to transformers pipeline with just a few differences which are explained below. Just provide the path/url to the model, and itâll download the model if needed from the hub and automatically create onnx graph and run inference.
Read more