Accelerated NLP pipelines for fast inference on CPU and GPU. Built with Transformers, Optimum and ONNX Runtime
Accelerated NLP pipelines for fast inference 🚀 on CPU and GPU. Built with 🤗Transformers, Optimum and ONNX runtime.
Installation:
With PyPI:
pip install optimum-transformers
Or directly from GitHub:
pip install git+https://github.com/AlekseyKorshuk/optimum-transformers
Usage:
The pipeline API is similar to transformers pipeline
with just a few differences which are explained below.
Just provide the path/url to the model, and it’ll download the model if needed from
the hub and automatically create onnx graph and run inference.