Given a model and targeted hardware, Olive (abbreviation of Onnx LIVE) composes the best suitable optimization techniques to output the most efficient ONNX model(s) for inferencing on the cloud or edge, while taking a set of constraints such as accuracy and latency into consideration.
- Reduce frustration of manual trial-and-error model optimization experimentation. Define your target and precision and let Olive automatically produce the best model for you.
- 40+ built-in model optimization components covering industry-leading techniques across model compression, optimization, finetuning, and compilation.
- Easy-to-use CLI for common model optimization tasks.
- Workflows to orchestrate model transformations and optimizations steps.
- Support for compiling LoRA adapters for MultiLoRA serving.
- Seamless integration with Hugging Face and Azure AI.
- Built-in caching mechanism to improve productivity.
Here are some recent videos, blog articles and labs that highlight Olive:
- [ Nov 2024 ] Democratizing AI Model optimization with the new Olive CLI
- [ Nov 2024 ] Unlocking NLP Potential: Fine-Tuning with Microsoft Olive (Ignite Pre-Day Lab PRE016)
- [ Nov 2024 ] Olive supports generating models for MultiLoRA serving on the ONNX Runtime
- [ Oct 2024 ] Windows Dev Chat: Optimizing models from Hugging Face for the ONNX Runtime (video)
- [ May 2024 ] AI Toolkit - VS Code Extension that uses Olive to fine tune models
For a full list of news and blogs, read the news archive.
The following notebooks are available that demonstrate key optimization workflows with Olive and include the application code to inference the optimized models on the ONNX Runtime.
Title | Task | Description | Time Required | Notebook Links |
---|---|---|---|---|
Quickstart | Text Generation | Learn how to quantize & optimize an SLM for the ONNX Runtime using a single Olive command. | 5mins | Download / Open in Colab |
Optimizing popular SLMs | Text Generation | Choose from a curated list of over 20 popular SLMs to quantize & optimize for the ONNX runtime. | 5mins | Download / Open in Colab |
How to finetune models for on-device inference | Text Generation | Learn how to Quantize (using AWQ method), fine-tune, and optimize an SLM for on-device inference. | 15mins | Download / Open in Colab |
If you prefer using the command line directly instead of Jupyter notebooks, we've outlined the quickstart commands here.
We recommend installing Olive in a virtual environment or a conda environment.
pip install olive-ai[ort-genai,auto-opt]
pip install transformers==4.44.2
Note
Olive has optional dependencies that can be installed to enable additional features. Please refer to Olive package config for the list of extras and their dependencies.
In this quickstart you'll be optimizing HuggingFaceTB/SmolLM2-135M-Instruct, which has many model files in the Hugging Face repo for different precisions that are not required by Olive. To minimize the download, cache the original Hugging Face model files (safetensors and configuration) in the main folder of the Hugging Face repo using:
huggingface-cli download HuggingFaceTB/SmolLM2-135M-Instruct *.json *.safetensors *.txt
Next, run the automatic optimization:
olive auto-opt \
--model_name_or_path HuggingFaceTB/SmolLM2-135M-Instruct \
--output_path models/smolm2 \
--device cpu \
--provider CPUExecutionProvider \
--use_ort_genai \
--precision int4 \
--log_level 1
Tip
PowerShell Users
Line continuation between Bash and PowerShell are not interchangable. If you are using PowerShell, then you can copy-and-paste the following command that uses compatible line continuation.olive auto-opt `
--model_name_or_path HuggingFaceTB/SmolLM2-135M-Instruct `
--output_path models/smolm2 `
--device cpu `
--provider CPUExecutionProvider `
--use_ort_genai `
--precision int4 `
--log_level 1
The automatic optimizer will:
- Acquire the model from the local cache (note: if you skipped the model download step then the entire contents of the Hugging Face model repo will be downloaded).
- Capture the ONNX Graph and store the weights in an ONNX data file.
- Optimize the ONNX Graph.
- Quantize the model to
int4
using RTN method.
Olive can automatically optimize popular model architectures like Llama, Phi, Qwen, Gemma, etc out-of-the-box - see detailed list here. Also, you can optimize other model architectures by providing details on the input/outputs of the model (io_config
).
The ONNX Runtime (ORT) is a fast and light-weight cross-platform inference engine with bindings for popular programming language such as Python, C/C++, C#, Java, JavaScript, etc. ORT enables you to infuse AI models into your applications so that inference is handled on-device.
The following code creates a simple console-based chat interface that inferences your optimized model - select Python and/or C# to expand the code:
Python
Create a Python file calledapp.py
and copy and paste the following code:
# app.py
import onnxruntime_genai as og
model_folder = "models/smolm2/model"
# Load the base model and tokenizer
model = og.Model(model_folder)
tokenizer = og.Tokenizer(model)
tokenizer_stream = tokenizer.create_stream()
# Set the max length to something sensible by default,
# since otherwise it will be set to the entire context length
search_options = {}
search_options['max_length'] = 200
search_options['past_present_share_buffer'] = False
chat_template = "<|im_start|>user\n{input}<|im_end|>\n<|im_start|>assistant\n"
text = input("Input: ")
# Keep asking for input phrases
while text != "exit":
if not text:
print("Error, input cannot be empty")
exit
# generate prompt (prompt template + input)
prompt = f'{chat_template.format(input=text)}'
# encode the prompt using the tokenizer
input_tokens = tokenizer.encode(prompt)
params = og.GeneratorParams(model)
params.set_search_options(**search_options)
params.input_ids = input_tokens
generator = og.Generator(model, params)
print("Output: ", end='', flush=True)
# stream the output
try:
while not generator.is_done():
generator.compute_logits()
generator.generate_next_token()
new_token = generator.get_next_tokens()[0]
print(tokenizer_stream.decode(new_token), end='', flush=True)
except KeyboardInterrupt:
print(" --control+c pressed, aborting generation--")
print()
text = input("Input: ")
To run the code, execute python app.py
. You'll be prompted to enter a message to the SLM - for example, you could ask what is the golden ratio, or def print_hello_world():. To exit type exit in the chat interface.
C#
Create a new C# Console app and install the Microsoft.ML.OnnxRuntimeGenAI Nuget package into your project:
mkdir ortapp
cd ortapp
dotnet new console
dotnet add package Microsoft.ML.OnnxRuntimeGenAI --version 0.5.2
Next, copy-and-paste the following code into your Program.cs
file and update modelPath
variable to be the absolute path of where you stored your optimized model.
// Program.cs
using Microsoft.ML.OnnxRuntimeGenAI;
internal class Program
{
private static void Main(string[] args)
{
string modelPath @"models/smolm2/model";
Console.Write("Loading model from " + modelPath + "...");
using Model model = new(modelPath);
Console.Write("Done\n");
using Tokenizer tokenizer = new(model);
using TokenizerStream tokenizerStream = tokenizer.CreateStream();
while (true)
{
Console.Write("User:");
string prompt = "<|im_start|>user\n" +
Console.ReadLine() +
"<|im_end|>\n<|im_start|>assistant\n";
var sequences = tokenizer.Encode(prompt);
using GeneratorParams gParams = new GeneratorParams(model);
gParams.SetSearchOption("max_length", 200);
gParams.SetInputSequences(sequences);
gParams.SetSearchOption("past_present_share_buffer", false);
Console.Out.Write("\nAI:");
using Generator generator = new(model, gParams);
while (!generator.IsDone())
{
generator.ComputeLogits();
generator.GenerateNextToken();
var token = generator.GetSequence(0)[^1];
Console.Out.Write(tokenizerStream.Decode(token));
Console.Out.Flush();
}
Console.WriteLine();
}
}
}
Run the application:
dotnet run
You'll be prompted to enter a message to the SLM - for example, you could ask what is the golden ratio, or def print_hello_world():. To exit type exit in the chat interface.
- We welcome contributions! Please read the contribution guidelines for more details on how to contribute to the Olive project.
- For feature requests or bug reports, file a GitHub Issue.
- For general discussion or questions, use GitHub Discussions.
Copyright (c) Microsoft Corporation. All rights reserved.
Licensed under the MIT License.