Olive
latest

OVERVIEW

  • Olive
  • Quick Tour
  • Olive Options

GET STARTED

  • Installation
  • Quickstart Examples

EXAMPLES

  • Acceleration on CPU
  • Acceleration on GPU
  • Acceleration on NPU
  • Acceleration on DPU

FEATURES

  • Model Transformations and Optimizations
  • Huggingface Model Optimization
  • Packaging Olive artifacts
  • Custom Scripts
  • Azure ML integration

EXTENDING OLIVE

  • Design
  • How to add new optimization Pass

TUTORIALS

  • How To Configure System
  • How To Configure Metric
  • How To Configure Pass
  • How To Configure Data
  • How To Set Model Path
  • Advanced User Tour
  • Self-hosted Kubernetes cluster
  • Azure ML scripts

API REFERENCE

  • OliveModels
  • ResourcePath
  • OliveSystems
  • OliveEvaluator
  • Metric
  • SearchAlgorithms
  • Engine
  • Passes
Olive
  • Acceleration on CPU
  • View page source

Acceleration on CPU¶

Whisper optimization with PTQ and pre/post processing

BERT optimization with Intel® Neural Compressor Post Training Quantization

BERT optimization with QAT Customized Training Loop

ResNet optimization with QAT Default Training Loop

ResNet optimization with QAT PyTorch Lightning Module

Cifar10 optimization with OpenVINO for Intel HW

Acceleration on GPU¶

Bert optimization with CUDA/TensorRT

SqueezeNet latency optimization with DirectML

Stable Diffusion optimization with DirectML

Dolly V2 optimization with DirectML

Acceleration on NPU¶

Inception model optimization on Qualcomm NPU

Acceleration on DPU¶

ResNet optimization with Vitis-AI Post Training Quantization for AMD DPU

Previous Next

© Copyright 2023, olivedevteam@microsoft.com.

Built with Sphinx using a theme provided by Read the Docs.