Giter VIP home page Giter VIP logo

ai-on-z-triton-is-examples's Introduction

Demonstrating inference pipeline creation using triton inference server on IBM zSystems and LinuxONE.

1. Scope

The purpose of this project is to provide to sample artifacts to create inference pipeline using triton inference server on IBM zSystems and LinuxONE.

All the materials are provided as examples. Provided dockerfiles build open-source based (not proprietary) images.

The maintainers of this repository do not assert to be experts in containers or container security. Resources include:

2. Usage

These build files commonly rely on base images from the IBM Z and LinuxONE Container Image Registry (ICR) . This will require free basic authentication. Details can be found at the ICR link above.

Steps to build and run docker image.

  • Run docker build using the provided dockerfile i.e docker build -f ./Dockerfile .. Using a Ubuntu base image, this will create an environment with both specific Triton Server, Python Backend release with SnapML version.

  • Create and run a docker container using the image created on the prior step. As part of this step, you should map the required triton server ports to a port on your local system. An example follows: docker run --shm-size=1g --ulimit memlock=-1 --ulimit stack=67108864 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v//<path>/models:/models <imageid> tritonserver --model-repository=/models. This states the image in interactive mode, tells docker to delete the container upon exit, and publishes container HTTP service port at 8000, GRPC Inference Service at 8001, Metrics Service at 8002.

    Note: Please do refer each topic(e.g snapml-examples ) inside ai-on-z-triton-is-examples to create model reposity for the specific use cases.

3. Content

Folder(topic) Description
snapml-examples Serving SnapML models with Triton Inference Server on Linux on IBM zSystems

Note: We also support onnxmlir-triton-backend on Linux on IBM zSystems. This backend allows the usage of ONNX MLIR compiled models (model.so) with the Triton Inference Server.

4. Additional resources

Find out additional resources about Triton Inference Server in the below links.

  1. Triton Inference server user guide
  2. Triton Inference server github organization

5. License

If you would like to see the detailed LICENSE click here.

ai-on-z-triton-is-examples's People

Contributors

ibm-open-source-bot avatar pradghos avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

ai-on-z-triton-is-examples's Issues

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.