All Projects → IntelAI → inference-model-manager

IntelAI / inference-model-manager

Licence: Apache-2.0 license
Inference Model Manager for Kubernetes

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects
go
31211 projects - #10 most used programming language

Projects that are alternatives of or similar to inference-model-manager

Sparktorch
Train and run Pytorch models on Apache Spark.
Stars: ✭ 195 (+323.91%)
Mutual labels:  inference
Bmw Yolov4 Inference Api Gpu
This is a repository for an nocode object detection inference API using the Yolov3 and Yolov4 Darknet framework.
Stars: ✭ 237 (+415.22%)
Mutual labels:  inference
graphsignal
Graphsignal Python agent
Stars: ✭ 158 (+243.48%)
Mutual labels:  inference
Volksdep
volksdep is an open-source toolbox for deploying and accelerating PyTorch, ONNX and TensorFlow models with TensorRT.
Stars: ✭ 195 (+323.91%)
Mutual labels:  inference
Libonnx
A lightweight, portable pure C99 onnx inference engine for embedded devices with hardware acceleration support.
Stars: ✭ 217 (+371.74%)
Mutual labels:  inference
Grakn
TypeDB: a strongly-typed database
Stars: ✭ 2,947 (+6306.52%)
Mutual labels:  inference
Torch2trt
An easy to use PyTorch to TensorRT converter
Stars: ✭ 2,974 (+6365.22%)
Mutual labels:  inference
go-ml-benchmarks
⏱ Benchmarks of machine learning inference for Go
Stars: ✭ 27 (-41.3%)
Mutual labels:  inference
Tensorrt Laboratory
Explore the Capabilities of the TensorRT Platform
Stars: ✭ 236 (+413.04%)
Mutual labels:  inference
awesome-ml-demos-with-ios
The challenge projects for Inferencing machine learning models on iOS
Stars: ✭ 1,040 (+2160.87%)
Mutual labels:  inference
Elfi
ELFI - Engine for Likelihood-Free Inference
Stars: ✭ 208 (+352.17%)
Mutual labels:  inference
Frozen Graph Tensorflow
Save, Load Frozen Graph and Run Inference From Frozen Graph in TensorFlow 1.x and 2.x
Stars: ✭ 212 (+360.87%)
Mutual labels:  inference
Models
Model Zoo for Intel® Architecture: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors
Stars: ✭ 248 (+439.13%)
Mutual labels:  inference
Pytorch Cpp Inference
Serving PyTorch 1.0 Models as a Web Server in C++
Stars: ✭ 194 (+321.74%)
Mutual labels:  inference
object-flaw-detector-cpp
Detect various irregularities of a product as it moves along a conveyor belt.
Stars: ✭ 19 (-58.7%)
Mutual labels:  inference
Pytorch Cpp
PyTorch C++ inference with LibTorch
Stars: ✭ 194 (+321.74%)
Mutual labels:  inference
Adlik
Adlik: Toolkit for Accelerating Deep Learning Inference
Stars: ✭ 237 (+415.22%)
Mutual labels:  inference
approxposterior
A Python package for approximate Bayesian inference and optimization using Gaussian processes
Stars: ✭ 36 (-21.74%)
Mutual labels:  inference
dbnR
Gaussian dynamic Bayesian networks structure learning and inference based on the bnlearn package
Stars: ✭ 33 (-28.26%)
Mutual labels:  inference
Dawn Bench Entries
DAWNBench: An End-to-End Deep Learning Benchmark and Competition
Stars: ✭ 254 (+452.17%)
Mutual labels:  inference

Inference Model Manager for Kubernetes

CircleCI

Inference Model Manager for Kubernetes is an open source SW platform intended to provide convenient solution for hosting, management and scaling inference processing endpoints exposed over gRPC protocol.

It is built on top of Kubernetes and integrates with TensorFlow Serving* and OpenVINO Model Server* for exposing the inference services via gRPC endpoints.

It is intended for organizations who dynamically deploy and scale inference endpoints.

  • Users are organized into tenants
  • Multiple tenants are supported with “soft” isolation

Inference Model Manager for Kubernetes includes a custom REST API which simplifies the configuration and management of hosted inference services. Inference Model Manager integrates with Minio or other S3 compatible components used for storage of the AI models.

Inference Model Manager for Kubernetes conjoins inference services scalability and easy management with security features like:

  • limiting access to inference endpoints to authorized clients only
  • preventing unauthorized access to management API
  • limiting access to tenant data based on group membership information from external identity provider.

Fully customizable serving templates provide predefined and optimized Kubernetes configurations of inference services. It ensures well tuned performance with maximum simplicity on the users side. Templates can enable additional model servers or adjust them to application needs and the infrastructure configuration.

Installation quicksheet (beta)

Architecture overview

Prerequisites and requirements

Building platform components

Deployment guide

Platform admin guide

Platform user guide

Example grpc client

Security recommendation for Kubernetes

Troubleshooting

Serving templates

Troubleshooting

Collaboration

By contributing to the project software, you agree that your contributions will be licensed under the Apache 2.0 license that is included in the LICENSE file in the root directory of this source tree. The user materials are licensed under CC-BY-ND 4.0.

Contact

Submit Github issue to ask a question, submit a request or report a bug.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].