youtube image
From YouTube: Model Serving at the Edge Made Easier - Paul Van Eck & Animesh Singh, IBM

Description

Model Serving at the Edge Made Easier - Paul Van Eck & Animesh Singh, IBM

As edge devices consume the world, the ability to deploy AI models on these devices becomes increasingly vital. Challenges surrounding the management of numerous models across a multitude of edge hosts can be tricky. Not only that, the limited compute power that edge hosts provide makes it necessary to eliminate as much overhead as possible. These are common pain points holding users back from large scale adoption. However, with the combination of ModelMesh with technologies like K3s and MicroShift, the practicality of employing such a system has increased dramatically. As the multi-model serving backend of KServe, ModelMesh offers a small-footprint control-plane for managing model deployments on Kubernetes. Using multi-model runtimes with intelligent model loading/unloading, ModelMesh is able to make the most out of a limited set of resources while still providing the capability to serve many models for inference. Come to this talk to get the edge on edge model serving!