MATIH Platform is in active MVP development. Documentation reflects current implementation status.
17. Kubernetes & Helm
ML Infrastructure
Overview

ML Infrastructure Overview

MATIH deploys a comprehensive ML infrastructure stack for experiment tracking, feature management, distributed training, and model serving. These components integrate with the ML Service and AI Service to provide end-to-end ML lifecycle management.


Component Summary

ComponentPurposeGPU Required
MLflowExperiment tracking, model registry, artifact storeNo
FeastFeature store (online + offline)No
RayDistributed training, hyperparameter tuningOptional
vLLMHigh-throughput LLM inference serverYes
TritonMulti-framework model servingYes
JupyterHubInteractive notebook environmentOptional

Section Contents

PageDescription
MLflowExperiment tracking and model registry
FeastFeature store with online and offline stores
RayDistributed computing cluster
vLLMLLM inference server
TritonMulti-framework inference server
JupyterHubNotebook environment