Yury Babak created IGNITE-10286:
-----------------------------------
Summary: [ML] Umbrella: Model serving
Key: IGNITE-10286
URL: https://issues.apache.org/jira/browse/IGNITE-10286
Project: Ignite
Issue Type: New Feature
Components: ml
Reporter: Yury Babak
Assignee: Yury Babak
We want to have convenient API for model serving. It means that we need a
mechanism for storing models and infer them inside Apache Ignite.
For now, I see 2 important features - distributed storage for any models and
inference.
>From my point of view, we could use some built-in(predefined) cache as model
>storage. And use service grid for model inference. We could implement some
>"ModelService" for access to our storage, receive the list of all suitable
>model(including model metrics and some other information about a model),
>choose one(or several) and infer it from this service.
Model from TF should also use the same mechanisms for storing and inference.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)