Yury Babak created IGNITE-10286:
-----------------------------------

             Summary: [ML] Umbrella: Model serving
                 Key: IGNITE-10286
                 URL: https://issues.apache.org/jira/browse/IGNITE-10286
             Project: Ignite
          Issue Type: New Feature
          Components: ml
            Reporter: Yury Babak
            Assignee: Yury Babak


We want to have convenient API for model serving. It means that we need a 
mechanism for storing models and infer them inside Apache Ignite.

For now, I see 2 important features - distributed storage for any models and 
inference.

>From my point of view, we could use some built-in(predefined) cache as model 
>storage. And use service grid for model inference. We could implement some 
>"ModelService" for access to our storage, receive the list of all suitable 
>model(including model metrics and some other information about a model), 
>choose one(or several) and infer it from this service.

Model from TF should also use the same mechanisms for storing and inference.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to