[ https://issues.apache.org/jira/browse/CAMEL-21019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Tadayoshi Sato updated CAMEL-21019: ----------------------------------- Fix Version/s: 4.10.0 > Add a component for TensorFlow Serving > -------------------------------------- > > Key: CAMEL-21019 > URL: https://issues.apache.org/jira/browse/CAMEL-21019 > Project: Camel > Issue Type: New Feature > Components: camel-ai > Affects Versions: 4.7.0 > Reporter: Tadayoshi Sato > Assignee: Tadayoshi Sato > Priority: Major > Fix For: 4.10.0 > > > Running a TensorFlow model is already supported through Camel DJL component. > However, Camel users might prefer to externalise inferencing to an external > server instead of running it inside the Camel route. For TensorFlow models, > it is generally done with [TensorFlow > Serving|https://www.tensorflow.org/tfx/guide/serving], which is a REST API > server for inferencing with TensorFlow. Camel should provide a producer > component that makes it easy to invoke the TensorFlow specific REST API from > the routes. -- This message was sent by Atlassian Jira (v8.20.10#820010)