TensorFlow Serving is an open source, high-performance, serving system that provides a system to handle the inference aspect of machine learning.
On IBM® z16™ and later (running Linux on IBM Z or IBM® z/OS® Container Extensions (IBM zCX)), TensorFlow core Graph Execution will leverage new inference acceleration capabilities that transparently target the IBM Integrated Accelerator for AI through the IBM z Deep Neural Network (zDNN) library. The IBM zDNN library contains a set of primitives that support Deep Neural Networks. These primitives transparently target the IBM Integrated Accelerator for AI on IBM z16™ and later. No changes to the original model are needed to take advantage of the new inference acceleration capabilities.
Note. When using IBM Z Accelerated Serving for TensorFlow on either an IBM z14™ or an IBM z15™, TensorFlow will transparently target the CPU with no changes to the model. See IBM Z Accelerated Serving for TensorFlow for more information This image is built by IBM to run on the IBM Z architecture and is not affiliated with any other community that provides a version of this image.1.2.0 | docker pull icr.io/ibmz/ibmz-accelerated-serving-for-tensorflow@sha256:27c818141999106da13af9df42d7fc917ee434a27f7442bf57b025a83d96a86d | Vulnerability Report | 06-11-2024 |
Version | Pull String | Security (IBM Cloud) | Created |
---|---|---|---|
For documentation and samples for the IBM Z Accelerated Serving for TensorFlow container image, please visit the GitHub Repository here.