Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com Professional-Machine-Learning-Engineer exam questions have been updated and answers have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-Engineer dumps with Test Engine here:
You are working on a system log anomaly detection model for a cybersecurity organization. You have developed the model using TensorFlow, and you plan to use it for real-time prediction. You need to create a Dataflow pipeline to ingest data via Pub/Sub and write the results to BigQuery. You want to minimize the serving latency as much as possible. What should you do?
Correct Answer: A
Containerizing the model prediction logic in Cloud Run allows for easy and efficient deployment of the model, and allows it to be invoked by Dataflow. Cloud Run is a fully managed service that allows you to run stateless containers in a serverless environment. It automatically scales instances up and down based on the traffic, which can minimize the serving latency. Additionally, Dataflow can easily invoke Cloud Run services via HTTP requests, making it simple to integrate into your pipeline. This allows the Dataflow pipeline to focus on data ingestion and processing, while the Cloud Run service handles the real-time predictions. While it is possible to load the model directly into the Dataflow job as a dependency, this approach can increase the complexity of the pipeline and could lead to increased latency. Other options, such as deploying the model to a Vertex AI endpoint or a TFServing container on GKE, would also work but this option is the most optimal for minimizing the serving latency.