Deal of The Day! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Google Exam Professional Machine Learning Engineer Topic 6 Question 66 Discussion

Actual exam question for Google's Professional Machine Learning Engineer exam
Question #: 66
Topic #: 6
[All Professional Machine Learning Engineer Questions]

You work for a social media company. You want to create a no-code image classification model for an iOS mobile application to identify fashion accessories You have a labeled dataset in Cloud Storage You need to configure a training workflow that minimizes cost and serves predictions with the lowest possible latency What should you do?

Show Suggested Answer Hide Answer
Suggested Answer: D

Applying quantization to your SavedModel by reducing the floating point precision can help reduce the serving latency by decreasing the amount of memory and computation required to make a prediction. TensorFlow provides tools such as the tf.quantization module that can be used to quantize models and reduce their precision, which can significantly reduce serving latency without a significant decrease in model performance.


Contribute your Thoughts:

Kimberely
4 days ago
That's a valid point, but I still think option A provides better scalability and flexibility for future updates in the model.
upvoted 0 times
...
Bulah
5 days ago
I disagree, I believe option B is more suitable as it utilizes AutoML Edge and Core ML model for direct integration with the mobile application.
upvoted 0 times
...
Kimberely
6 days ago
I think option A is the best choice because it involves using AutoML and Vertex AI Model Registry for efficient model training and prediction.
upvoted 0 times
...

Save Cancel