New Year Sale 2026! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Google Professional Machine Learning Engineer Exam - Topic 6 Question 87 Discussion

Actual exam question for Google's Professional Machine Learning Engineer exam
Question #: 87
Topic #: 6
[All Professional Machine Learning Engineer Questions]

Your data science team needs to rapidly experiment with various features, model architectures, and hyperparameters. They need to track the accuracy metrics for various experiments and use an API to query the metrics over time. What should they use to track and report their experiments while minimizing manual effort?

Show Suggested Answer Hide Answer
Suggested Answer: D

The best option to build a comprehensive system that recommends images to users that are similar in appearance to their own uploaded images is to download a pretrained convolutional neural network (CNN), and use the model to generate embeddings of the input images. Embeddings are low-dimensional representations of high-dimensional data that capture the essential features and semantics of the data. By using a pretrained CNN, you can leverage the knowledge learned from large-scale image datasets, such as ImageNet, and apply it to your own domain. A pretrained CNN can be used as a feature extractor, where the output of the last hidden layer (or any intermediate layer) is taken as the embedding vector for the input image. You can then measure the similarity between embeddings using a distance metric, such as cosine similarity or Euclidean distance, and recommend images that have the highest similarity scores to the user's uploaded image. Option A is incorrect because downloading a pretrained CNN and fine-tuning the model to predict hashtags based on the input images may not capture the visual similarity of the images, as hashtags may not reflect the appearance of the images accurately. For example, two images of different breeds of dogs may have the same hashtag #dog, but they may not look similar to each other. Moreover, fine-tuning the model may require additional data and computational resources, and it may not generalize well to new images that have different or missing hashtags. Option B is incorrect because retrieving image labels and dominant colors from the input images using the Vision API may not capture the visual similarity of the images, as labels and colors may not reflect the fine-grained details of the images. For example, two images of the same breed of dog may have different labels and colors depending on the background, lighting, and angle of the image. Moreover, using the Vision API may incur additional costs and latency, and it may not be able to handle custom or domain-specific labels. Option C is incorrect because using the provided hashtags to create a collaborative filtering algorithm may not capture the visual similarity of the images, as collaborative filtering relies on the ratings or preferences of users, not the features of the images. For example, two images of different animals may have similar ratings or preferences from users, but they may not look similar to each other. Moreover, collaborative filtering may suffer from the cold start problem, where new images or users that have no ratings or preferences cannot be recommended.Reference:

Image similarity search with TensorFlow

Image embeddings documentation

Pretrained models documentation

Similarity metrics documentation


Contribute your Thoughts:

0/2000 characters
Norah
3 months ago
Kubeflow Pipelines (A) is great for automation, though!
upvoted 0 times
...
Domingo
4 months ago
Wait, using Google Sheets? That sounds risky for data accuracy!
upvoted 0 times
...
Shawnda
4 months ago
Not sure about C, Cloud Monitoring feels a bit limited for this.
upvoted 0 times
...
Marica
4 months ago
Definitely agree, B seems more scalable!
upvoted 0 times
...
Peggy
4 months ago
I think option B is solid for tracking metrics in BigQuery.
upvoted 0 times
...
Angelo
4 months ago
I don't think using Google Sheets is ideal for this scenario. It seems too manual compared to the other options we reviewed in class.
upvoted 0 times
...
Silva
5 months ago
I’m a bit confused about Cloud Monitoring versus BigQuery for tracking metrics. I feel like both could work, but I can't recall the specifics.
upvoted 0 times
...
Vannessa
5 months ago
I think using BigQuery with AI Platform Training sounds familiar. We practiced a similar question where we had to store metrics efficiently.
upvoted 0 times
...
Lenita
5 months ago
I remember we discussed using Kubeflow Pipelines for tracking experiments, but I'm not entirely sure if it's the best option for minimizing manual effort.
upvoted 0 times
...
Avery
5 months ago
This is a great question to test our understanding of cloud-based experiment tracking. I feel pretty confident that the Kubeflow Pipelines approach is the way to go - it's designed specifically for this kind of ML experimentation workflow, and the API integration sounds like it will make the reporting really seamless.
upvoted 0 times
...
Kimberely
5 months ago
Okay, I've got a strategy here. I'm going to eliminate the options that require more manual effort, like exporting to a spreadsheet. That leaves the cloud platform options that can integrate the metrics tracking. I think I'll go with the one that uses the most robust API for querying the results over time.
upvoted 0 times
...
Weldon
5 months ago
Hmm, I'm a bit unsure about this one. There are a few different cloud services mentioned, and I'm not super familiar with all of their capabilities. I'll need to think through the pros and cons of each approach to decide which one seems most efficient.
upvoted 0 times
...
Vannessa
5 months ago
This seems like a straightforward question about tracking experiment metrics. I think the key is finding a solution that minimizes manual effort, so I'd focus on the options that leverage APIs to automate the process.
upvoted 0 times
...
Royal
5 months ago
This question seems straightforward, but I want to make sure I understand the key differences between the Hardware and Virtual offerings of Panorama.
upvoted 0 times
...
Blondell
10 months ago
Hold up, did someone say 'Cloud Monitoring'? I thought we were trying to automate things, not create more manual work. Kubeflow Pipelines is the way to go, no doubt about it.
upvoted 0 times
Stefan
9 months ago
C) Yeah, Cloud Monitoring sounds like it would add more manual work. Kubeflow Pipelines is the best option for tracking and reporting experiments.
upvoted 0 times
...
Adaline
9 months ago
B) I agree, Kubeflow Pipelines will definitely help automate the process and save time.
upvoted 0 times
...
Moon
9 months ago
A) Use Kubeflow Pipelines to execute the experiments Export the metrics file, and query the results using the Kubeflow Pipelines API.
upvoted 0 times
...
...
Maryann
10 months ago
A shared Google Sheets file? Really? That's like using a water pistol to put out a forest fire. Kubeflow Pipelines is clearly the superior choice here.
upvoted 0 times
Ben
9 months ago
C) Use Al Platform Training to execute the experiments Write the accuracy metrics to Cloud Monitoring, and query the results using the Monitoring API.
upvoted 0 times
...
Precious
10 months ago
B) Use Al Platform Training to execute the experiments Write the accuracy metrics to BigQuery, and query the results using the BigQueryAPI.
upvoted 0 times
...
Novella
10 months ago
A) Use Kubeflow Pipelines to execute the experiments Export the metrics file, and query the results using the Kubeflow Pipelines API.
upvoted 0 times
...
...
Glory
11 months ago
I'm all about that BigQuery life. Storing the metrics in BigQuery and querying them through the API seems like a solid approach. Plus, I hear it's great for scaling.
upvoted 0 times
Margret
9 months ago
I agree, using BigQuery for storing and querying metrics can really streamline the process.
upvoted 0 times
...
Rupert
9 months ago
B) Use Al Platform Training to execute the experiments Write the accuracy metrics to BigQuery, and query the results using the BigQuery API.
upvoted 0 times
...
Bronwyn
10 months ago
That's a smart choice! BigQuery is definitely great for handling large datasets and scaling.
upvoted 0 times
...
Santos
10 months ago
B) Use Al Platform Training to execute the experiments Write the accuracy metrics to BigQuery, and query the results using the BigQuery API.
upvoted 0 times
...
...
Eun
11 months ago
I prefer using Al Platform Training and writing the accuracy metrics to BigQuery. It's more efficient for our team.
upvoted 0 times
...
Annita
11 months ago
I agree with Nickolas. Kubeflow Pipelines will help us minimize manual effort and easily query the metrics.
upvoted 0 times
...
Kate
11 months ago
I agree, Kubeflow Pipelines is the way to go. Anything that minimizes manual effort is a win in my book. Plus, it just sounds cooler than the other options.
upvoted 0 times
Jerry
10 months ago
User 2: I agree, anything that minimizes manual effort is a win. Kubeflow Pipelines does sound cooler than the other options.
upvoted 0 times
...
Felix
10 months ago
User 1: I think we should use Kubeflow Pipelines for tracking and reporting experiments.
upvoted 0 times
...
...
Mona
11 months ago
Kubeflow Pipelines seems like the right choice here. It's designed for ML experimentation and tracking, and the API makes it easy to query the results.
upvoted 0 times
Daisy
10 months ago
A) Use Kubeflow Pipelines to execute the experiments Export the metrics file, and query the results using the Kubeflow Pipelines API.
upvoted 0 times
...
Haydee
10 months ago
BigQuery is powerful for storing metrics, but Kubeflow Pipelines might be more efficient for tracking.
upvoted 0 times
...
Lauryn
10 months ago
B) Use Al Platform Training to execute the experiments Write the accuracy metrics to BigQuery, and query the results using the BigQueryAPI.
upvoted 0 times
...
Jin
10 months ago
Kubeflow Pipelines is great for tracking experiments and the API is convenient for querying.
upvoted 0 times
...
Billy
10 months ago
A) Use Kubeflow Pipelines to execute the experiments Export the metrics file, and query the results using the Kubeflow Pipelines API.
upvoted 0 times
...
...
Nickolas
11 months ago
I think we should use Kubeflow Pipelines to track and report our experiments.
upvoted 0 times
...

Save Cancel