You recently deployed a model lo a Vertex Al endpoint and set up online serving in Vertex Al Feature Store. You have configured a daily batch ingestion job to update your featurestore During the batch ingestion jobs you discover that CPU utilization is high in your featurestores online serving nodes and that feature retrieval latency is high. You need to improve online serving performance during the daily batch ingestion. What should you do?
Vertex AI Feature Store provides two options for online serving: Bigtable and optimized online serving. Both options support autoscaling, which means that the number of online serving nodes can automatically adjust to the traffic demand. By enabling autoscaling, you can improve the online serving performance and reduce the feature retrieval latency during the daily batch ingestion. Autoscaling also helps you optimize the cost and resource utilization of your featurestore.Reference:
Online serving | Vertex AI | Google Cloud
New Vertex AI Feature Store: BigQuery-Powered, GenAI-Ready | Google Cloud Blog
Sanjuana
10 months agoSlyvia
10 months agoIsadora
11 months agoLarue
11 months agoJina
1 years agoWalton
11 months agoVi
12 months agoPercy
12 months agoShelba
12 months agoOdette
12 months agoMitzie
12 months agoMaia
1 years agoSherly
1 years agoCorrinne
1 years ago