in

Improved ANN or similarity matching based on vectors with Vertex AI


In order to power these online, mission-critical applications, developers need a reliable service they can trust to be fast and handle the load. For this, we offer vector search capability as part of the Vertex AI Search platform. Vector search (formerly Vertex Matching Engine) finds the most relevant embeddings at scale, blazingly fast. It is based on the same technology that powers core Google services. Today, we’re introducing new features and improvements to make vector search even more useful to developers.

Inside vector search

Vector search’s previously-available features cover a wide range of developer needs and enterprise requirements:

Scales to match your needs
With vector search, developers don’t need to worry about scaling the service up and down; the service auto-scales based on the load. Vector search also enables customization and tunability. For example, developers can easily tune between recall rate and latency, adjusting to match their use case.

Keeps your vector data up to date
Your business data might change over time and vector search can quickly adapt to these changes. With incremental streaming updates, developers don’t have to wait for the entire index to be rebuilt. You can stream your embeddings into vector search and have them ready to query within a few seconds.

Private and secure
Vector search offers enterprise users peace of mind with security and compliance features, such as VPC Service Controls, Customer Managed Encryption Keys (CMEK), and Access Transparency.

These capabilities help meet the security, privacy, and compliance requirements for developers’ mission-critical workloads.

While vector search supports easy-to-use public endpoint deployment, developers can also choose to set up VPC or Private Service Connect (in Public Preview) endpoints, for added data security.

Easy to integrate
Vector search pairs well with other Vertex AI platform offerings. For example, in order to easily build a highly relevant gen AI or search user experience, developers can use LLMs from Vertex AI Model Garden to generate embeddings from their business data, and index them into vector search for fast retrieval.

What’s new: easier to get started and new capabilities

Today, we announce new search features for vector search, and a set of improvements that make it easier for developers to get up and running. With these improvements, Vector search makes it simple to pair LLMs and other embedding foundation models with business data to power fast and relevant user experiences.

  • Vector search UI: With the new UI, now available in Public Preview, developers can get started more easily, as well as monitor index and vector performance. Developers can also create and deploy their indexes directly from the UI – no coding required.
  • Faster to get started: Now generally available, new enhancements reduce the index build time for smaller indexes from hours to minutes, minimizing development friction and getting developers up and running faster.
  • New filtering capabilities: With new filtering capabilities coming soon to Public Preview, app developers can define and filter on numerical range metadata at query time, in addition to the tag-based filtering available today. This improvement opens up support for new vector-based application use cases.

Improved documentation: Improved documentation makes it easier to learn how to get started, explore different capabilities, and follow along with step-by-step examples of building apps with vector search.

Azure OpenAI Series – Azure AI and Business Intelligence: Unlocking Insights

Vertex AI Search adds new generative AI capabilities