Large Language Model (LLM)-driven grounding methodology with Google Cloud Platform (GCP)

Grounding Case Study with PaLM 2
Text Bison & Vertex AI search

Grounding stands out by creating a dynamic link between models and external data, ensuring real-time adaptability and reducing inaccuracies.

Grounding - Executive Summary

Grounding stands out by creating a dynamic link between models and external data, ensuring real-time adaptability and reducing inaccuracies. This strategy is emphasized in the project documentation to enhance the precision and reliability of language models, especially through the integration of Vertex AI Search and Conversation.

The grounding strategy focuses on text-bison and chat-bison models, anchoring generated content in designated data stores within Vertex AI Search. This approach mitigates model hallucinations, secures anchored responses, and increases the credibility of the generated content.

Grounding transforms AI models, making them more focused, accurate, and efficient by reducing hallucinations, anchoring responses, and enhancing trustworthiness. Grounded models differ from non-grounded ones in data dependence, response accuracy, and adaptability.

Implementing grounding involves activating Vertex AI Search, creating a data store, and grounding the model using the API. Regular monitoring and iteration to enhance grounding, and ensure compliance with data use terms and security standards are vital considerations.

In conclusion, integrating grounding in Vertex AI highlights the commitment to delivering accurate, relevant, and trustworthy content. The documentation serves as a comprehensive guide for implementing, monitoring, and optimizing grounding features, aligning with the broader goal of providing high-quality, reliable, and innovative solutions in the dynamic world of generative AI projects.

Steps for Crafting Grounding:

Utilizing Vertex AI Search App to Ground Model Responses:

Vertex AI Search and Conversation lets developers tap into the power of Google’s foundation models, search expertise, and conversational AI technologies to create enterprise grade generative AI applications.

Three types of applications can be created from the service
1. Search – an LLM-powered service that enables it to give a Google-level search experience from random data.

2. Chat – this service is empowered with chatbots and apps (search apps) to have a dialog flow with the end user.

3. Recommendations – this system recommends media (related videos, images, news) to have personalized content and generic recommendations to recommend non-media content.
The Vertex AI Search enables 4 different data sources to crawl and retrieve the data:
  • Website URLs
  • BigQuery table
  • Cloud Storage (Cloud Buckets)
  • API – calling an API to import the data manually.

Grounding ‘text-bison’ to website URLs

Creating Search application

Three types of applications can be created from the service
Note : Adding ‘ /* ’ at the end of the main page URL allows the service to navigate to all the web pages available in it.

Pointing data store to the model

Fill the path in the format:
projects/{project_id}/locations/global/collections/default_collection/dataStores/{data_store_id}

Get the data store ID from the created search application.