| # Vertex Generative AI SDK for Python | |
| The Vertex Generative AI SDK helps developers use Google's generative AI | |
| [Gemini models](http://cloud.google.com/vertex-ai/docs/generative-ai/multimodal/overview) | |
| to build AI-powered features and applications. | |
| The SDKs support use cases like the following: | |
| - Generate text from texts, images and videos (multimodal generation) | |
| - Build stateful multi-turn conversations (chat) | |
| - Function calling | |
| ## Installation | |
| To install the | |
| [google-cloud-aiplatform](https://pypi.org/project/google-cloud-aiplatform/) | |
| Python package, run the following command: | |
| ```shell | |
| pip3 install --upgrade --user "google-cloud-aiplatform>=1.38" | |
| ``` | |
| ## Usage | |
| For detailed instructions, see [quickstart](http://cloud.google.com/vertex-ai/docs/generative-ai/start/quickstarts/quickstart-multimodal) and [Introduction to multimodal classes in the Vertex AI SDK](http://cloud.google.com/vertex-ai/docs/generative-ai/multimodal/sdk-for-gemini/gemini-sdk-overview-reference). | |
| #### Imports: | |
| ```python | |
| import vertexai | |
| ``` | |
| #### Initialization: | |
| ```python | |
| vertexai.init(project='my-project', location='us-central1') | |
| ``` | |
| #### Basic generation: | |
| ```python | |
| from vertexai.generative_models import GenerativeModel | |
| model = GenerativeModel("gemini-pro") | |
| print(model.generate_content("Why is sky blue?")) | |
| ``` | |
| #### Using images and videos | |
| ```python | |
| from vertexai.generative_models import GenerativeModel, Image | |
| vision_model = GenerativeModel("gemini-pro-vision") | |
| # Local image | |
| image = Image.load_from_file("image.jpg") | |
| print(vision_model.generate_content(["What is shown in this image?", image])) | |
| # Image from Cloud Storage | |
| image_part = generative_models.Part.from_uri("gs://download.tensorflow.org/example_images/320px-Felis_catus-cat_on_snow.jpg", mime_type="image/jpeg") | |
| print(vision_model.generate_content([image_part, "Describe this image?"])) | |
| # Text and video | |
| video_part = Part.from_uri("gs://cloud-samples-data/video/animals.mp4", mime_type="video/mp4") | |
| print(vision_model.generate_content(["What is in the video? ", video_part])) | |
| ``` | |
| #### Chat | |
| ```python | |
| from vertexai.generative_models import GenerativeModel, Image | |
| vision_model = GenerativeModel("gemini-ultra-vision") | |
| vision_chat = vision_model.start_chat() | |
| image = Image.load_from_file("image.jpg") | |
| print(vision_chat.send_message(["I like this image.", image])) | |
| print(vision_chat.send_message("What things do I like?.")) | |
| ``` | |
| #### System instructions | |
| ```python | |
| from vertexai.generative_models import GenerativeModel | |
| model = GenerativeModel( | |
| "gemini-1.0-pro", | |
| system_instruction=[ | |
| "Talk like a pirate.", | |
| "Don't use rude words.", | |
| ], | |
| ) | |
| print(model.generate_content("Why is sky blue?")) | |
| ``` | |
| #### Function calling | |
| ```python | |
| # First, create tools that the model is can use to answer your questions. | |
| # Describe a function by specifying it's schema (JsonSchema format) | |
| get_current_weather_func = generative_models.FunctionDeclaration( | |
| name="get_current_weather", | |
| description="Get the current weather in a given location", | |
| parameters={ | |
| "type": "object", | |
| "properties": { | |
| "location": { | |
| "type": "string", | |
| "description": "The city and state, e.g. San Francisco, CA" | |
| }, | |
| "unit": { | |
| "type": "string", | |
| "enum": [ | |
| "celsius", | |
| "fahrenheit", | |
| ] | |
| } | |
| }, | |
| "required": [ | |
| "location" | |
| ] | |
| }, | |
| ) | |
| # Tool is a collection of related functions | |
| weather_tool = generative_models.Tool( | |
| function_declarations=[get_current_weather_func], | |
| ) | |
| # Use tools in chat: | |
| model = GenerativeModel( | |
| "gemini-pro", | |
| # You can specify tools when creating a model to avoid having to send them with every request. | |
| tools=[weather_tool], | |
| ) | |
| chat = model.start_chat() | |
| # Send a message to the model. The model will respond with a function call. | |
| print(chat.send_message("What is the weather like in Boston?")) | |
| # Then send a function response to the model. The model will use it to answer. | |
| print(chat.send_message( | |
| Part.from_function_response( | |
| name="get_current_weather", | |
| response={ | |
| "content": {"weather": "super nice"}, | |
| } | |
| ), | |
| )) | |
| ``` | |
| #### Automatic Function calling | |
| Note: The `FunctionDeclaration.from_func` converter does not support nested types for parameters. Please provide full `FunctionDeclaration` instead. | |
| ```python | |
| from vertexai.preview.generative_models import GenerativeModel, Tool, FunctionDeclaration, AutomaticFunctionCallingResponder | |
| # First, create functions that the model can use to answer your questions. | |
| def get_current_weather(location: str, unit: str = "centigrade"): | |
| """Gets weather in the specified location. | |
| Args: | |
| location: The location for which to get the weather. | |
| unit: Optional. Temperature unit. Can be Centigrade or Fahrenheit. Defaults to Centigrade. | |
| """ | |
| return dict( | |
| location=location, | |
| unit=unit, | |
| weather="Super nice, but maybe a bit hot.", | |
| ) | |
| # Infer function schema | |
| get_current_weather_func = FunctionDeclaration.from_func(get_current_weather) | |
| # Tool is a collection of related functions | |
| weather_tool = Tool( | |
| function_declarations=[get_current_weather_func], | |
| ) | |
| # Use tools in chat: | |
| model = GenerativeModel( | |
| "gemini-pro", | |
| # You can specify tools when creating a model to avoid having to send them with every request. | |
| tools=[weather_tool], | |
| ) | |
| # Activate automatic function calling: | |
| afc_responder = AutomaticFunctionCallingResponder( | |
| # Optional: | |
| max_automatic_function_calls=5, | |
| ) | |
| chat = model.start_chat(responder=afc_responder) | |
| # Send a message to the model. The model will respond with a function call. | |
| # The SDK will automatically call the requested function and respond to the model. | |
| # The model will use the function call response to answer the original question. | |
| print(chat.send_message("What is the weather like in Boston?")) | |
| ``` | |
| #### Evaluation | |
| - To perform bring-your-own-response(BYOR) evaluation, provide the model responses in the `response` column in the dataset. If a pairwise metric is used for BYOR evaluation, provide the baseline model responses in the `baseline_model_response` column. | |
| ```python | |
| import pandas as pd | |
| from vertexai.evaluation import EvalTask, MetricPromptTemplateExamples | |
| eval_dataset = pd.DataFrame({ | |
| "prompt" : [...], | |
| "reference": [...], | |
| "response" : [...], | |
| "baseline_model_response": [...], | |
| }) | |
| eval_task = EvalTask( | |
| dataset=eval_dataset, | |
| metrics=[ | |
| "bleu", | |
| "rouge_l_sum", | |
| MetricPromptTemplateExamples.Pointwise.FLUENCY, | |
| MetricPromptTemplateExamples.Pairwise.SAFETY | |
| ], | |
| experiment="my-experiment", | |
| ) | |
| eval_result = eval_task.evaluate(experiment_run_name="eval-experiment-run") | |
| ``` | |
| - To perform evaluation with Gemini model inference, specify the `model` parameter with a `GenerativeModel` instance. The input column name to the model is `prompt` and must be present in the dataset. | |
| ```python | |
| from vertexai.evaluation import EvalTask | |
| from vertexai.generative_models import GenerativeModel | |
| eval_dataset = pd.DataFrame({ | |
| "reference": [...], | |
| "prompt" : [...], | |
| }) | |
| result = EvalTask( | |
| dataset=eval_dataset, | |
| metrics=["exact_match", "bleu", "rouge_1", "rouge_l_sum"], | |
| experiment="my-experiment", | |
| ).evaluate( | |
| model=GenerativeModel("gemini-1.5-pro"), | |
| experiment_run_name="gemini-eval-run" | |
| ) | |
| ``` | |
| - If a `prompt_template` is specified, the `prompt` column is not required. Prompts can be assembled from the evaluation dataset, and all prompt template variable names must be present in the dataset columns. | |
| ```python | |
| import pandas as pd | |
| from vertexai.evaluation import EvalTask, MetricPromptTemplateExamples | |
| from vertexai.generative_models import GenerativeModel | |
| eval_dataset = pd.DataFrame({ | |
| "context" : [...], | |
| "instruction": [...], | |
| }) | |
| result = EvalTask( | |
| dataset=eval_dataset, | |
| metrics=[MetricPromptTemplateExamples.Pointwise.SUMMARIZATION_QUALITY], | |
| ).evaluate( | |
| model=GenerativeModel("gemini-1.5-pro"), | |
| prompt_template="{instruction}. Article: {context}. Summary:", | |
| ) | |
| ``` | |
| - To perform evaluation with custom model inference, specify the `model` | |
| parameter with a custom inference function. The input column name to the | |
| custom inference function is `prompt` and must be present in the dataset. | |
| ```python | |
| from openai import OpenAI | |
| from vertexai.evaluation import EvalTask, MetricPromptTemplateExamples | |
| client = OpenAI() | |
| def custom_model_fn(input: str) -> str: | |
| response = client.chat.completions.create( | |
| model="gpt-3.5-turbo", | |
| messages=[ | |
| {"role": "user", "content": input} | |
| ] | |
| ) | |
| return response.choices[0].message.content | |
| eval_dataset = pd.DataFrame({ | |
| "prompt" : [...], | |
| "reference": [...], | |
| }) | |
| result = EvalTask( | |
| dataset=eval_dataset, | |
| metrics=[MetricPromptTemplateExamples.Pointwise.SAFETY], | |
| experiment="my-experiment", | |
| ).evaluate( | |
| model=custom_model_fn, | |
| experiment_run_name="gpt-eval-run" | |
| ) | |
| ``` | |
| - To perform pairwise metric evaluation with model inference step, specify | |
| the `baseline_model` input to a `PairwiseMetric` instance and the candidate | |
| `model` input to the `EvalTask.evaluate()` function. The input column name | |
| to both models is `prompt` and must be present in the dataset. | |
| ```python | |
| import pandas as pd | |
| from vertexai.evaluation import EvalTask, MetricPromptTemplateExamples, PairwiseMetric | |
| from vertexai.generative_models import GenerativeModel | |
| baseline_model = GenerativeModel("gemini-1.0-pro") | |
| candidate_model = GenerativeModel("gemini-1.5-pro") | |
| pairwise_groundedness = PairwiseMetric( | |
| metric_prompt_template=MetricPromptTemplateExamples.get_prompt_template( | |
| "pairwise_groundedness" | |
| ), | |
| baseline_model=baseline_model, | |
| ) | |
| eval_dataset = pd.DataFrame({ | |
| "prompt" : [...], | |
| }) | |
| result = EvalTask( | |
| dataset=eval_dataset, | |
| metrics=[pairwise_groundedness], | |
| experiment="my-pairwise-experiment", | |
| ).evaluate( | |
| model=candidate_model, | |
| experiment_run_name="gemini-pairwise-eval-run", | |
| ) | |
| ``` | |
| ## Documentation | |
| You can find complete documentation for the Vertex AI SDKs and the Gemini model in the Google Cloud [documentation](https://cloud.google.com/vertex-ai/docs/generative-ai/learn/overview) | |
| ## Contributing | |
| See [Contributing](https://github.com/googleapis/python-aiplatform/blob/main/CONTRIBUTING.rst) for more information on contributing to the Vertex AI Python SDK. | |
| ## License | |
| The contents of this repository are licensed under the [Apache License, version 2.0](http://www.apache.org/licenses/LICENSE-2.0). |