diff --git a/tests/system/aiplatform/test_dataset.py b/tests/system/aiplatform/test_dataset.py index b902d54766..ba5ff69183 100644 --- a/tests/system/aiplatform/test_dataset.py +++ b/tests/system/aiplatform/test_dataset.py @@ -196,6 +196,9 @@ def test_get_nonexistent_dataset(self): with pytest.raises(exceptions.NotFound): aiplatform.ImageDataset(dataset_name="0") + @pytest.mark.skip( + reason="Creating a Text Dataset/training AutoML TEXT models is no longer supported" + ) def test_get_new_dataset_and_import(self, dataset_gapic_client): """Retrieve new, empty dataset and import a text dataset using import(). Then verify data items were successfully imported.""" diff --git a/tests/system/aiplatform/test_telemetry.py b/tests/system/aiplatform/test_telemetry.py index b992c025a7..4bb6a05193 100644 --- a/tests/system/aiplatform/test_telemetry.py +++ b/tests/system/aiplatform/test_telemetry.py @@ -23,7 +23,7 @@ from vertexai.generative_models import GenerativeModel -GEMINI_MODEL_NAME = "gemini-1.5-pro-002" +GEMINI_MODEL_NAME = "gemini-2.5-flash" class TestTelemetry(e2e_base.TestEndToEnd): diff --git a/tests/system/vertexai/test_prompts.py b/tests/system/vertexai/test_prompts.py index fa9ad6a8b6..60c1da0047 100644 --- a/tests/system/vertexai/test_prompts.py +++ b/tests/system/vertexai/test_prompts.py @@ -30,6 +30,8 @@ from tests.system.aiplatform import e2e_base from google import auth +MODEL_NAME = "gemini-2.5-flash" + _REQUEST_FUNCTION_PARAMETER_SCHEMA_STRUCT = { "type": "object", "properties": { @@ -74,7 +76,7 @@ def test_create_prompt_with_variables(self): {"name": "Bob", "day": "Tuesday"}, ], generation_config=GenerationConfig(temperature=0.1), - model_name="gemini-1.5-pro-002", + model_name=MODEL_NAME, safety_settings=[ SafetySetting( category=SafetySetting.HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT, @@ -140,11 +142,11 @@ def test_create_prompt_with_function_calling(self): prompt_data="What is the weather like in Boston?", tools=[weather_tool], tool_config=tool_config, - model_name="gemini-1.5-pro-002", + model_name=MODEL_NAME, ) # (Optional) Create a separate prompt resource to save the version to - prompt_temp = Prompt(model_name="gemini-1.5-pro-002") + prompt_temp = Prompt(model_name=MODEL_NAME) prompt_temp1 = prompts.create_version(prompt=prompt_temp, version_name="empty") # Create a new version to an existing prompt @@ -190,6 +192,6 @@ def test_get_prompt_with_function_calling(self): # Generate content using the prompt response = prompt.generate_content( - model_name="gemini-1.5-pro-002", contents=prompt.assemble_contents() + model_name=MODEL_NAME, contents=prompt.assemble_contents() ) assert response diff --git a/tests/system/vertexai/test_tokenization.py b/tests/system/vertexai/test_tokenization.py index d586e53fd1..098da28d21 100644 --- a/tests/system/vertexai/test_tokenization.py +++ b/tests/system/vertexai/test_tokenization.py @@ -107,6 +107,7 @@ @pytest.mark.parametrize("api_endpoint_env_name", [PROD_API_ENDPOINT]) +@pytest.mark.skip(reason="Models are deprecated.") class TestTokenization(e2e_base.TestEndToEnd): """System tests for tokenization."""