Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions tests/system/aiplatform/test_dataset.py
Original file line number Diff line number Diff line change
Expand Up @@ -196,6 +196,9 @@ def test_get_nonexistent_dataset(self):
with pytest.raises(exceptions.NotFound):
aiplatform.ImageDataset(dataset_name="0")

@pytest.mark.skip(
reason="Creating a Text Dataset/training AutoML TEXT models is no longer supported"
)
def test_get_new_dataset_and_import(self, dataset_gapic_client):
"""Retrieve new, empty dataset and import a text dataset using import().
Then verify data items were successfully imported."""
Expand Down
2 changes: 1 addition & 1 deletion tests/system/aiplatform/test_telemetry.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@

from vertexai.generative_models import GenerativeModel

GEMINI_MODEL_NAME = "gemini-1.5-pro-002"
GEMINI_MODEL_NAME = "gemini-2.5-flash"


class TestTelemetry(e2e_base.TestEndToEnd):
Expand Down
10 changes: 6 additions & 4 deletions tests/system/vertexai/test_prompts.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,6 +30,8 @@
from tests.system.aiplatform import e2e_base
from google import auth

MODEL_NAME = "gemini-2.5-flash"

_REQUEST_FUNCTION_PARAMETER_SCHEMA_STRUCT = {
"type": "object",
"properties": {
Expand Down Expand Up @@ -74,7 +76,7 @@ def test_create_prompt_with_variables(self):
{"name": "Bob", "day": "Tuesday"},
],
generation_config=GenerationConfig(temperature=0.1),
model_name="gemini-1.5-pro-002",
model_name=MODEL_NAME,
safety_settings=[
SafetySetting(
category=SafetySetting.HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT,
Expand Down Expand Up @@ -140,11 +142,11 @@ def test_create_prompt_with_function_calling(self):
prompt_data="What is the weather like in Boston?",
tools=[weather_tool],
tool_config=tool_config,
model_name="gemini-1.5-pro-002",
model_name=MODEL_NAME,
)

# (Optional) Create a separate prompt resource to save the version to
prompt_temp = Prompt(model_name="gemini-1.5-pro-002")
prompt_temp = Prompt(model_name=MODEL_NAME)
prompt_temp1 = prompts.create_version(prompt=prompt_temp, version_name="empty")

# Create a new version to an existing prompt
Expand Down Expand Up @@ -190,6 +192,6 @@ def test_get_prompt_with_function_calling(self):

# Generate content using the prompt
response = prompt.generate_content(
model_name="gemini-1.5-pro-002", contents=prompt.assemble_contents()
model_name=MODEL_NAME, contents=prompt.assemble_contents()
)
assert response
1 change: 1 addition & 0 deletions tests/system/vertexai/test_tokenization.py
Original file line number Diff line number Diff line change
Expand Up @@ -107,6 +107,7 @@


@pytest.mark.parametrize("api_endpoint_env_name", [PROD_API_ENDPOINT])
@pytest.mark.skip(reason="Models are deprecated.")
class TestTokenization(e2e_base.TestEndToEnd):
"""System tests for tokenization."""

Expand Down
Loading