Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
"gs://google-cloud-aiplatform/schema/dataset/metadata/multimodal_1.0.0.yaml"
)
BIGQUERY_TABLE_NAME = "vertex-sdk-dev.multimodal_dataset.test-table"
DATASET = "8810841321427173376"
DATASET = "projects/vertex-sdk-dev/locations/us-central1/datasets/8810841321427173376"


def test_assemble_dataset(client):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
"gs://google-cloud-aiplatform/schema/dataset/metadata/multimodal_1.0.0.yaml"
)
BIGQUERY_TABLE_NAME = "vertex-sdk-dev.multimodal_dataset.test-table"
DATASET = "8810841321427173376"
DATASET = "projects/vertex-sdk-dev/locations/us-central1/datasets/8810841321427173376"


def test_assess_dataset(client):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -33,10 +33,9 @@ def test_delete_dataset(client):
},
}
)
name = dataset.name.split("/datasets/")[1]

operation = client.datasets._delete_multimodal_dataset(
name=name,
name=dataset.name,
)
assert isinstance(operation, types.MultimodalDatasetOperation)
assert operation.done
Expand All @@ -53,10 +52,9 @@ def test_delete_dataset_with_public_method(client):
},
}
)
name = dataset.name.split("/datasets/")[1]

operation = client.datasets.delete_multimodal_dataset(
name=name,
name=dataset.name,
)
assert isinstance(operation, types.MultimodalDatasetOperation)
assert operation.done
Expand All @@ -82,13 +80,12 @@ async def test_delete_dataset_async(client):
},
}
)
name = dataset.name.split("/datasets/")[1]

operation = await client.aio.datasets._delete_multimodal_dataset(
name=name,
name=dataset.name,
)
assert isinstance(operation, types.MultimodalDatasetOperation)
assert operation
assert operation.done


@pytest.mark.asyncio
Expand All @@ -103,10 +100,9 @@ async def test_delete_dataset_with_public_method_async(client):
},
}
)
name = dataset.name.split("/datasets/")[1]

operation = await client.aio.datasets.delete_multimodal_dataset(
name=name,
name=dataset.name,
)
assert isinstance(operation, types.MultimodalDatasetOperation)
assert operation.done
Original file line number Diff line number Diff line change
Expand Up @@ -20,15 +20,15 @@
import pytest

BIGQUERY_TABLE_NAME = "vertex-sdk-dev.multimodal_dataset.test-table"
DATASET = "8810841321427173376"
DATASET = "projects/964831358985/locations/us-central1/datasets/8810841321427173376"


def test_get_dataset(client):
dataset = client.datasets._get_multimodal_dataset(
name=DATASET,
)
assert isinstance(dataset, types.MultimodalDataset)
assert dataset.name.endswith(DATASET)
assert dataset.name == DATASET
assert dataset.display_name == "test-display-name"


Expand All @@ -37,7 +37,7 @@ def test_get_dataset_from_public_method(client):
name=DATASET,
)
assert isinstance(dataset, types.MultimodalDataset)
assert dataset.name.endswith(DATASET)
assert dataset.name == DATASET
assert dataset.display_name == "test-display-name"


Expand All @@ -55,7 +55,7 @@ async def test_get_dataset_async(client):
name=DATASET,
)
assert isinstance(dataset, types.MultimodalDataset)
assert dataset.name.endswith(DATASET)
assert dataset.name == DATASET
assert dataset.display_name == "test-display-name"


Expand All @@ -65,5 +65,5 @@ async def test_get_dataset_from_public_method_async(client):
name=DATASET,
)
assert isinstance(dataset, types.MultimodalDataset)
assert dataset.name.endswith(DATASET)
assert dataset.name == DATASET
assert dataset.display_name == "test-display-name"
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
"gs://google-cloud-aiplatform/schema/dataset/metadata/multimodal_1.0.0.yaml"
)
BIGQUERY_TABLE_NAME = "vertex-sdk-dev.multimodal_dataset.test-table"
DATASET = "8810841321427173376"
DATASET = "projects/vertex-sdk-dev/locations/us-central1/datasets/8810841321427173376"


def test_update_dataset(client):
Expand Down
40 changes: 20 additions & 20 deletions vertexai/_genai/datasets.py
Original file line number Diff line number Diff line change
Expand Up @@ -274,9 +274,9 @@ def _assemble_multimodal_dataset(
request_dict = _AssembleDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}:assemble".format_map(request_url_dict)
path = "{name}:assemble".format_map(request_url_dict)
else:
path = "datasets/{name}:assemble"
path = "{name}:assemble"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -364,9 +364,9 @@ def _assess_multimodal_dataset(
request_dict = _AssessDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}:assess".format_map(request_url_dict)
path = "{name}:assess".format_map(request_url_dict)
else:
path = "datasets/{name}:assess"
path = "{name}:assess"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -513,9 +513,9 @@ def _delete_multimodal_dataset(
)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}".format_map(request_url_dict)
path = "{name}".format_map(request_url_dict)
else:
path = "datasets/{name}"
path = "{name}"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -583,9 +583,9 @@ def _get_multimodal_dataset(
request_dict = _GetMultimodalDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}".format_map(request_url_dict)
path = "{name}".format_map(request_url_dict)
else:
path = "datasets/{name}"
path = "{name}"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -805,9 +805,9 @@ def _update_multimodal_dataset(
request_dict = _UpdateMultimodalDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}".format_map(request_url_dict)
path = "{name}".format_map(request_url_dict)
else:
path = "datasets/{name}"
path = "{name}"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -1475,9 +1475,9 @@ async def _assemble_multimodal_dataset(
request_dict = _AssembleDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}:assemble".format_map(request_url_dict)
path = "{name}:assemble".format_map(request_url_dict)
else:
path = "datasets/{name}:assemble"
path = "{name}:assemble"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -1567,9 +1567,9 @@ async def _assess_multimodal_dataset(
request_dict = _AssessDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}:assess".format_map(request_url_dict)
path = "{name}:assess".format_map(request_url_dict)
else:
path = "datasets/{name}:assess"
path = "{name}:assess"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -1720,9 +1720,9 @@ async def _delete_multimodal_dataset(
)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}".format_map(request_url_dict)
path = "{name}".format_map(request_url_dict)
else:
path = "datasets/{name}"
path = "{name}"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -1792,9 +1792,9 @@ async def _get_multimodal_dataset(
request_dict = _GetMultimodalDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}".format_map(request_url_dict)
path = "{name}".format_map(request_url_dict)
else:
path = "datasets/{name}"
path = "{name}"

query_params = request_dict.get("_query")
if query_params:
Expand Down Expand Up @@ -2020,9 +2020,9 @@ async def _update_multimodal_dataset(
request_dict = _UpdateMultimodalDatasetParameters_to_vertex(parameter_model)
request_url_dict = request_dict.get("_url")
if request_url_dict:
path = "datasets/{name}".format_map(request_url_dict)
path = "{name}".format_map(request_url_dict)
else:
path = "datasets/{name}"
path = "{name}"

query_params = request_dict.get("_query")
if query_params:
Expand Down
Loading