3030
3131from kolena ._api .v1 .event import EventAPI
3232from kolena ._api .v2 .dataset import CommitData
33- from kolena ._api .v2 .dataset import EntityData
33+ from kolena ._api .v2 .dataset import DatasetEntity
3434from kolena ._api .v2 .dataset import Filters
3535from kolena ._api .v2 .dataset import ListCommitHistoryRequest
3636from kolena ._api .v2 .dataset import ListCommitHistoryResponse
@@ -189,7 +189,7 @@ def _upload_dataset_chunk(df: pd.DataFrame, load_uuid: str, id_fields: List[str]
189189 upload_data_frame (df = df_serialized , load_uuid = load_uuid )
190190
191191
192- def _load_dataset_metadata (name : str , raise_error_if_not_found : bool = True ) -> Optional [EntityData ]:
192+ def _load_dataset_metadata (name : str , raise_error_if_not_found : bool = True ) -> Optional [DatasetEntity ]:
193193 """
194194 Load the metadata of a given dataset.
195195
@@ -210,13 +210,13 @@ def _load_dataset_metadata(name: str, raise_error_if_not_found: bool = True) ->
210210 return None
211211 response .raise_for_status ()
212212
213- return from_dict (EntityData , response .json ())
213+ return from_dict (DatasetEntity , response .json ())
214214
215215
216216def _resolve_id_fields (
217217 df : pd .DataFrame ,
218218 id_fields : Optional [List [str ]],
219- existing_dataset : Optional [EntityData ],
219+ existing_dataset : Optional [DatasetEntity ],
220220) -> List [str ]:
221221 existing_id_fields = []
222222 if existing_dataset :
@@ -269,7 +269,7 @@ def _send_upload_dataset_request(
269269 commit_tags : Optional [List [str ]] = None ,
270270 dataset_tags : Optional [List [str ]] = None ,
271271 description : Optional [str ] = None ,
272- ) -> EntityData :
272+ ) -> DatasetEntity :
273273 request = RegisterRequest (
274274 name = name ,
275275 id_fields = id_fields ,
@@ -282,8 +282,8 @@ def _send_upload_dataset_request(
282282 )
283283 response = krequests .post (Path .REGISTER , json = asdict (request ))
284284 krequests .raise_for_status (response )
285- data = from_dict (EntityData , response .json ())
286- return data
285+ dataset_entity = from_dict (DatasetEntity , response .json ())
286+ return dataset_entity
287287
288288
289289def _upload_dataset (
@@ -296,10 +296,10 @@ def _upload_dataset(
296296 commit_tags : Optional [List [str ]] = None ,
297297 dataset_tags : Optional [List [str ]] = None ,
298298 description : Optional [str ] = None ,
299- ) -> None :
299+ ) -> DatasetEntity :
300300 prepared_id_fields , load_uuid = _prepare_upload_dataset_request (name , df , id_fields = id_fields )
301301
302- data = _send_upload_dataset_request (
302+ dataset_entity = _send_upload_dataset_request (
303303 name ,
304304 prepared_id_fields ,
305305 load_uuid ,
@@ -309,7 +309,8 @@ def _upload_dataset(
309309 dataset_tags = dataset_tags ,
310310 description = description ,
311311 )
312- log .info (f"uploaded dataset '{ name } ' ({ get_dataset_url (dataset_id = data .id )} )" )
312+ log .info (f"uploaded dataset '{ name } ' ({ get_dataset_url (dataset_id = dataset_entity .id )} )" )
313+ return dataset_entity
313314
314315
315316@with_event (event_name = EventAPI .Event .REGISTER_DATASET )
@@ -322,7 +323,7 @@ def upload_dataset(
322323 dataset_tags : Optional [List [str ]] = None ,
323324 append_only : bool = False ,
324325 description : Optional [str ] = None ,
325- ) -> None :
326+ ) -> DatasetEntity :
326327 """
327328 Create or update a dataset with the contents of the provided DataFrame `df`.
328329
@@ -343,8 +344,10 @@ def upload_dataset(
343344 datapoints from the input dataframe will be added, and existing datapoints will be modified if present in the
344345 input dataframe, but no datapoints will be deleted from the datasets. This behaves like an `UPSERT` operation.
345346 :param description: Optionally specify the description of the dataset.
347+
348+ :return: The dataset as a [`DatasetEntity`][kolena.dataset.DatasetEntity] object.
346349 """
347- _upload_dataset (
350+ return _upload_dataset (
348351 name ,
349352 df ,
350353 id_fields = id_fields ,
0 commit comments