data_transfer Package
Classes
| DataTransferCopy | Note This is an experimental class, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information. Base class for data transfer copy node. You should not instantiate this class directly. Instead, you should create from builder function: copy_data. | 
| DataTransferCopyComponent | Note This is an experimental class, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information. DataTransfer copy component version, used to define a data transfer copy component. | 
| DataTransferExport | Note This is an experimental class, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information. Base class for data transfer export node. You should not instantiate this class directly. Instead, you should create from builder function: export_data. | 
| DataTransferExportComponent | Note This is an experimental class, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information. DataTransfer export component version, used to define a data transfer export component. | 
| DataTransferImport | Note This is an experimental class, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information. Base class for data transfer import node. You should not instantiate this class directly. Instead, you should create from builder function: import_data. | 
| DataTransferImportComponent | Note This is an experimental class, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information. DataTransfer import component version, used to define a data transfer import component. | 
| Database | Define a database class for a DataTransfer Component or Job. | 
| FileSystem | Define a file system class of a DataTransfer Component or Job. e.g. source_s3 = FileSystem(path='s3://my_bucket/my_folder', connection='azureml:my_s3_connection') | 
Functions
copy_data
Note
This is an experimental method, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information.
Create a DataTransferCopy object which can be used inside dsl.pipeline as a function.
copy_data(*, name: str | None = None, description: str | None = None, tags: Dict | None = None, display_name: str | None = None, experiment_name: str | None = None, compute: str | None = None, inputs: Dict | None = None, outputs: Dict | None = None, is_deterministic: bool = True, data_copy_mode: str | None = None, **kwargs: Any) -> DataTransferCopyKeyword-Only Parameters
| Name | Description | 
|---|---|
| name | The name of the job. Default value: None | 
| description | Description of the job. Default value: None | 
| tags | Tag dictionary. Tags can be added, removed, and updated. Default value: None | 
| display_name | Display name of the job. Default value: None | 
| experiment_name | Name of the experiment the job will be created under. Default value: None | 
| compute | The compute resource the job runs on. Default value: None | 
| inputs | Mapping of inputs data bindings used in the job. Default value: None | 
| outputs | Mapping of outputs data bindings used in the job. Default value: None | 
| is_deterministic | Specify whether the command will return same output given same input. If a command (component) is deterministic, when use it as a node/step in a pipeline, it will reuse results from a previous submitted job in current workspace which has same inputs and settings. In this case, this step will not use any compute resource. Default to be True, specify is_deterministic=False if you would like to avoid such reuse behavior. Default value: True | 
| data_copy_mode | data copy mode in copy task, possible value is "merge_with_overwrite", "fail_if_conflict". Default value: None | 
Returns
| Type | Description | 
|---|---|
| A DataTransferCopy object. | 
export_data
Note
This is an experimental method, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information.
Create a DataTransferExport object which can be used inside dsl.pipeline.
export_data(*, name: str | None = None, description: str | None = None, tags: Dict | None = None, display_name: str | None = None, experiment_name: str | None = None, compute: str | None = None, sink: Dict | Database | FileSystem | None = None, inputs: Dict | None = None, **kwargs: Any) -> DataTransferExportKeyword-Only Parameters
| Name | Description | 
|---|---|
| name | The name of the job. Default value: None | 
| description | Description of the job. Default value: None | 
| tags | Tag dictionary. Tags can be added, removed, and updated. Default value: None | 
| display_name | Display name of the job. Default value: None | 
| experiment_name | Name of the experiment the job will be created under. Default value: None | 
| compute | The compute resource the job runs on. Default value: None | 
| sink | The sink of external data and databases. Default value: None | 
| inputs | Mapping of inputs data bindings used in the job. Default value: None | 
Returns
| Type | Description | 
|---|---|
| 
							<xref:azure.ai.ml.entities._job.pipeline._component_translatable.DataTransferExport>
						 | A DataTransferExport object. | 
Exceptions
| Type | Description | 
|---|---|
| If sink is not provided or exporting file system is not supported. | 
import_data
Note
This is an experimental method, and may change at any time. Please see https://aka.ms/azuremlexperimental for more information.
Create a DataTransferImport object which can be used inside dsl.pipeline.
import_data(*, name: str | None = None, description: str | None = None, tags: Dict | None = None, display_name: str | None = None, experiment_name: str | None = None, compute: str | None = None, source: Dict | Database | FileSystem | None = None, outputs: Dict | None = None, **kwargs: Any) -> DataTransferImportKeyword-Only Parameters
| Name | Description | 
|---|---|
| name | The name of the job. Default value: None | 
| description | Description of the job. Default value: None | 
| tags | Tag dictionary. Tags can be added, removed, and updated. Default value: None | 
| display_name | Display name of the job. Default value: None | 
| experiment_name | Name of the experiment the job will be created under. Default value: None | 
| compute | The compute resource the job runs on. Default value: None | 
| source | The data source of file system or database. Default value: None | 
| outputs | Mapping of outputs data bindings used in the job. The default will be an output port with the key "sink" and type "mltable". Default value: None | 
Returns
| Type | Description | 
|---|---|
| 
							<xref:azure.ai.ml.entities._job.pipeline._component_translatable.DataTransferImport>
						 | A DataTransferImport object. |