A B C D E F G H I K L M N P Q R S T U V W
| aci_webservice_deployment_config | Create a deployment config for deploying an ACI web service |
| aks_webservice_deployment_config | Create a deployment config for deploying an AKS web service |
| attach_aks_compute | Attach an existing AKS cluster to a workspace |
| azureml | azureml module User can access functions/modules in azureml that are not exposed through the exported R functions. |
| bandit_policy | Define a Bandit policy for early termination of HyperDrive runs |
| bayesian_parameter_sampling | Define Bayesian sampling over a hyperparameter search space |
| cancel_run | Cancel a run |
| choice | Specify a discrete set of options to sample from |
| complete_run | Mark a run as completed. |
| container_registry | Specify Azure Container Registry details |
| convert_to_dataset_with_csv_files | Convert the current dataset into a FileDataset containing CSV files. |
| convert_to_dataset_with_parquet_files | Convert the current dataset into a FileDataset containing Parquet files. |
| cran_package | Specifies a CRAN package to install in environment |
| create_aks_compute | Create an AksCompute cluster |
| create_aml_compute | Create an AmlCompute cluster |
| create_child_run | Create a child run |
| create_child_runs | Create one or many child runs |
| create_file_dataset_from_files | Create a FileDataset to represent file streams. |
| create_tabular_dataset_from_delimited_files | Create an unregistered, in-memory Dataset from delimited files. |
| create_tabular_dataset_from_json_lines_files | Create a TabularDataset to represent tabular data in JSON Lines files (http://jsonlines.org/). |
| create_tabular_dataset_from_parquet_files | Create an unregistered, in-memory Dataset from parquet files. |
| create_tabular_dataset_from_sql_query | Create a TabularDataset to represent tabular data in SQL databases. |
| create_workspace | Create a new Azure Machine Learning workspace |
| dataset_consumption_config | Represent how to deliver the dataset to a compute target. |
| data_path | Represents a path to data in a datastore. |
| data_type_bool | Configure conversion to bool. |
| data_type_datetime | Configure conversion to datetime. |
| data_type_double | Configure conversion to 53-bit double. |
| data_type_long | Configure conversion to 64-bit integer. |
| data_type_string | Configure conversion to string. |
| define_timestamp_columns_for_dataset | Define timestamp columns for the dataset. |
| delete_compute | Delete a cluster |
| delete_local_webservice | Delete a local web service from the local machine |
| delete_model | Delete a model from its associated workspace |
| delete_secrets | Delete secrets from a keyvault |
| delete_webservice | Delete a web service from a given workspace |
| delete_workspace | Delete a workspace |
| deploy_model | Deploy a web service from registered model(s) |
| detach_aks_compute | Detach an AksCompute cluster from its associated workspace |
| download_files_from_run | Download files from a run |
| download_file_from_run | Download a file from a run |
| download_from_datastore | Download data from a datastore to the local file system |
| download_from_file_dataset | Download file streams defined by the dataset as local files. |
| download_model | Download a model to the local file system |
| drop_columns_from_dataset | Drop the specified columns from the dataset. |
| estimator | Create an estimator |
| experiment | Create an Azure Machine Learning experiment |
| filter_dataset_after_time | Filter Tabular Dataset with time stamp columns after a specified start time. |
| filter_dataset_before_time | Filter Tabular Dataset with time stamp columns before a specified end time. |
| filter_dataset_between_time | Filter Tabular Dataset between a specified start and end time. |
| filter_dataset_from_recent_time | Filter Tabular Dataset to contain only the specified duration (amount) of recent data. |
| generate_entry_script | Generates the control script for the experiment. |
| generate_new_webservice_key | Regenerate one of a web service's keys |
| get_aks_compute_credentials | Get the credentials for an AksCompute cluster |
| get_best_run_by_primary_metric | Return the best performing run amongst all completed runs |
| get_child_runs | Get all children for the current run selected by specified filters |
| get_child_runs_sorted_by_primary_metric | Get the child runs sorted in descending order by best primary metric |
| get_child_run_hyperparameters | Get the hyperparameters for all child runs |
| get_child_run_metrics | Get the metrics from all child runs |
| get_compute | Get an existing compute cluster |
| get_current_run | Get the context object for a run |
| get_dataset_by_id | Get Dataset by ID. |
| get_dataset_by_name | Get a registered Dataset from the workspace by its registration name. |
| get_datastore | Get an existing datastore |
| get_default_datastore | Get the default datastore for a workspace |
| get_default_keyvault | Get the default keyvault for a workspace |
| get_environment | Get an existing environment |
| get_file_dataset_paths | Get a list of file paths for each file stream defined by the dataset. |
| get_input_dataset_from_run | Return the named list for input datasets. |
| get_model | Get a registered model |
| get_model_package_container_registry | Get the Azure container registry that a packaged model uses |
| get_model_package_creation_logs | Get the model package creation logs |
| get_run | Get an experiment run |
| get_runs_in_experiment | Return a generator of the runs for an experiment |
| get_run_details | Get the details of a run |
| get_run_details_with_logs | Get the details of a run along with the log files' contents |
| get_run_file_names | List the files that are stored in association with a run |
| get_run_metrics | Get the metrics logged to a run |
| get_secrets | Get secrets from a keyvault |
| get_secrets_from_run | Get secrets from the keyvault associated with a run's workspace |
| get_webservice | Get a deployed web service |
| get_webservice_keys | Retrieve auth keys for a web service |
| get_webservice_logs | Retrieve the logs for a web service |
| get_webservice_token | Retrieve the auth token for a web service |
| get_workspace | Get an existing workspace |
| get_workspace_details | Get the details of a workspace |
| github_package | Specifies a Github package to install in environment |
| grid_parameter_sampling | Define grid sampling over a hyperparameter search space |
| hyperdrive_config | Create a configuration for a HyperDrive run |
| inference_config | Create an inference configuration for model deployments |
| install_azureml | Install azureml sdk package |
| interactive_login_authentication | Manages authentication and acquires an authorization token in interactive login workflows. |
| invoke_webservice | Call a web service with the provided input |
| keep_columns_from_dataset | Keep the specified columns and drops all others from the dataset. |
| list_nodes_in_aml_compute | Get the details (e.g IP address, port etc) of all the compute nodes in the compute target |
| list_secrets | List the secrets in a keyvault |
| list_supported_vm_sizes | List the supported VM sizes in a region |
| list_workspaces | List all workspaces that the user has access to in a subscription ID |
| load_dataset_into_data_frame | Load all records from the dataset into a dataframe. |
| load_workspace_from_config | Load workspace configuration details from a config file |
| local_webservice_deployment_config | Create a deployment config for deploying a local web service |
| lognormal | Specify a normal distribution of the form 'exp(normal(mu, sigma))' |
| loguniform | Specify a log uniform distribution |
| log_accuracy_table_to_run | Log an accuracy table metric to a run |
| log_confusion_matrix_to_run | Log a confusion matrix metric to a run |
| log_image_to_run | Log an image metric to a run |
| log_list_to_run | Log a vector metric value to a run |
| log_metric_to_run | Log a metric to a run |
| log_predictions_to_run | Log a predictions metric to a run |
| log_residuals_to_run | Log a residuals metric to a run |
| log_row_to_run | Log a row metric to a run |
| log_table_to_run | Log a table metric to a run |
| median_stopping_policy | Define a median stopping policy for early termination of HyperDrive runs |
| merge_results | Combine the results from the parallel training. |
| mount_file_dataset | Create a context manager for mounting file streams defined by the dataset as local files. |
| normal | Specify a real value that is normally-distributed with mean 'mu' and standard deviation 'sigma' |
| package_model | Create a model package that packages all the assets needed to host a model as a web service |
| plot_run_details | Generate table of run details |
| primary_metric_goal | Define supported metric goals for hyperparameter tuning |
| promote_headers_behavior | Defines options for how column headers are processed when reading data from files to create a dataset. |
| pull_model_package_image | Pull the Docker image from a 'ModelPackage' to your local Docker environment |
| qlognormal | Specify a normal distribution of the form 'round(exp(normal(mu, sigma)) / q) * q' |
| qloguniform | Specify a uniform distribution of the form round(exp(uniform(min_value, max_value) / q) * q |
| qnormal | Specify a normal distribution of the form round(normal(mu, sigma) / q) * q |
| quniform | Specify a uniform distribution of the form 'round(uniform(min_value, max_value) / q) * q' |
| randint | Specify a set of random integers in the range [0, upper) |
| random_parameter_sampling | Define random sampling over a hyperparameter search space |
| random_split_dataset | Split file streams in the dataset into two parts randomly and approximately by the percentage specified. |
| register_azure_blob_container_datastore | Register an Azure blob container as a datastore |
| register_azure_data_lake_gen2_datastore | Initialize a new Azure Data Lake Gen2 Datastore. |
| register_azure_file_share_datastore | Register an Azure file share as a datastore |
| register_azure_postgre_sql_datastore | Initialize a new Azure PostgreSQL Datastore. |
| register_azure_sql_database_datastore | Initialize a new Azure SQL database Datastore. |
| register_dataset | Register a Dataset in the workspace |
| register_do_azureml_parallel | Registers AMLCompute as a parallel backend with the foreach package. |
| register_environment | Register an environment in the workspace |
| register_model | Register a model to a given workspace |
| register_model_from_run | Register a model for operationalization. |
| reload_local_webservice_assets | Reload a local web service's entry script and dependencies |
| resource_configuration | Initialize the ResourceConfiguration. |
| r_environment | Create an environment |
| save_model_package_files | Save a Dockerfile and dependencies from a 'ModelPackage' to your local file system |
| service_principal_authentication | Manages authentication using a service principle instead of a user identity. |
| set_default_datastore | Set the default datastore for a workspace |
| set_secrets | Add secrets to a keyvault |
| skip_from_dataset | Skip file streams from the top of the dataset by the specified count. |
| split_tasks | Splits the job into parallel tasks. |
| start_logging_run | Create an interactive logging run |
| submit_child_run | Submit an experiment and return the active child run |
| submit_experiment | Submit an experiment and return the active created run |
| take_from_dataset | Take a sample of file streams from top of the dataset by the specified count. |
| take_sample_from_dataset | Take a random sample of file streams in the dataset approximately by the probability specified. |
| truncation_selection_policy | Define a truncation selection policy for early termination of HyperDrive runs |
| uniform | Specify a uniform distribution of options to sample from |
| unregister_all_dataset_versions | Unregister all versions under the registration name of this dataset from the workspace. |
| unregister_datastore | Unregister a datastore from its associated workspace |
| update_aci_webservice | Update a deployed ACI web service |
| update_aks_webservice | Update a deployed AKS web service |
| update_aml_compute | Update scale settings for an AmlCompute cluster |
| update_local_webservice | Update a local web service |
| upload_files_to_datastore | Upload files to the Azure storage a datastore points to |
| upload_files_to_run | Upload files to a run |
| upload_folder_to_run | Upload a folder to a run |
| upload_to_datastore | Upload a local directory to the Azure storage a datastore points to |
| view_run_details | Initialize run details widget |
| wait_for_deployment | Wait for a web service to finish deploying |
| wait_for_model_package_creation | Wait for a model package to finish creating |
| wait_for_provisioning_completion | Wait for a cluster to finish provisioning |
| wait_for_run_completion | Wait for the completion of a run |
| write_workspace_config | Write out the workspace configuration details to a config file |