Skip to content

autogluon/autogluon-bench

AutoGluon-Bench

Welcome to AutoGluon-Bench, a suite for benchmarking your AutoML frameworks.

Setup

Follow the steps below to set up autogluon-bench:

# create virtual env and update pip
python3 -m venv .venv_agbench
source .venv_agbench/bin/activate
python3 -m pip install --upgrade pip

Install autogloun-bench from PyPI:

python3 -m pip install autogluon.bench

Install autogluon-bench from source for development:

git clone https://github.com/autogluon/autogluon-bench.git
cd autogluon-bench

# install from source in editable mode
pip install -e ".[tests]"

Run benchmarks locally

To run the benchmarks on your local machine, use the following command:

agbench run path/to/local_config_file

Check out our sample local configuration files for local runs.

The results are stored in the following directory: {WORKING_DIR}/{root_dir}/{module}/{benchmark_name}_{timestamp}.

Tabular and Timeseries Benchmark

To perform tabular or timeseries benchmarking, set the module to 'tabular' or 'timeseries'. You must set both Benchmark Configurations and Tabular/Timeseries Specific configurations, and each should have a single value. Refer to the sample configuration file for more details.

The tabular/timeseires module leverages the AMLB benchmarking framework. Required and optional AMLB arguments are specified via the configuration file mentioned previously.

Custom configuration is supported by providing a local directory to amlb_user_dir in the config, by which custom frameworks, constraints and datasets can be overriden. We have a minimum working custom config setup for benchmarking on a custom framework (an AutoGluon dev branch). In the sample configuration file, change the following field to:

framework: AutoGluon_dev:example
amlb_user_dir: path_to/sample_configs/amlb_configs

For more customizations, please follow the example custom configuration folder provided by AMLB and their documentation.

Multimodal Benchmark

For multimodal benchmarking, set the module to multimodal. Note that multimodal benchmarking directly calls the MultiModalPredictor, bypassing the extra layer of AMLB. Therefore, the required arguments are different from those for tabular or timeseries. Please refer to the sample multimodal local run configuration file.

We also support customizations on benchmarking framework, datasets, and metrics by providing custom_resource_dir, custom_dataloader, custom_metrics.

To define custom frameworks, you can follow the examples.

  1. Create a folder under working directory, e.g. custom_resources/
  2. Create a yaml file named multimodal_frameworks.yaml
  3. Add an entry to the file with repo as the GitHub URL, version as the branch or tag name, params to be used by MultiModalPredictor.
  4. Add custom_resource_dir: custom/resources/ in the run configuration file.

To add more datasets to your benchmarking jobs. We support custom datasets with custom defined data loaders. Follow these steps:

  1. Create a folder under the working directory, e.g. custom_dataloader/
  2. Create a dataset yaml file, custom_dataloader/datasets.yaml which includes all required properties for your problem type, please refer to the function.
  3. Create a dataset loader class, custom_dataloader/dataloader.py, which downloads and loads the dataset as a dataframe. Please set the required properties as mentioned above.
  4. Add custom_dataloader in the agbench run configuration, where dataloader_file, class_name and dataset_config_file are required.
  5. Make sure you have the proper permission to download the dataset. If running in AWS mode, we support downloading from the S3 bucket specified as DATA_BUCKET in the agbench run configuration under the same AWS Batch deployment account.

Please refer to here for more examples.

Adding custom metrics is similar as adding data loaders. Internally, we convert the custom metrics into an AutoGluon Scorer using the autogluon.core.metrics.make_scorer function. Follow these steps to set up:

  1. Create a folder under the working directory, e.g. custom_metrics/
  2. Create a metrics script, custom_metrics/metrics.py which has a function defined that returns a metrics score.
  3. Add custom_metrics in the agbench run configuration, where metrics_path, function_name are required. Aditional arguments can be added for the make_scorer function.

Please refer to here for more examples.

Run benchmarks on AWS

AutoGluon-Bench uses the AWS CDK to build an AWS Batch compute environment for benchmarking.

To get started, install Node.js and AWS CDK with the following instructions:

  1. Install Node Version Manager.
  2. Source profile or restart the terminal.
  3. Follow the Prerequisites section on the AWS CDK Guide and install an appropriate Node.js version for your system:
nvm install $VERSION  # install Node.js
npm install -g aws-cdk  # install aws-cdk
cdk --version  # verify the installation, you might need to update the Node.js version depending on the log.
  1. Follow the AWS CLI Installation Guide to install awscliv2.

If it is the first time using CDK to deploy to an AWS environment (An AWS environment is a combination of an AWS account and Region), please run the following:

cdk bootstrap aws://CDK_DEPLOY_ACCOUNT/CDK_DEPLOY_REGION

You will need a cloud configuration file to run the benchmarks. You can edit the provided sample cloud config files, or use the CLI tool to generate the cloud config files locally.

For multimodal:

agbench generate-cloud-config --module multimodal --cdk-deploy-account <AWS_ACCOUNT_ID> --cdk-deploy-region <AWS_ACCOUNT_REGION> --prefix <PREFIX> --metrics-bucket <METRICS_BUCKET> --data-bucket <DATA_BUCKET> --dataset-names DATASET_1,DATASET_2 --custom-resource-dir <CUSTOM_RESOURCE_DIR> --custom-dataloader "dataloader_file:value1;class_name:value2;dataset_config_file:value3"

For tabular or timeseries:

agbench generate-cloud-config --module <MODULE> --cdk-deploy-account <AWS_ACCOUNT_ID> --cdk-deploy-region <AWS_ACCOUNT_REGION> --prefix <PREFIX> --metrics-bucket <METRICS_BUCKET> --git-uri-branch <AMLB_GIT_URI_BRANCH> --framework <AMLB_FRAMEWORK> --amlb-benchmark <BENCHMARK1>,<BENCHMARK2> --amlb-task "BENCHMARK1:DATASET1,DATASET2;BENCHMARK2:DATASET3" --amlb-constraint <CONSTRAINT> --amlb-fold-to-run "BENCHMARK1:DATASET1:fold1/fold2,DATASET2:fold1/fold2;BENCHMARK1:DATASET3:fold1/fold2" --amlb-user-dir <AMLB_USER_DIR>

For more details, you can run

agbench generate-cloud-config --help

After having the configuration file ready, use the command below to initiate benchmark runs on cloud:

agbench run /path/to/cloud_config_file

This command automatically sets up an AWS Batch environment using instance specifications defined in the cloud config files. It also creates a lambda function named with your chosen LAMBDA_FUNCTION_NAME. This lambda function is automatically invoked with the cloud config file you provided, submitting a single AWS Batch job or a parent job for Array jobs to the job queue (named with the PREFIX you provided).

In order for the Lambda function to submit multiple Array child jobs simultaneously, you need to specify a list of values for each module-specific key. Each combination of configurations is saved and uploaded to your specified METRICS_BUCKET in S3, stored under S3://{METRICS_BUCKET}/configs/{module}/{BENCHMARK_NAME}_{timestamp}/{BENCHMARK_NAME}_split_{UID}.yaml. Here, UID is a unique ID assigned to the split.

The AWS infrastructure configurations and submitted job ID is saved locally at {WORKING_DIR}/{root_dir}/{module}/{benchmark_name}_{timestamp}/aws_configs.yaml. You can use this file to check the job status at any time:

agbench get-job-status --config-file /path/to/aws_configs.yaml

You can also check the job status using job IDs:

agbench get-job-status --job-ids JOB_ID_1 --job-ids JOB_ID_2 —cdk_deploy_region AWS_REGION

Job logs can be viewed on the AWS console. Each job has an UID attached to the name, which you can use to identify the respective config split. After the jobs are completed and reach the SUCCEEDED status in the job queue, you'll find metrics saved under S3://{METRICS_BUCKET}/{module}/{benchmark_name}_{timestamp}/{benchmark_name}_{timestamp}_{UID}.

A cloud configuration file with time-stamped benchmark_name is also saved under {WORKING_DIR}/{root_dir}/{module}/{benchmark_name}_{timestamp}/{module}_cloud_configs.yaml

By default, the infrastructure created is retained for future use. To automatically remove resources after the run, use the --remove_resources option:

agbench run path/to/cloud_config_file --remove-resources

This will check the job status every 2 minutes and remove resources after all jobs succeed. If any job fails, resources will be kept.

If you want to manually remove resources later, use:

agbench destroy-stack --config-file `{WORKING_DIR}/{root_dir}/{module}/{benchmark_name}_{timestamp}/aws_configs.yaml`

Or you can remove specific stacks by running:

agbench destroy-stack --static-resource-stack STATIC_RESOURCE_STACK_NAME --batch-stack BATCH_STACK_NAME --cdk-deploy-account CDK_DEPLOY_ACCOUNT --cdk-deploy-region CDK_DEPLOY_REGION

where you can find all argument values in {WORKING_DIR}/{root_dir}/{module}/{benchmark_name}_{timestamp}/aws_configs.yaml.

Configure the AWS infrastructure

The default infrastructure configurations are located here. CDK_DEPLOY_ACCOUNT: dummy CDK_DEPLOY_REGION: dummy PREFIX: ag-bench MAX_MACHINE_NUM: 20 BLOCK_DEVICE_VOLUME: 100 TIME_LIMIT: 3600 RESERVED_MEMORY_SIZE: 15000 INSTANCE: g4dn.2xlarge LAMBDA_FUNCTION_NAME: ag-bench-job

where:

  • CDK_DEPLOY_ACCOUNT and CDK_DEPLOY_REGION should be overridden with your AWS account ID and desired region to create the stack.
  • PREFIX is used as an identifier for the stack and resources created.
  • MAX_MACHINE_NUM is the maximum number of EC2 instances can be started for AWS Batch.
  • BLOCK_DEVICE_VOLUME is the size of storage device attached to instance.
  • TIME_LIMIT is the timeout of AWS Batch job, i.e. the maximum time the instance will run. There is a buffer of 3600s added on top of it to account for instance startup time and dataset download time.
  • RESERVED_MEMORY_SIZE is used together with the instance memory size to calculate the container shm_size.
  • INSTANCE is the EC2 instance type.
  • LAMBDA_FUNCTION_NAME is the lambda function prefix to submit jobs to AWS Batch.

To override these configurations, use the cdk_context key in your custom config file. See our sample cloud config for reference.

For multimodal module, these will also be overridden by a constraint defined here or a custom constraint specified in multimodal_constraints.yaml under custom_resource_dir. See sample custom constraints file

Monitoring metrics for your instances on AWS

A variety of metrics are available for the EC2 instances that are launched during benchmarking. These can be accessed through the AWS Console by following this navigation path: CloudWatch -> All metrics -> AWS namespaces -> EC2. For a comprehensive list of these metrics, refer to the official AWS documentation.

In addition to the standard metrics, we also provide a custom metric for GPUUtilization. This can be found in the CloudWatch section under All metrics -> Custom namespaces -> EC2. Please note that the GPUUtilization metric is also updated every five minutes.

We provide an option to save aggregated (average) custom hardware metrics (GPUUtilization and CPUUtilization logged in 5s intervals) to the benchmark directory under the provided S3 bucket, simply use the option when running benchmark:

agbench run --save-hardware-metrics

Note that currently this command waits for all jobs to become successful to pull the hardware metrics.

Evaluating benchmark runs

Benchmark results can be evaluated using the tools in src/autogluon/bench/eval/. The evaluation logic will aggregate, clean, and produce evaluation results for runs stored in S3. In a future release, we intend to add evaluation support for multimodal benchmark results.

Evaluation Steps

Begin by setting up AWS credentials for the default profile for the AWS account that has the benchmark results in S3.

Step 1: Aggregate AMLB results on S3. After running the benchmark in AWS mode, take note of the benchmark_name with timestamp in {WORKING_DIR}/{root_dir}/{module}/{benchmark_name}_{timestamp}/{module}_cloud_configs.yaml and run the command below:

agbench aggregate-amlb-results {METRICS_BUCKET} {module} {benchmark_name} --constraint {constraint}

This will create a new file on S3 with this signature:

s3://{METRICS_BUCKET}/aggregated/{module}/{benchmark_name}/results_automlbenchmark_{constraint}_{benchmark_name}.csv

Currently, aggregation is also supported for multimodal benchmark results without the --constratint option.

For more details, run:

agbench aggregate-amlb-results --help

Step 2: Further clean the aggregated results.

If the file is still on S3 from the previous step, run:

agbench clean-amlb-results {benchmark_name} --results-dir-input s3://{METRICS_BUCKET}/aggregated/{module}/{benchmark_name}/ --benchmark-name-in-input-path --constraints constratint_1 --constraints constratint_2 --results-dir-output {results_dir_output}
--out-path-prefix {out_path_prefix} --out-path-suffix {out_path_suffix}

where {results_dir_input} can also be a local directory. This will create a local file {results_dir_output}/{out_path_prefix}{benchmark_name}{out_path_suffix}.

For more details, run:

agbench clean-amlb-results --help

Step 3: Run evaluation on multiple cleaned files from Step 2

agbench evaluate-amlb-results --frameworks-run framework_1 --frameworks-run framework_2 --results-dir-input data/results/input/prepared/openml/ --paths file_name_1.csv --paths file_name_2.csv --output-suffix f"{module}_{preset}_{constraint}_{date}", --no-clean-data --no-use-tid-as-dataset-name