EvalAI supports hosting challenges with different configurations. Challenge organizers can choose to customize most aspects of the challenge but not limited to:
- Evaluation metrics
- Language/Framework to implement the metric
- Number of phases and data-splits
- Daily / monthly / overall submission limit
- Number of workers evaluating submissions
- Evaluation on remote machines
- Provide your AWS credentials to host code-upload based challenge
- Show / hide error bars on leaderboard
- Public / private leaderboards
- Allow / block certain email addresses to participate in the challenge or phase
- Choose which fields to export while downloading challenge submissions
We have hosted challenges from different domains such as:
- Machine learning (2019 SIOP Machine Learning Competition)
- Deep learning (Visual Dialog Challenge 2019 )
- Computer vision (Vision and Language Navigation)
- Natural language processing (VQA Challenge 2019)
- Healthcare (fastMRI Image Reconstruction )
- Self-driving cars (CARLA Autonomous Driving Challenge)
We categorize the challenges in two categories:
-
Prediction upload based challenges: Participants upload predictions corresponding to ground truth labels in the form of a file (could be any format:
json,npy,csv,txtetc.)Some of the popular prediction upload based challenges that we have hosted are shown below:
If you are interested in hosting prediction upload based challenges, then click here.
-
Code-upload based challenges: In these kind of challenges, participants upload their training code in the form of docker images using EvalAI-CLI.
We support two types of code-upload based challenges -
- Code-Upload Based Challenge (without Static Dataset): These are usually reinforcement learning challenges which involve uploading a trained model in form of docker images and the environment is also saved in form of a docker image.
- Static Code-Upload Based Challenge: These are challenges where the host might want the participants to upload models and they have static dataset on which they want to run the models and perform evaluations. This kind of challenge is especially useful in case of data privacy concerns.
We support two types of code-upload based challenges -
- Code-Upload Based Challenge (without Static Dataset): These are usually reinforcement learning challenges which involve uploading a trained agent in form of docker images and the environment is also saved in form of a docker image.
- Static Code-Upload Based Challenge: These are challenges where the host might want the participants to upload models and they have static dataset on which they want to run the models and perform evaluations. This kind of challenge is especially useful in case of data privacy concerns.
Some of the popular code-upload based challenges that we have hosted are shown below:
If you are interested in hosting code-upload based challenges, then click here. If you are interested in hosting static code-upload based challenges, then click here.
A good reference would be the Habitat Re-arrangement Challenge 2022.
Use EvalAI-Starters template. See this on how to use a repository as template.
Generate your github personal access token and copy it in clipboard.
Add the github personal access token in the forked repository's secrets with the name AUTH_TOKEN.
Now, go to EvalAI to fetch the following details -
evalai_user_auth_token- Go to profile page after logging in and click onGet your Auth Tokento copy your auth token.host_team_pk- Go to host team page and copy theIDfor the team you want to use for challenge creation.evalai_host_url- Usehttps://eval.aifor production server andhttps://staging.eval.aifor staging server.
Create a branch with name challenge in the forked repository from the master branch.
Note: Only changes in challenge branch will be synchronized with challenge on EvalAI.
Add evalai_user_auth_token and host_team_pk in github/host_config.json.

Read EvalAI challenge creation documentation to know more about how you want to structure your challenge. Once you are ready, start making changes in the yaml file, HTML templates, evaluation script according to your need.
Commit the changes and push the challenge branch in the repository and wait for the build to complete. View the logs of your build.


If challenge config contains errors then a issue will be opened automatically in the repository with the errors otherwise the challenge will be created on EvalAI.
Go to Hosted Challenges to view your challenge. The challenge will be publicly available once EvalAI admin approves the challenge.
To update the challenge on EvalAI, make changes in the repository and push on challenge branch and wait for the build to complete.
We have created a sample challenge configuration that we recommend you to use to get started. Use EvalAI-Starters template to start. See this on how to use a repository as template.
Open challenge_config.yml from the repository that you cloned in step-1. This file defines all the different settings of your challenge such as start date, end date, number of phases, and submission limits etc.
Edit this file based on your requirement. For reference to the fields, refer to the challenge configuration reference section.
Next step is to edit the challenge evaluation script that decides what metrics the submissions are going to be evaluated on for different phases.
Please refer to the writing evaluation script to complete this step.
Almost there. You just need to update the HTML templates in the templates/ directory of the bundle that you cloned.
EvalAI supports all kinds of HTML tags which means you can add images, videos, tables etc. Moreover, you can add inline CSS to add custom styling to your challenge details.
Congratulations! you have submitted your challenge configuration for review and EvalAI team has notified about this. EvalAI team will review and will approve the challenge.
If you have issues in creating a challenge on EvalAI, please feel free to contact us at team@cloudcv.org create an issue on our GitHub issues page.
Follow host challenge using github section to set up a challenge on EvalAI.
Set the remote_evaluation parameter to True in challenge_config.yaml. This challenge config file defines all the different settings of your challenge such as start date, end date, number of phases, and submission limits etc.
Edit this file based on your requirement. For reference to the fields, refer to the challenge configuration reference section.
Please ensure the following fields are set to the following values:
remote_evaluation : True
Refer to the following documentation for details on challenge configuration.
Next step is to edit the challenge evaluation script that decides what metrics the submissions are going to be evaluated on for different phases. Please refer to Writing Remote Evaluation Script section to complete this step.
-
Create conda environment to run the evaluation worker. Refer to conda's create environment section to set up a virtual environment.
-
Install the worker requirements from the
EvalAI-Starters/remote_challenge_evaluationpresent here:cd EvalAI-Starters/ pip install remote_challenge_evaluation/requirements.txt -
Start evaluation worker:
cd EvalAI-Starters/remote_challenge_evaluation python main.py
If you have issues in creating a challenge on EvalAI, please feel free to contact us at team@cloudcv.org create an issue on our GitHub issues page.




