SCMR Logo

Tasks


The CMRxRecon2025 challenge includes two regular tasks (announced during MICCAI anual meeting) and two special tasks (announced during SCMR anual meeting) . The tasks are awarded independently, so each team can choose to participate any one of them. For each task, participants can submit one model each (which can be four different models, or they can submit just one model, but they must submit it separately for each task). Please note that for the four tasks, the training dataset we provide is the same; however, the validation and test datasets are different for each of the four tasks.



Objective:

To evaluate the model's generalization performance across diverse centers and scanners (regular task 1), diseases (regular task 2), magnetic fields (special task 1), and populations (special task 2).




Regular TASK 1: CMR reconstruction model for multi-center evaluation

This task primarily focuses on addressing the issue of declining generalization performance between multiple centers. Participants are required to train the reconstruction model on the training dataset and achieve good multi-contrast cardiac image reconstruction results on the validation and test datasets. It is important to note that for this task, we will include data from two entirely new centers in the validation set (not present in the training set), and the test set will contain data from five entirely new centers (not present in the training set, including the two centers that appeared in the validation set).



Regular TASK 2: CMR reconstruction model for multiple diseases evaluation

This task primarily focuses on evaluating the reliability of the model in applications involving different cardiovascular diseases. Participants are required to train the reconstruction model on the training dataset and achieve good performance in disease applications on the validation and test datasets. It is important to note that for this task, we will include data for two diseases that have not appeared in the training set in the validation set, and the test set will contain data for five diseases that have not appeared in the training set (including the two diseases that appeared in the validation set).
Please note that to ensure the model training process is not biased by the type of disease, we will not disclose the disease information for each data point in the training and validation dataset.



Special TASK 1: CMR reconstruction model for 5T evaluation

This task primarily focuses on addressing the issue of declining reconstruction generalization performance under different magnetic field strengths, especially those not included in the training data. Participants are required to train the reconstruction model on the training dataset (mainly consisting of 1.5T and 3.0T) and achieve good multi-contrast cardiac image reconstruction results on the validation and test datasets (5.0T).



Special TASK 2: CMR reconstruction model for pediatric imaging evaluation

This task primarily focuses on addressing application issues in pediatric cardiac imaging. Participants are required to train the reconstruction model on the training dataset (mainly consisting of adults over 20 years old) and achieve good multi-contrast cardiac image reconstruction results on the validation and test datasets (minors under 18 years old). Please note that to ensure the model training process is not biased by age information, we will not disclose the age of each data point in the training dataset.



Metrics & Ranking


Metrics:

For the four tasks, we will use SSIM, PSNR, and NMSE as objective evaluation metrics, with SSIM being the primary metric for ranking. During the test phase, we will conduct an additional round of scoring (1 to 5 points) by experienced radiologists for the top 5 teams in each task, and subsequent evaluations will be based on these subjective scores. The scoring will cover three aspects: image quality, image artifacts, and clinical utility. The final ranking will be determined by the average of the objective and subjective score rankings, and this average ranking will serve as the final competition ranking.
Please note that during the validation phase, we will only conduct objective scoring and will not involve radiologist scoring.


Ranking methods:

1.When evaluating SSIM, we will narrow down the assessment field-of-view to the region where the heart is located, to avoid interference from the background area.
2. Regarding the sampling patterns and accelerations, all paired data are assigned equal weight when calculating the final ranking metrics.
3. Participating teams are required to submit docker containers and process all the cases in the test set on our server. For the cases without valid output, we will assign it to the lowest value of metric.



Awards

The top 5 winners in each task will receive monetary awards. The bonus distribution plan is shown in the table below.


Winner Monetary Awards Certificate Oral Presentation Summary Paper Involved
Top 1 $600
Top 2 $400
Top 3 $300
Top 4 $200
Top 5 $100

All submissions will be reported in the leaderboard. Each participating team can engage in any tasks or all four tasks. Prize-winning methods in Regular Task 1 and Task 2 will be announced publicly as part of a scientific session at the MICCAI annual meeting. Prize-winning methods in Special Task 1 and Task 2 will be announced publicly as part of a scientific session at the SCMR annual meeting.



Task Submission:

Validation submission tutorial: https://www.synapse.org/#!Synapse:syn59814210/wiki/628454
Test submission tutorial: https://www.synapse.org/#!Synapse:syn59814210/wiki/628454


Challenge Platform:

Hosted on Synapse platform.
Visit the Synapse Project Page



Principal of participation


Note: Participants are not required to upload the complete training code. But teams willing to upload the original training code will be automatically entered into the code-sharing pool.



Timeline

The schedule of the challenge is as follows. All deadlines are Pacific Standard Time (PST +11:59).


[Mar. 1, 2025] Website opens for registration
[Mar. 1, 2025] Release training data
[Apr. 1, 2025] Release validation data
[Apr. 1, 2025] Submission system opens for validation
[Jun. 1, 2025] Submission system opens for testing
[Jun. 30, 2025] STACOM paper submission deadline
[Jul. 20, 2025] Testing docker submission deadline for regular tasks
[Sep. 20, 2025] Testing docker submission deadline for special tasks
[Sep. 27, 2025] Release final results of regular tasks during the MICCAI annual meeting
[Feb. 4-7, 2026] Release final results of special tasks during the SCMR annual meeting



Rules

1) It should be restricted to the data provided by the previous CMRxRecon challenge as well as data from the 'fastMRI' challenge (the most related public dataset), under the terms and conditions associated with the data usage.
2) For each task, participants are allowed to train only one model to reconstruct various images at the aforementioned different undersampling scenarios.



Submission

The submission instructions will be released on the  Synapse platform.