Participation Guidelines

iDPP CLEF 2024

Tasks

iDPP CLEF 2024 offers two evaluation tasks focused on predicting the progression of Amyotrophic Lateral Sclerosis (ALS) from prospective data and one task on impact of exposure to pollutants on predicting the progression of Multiple Sclerosis (MS) from retrospective data.

For Tasks 1 and 2 on ALS participants are given prospective patient data collected over an average of nine months via a dedicated app developed by the BRAINTEASER project and sensor data collected from the sensors of a fitness smartwatch in the context of clinical trials in Turin, Pavia, Lisbon, and Madrid, fully anonymized.

For Task 3 on MS, participants are given a retrospective dataset containing roughly 1.5 years of visits and environmental data. This dataset comes from two clinical institutions, one in Pavia, Italy, and the other in Turin, Italy, and it contains data about real patients, fully anonymized.

All the datasets are highly curated and they are produced from the BRAINTEASER Ontology (BTO), developed by the BRAINTEASER project, which ensures the consistency of the data represented. Moreover, several checks have been performed to ensure that all the instances are clean, contain proper values in the expected ranges, and do not have contradictions.

Task 1 - Predicting ALSFRS-R Score from Sensor Data (ALS)

It focuses on predicting the twelve scores of the ALSFRS-R (ALS Functional Rating Scale - Revised), assigned by medical doctors roughly every three months, from the sensor data collected via the app. The ALSFRS-R is a somehow “subjective” evaluation usually performed by a medical doctor and this task will help in answering a currently open question in the research community, i.e. whether it could be derived from objective factors.

Participants will be given the ALSFRS-R questionnaire at the first visit with the scores for each question together with the time (number of days from diagnosis) at which the questionnaire was taken.

Participants will have to predict the values of the ALSFRS-R sub-scores at the second visit.

Participants will be given the time of the second visit (number of days from diagnosis) together with all the sensor data up to the time of the second visit.

The training data are available upon registration for the challenge and the test data will be available one week before the run submission deadline. Please see the Datasets and Important Dates sections for more information.

Task 2 - Predicting Patient Self-assessment Score from Sensor Data

It focuses on predicting the self-assessment score assigned by patients from the sensor data collected via the app. Self-assessment scores correspond to each of the ALSFRS-R scores but, while the latter ones are assigned by medical doctors during visits, the former ones are assigned via auto-evaluation by patients themselves using the provided app.

If the self-assessment performed by patients, more frequently than the assessment performed by medical doctors every three months or so, can be reliably predicted by sensor and app data, we can imagine a proactive application which, monitoring the sensor data, alerts the patient if an assessment is needed.

Participants will be given the first set of self-assessed scores together with the time (number of days from diagnosis) at which the questionnaire was taken.

Participants will have to predict the values of the self-assessed scores at the second auto-evaluation, happening one or two months after the first one.

Participants will be given the time of the second auto-evaluation (number of days from diagnosis) together with all the sensor data up to the time of the second auto-evaluation.

The training data are available upon registration for the challenge and the test data will be available one week before the run submission deadline. Please see the Datasets and Important Dates sections for more information.

Task 3 - Predicting Relapses from EDDS Sub-scores and Environmental Data (MS)

It focuses on predicting a relapse using environmental data and EDSS (Expanded Disability Status Scale) sub-scores. This task allows us to assess if exposure to different pollutants is a useful variable in predicting a relapse.

Participants will be asked to predict the week of the first relapse after the baseline considering environmental data based on a weekly granularity, given the status of the patient at the baseline, which is the first visit available in the considered time span. For each patient, the date of the baseline will be week 0 and all the other weeks will be relative to it.

Participants will be given all the observations and environmental data about a patient, i.e. also observations which may happen after the relapse to be predicted. All the patients are guaranteed to experience, at least, one relapse after the baseline.

The training data are available upon registration for the challenge and the test data will be available one week before the run submission deadline. Please see the Datasets and Important Dates sections for more information.

Participating

To participate in iDPP CLEF 2024, groups need to register at the following link:

Register

Important Dates

Datasets

Tasks 1 & 2 Datasets (ALS)

Tasks 1 and 2 share the same training dataset, comprising the following data:

All times are expressed in days between the diagnosis and the collection. Note that sensor data is not available for every day and may contain gaps due to lack of adherence from the patient and/or technical problems in the data collection. However, finding ways to work around these limitations is part of the challenge.

For a reference on the ALSFRS-R scores, please refer to LINK.

The test data will contain the same static and sensor data for other patients not in the training set. Only the first ALSFRS-R evaluation will be available, together with the time of the target ALSFRS-R evaluation whose scores are to be predicted.

Training Dataset

52 patients, 189 clinical ALSFRS-R, 301 self-assessed ALSFRS-R, 13946 days of sensor data.

Training datasets can be downloaded here:

Check out the legend for a description of the data.

Test Dataset

Details on the Datasets

Each of the training datasets consists of the following files:

Each of the test datasets consists of the following files:

Tasks 3 Dataset (MS)

Task 3 encompasses the following data:

All time intervals are expressed in weeks from the baseline. EDSS visits occur between the baseline and the occurrence of the first relapse, while environmental measurements span from January 1st, 2013, to 2023. It's important to note that environmental measurements may not be available for every week and contain missing data.

For more information on the EDSS score, please refer to this link.

Training and test datasets are divided in a 70-30% proportion. The test data will include the same static, EDSS, and environmental data for patients not included in the training set.

Specifically:

Details on the Dataset

The training dataset consists of the following files:

The test dataset consists of the following files:

The test dataset does not contain the test_outcome.csv file as it is the target that needs to be predicted by the challenge participants.

Check out the legend for a description of the data

More in detail, the train_outcome.csv file adopts the following format:


67396654612589370083623092407810766693,190
267505643482073532971907590162817517438,96
292211325266845962706770530341466085902,30
272084949450943266236977822325902693890,216
291361686090369065397020169907588173057,13
...
					

where:

Participant Repository

Participants are provided with a single repository for all the tasks they take part in. The repository contains the runs, resources, and possibly the code produced by each participant in order to promote reproducibility and open science.

The repository is organised as follows:

The submission and score folders are organized into sub-folders for each task as follows:

Participants which do not take part in a given task can simply delete the corresponding sub-folders.

Here, you can find a sample participant repository to get a better idea of its layout.

The goal of iDPP CLEF 2024 is to speed up the creation of systems and resources for MS and ALS progression prediction as well as openly share these systems and resources as much as possible. Therefore, participants are more than encouraged to share their code and any additional resources they have used or created.

All the contents of these repositories are released under the Creative Commons Attribution-ShareAlike 4.0 International License.

Run Submission Guidelines

Participating teams should satisfy the following guidelines:

Task 1 Run Format

Runs should be submitted as a text file (.txt) with the following format:


100619256189067386770484450960632124211 1 2 3 4 1 2 3 4 1 2 3 4 upd_T1_myDesc
101600333961427115125266345521826407539 1 2 3 4 1 2 3 4 1 2 3 4 upd_T1_myDesc
102874795308599532461878597137083911508 1 2 3 4 1 2 3 4 1 2 3 4 upd_T1_myDesc
123988288044597922158182615705447150224 1 2 3 4 1 2 3 4 1 2 3 4 upd_T1_myDesc
100381996772220382021070974955176218231 1 2 3 4 1 2 3 4 1 2 3 4 upd_T1_myDesc
...
					

where:

It is important to include all the columns and have a white space delimiter between the columns. No specific ordering is expected among patients (rows) in the submission file.

Task 2 Run Format

Runs should be submitted as a text file (.txt) with the following format:


100619256189067386770484450960632124211 1 2 3 4 1 2 3 4 1 2 3 4 upd_T2_myDesc
101600333961427115125266345521826407539 1 2 3 4 1 2 3 4 1 2 3 4 upd_T2_myDesc
102874795308599532461878597137083911508 1 2 3 4 1 2 3 4 1 2 3 4 upd_T2_myDesc
123988288044597922158182615705447150224 1 2 3 4 1 2 3 4 1 2 3 4 upd_T2_myDesc
100381996772220382021070974955176218231 1 2 3 4 1 2 3 4 1 2 3 4 upd_T2_myDesc
...
					

where:

It is important to include all the columns and have a white space delimiter between the columns. No specific ordering is expected among patients (rows) in the submission file.

Task 3 Run Format

Runs should be submitted as a text file (.txt) with the following format:


100619256189067386770484450960632124211 10 upd_T3_myDesc
101600333961427115125266345521826407539 47 upd_T3_myDesc
102874795308599532461878597137083911508 13 upd_T3_myDesc
123988288044597922158182615705447150224 1 upd_T3_myDesc
100381996772220382021070974955176218231 9 upd_T3_myDesc
...
					

where:

It is important to include all the columns and have a white space delimiter between the columns. No specific ordering is expected among patients (rows) in the submission file.

Submission Upload

Runs should be uploaded in the repository provided by the organizers. Following the repository structure discussed above, for example, a run submitted for the first task should be included in submission/task1.

Runs should be uploaded using the following name convention for their identifiers: <teamname>_T<1|2|3>_<freefield>

where:

For example, a complete run identifier may look like upd_T1_myDesc

where

The name of the text file containing the run must be the identifier of the run followed by the txt extension. In the above example upd_T1_myDesc.txt

Run Scores

Performance scores for the submitted runs will be returned by the organizers in the score folder, which follows the same structure as the submission folder. For each submitted run, participants will find a file named <teamname>_T<1|2|3>_<freefield>.score.txt where <teamname>_T<1|2|3>_<freefield> matches the corresponding run. The file will contain performance scores for each of the evaluation measures described below. In the above example upd_T1_myDesc.score.txt

Evaluation Measures

Task 1 - Predicting ALSFRS-R Score from Sensor Data (ALS)

The effectiveness of the submitted runs will be evaluated using MAE (Mean Absolute Error) and RMSE (Root Mean Square Error) of the predicted sub-scores with respect to the actual sub-scores.

Task 2 - Predicting patient self-assessment score from sensor (ALS)

The effectiveness of the submitted runs will be evaluated using MAE (Mean Absolute Error) and RMSE (Root Mean Square Error) of the predicted self-assed sub-scores with respect to the actual sub-scores.

Task 3 - Predicting relapses from EDDS sub-scores and environmental data (MS)

The effectiveness of the submitted runs will be evaluated using MAE (Mean Absolute Error) and RMSE (Root Mean Square Error) of the predicted week of relapse with respect to the actual one.

Participant Papers

Participants are expected to write a report describing their participation in iDPP CLEF 2024, their proposed solution, what features have been used for prediction, the analysis of the experimental results and insights derived from them.

Participant reports will be peer-reviewed and accepted reports will be published in the CLEF 2023 Working Notes at CEUR-WS, indexed in DBLP and Scopus.

A template for the report is contained in the report folder of each participant repository. The template does not only define the layout of the report but it also provides a suggested structure for it and illustrates the kind of expected contents.
Participants are strongly encouraged to use LaTeX, whose template provides all the details about the report. However, it is possible to also use Word and a basic template is provided as well.

Participant papers are expected to be 10-20 pages in length, excluding references.

The schedule for submission, revision, and camera ready of the participant report is detailed above in the Important Dates section.

Participant papers can be submitted via Easychair at the following link https://easychair.org/conferences/?conf=clef2024 by choosing the Intelligent Disease Progression Prediction (iDPP) track.

Get in Contact

If you need any additional information, please get in contact with us writing to Nicola Ferro, ferrodei.unipd.it.

This project has received funding from the European Union’s Horizon 2020 research and innovation programme under the grant agreement No GA101017598. The European Commission’s support for this project and the production of this website does not constitute an endorsement of the contents, which reflect the views only of the authors, and the Commission cannot be held responsible for any use which may be made of the information contained therein