eRisk 2026:

Early risk prediction on the Internet


CLEF 2026 Workshop

Jena, Germany 馃嚛馃嚜, 21-24 September 2026

Find Out More How To Obtain the 2025 Datasets

CLEF eRisk 2026:

Early risk prediction on the Internet


eRisk explores the evaluation methodology, effectiveness metrics and practical applications (particularly those related to health and safety) of early risk detection on the Internet. Early detection technologies can be employed in different areas, particularly those related to health and safety. For instance, early alerts could be sent when a predator starts interacting with a child for sexual purposes, or when a potential offender starts publishing antisocial threats on a blog, forum or social network. Our main goal is to pioneer a new interdisciplinary research area that would be potentially applicable to a wide variety of situations and to many different personal profiles. Examples include potential paedophiles, stalkers, individuals that could fall into the hands of criminal organisations, people with suicidal inclinations, or people susceptible to depression.

Participate


This is the tenth year of eRisk and the lab plans to organize three tasks:

Task 1: Conversational Depression Detection

This task extends last year's pilot by detecting depression through conversational agents while improving access and reproducibility. Participants will interact with LLM personas fine-tuned with diverse user histories and released on Hugging Face. Each model will be released on a different day, and participants will have limited days before giving their predictions.

The challenge is to determine whether each persona exhibits signs of depression and, within a limited conversational window, identify active depressive symptoms and the overall depression level. The LLM personas will reflect different severity levels guided by the BDI-II questionnaire, allowing systems to be evaluated across a spectrum of simulated depression.

Teams will download the released persona models, conduct their interactions, and submit predictions a few days later. Evaluation will focus on two key aspects: (i) accurate identification of depressive symptoms present in the persona (if any) and (ii) the overall depression level of the persona, following BDI-II standards.

A limited number of runs will be accepted, with both fully automated and manual-in-the-loop variants permitted, encouraging exploration of conversational strategies while maintaining comparability across submissions.

The proceedings of the lab will be published in the online CEUR-WS Proceedings and on the conference website.

To have access to the collection, all participants have to fill, sign, and send a user agreement form (follow the instructions provided here). Once you have submitted the signed copyright form, you can proceed to register for the lab at CLEF 2026 Labs Registration site.

Important Dates

Task 2: Contextualized Early Detection of Depression

This is the second edition of the contextualized early detection task, first introduced in eRisk 2025.

This task focuses on detecting early signs of depression by analyzing full conversational contexts. Unlike previous tasks that focused on isolated user posts, this challenge considers the broader dynamics of interactions by incorporating writings from all individuals involved in the conversation. Participants must process user interactions sequentially, analyze natural dialogues, and detect signs of depression within these rich contexts. Texts will be processed chronologically to simulate real-world conditions, making the task applicable to monitoring user interactions in blogs, social networks, or other types of online media.

The test collection for this task follows the format described in Losada & Crestani, 2016 and is derived from the same data sources as previous eRisk tasks. The dataset includes:

  • 1. Writing history from the specific user: Posts from specific target social media users for depression estimation.
  • 2. Full conversational contexts:
    • The discussion title and comment.
    • Comments from all users participating in the conversation, in chronological order.
    • Messages exchanged between the specific target user to classify and others in the discussion.

There are two categories of users: individuals suffering depression and control users. For each user, the collection contains a sequence of writings from that specific user along with the rest of the users that participated in the conversation (in chronological order). This approach allows systems to monitor ongoing interactions and make timely decisions based on the evolution of the conversation.

The task is organized into two different stages:

  • Training Stage. Participants will be provided with the contextualized dataset from eRisk 2025, which includes full conversational contexts. This collection serves as training data to enable reproducible development and validation.

    • The training data includes both isolated writings from users and their full conversation contexts.
    • For each user, the dataset indicates whether they have explicitly mentioned a depression diagnosis.
    • Training data may also include users from prior early depression detection tasks, allowing teams to train their systems effectively.
  • Test Stage. During the test phase, participants will connect to our server that provides user writings iteratively, including full conversational contexts (e.g., the discussion title and other users' comments).

    Participants have to:

    • Analyze the data in real time and send their predictions after processing each writing.
    • Use the provided full conversational contexts for each user interaction, simulating real-world scenarios.

Evaluation: The evaluation will consider not only the correctness of the system's output (i.e., whether or not the user is depressed) but also the delay taken to emit its decision. To meet this aim, we will consider the ERDE metric proposed in Losada & Crestani, 2016 and other alternative evaluation measures. A full description of the evaluation metrics can be found in 2021's eRisk overview.

The proceedings of the lab will be published in the online CEUR-WS Proceedings and on the conference website.

To have access to the collection, all participants must fill, sign, and send a user agreement form (follow the instructions provided here). Once you have submitted the signed copyright form, you can proceed to register for the lab at CLEF 2026 Labs Registration site.

Important Dates

Task 3: ADHD Symptom Sentence Ranking

This new task targets sentence-level retrieval for the 18 symptoms defined in the Adult ADHD Self-Report Scale (ASRS-v1.1). Participants must rank candidate sentences by their relevance to each symptom. A sentence is considered relevant when it conveys information about the user's state with respect to the target ADHD symptom (irrespective of polarity or stance), encouraging models to capture clinically meaningful evidence rather than surface keywords.

You can view the ADHD questionnaire here or download the official PDF from this link.

We will release a sentence-tagged dataset derived from publicly available social media writings, collected to contain ADHD-related expressions. As this is the first edition of the ADHD ranking task, no annotated training data will be provided. The release will consist solely of the test inputs, following the formatting conventions of recent eRisk ranking tasks.

Participants will submit 18 rankings (one per ADHD symptom) ordering candidate sentences by decreasing likelihood of relevance. Relevance assessments will be produced via top-k pooling and expert annotation, and systems will be evaluated using standard IR metrics such as MAP, nDCG (e.g., @100), and P@10.

The task is organized into two different stages:

  • Submission stage. After the release of the datasets, the participants will have time to produce and upload to our FTP server their TREC-formatted runs. Each participant may upload up to 5 files corresponding to 5 systems to the FTP.
    The required submission TREC format is as follows:
                symptom_number   Q0   sentence-id   position_in_ranking   score   system_name
                    

    An example of the format of your runs should be as follows:

                1   Q0   sentence-id-121   0001   10    myGroupNameMyMethodName
                1   Q0   sentence-id-234   0002   9.5   myGroupNameMyMethodName
                1   Q0   sentence-id-345   0003   9     myGroupNameMyMethodName
                ...
                18  Q0   sentence-id-456   0998   1.25  myGroupNameMyMethodName
                18  Q0   sentence-id-242   0999   1     myGroupNameMyMethodName
                18  Q0   sentence-id-347   1000   0.9   myGroupNameMyMethodName
                    

    Participants should submit up to 1000 results sorted by estimated relevance for each of the 18 symptoms of the ADHD questionnaire (ASRS-v1.1). Each line contains: symptom_number, Q0, sentence-id, position_in_ranking, score, system_name.

  • Evaluation stage. Once the submission stage is closed, the submitted runs will be used for obtaining the relevance judgments using classical pooling strategies with human assessors. With those judgments, systems will be evaluated.

By extending symptom-oriented retrieval beyond depression to ADHD, this task advances interpretable, symptom-aware retrieval and supports cross-condition generalisation at sentence granularity.

The proceedings of the lab will be published in the online CEUR-WS Proceedings and on the conference website.

To have access to the collection, all participants must fill, sign, and send a user agreement form (follow the instructions provided here). Once you have submitted the signed copyright form, you can proceed to register for the lab at CLEF 2026 Labs Registration site.

Important Dates
Ongoing schedule



18 NOV
  • Registration for lab opens
  • 18/11/2025

01 DEC
  • Release of the training data (T1,T2) and test dataset for T2 and T3
  • 19/12/2025

05 FEB
  • T2 and T3: Beginning of test stage (servers are open)
  • 05/02/2026

12 APR
  • T2 and T3: End of test stage (server closes).
  • 12/04/2026

10 MAY
  • Release of evaluation results to all participants
  • 10/05/2026

30 MAY
  • Submission of Participant Papers [CEUR-WS]
  • 28/05/2026

27 JUN
  • Notification of acceptance
  • 30/06/2026

07 JUL
  • Camera ready. Participant Papers [CEUR-WS]
  • 06/07/2026

Programme


To Be Announced

The programme for eRisk 2026 will be announced closer to the conference date.



Organizers


More information


+34 881 016 027

CLEF 2026 Conference & CLEF initiative:

CLEF 2026
CLEF

Funded by

  • Big-eRisk: Predicci贸n temprana de riesgos personales en conjuntos de datos masivos. Ministerio de Ciencia e Innovaci贸n, Agencia Estatal de Investigaci贸n, Plan de Recuperaci贸n, Transformaci贸n y Resiliencia, Uni贸n Europea-Next Generation EU PLEC2021-007662
  • Projects PID2022-137061OBC21 (Ministerio de Ciencia e Innovaci贸n supported by the European Regional Development Fund)
Ministerio de Ciencia e Innovaci贸n, Agencia Estatal de Investigaci贸n, Plan de Recuperaci贸n, Transformaci贸n y Resiliencia, Uni贸n Europea-Next Generation EU