CLEF 2026 Workshop
Jena, Germany 馃嚛馃嚜, 21-24 September 2026
Find Out More How To Obtain the 2025 DatasetseRisk explores the evaluation methodology, effectiveness metrics and practical applications (particularly those related to health and safety) of early risk detection on the Internet. Early detection technologies can be employed in different areas, particularly those related to health and safety. For instance, early alerts could be sent when a predator starts interacting with a child for sexual purposes, or when a potential offender starts publishing antisocial threats on a blog, forum or social network. Our main goal is to pioneer a new interdisciplinary research area that would be potentially applicable to a wide variety of situations and to many different personal profiles. Examples include potential paedophiles, stalkers, individuals that could fall into the hands of criminal organisations, people with suicidal inclinations, or people susceptible to depression.
This is the tenth year of eRisk and the lab plans to organize three tasks:
This task extends last year's pilot by detecting depression through conversational agents while improving access and reproducibility. Participants will interact with LLM personas fine-tuned with diverse user histories and released on Hugging Face. Each model will be released on a different day, and participants will have limited days before giving their predictions.
The challenge is to determine whether each persona exhibits signs of depression and, within a limited conversational window, identify active depressive symptoms and the overall depression level. The LLM personas will reflect different severity levels guided by the BDI-II questionnaire, allowing systems to be evaluated across a spectrum of simulated depression.
Teams will download the released persona models, conduct their interactions, and submit predictions a few days later. Evaluation will focus on two key aspects: (i) accurate identification of depressive symptoms present in the persona (if any) and (ii) the overall depression level of the persona, following BDI-II standards.
A limited number of runs will be accepted, with both fully automated and manual-in-the-loop variants permitted, encouraging exploration of conversational strategies while maintaining comparability across submissions.
The proceedings of the lab will be published in the online CEUR-WS Proceedings and on the conference website.
To have access to the collection, all participants have to fill, sign, and send a user agreement form (follow the instructions provided here). Once you have submitted the signed copyright form, you can proceed to register for the lab at CLEF 2026 Labs Registration site.
Important DatesThis is the second edition of the contextualized early detection task, first introduced in eRisk 2025.
This task focuses on detecting early signs of depression by analyzing full conversational contexts. Unlike previous tasks that focused on isolated user posts, this challenge considers the broader dynamics of interactions by incorporating writings from all individuals involved in the conversation. Participants must process user interactions sequentially, analyze natural dialogues, and detect signs of depression within these rich contexts. Texts will be processed chronologically to simulate real-world conditions, making the task applicable to monitoring user interactions in blogs, social networks, or other types of online media.
The test collection for this task follows the format described in Losada & Crestani, 2016 and is derived from the same data sources as previous eRisk tasks. The dataset includes:
There are two categories of users: individuals suffering depression and control users. For each user, the collection contains a sequence of writings from that specific user along with the rest of the users that participated in the conversation (in chronological order). This approach allows systems to monitor ongoing interactions and make timely decisions based on the evolution of the conversation.
The task is organized into two different stages:
Participants have to:
Evaluation: The evaluation will consider not only the correctness of the system's output (i.e., whether or not the user is depressed) but also the delay taken to emit its decision. To meet this aim, we will consider the ERDE metric proposed in Losada & Crestani, 2016 and other alternative evaluation measures. A full description of the evaluation metrics can be found in 2021's eRisk overview.
The proceedings of the lab will be published in the online CEUR-WS Proceedings and on the conference website.
To have access to the collection, all participants must fill, sign, and send a user agreement form (follow the instructions provided here). Once you have submitted the signed copyright form, you can proceed to register for the lab at CLEF 2026 Labs Registration site.
Important DatesThis new task targets sentence-level retrieval for the 18 symptoms defined in the Adult ADHD Self-Report Scale (ASRS-v1.1). Participants must rank candidate sentences by their relevance to each symptom. A sentence is considered relevant when it conveys information about the user's state with respect to the target ADHD symptom (irrespective of polarity or stance), encouraging models to capture clinically meaningful evidence rather than surface keywords.
You can view the ADHD questionnaire here or download the official PDF from this link.
We will release a sentence-tagged dataset derived from publicly available social media writings, collected to contain ADHD-related expressions. As this is the first edition of the ADHD ranking task, no annotated training data will be provided. The release will consist solely of the test inputs, following the formatting conventions of recent eRisk ranking tasks.
Participants will submit 18 rankings (one per ADHD symptom) ordering candidate sentences by decreasing likelihood of relevance. Relevance assessments will be produced via top-k pooling and expert annotation, and systems will be evaluated using standard IR metrics such as MAP, nDCG (e.g., @100), and P@10.
The task is organized into two different stages:
symptom_number Q0 sentence-id position_in_ranking score system_name
An example of the format of your runs should be as follows:
1 Q0 sentence-id-121 0001 10 myGroupNameMyMethodName
1 Q0 sentence-id-234 0002 9.5 myGroupNameMyMethodName
1 Q0 sentence-id-345 0003 9 myGroupNameMyMethodName
...
18 Q0 sentence-id-456 0998 1.25 myGroupNameMyMethodName
18 Q0 sentence-id-242 0999 1 myGroupNameMyMethodName
18 Q0 sentence-id-347 1000 0.9 myGroupNameMyMethodName
Participants should submit up to 1000 results sorted by estimated relevance for each of the 18 symptoms of the ADHD questionnaire (ASRS-v1.1). Each line contains: symptom_number, Q0, sentence-id, position_in_ranking, score, system_name.
By extending symptom-oriented retrieval beyond depression to ADHD, this task advances interpretable, symptom-aware retrieval and supports cross-condition generalisation at sentence granularity.
The proceedings of the lab will be published in the online CEUR-WS Proceedings and on the conference website.
To have access to the collection, all participants must fill, sign, and send a user agreement form (follow the instructions provided here). Once you have submitted the signed copyright form, you can proceed to register for the lab at CLEF 2026 Labs Registration site.
Important Dates18/11/2025
19/12/2025
05/02/2026
12/04/2026
10/05/2026
28/05/2026
30/06/2026
06/07/2026
The programme for eRisk 2026 will be announced closer to the conference date.