Call for Papers

Home » Submit » Call for Papers

Author Information

The AHLI Conference on Health, Inference, and Learning (CHIL) solicits work across a variety of disciplines at the intersection of machine learning and healthcare. CHIL 2025 invites submissions focused on artificial intelligence and machine learning (AI/ML) techniques that address challenges in health, which we view broadly as including clinical healthcare, public health, population health, and beyond.

Specifically, authors are invited to submit 8-10 page papers (with unlimited pages for references) to one of 3 possible tracks: Models and Methods, Applications and Practice, or Impact and Society. Each track is described in detail below. Authors will select exactly one primary track when they register each submission, in addition to one or more sub-disciplines. Appropriate track and sub-discipline selection will ensure that each submission is reviewed by a knowledgeable set of reviewers.

Important Dates

Wednesday, January 15, 2025: Submissions site opens
Monday, February 10, 2025: Submissions due

Early February 2025: Bidding opens for reviewers
Early to mid-February 2025: Bidding closes for reviewers
Saturday, February 15, 2025: Reviews assigned
Monday, March 3, 2025: Reviews due
Tuesday, March 11, 2025: Reviews released
Monday, March 24, 2025: Author/reviewer discussion period
Monday, March 31, 2025: Meta-review deadline
Wednesday, April 9, 2025: Author notification
Wednesday, June 25 – Friday, June 27, 2025: Conference dates

Quick Links

Tracks

  • Track 1: Models and Methods: Algorithms, Inference, and Estimation
  • Track 2: Applications and Practice: Investigation, Evaluation, Interpretation, and Deployment
  • Track 3: Impact and Society: Policy, Public Health, Social Outcomes, and Economics

Evaluation

Works submitted to CHIL will be reviewed by at least 3 reviewers. Detailed reviewer instructions and evaluation criteria will be posted later. Reviewers will be asked to primarily judge the work according to the following criteria:

Relevance: Is the submission relevant to health, broadly construed? Does the problem addressed fall into the domains of machine learning and healthcare? 

Quality: Is the submission technically sound? Are claims well supported by theoretical analysis or experimental results? Are the authors careful and honest about evaluating both the strengths and weaknesses of their work? Is the work complete rather than a work in progress? 

Originality: Are the tasks, methods and results novel? Is it clear how this work differs from previous contributions? Is related work adequately cited to provide context? Does the submission contribute unique data, unique conclusions about existing data, or a unique theoretical or experimental approach?

Clarity: Is the submission clearly written? Is it well-organized? Does it adequately provide enough information for readers to reproduce experiments or results? 

Significance: Is the contribution of the work important? Are other researchers or practitioners likely to use the ideas or build on them? Does the work advance the state of the art in a demonstrable way? 

Final decisions will be made by Track and Proceedings Chairs, taking into account reviewer comments, ratings of confidence and expertise, and our own editorial judgment. Reviewers will be able to recommend that submissions change tracks or flag submissions for ethical issues, relevance and suitability concerns.


Submission Format and Guidelines

Submission Site

Submissions should be made via the online submission system (Coming soon!). At least one author of each accepted paper is required to register for, attend, and present the work at the conference in order for the paper to appear in the conference proceedings.

Length and Formatting

Submitted papers must be 8-10 pages (including all figures and tables). Unlimited additional pages can be used for references and additional supplementary materials (e.g. appendices). Reviewers will not be required to read the supplementary materials.

Authors are required to use the LaTex template: Overleaf

Required Sections
Similar to last year, two sections will be required: 1) Data and Code Availability, and 2) Institutional Review Board (IRB). If either of the sections are missing, the conference organizers reserve the right to desk-reject the submission.

Data and Code Availability: This initial paragraph is required. Briefly state what data you use (including citations if appropriate) and whether the data are available to other researchers. If you are not sharing code, you must explicitly state that you are not making your code available. If you are making your code available, then at the time of submission for review, please include your code as supplemental material or as a code repository link; in either case, your code must be anonymized. If your paper is accepted, then you should de-anonymize your code for the camera-ready version of the paper. If you do not include this data and code availability statement for your paper, or you provide code that is not anonymized at the time of submission, then your paper will be desk-rejected. Your experiments later could refer to this initial data and code availability statement if it is helpful (e.g., to avoid restating what data you use).

Institutional Review Board (IRB): This endmatter section is required. If your research requires IRB approval or has been designated by your IRB as Not Human Subject Research, then for the camera-ready version of the paper, you must provide IRB information (and at the time of submission for review, you can say that this IRB information will be provided if the paper is accepted). If your research does not require IRB approval, then you must state this to be the case. This section does not count toward the paper page limit.

Archival Submissions

Submissions to the main conference are considered archival and will appear in the published proceedings of the conference, if accepted. Author notification of acceptance will be provided by the listed date under Important Dates.

Preprint Submission Policy

Submissions to preprint servers (such as ArXiv or MedRxiv) are allowed while the papers are under review. While reviewers will be encouraged not to search for the papers, you accept that uploading the paper may make your identity known. 

Peer Review

The review process is mutually anonymous (aka “double blind”). Your submitted paper, as well as any supporting text or revisions provided during the discussion period, should be completely anonymized (including links to code repositories such as Github). Please do not include any identifying information, and refrain from citing the authors’ own prior work in anything other than third-person. Violations of this anonymity policy at any stage before final manuscript acceptance decisions may result in rejection without further review.

Conference organizers and reviewers are required to maintain confidentiality of submitted material. Upon acceptance, the titles, authorship, and abstracts of papers will be released prior to the conference.

You may not submit papers that are identical, or substantially similar to versions that are currently under review at another conference or journal, have been previously published, or have been accepted for publication. Submissions to the main conference are considered archival and will appear in the published proceedings of the conference if accepted.

An exception to this rule is extensions of workshop papers that have previously appeared in non-archival venues, such as workshops, arXiv, or similar without formal proceedings. These works may be submitted as-is or in an extended form, though they must follow our manuscript formatting guidelines. CHIL also welcomes full paper submissions that extend previously published short papers or abstracts, so long as the previously published version does not exceed 4 pages in length. Note that the submission should not cite the workshop/report and preserve anonymity in the submitted manuscript.

Upon submission, authors will select one or more relevant sub-discipline(s). Peer reviewers for a paper will be experts in the sub-discipline(s) selected upon its submission. 

Open Access

CHIL is committed to open science and ensuring our proceedings are freely available.

Responsible and Ethical Research 

Computer software submissions should include an anonymized code link or code attached as supplementary material, licensing information, and provide documentation to facilitate use and reproducibility (e.g., package versions, README, intended use, and execution examples that facilitate execution by other researchers). 

Submissions that include analysis on public datasets need to include appropriate citations and data sequestration protocols, including train/validation/test splits, where appropriate. Submissions that include analysis of non-public datasets need to additionally include information about data source, collection sites, subject demographics and subgroups statistics, data acquisition protocols, informed consent, IRB and any other information supporting evidence of adherence to data collection and release protocols. Read our Review Policy.

Authors should discuss ethical implications and responsible uses of their work.

Submission Areas

Track Chairs will oversee the reviewing process. In case you are not sure which track your submission fits under, feel free to contact the Track or Proceedings Chairs for clarification. The Proceedings Chairs reserve the right to move submissions between tracks if they believe that a submission has been misclassified.

Track 1: Models and Methods: Algorithms, Inference, and Estimation

Description

Advances in machine learning are critical for a better understanding of health. This track seeks technical contributions in modeling, inference, and estimation in health-focused or health-inspired settings. We welcome submissions that develop novel methods and algorithms, introduce relevant machine learning tasks, identify challenges with prevalent approaches, or learn from multiple sources of data (e.g. non-clinical and clinical data).

Our focus on health is broadly construed, including clinical healthcare, public health, and population health. While submissions should be primarily motivated by problems relevant to health, the contributions themselves are not required to be directly applied to real health data. For example, authors may use synthetic datasets to demonstrate properties of their proposed algorithms.  

We welcome submissions from many perspectives, including but not limited to supervised learning, unsupervised learning, reinforcement learning, causal inference, representation learning, survival analysis, domain adaptation or generalization, interpretability, robustness, and algorithmic fairness. All kinds of health-relevant data types are in scope, including tabular health records, time series, text, images, videos, knowledge graphs, and more. We welcome all kinds of methodologies, from deep learning to probabilistic modeling to rigorous theory and beyond.

Example Papers

Tang, S., Dunnmon, J. A., Qu, L., Saab, K. K., Baykaner, T., Lee-Messer, C., Rubin, D. L. “Modeling Multivariate Biosignals With Graph Neural Networks and Structured State Space Models.” Proceedings of the Conference on Health, Inference, and Learning (CHIL), 2023.

Zhu, J., Darnell, G., Kumar, A., Zhao, D., Li, B., Nguyen, X., Ren, S. Y. “PhysioMTL: Personalizing Physiological Patterns using Optimal Transport Multi-Task Regression.” Proceedings of the Conference on Health, Inference, and Learning (CHIL), 2022.

Choi, E., Xiao, C., Stewart, W. F., Sun, J. “MiME: Multilevel medical embedding of electronic health records for predictive healthcare.” Advances in Neural Information Processing Systems (NeurIPS), 2018.

Shalit, U., Johansson, F.D., Sontag, D. “Estimating individual treatment effect: generalization bounds and algorithms.” Proceedings of the 34th International Conference on Machine Learning (ICML), 2017.

Futoma, J., Hariharan, S., Heller, K.. “Learning to detect sepsis with a multitask Gaussian process RNN classifier.” Proceedings of the 34th International Conference on Machine Learning (ICML), 2017.

Track 2: Applications and Practice: Investigation, Evaluation, Interpretation, and Deployment

Description

The goal of this track is to highlight works applying robust methods, models, or practices to identify, characterize, audit, evaluate, or benchmark ML approaches to healthcare problems. Additionally, we welcome work describing deployments and datasets used to empirically evaluate these systems in practice. Whereas the goal of Track 1 is to select papers that show significant algorithmic novelty, submit your work here if the contribution is describing an emerging or established innovative application of ML in healthcare. Areas of interest include but are not limited to:

  • Datasets and simulation frameworks for addressing gaps in ML healthcare applications, such as monitoring real-world performance of ML algorithms
  • Tools and platforms that facilitate the deployment and integration of AI algorithms for healthcare applications
  • Innovative ML-based approaches to solving practical problems grounded in a healthcare application
  • Surveys, benchmarks, evaluations and best practices of using ML in healthcare
  • Emerging applications of AI in healthcare

Introducing a new method is not prohibited by any means for this track, but the focus should be on the extent to which the proposed ideas contribute to addressing a practical limitation (e.g., robustness, computational scalability, improved performance). We encourage submissions in both more traditional clinical areas (e.g., electronic health records (EHR), medical image analysis),  as well as in emerging fields (e.g., remote and telehealth medicine, integration of omics). 

Example Papers

Kinyanjui, N.M., Johansson, F.D. ADCB: “An Alzheimer’s disease simulator for benchmarking observational estimators of causal effects”. Proceedings of the Conference on Health, Inference, and Learning (CHIL), 2022.

Zhou, H., Chen, Y., Lipton, Z. “Evaluating Model Performance in Medical Datasets Over Time”. Proceedings of the Conference on Health, Inference, and Learning (CHIL), 2023.

Wang, X., Peng, Y., Lu, L., Lu, Z., Bagheri, M., Summers, R. M. “Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases.” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.

Nestor B., McDermott, M.B., Boag W., Berner, G., Naumann, T., Hughes, M. C, Goldenberg, A., and Ghassemi, M. “Feature robustness in non-stationary health records: caveats to deployable model performance in common clinical machine learning tasks.” Machine Learning for Healthcare Conference (MLHC), 2019.

Cha, K. H., Petrick, N., Pezeshk, A., Graff, C. G., Sharma, D., Badal, A., Berkman Sahiner, B. “Evaluation of data augmentation via synthetic images for improved breast mass detection on mammograms using deep learning.Journal of Medical Imaging (2020).

Track 3: Impact and Society: Policy, Public Health, and Social Outcomes

Description

Algorithms do not exist in a vacuum: instead, they often explicitly aim for important social outcomes. This track considers issues at the intersection of algorithms and the societies they seek to impact, specifically for health. Submissions could include methodological contributions such as algorithmic development and performance evaluation for policy and public health applications, large-scale or challenging data collection, combining clinical and non-clinical data, as well as detecting and measuring bias. Submissions could also include impact-oriented research such as determining how algorithmic systems for health may introduce, exacerbate, or reduce inequities and inequalities, discrimination, and unjust outcomes, as well as evaluating the economic implications of these systems. We invite submissions tackling the responsible design of AI applications for healthcare and public health. System design for the implementation of such applications at scale is also welcome, which often requires balancing various tradeoffs in decision-making. Submissions related to understanding barriers to the deployment and adoption of algorithmic systems for societal-level health applications are also of interest. In addressing these problems, insights from social sciences, law, clinical medicine, and the humanities can be crucial.

Example Papers

Wu, K., Wu, E., Rodolfa, K., Ho, D., Zhou, J. “Regulating AI Adaptation: An Analysis of AI Medical Device Updates”. Proceedings of the Conference on Health, Inference, and Learning (CHIL), 2024.

Price, K.L., Amy Xia, H., Lakshminarayanan, M., Madigan, D., Manner, D., Scott, J., Stamey, J.D. and Thompson, L. “Bayesian methods for design and analysis of safety trials”. Pharmaceutical Statistics, 2024.

Seyyed-Kalantari, L., Zhang, H., McDermott M. A. B.,Chen, I. Y., Ghassemi, M. “Underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations.” Nature Medicine, 2021.

Yang, W., Lorch, L., Graule, M., Lakkaraju, H., Doshi-Velez, F. “Incorporating interpretable output constraints in Bayesian neural networks.” Advances in Neural Information Processing Systems (NeurIPS), 2020.

Obermeyer, Z.,Powers, B., Vogeli, C., Mullainathan, S. “Dissecting racial bias in an algorithm used to manage the health of populations“. Science, 2019.