Skip to content

Commit

Permalink
Merge pull request #34 from Inria-Empenn/hugoblox-import-publications
Browse files Browse the repository at this point in the history
Hugo Blox Builder - Import latest publications
  • Loading branch information
alexpron authored Oct 25, 2024
2 parents 7d7ab08 + 456fe67 commit fbdc8b4
Show file tree
Hide file tree
Showing 16 changed files with 600 additions and 0 deletions.
68 changes: 68 additions & 0 deletions content/publication/chen-sampling-2023/cite.bib
Original file line number Diff line number Diff line change
@@ -0,0 +1,68 @@
@article{chen_sampling_2023,
abstract = {Abstract
Background
The development of machine learning models for aiding in the diagnosis of mental disorder is recognized as a significant breakthrough in the field of psychiatry. However, clinical practice of such models remains a challenge, with poor generalizability being a major limitation.
Methods
Here, we conducted a pre-registered meta-research assessment on neuroimaging-based models in the psychiatric literature, quantitatively examining global and regional sampling issues over recent decades, from a view that has been relatively underexplored. A total of 476 studies (
n
= 118,137) were included in the current assessment. Based on these findings, we built a comprehensive 5-star rating system to quantitatively evaluate the quality of existing machine learning models for psychiatric diagnoses.
Results
A global sampling inequality in these models was revealed quantitatively (sampling Gini coefficient (
G
) = 0.81,
p
\textless .01), varying across different countries (regions) (e.g., China,
G
= 0.47; the USA,
G
= 0.58; Germany,
G
= 0.78; the UK,
G
= 0.87). Furthermore, the severity of this sampling inequality was significantly predicted by national economic levels (
β
=  − 2.75,
p
\textless .001,
R
2
adj
= 0.40;
r
=  − .84, 95% CI: − .41 to − .97), and was plausibly predictable for model performance, with higher sampling inequality for reporting higher classification accuracy. Further analyses showed that lack of independent testing (84.24% of models, 95% CI: 81.0–87.5%), improper cross-validation (51.68% of models, 95% CI: 47.2–56.2%), and poor technical transparency (87.8% of models, 95% CI: 84.9–90.8%)/availability (80.88% of models, 95% CI: 77.3–84.4%) are prevailing in current diagnostic classifiers despite improvements over time. Relating to these observations, model performances were found decreased in studies with independent cross-country sampling validations (all
p
\textless .001, BF
10
\textgreater 15). In light of this, we proposed a purpose-built quantitative assessment checklist, which demonstrated that the overall ratings of these models increased by publication year but were negatively associated with model performance.
Conclusions
Together, improving sampling economic equality and hence the quality of machine learning models may be a crucial facet to plausibly translating neuroimaging-based diagnostic classifiers into clinical practice.},
author = {Chen, Zhiyi and Hu, Bowen and Liu, Xuerong and Becker, Benjamin and Eickhoff, Simon B. and Miao, Kuan and Gu, Xingmei and Tang, Yancheng and Dai, Xin and Li, Chao and Leonov, Artemiy and Xiao, Zhibing and Feng, Zhengzhi and Chen, Ji and Chuan-Peng, Hu},
doi = {10.1186/s12916-023-02941-4},
file = {Chen et al. - 2023 - Sampling inequalities affect generalization of neu.pdf:/home/alpron/Zotero/storage/YTDHFA4C/Chen et al. - 2023 - Sampling inequalities affect generalization of neu.pdf:application/pdf},
issn = {1741-7015},
journal = {BMC Medicine},
language = {en},
month = {July},
number = {1},
pages = {241},
title = {Sampling inequalities affect generalization of neuroimaging-based diagnostic classifiers in psychiatry},
url = {https://bmcmedicine.biomedcentral.com/articles/10.1186/s12916-023-02941-4},
urldate = {2024-10-21},
volume = {21},
year = {2023}
}
58 changes: 58 additions & 0 deletions content/publication/chen-sampling-2023/index.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
---
title: Sampling inequalities affect generalization of neuroimaging-based diagnostic
classifiers in psychiatry
authors:
- Zhiyi Chen
- Bowen Hu
- Xuerong Liu
- Benjamin Becker
- Simon B. Eickhoff
- Kuan Miao
- Xingmei Gu
- Yancheng Tang
- Xin Dai
- Chao Li
- Artemiy Leonov
- Zhibing Xiao
- Zhengzhi Feng
- Ji Chen
- Hu Chuan-Peng
date: '2023-07-01'
publishDate: '2024-10-25T13:25:50.233832Z'
publication_types:
- article-journal
publication: '*BMC Medicine*'
doi: 10.1186/s12916-023-02941-4
abstract: 'Abstract Background The development of machine learning models for aiding
in the diagnosis of mental disorder is recognized as a significant breakthrough
in the field of psychiatry. However, clinical practice of such models remains a
challenge, with poor generalizability being a major limitation. Methods Here,
we conducted a pre-registered meta-research assessment on neuroimaging-based models
in the psychiatric literature, quantitatively examining global and regional sampling
issues over recent decades, from a view that has been relatively underexplored.
A total of 476 studies ( n = 118,137) were included in the current assessment. Based
on these findings, we built a comprehensive 5-star rating system to quantitatively
evaluate the quality of existing machine learning models for psychiatric diagnoses. Results A
global sampling inequality in these models was revealed quantitatively (sampling
Gini coefficient ( G ) = 0.81, p textless .01), varying across different countries
(regions) (e.g., China, G = 0.47; the USA, G = 0.58; Germany, G = 0.78; the UK,
G = 0.87). Furthermore, the severity of this sampling inequality was significantly
predicted by national economic levels ( β =  − 2.75, p textless .001, R 2 adj = 0.40;
r =  − .84, 95% CI: − .41 to − .97), and was plausibly predictable for model performance,
with higher sampling inequality for reporting higher classification accuracy. Further
analyses showed that lack of independent testing (84.24% of models, 95% CI: 81.0–87.5%),
improper cross-validation (51.68% of models, 95% CI: 47.2–56.2%), and poor technical
transparency (87.8% of models, 95% CI: 84.9–90.8%)/availability (80.88% of models,
95% CI: 77.3–84.4%) are prevailing in current diagnostic classifiers despite improvements
over time. Relating to these observations, model performances were found decreased
in studies with independent cross-country sampling validations (all p textless .001,
BF 10 textgreater 15). In light of this, we proposed a purpose-built quantitative
assessment checklist, which demonstrated that the overall ratings of these models
increased by publication year but were negatively associated with model performance. Conclusions
Together, improving sampling economic equality and hence the quality of machine
learning models may be a crucial facet to plausibly translating neuroimaging-based
diagnostic classifiers into clinical practice.'
links:
- name: URL
url: https://bmcmedicine.biomedcentral.com/articles/10.1186/s12916-023-02941-4
---
15 changes: 15 additions & 0 deletions content/publication/christodoulou-confidence-2024/cite.bib
Original file line number Diff line number Diff line change
@@ -0,0 +1,15 @@
@misc{christodoulou_confidence_2024,
abstract = {Medical imaging is spearheading the AI transformation of healthcare. Performance reporting is key to determine which methods should be translated into clinical practice. Frequently, broad conclusions are simply derived from mean performance values. In this paper, we argue that this common practice is often a misleading simplification as it ignores performance variability. Our contribution is threefold. (1) Analyzing all MICCAI segmentation papers (n = 221) published in 2023, we first observe that more than 50% of papers do not assess performance variability at all. Moreover, only one (0.5%) paper reported confidence intervals (CIs) for model performance. (2) To address the reporting bottleneck, we show that the unreported standard deviation (SD) in segmentation papers can be approximated by a second-order polynomial function of the mean Dice similarity coefficient (DSC). Based on external validation data from 56 previous MICCAI challenges, we demonstrate that this approximation can accurately reconstruct the CI of a method using information provided in publications. (3) Finally, we reconstructed 95% CIs around the mean DSC of MICCAI 2023 segmentation papers. The median CI width was 0.03 which is three times larger than the median performance gap between the first and second ranked method. For more than 60% of papers, the mean performance of the second-ranked method was within the CI of the first-ranked method. We conclude that current publications typically do not provide sufficient evidence to support which models could potentially be translated into clinical practice.},
author = {Christodoulou, Evangelia and Reinke, Annika and Houhou, Rola and Kalinowski, Piotr and Erkan, Selen and Sudre, Carole H. and Burgos, Ninon and Boutaj, Sofiène and Loizillon, Sophie and Solal, Maëlys and Rieke, Nicola and Cheplygina, Veronika and Antonelli, Michela and Mayer, Leon D. and Tizabi, Minu D. and Cardoso, M. Jorge and Simpson, Amber and Jäger, Paul F. and Kopp-Schneider, Annette and Varoquaux, Gaël and Colliot, Olivier and Maier-Hein, Lena},
file = {Christodoulou et al. - 2024 - Confidence intervals uncovered Are we ready for r.pdf:/home/alpron/Zotero/storage/DRN46WQP/Christodoulou et al. - 2024 - Confidence intervals uncovered Are we ready for r.pdf:application/pdf},
keywords = {Computer Science - Artificial Intelligence, Computer Science - Computer Vision and Pattern Recognition, Computer Science - Machine Learning},
language = {en},
month = {September},
note = {arXiv:2409.17763 [cs]},
publisher = {arXiv},
shorttitle = {Confidence intervals uncovered},
title = {Confidence intervals uncovered: Are we ready for real-world medical imaging AI?},
url = {http://arxiv.org/abs/2409.17763},
urldate = {2024-10-21},
year = {2024}
}
58 changes: 58 additions & 0 deletions content/publication/christodoulou-confidence-2024/index.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
---
title: 'Confidence intervals uncovered: Are we ready for real-world medical imaging
AI?'
authors:
- Evangelia Christodoulou
- Annika Reinke
- Rola Houhou
- Piotr Kalinowski
- Selen Erkan
- Carole H. Sudre
- Ninon Burgos
- Sofiène Boutaj
- Sophie Loizillon
- Maëlys Solal
- Nicola Rieke
- Veronika Cheplygina
- Michela Antonelli
- Leon D. Mayer
- Minu D. Tizabi
- M. Jorge Cardoso
- Amber Simpson
- Paul F. Jäger
- Annette Kopp-Schneider
- Gaël Varoquaux
- Olivier Colliot
- Lena Maier-Hein
date: '2024-09-01'
publishDate: '2024-10-25T13:25:50.225704Z'
publication_types:
- manuscript
publication: '*arXiv*'
abstract: Medical imaging is spearheading the AI transformation of healthcare. Performance
reporting is key to determine which methods should be translated into clinical practice.
Frequently, broad conclusions are simply derived from mean performance values. In
this paper, we argue that this common practice is often a misleading simplification
as it ignores performance variability. Our contribution is threefold. (1) Analyzing
all MICCAI segmentation papers (n = 221) published in 2023, we first observe that
more than 50% of papers do not assess performance variability at all. Moreover,
only one (0.5%) paper reported confidence intervals (CIs) for model performance.
(2) To address the reporting bottleneck, we show that the unreported standard deviation
(SD) in segmentation papers can be approximated by a second-order polynomial function
of the mean Dice similarity coefficient (DSC). Based on external validation data
from 56 previous MICCAI challenges, we demonstrate that this approximation can accurately
reconstruct the CI of a method using information provided in publications. (3) Finally,
we reconstructed 95% CIs around the mean DSC of MICCAI 2023 segmentation papers.
The median CI width was 0.03 which is three times larger than the median performance
gap between the first and second ranked method. For more than 60% of papers, the
mean performance of the second-ranked method was within the CI of the first-ranked
method. We conclude that current publications typically do not provide sufficient
evidence to support which models could potentially be translated into clinical practice.
tags:
- Computer Science - Artificial Intelligence
- Computer Science - Computer Vision and Pattern Recognition
- Computer Science - Machine Learning
links:
- name: URL
url: http://arxiv.org/abs/2409.17763
---
18 changes: 18 additions & 0 deletions content/publication/cokelaer-reprohackathons-2023/cite.bib
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
@article{cokelaer_reprohackathons_2023,
abstract = {The reproducibility crisis has highlighted the importance of improving the way bioinformatics data analyses are implemented, executed, and shared. To address this, various tools such as content versioning systems, workflow management systems, and software environment management systems have been developed. While these tools are becoming more widely used, there is still much work to be done to increase their adoption. The most effective way to ensure reproducibility becomes a standard part of most bioinformatics data analysis projects is to integrate it into the curriculum of bioinformatics Master’s programs.In this article, we present the Reprohackathon, a Master’s course that we have been running for the last 3 years at Université Paris-Saclay (France), and that has been attended by a total of 123 students. The course is divided into two parts. The first part includes lessons on the challenges related to reproducibility, content versioning systems, container management, and workflow systems. In the second part, students work on a data analysis project for 3–4 months, reanalyzing data from a previously published study. The Reprohackaton has taught us many valuable lessons, such as the fact that implementing reproducible analyses is a complex and challenging task that requires significant effort. However, providing in-depth teaching of the concepts and the tools during a Master’s degree program greatly improves students’ understanding and abilities in this area.},
author = {Cokelaer, Thomas and Cohen-Boulakia, Sarah and Lemoine, Frédéric},
doi = {10.1093/bioinformatics/btad227},
file = {Full Text PDF:/home/alpron/Zotero/storage/XGMI5F5J/Cokelaer et al. - 2023 - Reprohackathons promoting reproducibility in bioi.pdf:application/pdf;Snapshot:/home/alpron/Zotero/storage/5USUN9I3/7210451.html:text/html},
issn = {1367-4811},
journal = {Bioinformatics},
keywords = {hackathon, training},
month = {June},
number = {Supplement_1},
pages = {i11--i20},
shorttitle = {Reprohackathons},
title = {Reprohackathons: promoting reproducibility in bioinformatics through training},
url = {https://doi.org/10.1093/bioinformatics/btad227},
urldate = {2024-10-01},
volume = {39},
year = {2023}
}
37 changes: 37 additions & 0 deletions content/publication/cokelaer-reprohackathons-2023/index.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,37 @@
---
title: 'Reprohackathons: promoting reproducibility in bioinformatics through training'
authors:
- Thomas Cokelaer
- Sarah Cohen-Boulakia
- Frédéric Lemoine
date: '2023-06-01'
publishDate: '2024-10-25T13:25:50.211524Z'
publication_types:
- article-journal
publication: '*Bioinformatics*'
doi: 10.1093/bioinformatics/btad227
abstract: The reproducibility crisis has highlighted the importance of improving the
way bioinformatics data analyses are implemented, executed, and shared. To address
this, various tools such as content versioning systems, workflow management systems,
and software environment management systems have been developed. While these tools
are becoming more widely used, there is still much work to be done to increase their
adoption. The most effective way to ensure reproducibility becomes a standard part
of most bioinformatics data analysis projects is to integrate it into the curriculum
of bioinformatics Master’s programs.In this article, we present the Reprohackathon,
a Master’s course that we have been running for the last 3 years at Université Paris-Saclay
(France), and that has been attended by a total of 123 students. The course is divided
into two parts. The first part includes lessons on the challenges related to reproducibility,
content versioning systems, container management, and workflow systems. In the second
part, students work on a data analysis project for 3–4 months, reanalyzing data
from a previously published study. The Reprohackaton has taught us many valuable
lessons, such as the fact that implementing reproducible analyses is a complex and
challenging task that requires significant effort. However, providing in-depth teaching
of the concepts and the tools during a Master’s degree program greatly improves
students’ understanding and abilities in this area.
tags:
- hackathon
- training
links:
- name: URL
url: https://doi.org/10.1093/bioinformatics/btad227
---
18 changes: 18 additions & 0 deletions content/publication/cosmo-software-2023/cite.bib
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
@incollection{cosmo_software_2023,
abstract = {Software Heritage is the largest public archive of software source code and associated development history, as captured by modern version control systems. As of July 2023, it has archived more than 16 billion unique source code files coming from more than 250 million collaborative development projects. In this chapter, we describe the Software Heritage ecosystem, focusing on research and open science use cases.},
address = {Cham},
author = {Cosmo, Roberto Di and Zacchiroli, Stefano},
booktitle = {Software Ecosystems: Tooling and Analytics},
doi = {10.1007/978-3-031-36060-2_2},
editor = {Mens, Tom and De Roover, Coen and Cleve, Anthony},
file = {Cosmo and Zacchiroli - 2023 - The Software Heritage Open Science Ecosystem.pdf:/home/alpron/Zotero/storage/X9QREIGC/Cosmo and Zacchiroli - 2023 - The Software Heritage Open Science Ecosystem.pdf:application/pdf},
isbn = {978-3-031-36060-2},
keywords = {Reproducibility, FAIR, open scienc, swh, swh data model, swh features},
language = {en},
pages = {33--61},
publisher = {Springer International Publishing},
title = {The Software Heritage Open Science Ecosystem},
url = {https://doi.org/10.1007/978-3-031-36060-2_2},
urldate = {2024-07-12},
year = {2023}
}
28 changes: 28 additions & 0 deletions content/publication/cosmo-software-2023/index.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,28 @@
---
title: The Software Heritage Open Science Ecosystem
authors:
- Roberto Di Cosmo
- Stefano Zacchiroli
date: '2023-01-01'
publishDate: '2024-10-25T13:25:50.260016Z'
publication_types:
- chapter
publication: '*Software Ecosystems: Tooling and Analytics*'
doi: 10.1007/978-3-031-36060-2_2
abstract: Software Heritage is the largest public archive of software source code
and associated development history, as captured by modern version control systems.
As of July 2023, it has archived more than 16 billion unique source code files coming
from more than 250 million collaborative development projects. In this chapter,
we describe the Software Heritage ecosystem, focusing on research and open science
use cases.
tags:
- Reproducibility
- FAIR
- open scienc
- swh
- swh data model
- swh features
links:
- name: URL
url: https://doi.org/10.1007/978-3-031-36060-2_2
---
23 changes: 23 additions & 0 deletions content/publication/marelli-scrutinizing-2018/cite.bib
Original file line number Diff line number Diff line change
@@ -0,0 +1,23 @@
@article{marelli_scrutinizing_2018,
abstract = {How will new decentralized governance impact research?
,
On 25 May 2018, the European Union (EU) regulation 2016/679 on data protection, also known as the General Data Protection Regulation (GDPR), will take effect. The GDPR, which repeals previous European legislation on data protection (Directive 95/46/EC) (
1
), is bound to have major effects on biomedical research and digital health technologies, in Europe and beyond, given the global reach of EU-based research and the prominence of international research networks requiring interoperability of standards. Here we describe ways in which the GDPR will become a critical tool to structure flexible governance for data protection. As a timely forecast for its potential impact, we analyze the implications of the GDPR in an ongoing paradigmatic legal controversy involving the database originally assembled by one of the world's first genomic biobanks, Shardna.},
author = {Marelli, Luca and Testa, Giuseppe},
copyright = {http://www.sciencemag.org/about/science-licenses-journal-article-reuse},
doi = {10.1126/science.aar5419},
file = {Submitted Version:/home/alpron/Zotero/storage/LIQDLUX2/Marelli and Testa - 2018 - Scrutinizing the EU General Data Protection Regula.pdf:application/pdf},
issn = {0036-8075, 1095-9203},
journal = {Science},
language = {en},
month = {May},
number = {6388},
pages = {496--498},
title = {Scrutinizing the EU General Data Protection Regulation},
url = {https://www.science.org/doi/10.1126/science.aar5419},
urldate = {2024-10-22},
volume = {360},
year = {2018}
}
Loading

0 comments on commit fbdc8b4

Please sign in to comment.