Reference : Generalising from conventional pipelines using deep learning in high‑throughput scree... |
Scientific journals : Article | |||
Life sciences : Multidisciplinary, general & others Engineering, computing & technology : Multidisciplinary, general & others | |||
Systems Biomedicine | |||
http://hdl.handle.net/10993/48972 | |||
Generalising from conventional pipelines using deep learning in high‑throughput screening workfows | |
English | |
Garcia Santa Cruz, Beatriz ![]() | |
Sölter, Jan [] | |
Gomez Giro, Gemma [University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB) > Developmental and Cellular Biology >] | |
Saraiva, Claudia [University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB) > Developmental and Cellular Biology >] | |
Sabaté Soler, Sonia [University of Luxembourg > Faculty of Science, Technology and Medicine (FSTM) > Department of Life Sciences and Medicine (DLSM) >] | |
Modamio Chamarro, Jenifer [] | |
Barmpa, Kyriaki [University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB) > Developmental and Cellular Biology >] | |
Schwamborn, Jens Christian [University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB) > Developmental and Cellular Biology >] | |
Hertel, Frank [University of Luxembourg > Faculty of Science, Technology and Communication (FSTC) > >] | |
Jarazo, Javier [University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB) > >] | |
Husch, Andreas [University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB) > Interventional Neuroscience >] | |
6-Jul-2022 | |
Scientific Reports | |
Nature Publishing Group | |
Yes | |
International | |
2045-2322 | |
London | |
United Kingdom | |
[en] complex disease ; high-throughput screening ; image analysis ; deep learning approaches ; microscopy-image analysis | |
[en] The study of complex diseases relies on large amounts of data to build models toward precision
medicine. Such data acquisition is feasible in the context of high-throughput screening, in which the quality of the results relies on the accuracy of the image analysis. Although state-of-the-art solutions for image segmentation employ deep learning approaches, the high cost of manually generating ground truth labels for model training hampers the day-to-day application in experimental laboratories. Alternatively, traditional computer vision-based solutions do not need expensive labels for their implementation. Our work combines both approaches by training a deep learning network using weak training labels automatically generated with conventional computer vision methods. Our network surpasses the conventional segmentation quality by generalising beyond noisy labels, providing a 25% increase of mean intersection over union, and simultaneously reducing the development and inference times. Our solution was embedded into an easy-to-use graphical user interface that allows researchers to assess the predictions and correct potential inaccuracies with minimal human input. To demonstrate the feasibility of training a deep learning solution on a large dataset of noisy labels automatically generated by a conventional pipeline, we compared our solution against the common approach of training a model from a small manually curated dataset by several experts. Our work suggests that humans perform better in context interpretation, such as error assessment, while computers outperform in pixel-by-pixel fne segmentation. Such pipelines are illustrated with a case study on image segmentation for autophagy events. This work aims for better translation of new technologies to real-world settings in microscopy-image analysis. | |
Researchers ; Professionals ; General public | |
http://hdl.handle.net/10993/48972 | |
10.1038/s41598-022-15623-7 | |
https://www.nature.com/articles/s41598-022-15623-7 |
File(s) associated to this reference | ||||||||||||||
Fulltext file(s):
| ||||||||||||||
All documents in ORBilu are protected by a user license.