Publications
”Dual-feature selectivity enables bidirectional coding in visual cortical neurons”
Januar 2026
eLife (Reviewed Preprint)
Katrin FrankeNikos KarantzasKonstantin WillekeMaria DiamantakiKandan RamakrishnanHasan Atakan BedelPavithra ElumalaiKelli RestivoPaul FaheyCate NealleyTori ShinnGabrielle GarciaSaumil PatelAlexander EckerEdgar Y WalkerEmmanouil FroudarakisSophia SanbornFabian H SinzAndreas Tolias
CAIMed Groups:
Digital Twins
Biomedical Image Recognition
Sensory neurons are traditionally viewed as feature detectors that respond with an increase in firing rate to preferred stimuli while remaining unresponsive to others. Here, we identify a dual-feature encoding strategy in macaque visual cortex, wherein many neurons in areas V1 and V4 are selectively tuned to two distinct visual features—one that enhances and one that suppresses activity—around an elevated baseline firing rate. By combining neuronal recordings with functional digital twin models—deep learningbased predictive models of biological neurons—we were able to systematically identify each neuron’s preferred and non-preferred features. These feature pairs served as anchors for a continuous, low-dimensional axis in natural image similarity space, along which neuronal activity varied approximately linearly. Within a single visual area, visual features that strongly or weakly activated individual neurons also had a high probability of modulating the activity of other neurons, suggesting a shared feature selectivity across the population that structures stimulus encoding. We show that this encoding strategy is conserved across species, present in both primary and lateral visual areas of mouse cortex. Dual-feature selectivity is consistent with recent anatomical evidence for feature-specific inhibitory connectivity, complementing the feature-detector principle through circuit mechanisms in which selective excitation and inhibition may together enhance the representational capacity of the neuronal population.
The full text can be found here .
Funded by CAIMed
”A distinct monocyte transcriptional state links systemic immune dysregulation to pulmonary impairment in long COVID”
Januar 2026
nature immnunoloy
Saumya Kumar, Chaofan Li, Liang Zhou, Qiuyao Zhan, Ahmed Alaswad, Sonja Volland, Bibiana Costa, Simon Alexander Krooss, Isabel Klefenz, Hagen Schmaus, Antonia Zeuzem, Dorothee von Witzendorff, Helena Lickei, Lea Pueschel, Anke R. M. Kraft, Markus Cornberg, Andreas Rembert Koczulla, Isabell Pink, Marius M. Hoeper, Cheng-Jian Xu, Susanne Häussler, Miriam Wiestler, Mihai G. Netea, Thomas Illig, Jie Sun & Yang Li
CAIMed Groups:
AI & Bioinformatics
The mechanisms driving immune dysregulation in long COVID disease remain elusive. Here we integrated single-cell multiome data, immunological profiling and functional assays to investigate immune alterations across multiple cohorts. A transcriptional state in circulating monocytes (LC-Mo) was enriched in individuals with mild–moderate acute infection and accompanied by persistent elevations of plasma CCL2, CXCL11 and TNF. LC-Mo showed TGFβ and WNT–β-catenin signaling and correlated with fatigue severity. Protein markers of LC-Mo were increased in individuals with pronounced fatigue or dyspnea, and those with severe respiratory symptoms showed higher LC-Mo expression. Epigenetically, LC-Mo exhibited AP-1- and NF-κB1-driven profibrotic programs. LC-Mo-like macrophages in bronchoalveolar lavage samples from individuals with severe respiratory symptoms displayed a profibrotic profile, and individuals with a high LC-Mo transcriptional state showed impaired interferon responses after stimulation. Collectively, our findings define a pathogenic monocyte transcriptional state linking systemic immune dysfunction to persistent long COVID disease, providing mechanistic insights and potential therapeutic targets.
The full text can be found here .
Funded by CAIMed
“SPONGE: Competing Sparse Language Representations forEffective Cross-Lingual Knowledge Transfer in Healthcare”
Dezember 2025
Transactions on Machine Learning Research
Jens-Michalis Papaioannou, Alexei Figueroa Rosero, Conor Fallon, Anna Capilla, Alexandra Bekiaridou, Stavros Zanos, Wolfgang Nejdl and Alexander Löser
CAIMed Groups:
AI & Causality
In domains with privacy constraints, most knowledge resides in siloed datasets, hindering the development of a model with all relevant knowledge for a task. Clinical NLP is a prime example of these constraints in practice. Research in this area typically falls back to the canonical setting of sequential transfer learning, where a model pre-trained on large corporats finetuned on a smaller annotated dataset. An avenue for knowledge transfer among diverse clinics is multi-step sequential transfer learning since models1 are more likely to be shared than private clinical data. This setting poses challenges of cross-linguality, domain diversity, and varying label distributions which undermine generalisation. We propose SPONGE, an efficient prototypical architecture that leverages competing sparse language representations. These encompass distributed knowledge and create the necessary level of redundancy foreffective transfer learning across multiple datasets. We identify that prototypical classifiers are critically sensitive to label-recency bias which we mitigate with a novel strategy at inference time. SPONGE in combination with this strategy significantly boosts generalisation performance to unseen data. With the help of medical professionals, we show that the explainability of our models is clinically relevant. We make all source code2 available
The full text can be found here .
Funded by CAIMed
ITL-LIME: Instance-Based Transfer Learning for Enhancing Local Explanations in Low-Resource Data Settings
November 2025
CIKM ’25: Proceedings of the 34th ACM International Conference on Information and Knowledge Management
Rehan Raza, Guanjin Wang, Kok Wai Wong, Hamid Laga, Marco Fisichella
DOI:
https://doi.org/10.1145/3746252.3761183
CAIMed Groups:
AI & Causality
Explainable Artificial Intelligence (XAI) methods, such as Local Interpretable Model-Agnostic Explanations (LIME), have advanced the interpretability of black-box machine learning models by approximating their behavior locally using interpretable surrogate models. However, LIME’s inherent randomness in perturbation and sampling can lead to locality and instability issues, especially in scenarios with limited training data. In such cases, data scarcity can result in the generation of unrealistic variations and samples that deviate from the true data manifold. Consequently, the surrogate model may fail to accurately approximate the complex decision boundary of the original model. To address these challenges, we propose a novel Instance-based Transfer Learning LIME framework (ITL-LIME) that enhances explanation fidelity and stability in data-constrained environments. ITL-LIME introduces instance transfer learning into the LIME framework by leveraging relevant real instances from a related source domain to aid the explanation process in the target domain. Specifically, we employ clustering to partition the source domain into clusters with representative prototypes. Instead of generating random perturbations, our method retrieves pertinent real source instances from the source cluster whose prototype is most similar to the target instance. These are then combined with the target instance’s neighboring real instances. To define a compact locality, we further construct a contrastive learning-based encoder as a weighting mechanism to assign weights to the instances from the combined set based on their proximity to the target instance. Finally, these weighted source and target instances are used to train the surrogate model for explanation purposes. Experimental evaluation with real-world datasets demonstrates that ITL-LIME greatly improves the stability and fidelity of LIME explanations in scenarios with limited data. Our code is available at https://github.com/rehanrazaa/ITL-LIME.
The full text can be found here .
Funded by CAIMed
“Rare Structural Variants Uncovered by Optical Genome Mapping in Multisystem Inflammatory Syndrome in Children (MIS-C)”
August 2025
Advanced genetics (Hoboken, N.J.)
Catherine A. Brownstein, Caspar I. van der Made, Kristin Cabral, Shira Rockowitz, Donghun Kang, Maximilian Schieck, Andy Wing Chun Pang, Jeffrey M. Robinson, Alex R. Hastie, Alka Chaubey, The C.O.V.I.D19hostgenomeS.V. Consortium, Alexander Hoischen, Alan H. Beggs
Multisystem inflammatory syndrome in children (MIS-C) is a pediatric complication of SARS-CoV-2 infection characterized by multiorgan inflammation and frequently by cardiovascular dysfunction. In a single-center prospective cohort study, optical genome mapping (OGM) was performed on 14 patients, including 11 meeting CDC criteria for MIS-C and 3 with MIS-C–like (MIS-CL) presentations. SVs and CNVs were filtered against population and internal OGM control databases. Seven patients (50%) harbored prioritized variants within or near genes implicated in immune regulation or SARS-CoV-2 response. These included intronic insertions or deletions in ORAI1, STAT4, and ITPR1 (n = 4 patients); a heterozygous insertion disrupting BATF; a large deletion spanning exons 2–10 of CFHR5; and an upstream insertion near DOCK2. Application of OGM to patients with MIS-C and MIS-CL revealed SVs potentially impacting inflammation, COVID-19 severity, and Kawasaki Disease susceptibility. Although causality cannot yet be assigned, the identification of rare structural variants highlights biologically plausible mechanisms that may contribute to disease heterogeneity. These findings establish the feasibility and value of OGM in the assessment of complex pediatric syndromes, such as children with MIS-C or a severe course of SARS-CoV-2 infection.
The full text can be found here .
Funded by CAIMed
“Learning from the Right Patches: A Two-Stage Wavelet-Driven MaskedAutoencoder for Histopathology Representation Learning”
November 2025
arxiv.org
Maryam Badar, Raneen Younis, Sandipan Sikdar, Wolfgang Nejdl & Marco Fisichella
CAIMed Groups:
Human Centered AI
Federated learning (FL) is an emerging communication-efficient and collaborative learning paradigm of machine learning with privacy guarantees. As these advancements unfold, adapting FL for fairness-aware learning becomes crucial. In this context, we propose a pre-processing fairness and utility (balanced accuracy) enhancing agnostic federated learning framework (Fed-FUEL) that mitigates discrimination embedded in the non-independent identically distributed data. We contribute a novel adaptive data manipulation method that mitigates discrimination embedded in the data at client side during optimization, resulting in an optimized and fair centralized server. This pre-processing approach abstracts the model architecture from the equation, offering a significant advantage in a federated environment. This abstraction not only facilitates a broader application across diverse model architectures without necessitating modifications but also sidesteps the potential complexities and inefficiencies associated with model-specific in-processing methods. Extensive experiments with a range of publicly available datasets demonstrate that our method outperforms the competing baselines in terms of both discrimination mitigation and predictive performance. Our model effectively adapts to both statistical and causal fairness notions, as shown through our experiments.
The full text can be found here .
Funded by CAIMed
“In Silico Clinical Trials in Drug Development: A Systematic Review”
November 2025
Therapeutic Innovation & Regulatory Science
Bohua Chen, Lucia Chantal Schneider, Christian Röver, Emmanuelle Comets, Markus Christian Elze, Andrew Hooker, Joanna IntHout, Anne-Sophie Jannot, Daria Julkowska, Yanis Mimouni, Marina Savelieva, Nigel Stallard, Moreno Ursino, Marc Vandemeulebroecke, Sebastian Weber, Martin Posch, Sarah Zohar & Tim Friede
CAIMed Groups:
Statistical Evidence in AI Systems
In the context of clinical research, computational models have received increasing attention over the past decades. In this systematic review, we aimed to provide an overview of the role of so-called in silico clinical trials (ISCTs) in medical applications. Exemplary for the broad field of clinical medicine, we focused on in silico (IS) methods applied in drug development, sometimes also referred to as model informed drug development (MIDD). We searched PubMed and ClinicalTrials.gov for published articles and registered clinical trials related to ISCTs. We identified 202 articles and 48 trials, and of these, 76 articles and 19 trials were directly linked to drug development. We extracted information from all 202 articles and 48 clinical trials and conducted a more detailed review of the methods used in the 76 articles that are connected to drug development. Regarding application, most articles and trials focused on cancer and imaging-related research while rare and pediatric diseases were only addressed in 14 articles and 5 trials, respectively. While some models were informed combining mechanistic knowledge with clinical or preclinical (in-vivo or in-vitro) data, the majority of models were fully data-driven, illustrating that clinical data is a crucial part in the process of generating synthetic data in ISCTs. Regarding reproducibility, a more detailed analysis revealed that only 24% (18 out of 76) of the articles provided an open-source implementation of the applied models, and in only 20% of the articles the generated synthetic data were publicly available. Despite the widely raised interest, we also found that it is still uncommon for ISCTs to be part of a registered clinical trial and their application is restricted to specific diseases leaving potential benefits of ISCTs not fully exploited.
The full text can be found here .
Funded by CAIMed
“Fed-FUEL: fairness and utility enhancing agnostic federated learning framework”
September 2025
Data Mining and Knowledge Discovery
Maryam Badar, Raneen Younis, Sandipan Sikdar, Wolfgang Nejdl and Marco Fisichella
CAIMed Groups:
AI & Causality
Federated learning (FL) is an emerging communication-efficient and collaborative learning paradigm of machine learning with privacy guarantees. As these advancements unfold, adapting FL for fairness-aware learning becomes crucial. In this context, we propose a pre-processing fairness and utility (balanced accuracy) enhancing agnostic federated learning framework (Fed-FUEL) that mitigates discrimination embedded in the non-independent identically distributed data. We contribute a novel adaptive data manipulation method that mitigates discrimination embedded in the data at client side during optimization, resulting in an optimized and fair centralized server. This pre-processing approach abstracts the model architecture from the equation, offering a significant advantage in a federated environment. This abstraction not only facilitates a broader application across diverse model architectures without necessitating modifications but also sidesteps the potential complexities and inefficiencies associated with model-specific in-processing methods. Extensive experiments with a range of publicly available datasets demonstrate that our method outperforms the competing baselines in terms of both discrimination mitigation and predictive performance. Our model effectively adapts to both statistical and causal fairness notions, as shown through our experiments.
The full text can be found here .
Funded by CAIMed
”Characterizing Vision Backbones for Dense Prediction with Dense Attentive Probing”
September 2025
TMLR
Timo Lüddecke, Alexander S. Ecker
CAIMed Groups:
Biomedical Image Recognition
The paradigm of pretraining a backbone on a large set of (often unlabeled) images has gained popularity. The quality of the resulting features is commonly measured by freezing the backbone and training different task heads on top of it. However, current evaluations cover only classifications of whole images or require complex dense task heads which introduce a large number of parameters and add their own inductive biases. In this work, we propose dense attentive probing, a parameter-efficient readout method for dense prediction on arbitrary backbones – independent of the size and resolution of their feature volume. To this end, we extend cross-attention with distance-based masks of learnable sizes. We employ this method to evaluate 18 common backbones on dense predictions tasks in three dimensions: instance awareness, local semantics and spatial understanding. We find that DINOv2 outperforms all other backbones tested – including those supervised with masks and language – across all three task categories. Furthermore, our analysis suggests that self-supervised pretraining tends to yield features that separate object instances better than vision-language models.
Code is available at http://eckerlab.org/code/deap.
The full text can be found here .
Funded by CAIMed
”Learning to cluster neuronal function”
September 2025
NeurIPS 2025
Nina Nellen, Polina Turishcheva, Michaela Vystrčilová, Shashwat Sridhar, Tim Gollisch, Andreas Tolias and Alexander Ecker
CAIMed Groups:
Biomedical Image Recognition
Deep neural networks trained to predict neural activity from visual input and behaviour have shown great potential to serve as digital twins of the visual cortex. Per-neuron embeddings derived from these models could potentially be used to map the functional landscape or identify cell types. However, state-of-the-art predictive models of mouse V1 do not generate functional embeddings that exhibit clear clustering patterns which would correspond to cell types. This raises the question whether the lack of clustered structure is due to limitations of current models or a true feature of the functional organization of mouse V1. In this work, we introduce DECEMber — Deep Embedding Clustering via Expectation Maximization-based refinement — an explicit inductive bias into predictive models that enhances clustering by adding an auxiliary-distribution-inspired loss function that enforces structured organization among per-neuron embeddings. We jointly optimize both neuronal feature embeddings and clustering parameters, updating cluster centers and scale matrices using the EM-algorithm. We demonstrate that these modifications improve cluster consistency while preserving high predictive performance and surpassing standard clustering methods in terms of stability. Moreover, DECEMber generalizes well across species (mice, primates) and visual areas (retina, V1, V4). The code is available at https://github.com/Nisone2000/DECEMber, https://github.com/ecker-lab/cnn-training.
The full text can be found here .
Funded by CAIMed
”A Circular Argument: Does RoPE need to be Equivariant for Vision?”
September 2025
NeurIPS 2025
Chase van de Geijn, Timo Lüddecke, Polina Turishcheva and Alexander S. Ecker
DOI:
https://openreview.net/forum?id=WCenI6RU9s
CAIMed Groups:
Biomedical Image Recognition
Digital Twins
Rotary Positional Encodings (RoPE) have emerged as a highly effective technique for one-dimensional sequences in Natural Language Processing spurring recent progress towards generalizing RoPE to higher-dimensional data such as images and videos. The success of RoPE has been thought to be due to its positional equivariance, i.e. its status as a textit{relative} positional encoding. In this paper, we mathematically show RoPE to be one of the most general solutions for equivariant positional embedding in one-dimensional data. Moreover, we show Mixed RoPE to be the analogously general solution for M-dimensional data, if we require commutative generators — a property necessary for RoPE’s equivariance. However, we question the necessity of equivariance. We propose Spherical RoPE, a method analogous to Mixed RoPE, but with the assumption of anti-commutative generators — relaxing the equivariant condition. Empirically, we find Spherical RoPE to have the equivalent learning behavior as its equivariant analogues. This strongly suggests that relative positional embeddings are not as important as is commonly believed. We expect this discovery to facilitate future work in positional encodings for vision that are faster and generalize better by removing the preconception that they must be relative.
The full text can be found here .
Funded by CAIMed
“VANILLA: Validated knowledge graph completion—A Normalization-based framework for Integrity, Link prediction, and Logical Accuracy”
September 2025
Knowledge-Based Systems, 325, Article 113939
CAIMed Groups:
AI & Active Agents
Semantic Models
Knowledge graphs (KGs) are expressive data structures for integrating and describing heterogeneous data by unifying factual information and domain knowledge. However, under the Open World Assumption (OWA), the absence of facts does not imply falsity—only incompleteness. Inductive learning methods, particularly numerical techniques such as Knowledge Graph Embeddings (KGEs) and Graph Neural Networks (GNNs), are widely used for link prediction and classification tasks in KGs. These models excel at capturing latent patterns and exploiting structural properties at scale. Nevertheless, their performance can be significantly degraded by anomalies in KG representations—semantic inconsistencies and modeling artifacts that arise from unconstrained data integration. Such anomalies obscure the intended meaning of relations, introduce noise, and mislead numerical learning models. To address this issue, we introduce a normalization theory for KGs that enforces semantic consistency through normal forms. These forms restructure KGs to eliminate representational anomalies, ensuring that the data adheres to well-defined semantic constraints. We present VANILLA, a neuro-symbolic framework that combines symbolic rule learning, numerical inductive models, and constraint-based validation. By aligning inductive predictions with normalized, ontology-aware KG structures, VANILLA enables accurate and semantically grounded KG completion. Experimental results show that our approach significantly improves predictive performance while maintaining semantic integrity, demonstrating the value of normalization in hybrid KG learning systems. VANILLA is publicly available on GitHub https://github.com/SDM-TIB/VANILLA
The full text can be found here .
Funded by CAIMed
“Probabilistic Domain Adaptation for Biomedical Image Segmentation”
August 2025
ICCVW 2025
Anwai Archit, Constantin Pape
Segmentation is a crucial analysis task in biomedical imaging. Given the diverse experimental settings in this field, the lack of generalization limits the use of deep learning in practice. Domain adaptation is a promising remedy: it involves training a model for a given task on a source dataset with labels and adapts it to a target dataset without additional labels. We introduce a probabilistic domain adaptation method, building on self-training approaches and the Probabilistic UNet. We use the latter to sample multiple segmentation hypotheses to implement better pseudo-label filtering. We further study joint and separate source-target training strategies and evaluate our method on three challenging domain adaptation tasks for biomedical segmentation.
The full text can be found here .
Funded by CAIMed
„Agent-based modeling for realistic reproduction of human mobility and contact behavior to evaluate test and isolation strategies in epidemic infectious disease spread^“
July 2025
Computers in Biology and Medicine, 193, 110269
David Kerkmann, Sascha Korf, Khoa Nguyen, Daniel Abele, Alain Schengen, Carlotta Gerstein, Jens Henrik Göbbert, Achim Basermann, Martin J. Kühn, Michael Meyer-Hermann
CAIMed Groups:
AI & Active Agents
Human Centered AI
Agent-based models have proven to be useful tools in supporting decision-making processes in different application domains. The advent of modern computers and supercomputers has enabled these bottom-up approaches to realistically model human mobility and contact behavior.
The COVID-19 pandemic showcased the urgent need for detailed and informative models that can answer research questions on transmission dynamics. We present a sophisticated agent-based model to simulate the spread of respiratory diseases. The model is highly modularized and can be used on various scales, from a small collection of buildings up to cities or countries. Although not being the focus of this paper, the model has undergone performance engineering on a single core and provides an efficient intra- and inter-simulation parallelization for time-critical decision-making processes.
In order to allow answering research questions on individual level resolution, nonpharmaceutical intervention strategies such as face masks or venue closures can be implemented for particular locations or agents. In particular, we allow for sophisticated testing and isolation strategies to study the effects of minimal-invasive infectious disease mitigation.
With realistic human mobility patterns for the region of Brunswick, Germany, we study the effects of different interventions between March 1st and May 30, 2021 in the SARS-CoV-2 pandemic. Our analyses suggest that symptom-independent testing has limited impact on the mitigation of disease dynamics if the dark figure in symptomatic cases is high. Furthermore, we found that quarantine length is more important than quarantine efficiency but that, with sufficient symptomatic control, also short quarantines can have a substantial effect.
The full text can be found here .
Funded by CAIMed
“A rule-based clinical decision support system for detection of acute kidney injury after pediatric cardiac surgery”
July 2025
Computers in Biology and Medicine, 193, 110382
Janice Wachenbrunner, Marcel Mast, Julia Böhnke, Nicole Rübsamen, Louisa Bode, André Karch, Henning Rathert, Alexander Horke, Philipp Beerbaum, Michael Marschollek, Thomas Jack, Martin Böhne
CAIMed Groups:
AI & Decisions
Clinical Decision Support
Acute kidney injury (AKI) is common in children with congenital heart disease following open-heart surgery with cardiopulmonary bypass (CPB). Early AKI detection in critically ill children requires clinician expertise to compile various data from different sources within a stressful and time-sensitive environment. However, as electronic health records provide data in a machine-readable format, this process could be supported by computerized systems. Therefore, we developed a time-aware, rule-based clinical decision support system (CDSS) to detect, stage, and track temporal AKI progression in children.
The full text can be found here .
Funded by CAIMed
“Foundation model of neural activity predicts response to new stimulus types”
April 2025
Nature
Eric Y. Wang, Paul G. Fahey, Zhuokun Ding, Stelios Papadopoulos, Kayla Ponder, Marissa A. Weis, Andersen Chang, Taliah Muhammad, Saumil Patel, Zhiwei Ding, Dat Tran, Jiakun Fu, Casey M. Schneider-Mizell, MICrONS Consortium, R. Clay Reid, Forrest Collman, Nuno Maçarico da Costa, Katrin Franke, Alexander S. Ecker, Jacob Reimer, Xaq Pitkow, Fabian H. Sinz & Andreas S. Tolias
CAIMed Groups:
Biomedical Image Recognition
Digital Twins
The complexity of neural circuits makes it challenging to decipher the brain’s algorithms of intelligence. Recent breakthroughs in deep learning have produced models that accurately simulate brain activity, enhancing our understanding of the brain’s computational objectives and neural coding. However, it is difficult for such models to generalize beyond their training distribution, limiting their utility. The emergence of foundation models1 trained on vast datasets has introduced a new artificial intelligence paradigm with remarkable generalization capabilities. Here we collected large amounts of neural activity from visual cortices of multiple mice and trained a foundation model to accurately predict neuronal responses to arbitrary natural videos. This model generalized to new mice with minimal training and successfully predicted responses across various new stimulus domains, such as coherent motion and noise patterns. Beyond neural response prediction, the model also accurately predicted anatomical cell types, dendritic features and neuronal connectivity within the MICrONS functional connectomics dataset2. Our work is a crucial step towards building foundation models of the brain. As neuroscience accumulates larger, multimodal datasets, foundation models will reveal statistical regularities, enable rapid adaptation to new tasks and accelerate research.
The full text can be found here .
Funded by CAIMed
“An unsupervised map of excitatory neuron dendritic morphology in the mouse visual cortex”
April 2025
Nature communnications
Marissa A. Weis, Stelios Papadopoulos, Laura Hansel, Timo Lüddecke, Brendan Celii, Paul G. Fahey, Eric Y. Wang, J. Alexander Bae, Agnes L. Bodor, Derrick Brittain, JoAnn Buchanan, Daniel J. Bumbarger, Manuel A. Castro, Forrest Collman, Nuno Maçarico da Costa, Sven Dorkenwald, Leila Elabbady, Akhilesh Halageri, Zhen Jia, Chris Jordan, Dan Kapner, Nico Kemnitz, Sam Kinn, Kisuk Lee, Kai Li, Ran Lu, Thomas Macrina, Gayathri Mahalingam, Eric Mitchell, Shanka Subhra Mondal, Shang Mu, Barak Nehoran, Sergiy Popovych, R. Clay Reid, Casey M. Schneider-Mizell, H. Sebastian Seung, William Silversmith, Marc Takeno, Russel Torres, Nicholas L. Turner, William Wong, Jingpeng Wu, Wenjing Yin, Szi-chieh Yu, Jacob Reimer, Philipp Berens, Andreas S. Tolias and Alexander S. Ecker
CAIMed Groups:
Biomedical Image Recognition
Neurons in the neocortex exhibit astonishing morphological diversity, which is critical for properly wiring neural circuits and giving neurons their functional properties. However, the organizational principles underlying this morphological diversity remain an open question. Here, we took a data-driven approach using graph-based machine learning methods to obtain a low-dimensional morphological “bar code” describing more than 30,000 excitatory neurons in mouse visual areas V1, AL, and RL that were reconstructed from the millimeter scale MICrONS serial-section electron microscopy volume. Contrary to previous classifications into discrete morphological types (m-types), our data-driven approach suggests that the morphological landscape of cortical excitatory neurons is better described as a continuum, with a few notable exceptions in layers 5 and 6. Dendritic morphologies in layers 2–3 exhibited a trend towards a decreasing width of the dendritic arbor and a smaller tuft with increasing cortical depth. Inter-area differences were most evident in layer 4, where V1 contained more atufted neurons than higher visual areas. Moreover, we discovered neurons in V1 on the border to layer 5, which avoided deeper layers with their dendrites. In summary, we suggest that excitatory neurons’ morphological diversity is better understood by considering axes of variation than using distinct m-types.
The full text can be found here .
Funded by CAIMed
„Segment Anything for Histopathology”
March 2025
MIDL 2025
Titus Griebel, Anwai Archit, Constantin Pape
Nucleus segmentation is an important analysis task in digital pathology. However, methods for automatic segmentation often struggle with new data from a different distribution, requiring users to manually annotate nuclei and retrain data-specific models. Vision foundation models (VFMs), such as the Segment Anything Model (SAM), offer a more robust alternative for automatic and interactive segmentation. Despite their success in natural images, a foundation model for nucleus segmentation in histopathology is still missing. Initial efforts to adapt SAM have shown some success, but did not yet introduce a comprehensive model for diverse segmentation tasks. To close this gap, we introduce PathoSAM, a VFM for nucleus segmentation, based on training SAM on a diverse dataset. Our extensive experiments show that it is the new state-of-the-art model for automatic and interactive nucleus instance segmentation in histopathology. We also demonstrate how it can be adapted for other segmentation tasks, including semantic nucleus segmentation. For this task, we show that it yields results better than popular methods, while not yet beating the state-of-the-art, CellViT. Our models are open-source and compatible with popular tools for data annotation. We also provide scripts for whole-slide image segmentation.
The full text can be found here .
Funded by CAIMed
„Parameter Efficient Fine-Tuning of Segment Anything Model for Biomedical Imaging”
March 2025
MIDL 2025
Carolin Teuber, Anwai Archit, Constantin Pape
CAIMed Groups:
Segmentation is an important analysis task for biomedical images, enabling the study of individual organelles, cells or organs. Deep learning has massively improved segmentation methods, but challenges remain in generalization to new conditions, requiring costly data annotation. Vision foundation models, such as Segment Anything Model (SAM), address this issue through improved generalization. However, these models still require finetuning on annotated data, although with less annotations, to achieve optimal results for new conditions. As a downside, they require more computational resources. This makes parameter-efficient finetuning (PEFT) relevant. We contribute the first comprehensive study of PEFT for SAM applied to biomedical images. We find that the placement of PEFT layers is more important for efficiency than the type of layer for vision transformers and we provide a recipe for resource-efficient finetuning.
The full text can be found here .
Funded by CAIMed
“Using Photon-Counting CT Images for Lung Nodule Classification”
Leonie Thieme, Zahra Ahmadi, Steffen Oeltze-Jafra, Eike Petersen, Hoen-oh Shin, Andrea Schenk
CAIMed Groups:
AI & Decisions
Human-Centered AI
An automatic classification of the malignancy of lung nodules in computed tomography (CT) scans can support early detection of lung cancer, which is crucial for the treatment success. The novel photon-counting CT (PCCT) technology enables high image quality with a low radiation dose and provides additional spectral information. This research focuses on whether PCCT scans offer a benefit in the automatic classification of lung nodules. Establishing a dataset of PCCT images poses several challenges, such as the extraction of annotations or the data imbalance.
The full text can be found here .
Funded by CAIMed
“Transient silencing of hypermutation preserves B cell affinity during clonal bursting”
March 2025
Nature 641, 486–494
Juhee Pae, Niklas Schwan, Bertrand Ottino-Loffler, William S. DeWitt, Amar Garg, Juliana Bortolatto, Ashni A. Vora, Jin-Jie Shen, Alvaro Hobbs, Tiago B. R. Castro, Luka Mesin, Frederick A. Matsen IV, Michael Meyer-Hermann & Gabriel D. Victora
CAIMed Groups:
AI & Active Agents
Mathematical Models
In the course of antibody affinity maturation, germinal centre (GC) B cells mutate their immunoglobulin heavy- and light-chain genes in a process known as somatic hypermutation (SHM). Panels of mutant B cells with different binding affinities for antigens are then selected in a Darwinian manner, which leads to a progressive increase in affinity among the population. As with any Darwinian process, rare gain-of-fitness mutations must be identified and common loss-of-fitness mutations avoided. Progressive acquisition of mutations therefore poses a risk during large proliferative bursts, when GC B cells undergo several cell cycles in the absence of affinity-based selection. Using a combination of in vivo mouse experiments and mathematical modelling, here we show that GCs achieve this balance by strongly suppressing SHM during clonal-burst-type expansion, so that a large fraction of the progeny generated by these bursts does not deviate from their ancestral genotype. Intravital imaging and image-based cell sorting of a mouse strain carrying a reporter of cyclin-dependent kinase 2 (CDK2) activity showed that B cells that are actively undergoing proliferative bursts lack the transient CDK2low ‘G0-like’ phase of the cell cycle in which SHM takes place. We propose a model in which inertially cycling B cells mostly delay SHM until the G0-like phase that follows their final round of division in the GC dark zone, thus maintaining affinity as they clonally expand in the absence of selection.
The full text can be found here .
Funded by CAIMed
”What should a neuron aim for? Designing local objective functions based on information theory”
Januar 2025
ICLR 2025
Andreas Christian Schneider, Valentin Neuhaus, David Alexander Ehrlich, Abdullah Makkeh, Alexander S Ecker, Viola Priesemann, Michael Wibral
CAIMed Groups:
Digital Twins
Biomedical Image Recognition
In modern deep neural networks, the learning dynamics of individual neurons are often obscure, as the networks are trained via global optimization. Conversely, biological systems build on self-organized, local learning, achieving robustness and efficiency with limited global information. Here, we show how self-organization between individual artificial neurons can be achieved by designing abstract bio-inspired local learning goals. These goals are parameterized using a recent extension of information theory, Partial Information Decomposition (PID), which decomposes the information that a set of information sources holds about an outcome into unique, redundant and synergistic contributions. Our framework enables neurons to locally shape the integration of information from various input classes, i.e., feedforward, feedback, and lateral, by selecting which of the three inputs should contribute uniquely, redundantly or synergistically to the output. This selection is expressed as a weighted sum of PID terms, which, for a given problem, can be directly derived from intuitive reasoning or via numerical optimization, offering a window into understanding task-relevant local information processing. Achieving neuron-level interpretability while enabling strong performance using local learning, our work advances a principled information-theoretic foundation for local learning strategies.
The full text can be found here .
Funded by CAIMed
”Retrospective for the Dynamic Sensorium Competition for predicting large-scale mouse primary visual cortex activity from videos”
Dezember 2024
Advances in Neural Information Processing Systems (NeurIPS 2024)
Polina Turishcheva, Paul G. Fahey, Michaela Vystrčilová, Laura Hansel, Rachel Froebe, Kayla Ponder, Yongrong Qiu, Konstantin F. Willeke, Mohammad Bashiri, Ruslan Baikulov, Yu Zhu, Lei Ma, Shan Yu, Tiejun Huang, Bryan M. Li, Wolf De Wulf, Nina Kudryashova, Matthias H. Hennig, Nathalie L. Rochefort, Arno Onken, Eric Wang, Zhiwei Ding, Andreas S. Tolias, Fabian H. Sinz, Alexander S. Ecker
CAIMed Groups:
Digital Twins
Biomedical Image Recognition
Understanding how biological visual systems process information is challenging because of the nonlinear relationship between visual input and neuronal responses. Artificial neural networks allow computational neuroscientists to create predictive models that connect biological and machine vision.Machine learning has benefited tremendously from benchmarks that compare different models on the same task under standardized conditions. However, there was no standardized benchmark to identify state-of-the-art dynamic models of the mouse visual system.To address this gap, we established the SENSORIUM 2023 Benchmark Competition with dynamic input, featuring a new large-scale dataset from the primary visual cortex of ten mice. This dataset includes responses from 78,853 neurons to 2 hours of dynamic stimuli per neuron, together with behavioral measurements such as running speed, pupil dilation, and eye movements.The competition ranked models in two tracks based on predictive performance for neuronal responses on a held-out test set: one focusing on predicting in-domain natural stimuli and another on out-of-distribution (OOD) stimuli to assess model generalization.As part of the NeurIPS 2023 Competition Track, we received more than 160 model submissions from 22 teams. Several new architectures for predictive models were proposed, and the winning teams improved the previous state-of-the-art model by 50%. Access to the dataset as well as the benchmarking infrastructure will remain online at www.sensorium-competition.net.
The full text can be found here .
Funded by CAIMed
”Reproducibility of predictive networks for mouse visual cortex”
Dezember 2024
Advances in Neural Information Processing Systems (NeurIPS 2024)
Polina Turishcheva, Max F. Burg, Fabian H. Sinz, Alexander S. Ecker
CAIMed Groups:
Digital Twins
Biomedical Image Recognition
Deep predictive models of neuronal activity have recently enabled several new discoveries about the selectivity and invariance of neurons in the visual cortex.These models learn a shared set of nonlinear basis functions, which are linearly combined via a learned weight vector to represent a neuron’s function.Such weight vectors, which can be thought as embeddings of neuronal function, have been proposed to define functional cell types via unsupervised clustering.However, as deep models are usually highly overparameterized, the learning problem is unlikely to have a unique solution, which raises the question if such embeddings can be used in a meaningful way for downstream analysis.In this paper, we investigate how stable neuronal embeddings are with respect to changes in model architecture and initialization. We find that L1 regularization to be an important ingredient for structured embeddings and develop an adaptive regularization that adjusts the strength of regularization per neuron. This regularization improves both predictive performance and how consistently neuronal embeddings cluster across model fits compared to uniform regularization.To overcome overparametrization, we propose an iterative feature pruning strategy which reduces the dimensionality of performance-optimized models by half without loss of performance and improves the consistency of neuronal embeddings with respect to clustering neurons.Our results suggest that to achieve an objective taxonomy of cell types or a compact representation of the functional landscape, we need novel architectures or learning techniques that improve identifiability. The code is available https://github.com/pollytur/readout_reproducibility.
The full text can be found here .
Funded by CAIMed
”A chromatic feature detector in the retina signals visual context changes”
Oktober 2024
eLife
CAIMed Groups:
Biomedical Image Recognition
The retina transforms patterns of light into visual feature representations supporting behaviour. These representations are distributed across various types of retinal ganglion cells (RGCs), whose spatial and temporal tuning properties have been studied extensively in many model organisms, including the mouse. However, it has been difficult to link the potentially nonlinear retinal transformations of natural visual inputs to specific ethological purposes. Here, we discover a nonlinear selectivity to chromatic contrast in an RGC type that allows the detection of changes in visual context. We trained a convolutional neural network (CNN) model on large-scale functional recordings of RGC responses to natural mouse movies, and then used this model to search in silico for stimuli that maximally excite distinct types of RGCs. This procedure predicted centre colour opponency in transient suppressed-by-contrast (tSbC) RGCs, a cell type whose function is being debated. We confirmed experimentally that these cells indeed responded very selectively to Green-OFF, UV-ON contrasts. This type of chromatic contrast was characteristic of transitions from ground to sky in the visual scene, as might be elicited by head or eye movements across the horizon. Because tSbC cells performed best among all RGC types at reliably detecting these transitions, we suggest a role for this RGC type in providing contextual information (i.e. sky or ground) necessary for the selection of appropriate behavioural responses to other stimuli, such as looming objects. Our work showcases how a combination of experiments with natural stimuli and computational modelling allows discovering novel types of stimulus selectivity and identifying their potential ethological relevance.
The full text can be found here .
Funded by CAIMed
”Diverse task-driven modeling of macaque V4 reveals functional specialization towards semantic tasks”
Mai 2024
PLOS Computational Biology
Santiago A. Cadena, Konstantin F. Willeke, Kelli Restivo, George Denfield, Fabian H. Sinz, Matthias Bethge, Andreas S. Tolias, Alexander S. Ecker
CAIMed Groups:
Biomedical Image Recognition
The retina transforms patterns of light into visual feature representations supporting behaviour. These representations are distributed across various types of retinal ganglion cells (RGCs), whose spatial and temporal tuning properties have been studied extensively in many model organisms, including the mouse. However, it has been difficult to link the potentially nonlinear retinal transformations of natural visual inputs to specific ethological purposes. Here, we discover a nonlinear selectivity to chromatic contrast in an RGC type that allows the detection of changes in visual context. We trained a convolutional neural network (CNN) model on large-scale functional recordings of RGC responses to natural mouse movies, and then used this model to search in silico for stimuli that maximally excite distinct types of RGCs. This procedure predicted centre colour opponency in transient suppressed-by-contrast (tSbC) RGCs, a cell type whose function is being debated. We confirmed experimentally that these cells indeed responded very selectively to Green-OFF, UV-ON contrasts. This type of chromatic contrast was characteristic of transitions from ground to sky in the visual scene, as might be elicited by head or eye movements across the horizon. Because tSbC cells performed best among all RGC types at reliably detecting these transitions, we suggest a role for this RGC type in providing contextual information (i.e. sky or ground) necessary for the selection of appropriate behavioural responses to other stimuli, such as looming objects. Our work showcases how a combination of experiments with natural stimuli and computational modelling allows discovering novel types of stimulus selectivity and identifying their potential ethological relevance.
The full text can be found here .
Funded by CAIMed
”CXCR6 + CD69 + CD8 + T cells in ascites are associated with disease severity in patients with cirrhosis”
März 2024
JHEP reports : innovation in hepatology
Christian Niehaus, Sebastian Klein, Benedikt Strunz, Erich Freyer, Benjamin Maasoumy, Heiner Wedemeyer, Niklas K. Björkström, Anke R.M. Kraft, Markus Cornberg
CAIMed Groups:
Biomedical Image Recognition
The retina transforms patterns of light into visual feature representations supporting behaviour. These representations are distributed across various types of retinal ganglion cells (RGCs), whose spatial and temporal tuning properties have been studied extensively in many model organisms, including the mouse. However, it has been difficult to link the potentially nonlinear retinal transformations of natural visual inputs to specific ethological purposes. Here, we discover a nonlinear selectivity to chromatic contrast in an RGC type that allows the detection of changes in visual context. We trained a convolutional neural network (CNN) model on large-scale functional recordings of RGC responses to natural mouse movies, and then used this model to search in silico for stimuli that maximally excite distinct types of RGCs. This procedure predicted centre colour opponency in transient suppressed-by-contrast (tSbC) RGCs, a cell type whose function is being debated. We confirmed experimentally that these cells indeed responded very selectively to Green-OFF, UV-ON contrasts. This type of chromatic contrast was characteristic of transitions from ground to sky in the visual scene, as might be elicited by head or eye movements across the horizon. Because tSbC cells performed best among all RGC types at reliably detecting these transitions, we suggest a role for this RGC type in providing contextual information (i.e. sky or ground) necessary for the selection of appropriate behavioural responses to other stimuli, such as looming objects. Our work showcases how a combination of experiments with natural stimuli and computational modelling allows discovering novel types of stimulus selectivity and identifying their potential ethological relevance.
The full text can be found here .
Funded by CAIMed
“Continuous monitoring of physiological data using the patient vital status fusion score in septic critical care patients”
March 2024
Sci Rep 14, 7198
Philipp L. S. Ohland, Thomas Jack, Marcel Mast, Anette Melk, André Bleich & Steven R. Talbot
CAIMed Groups:
Clinical Decision Support
Accurate and standardized methods for assessing the vital status of patients are crucial for patient care and scientific research. This study introduces the Patient Vital Status (PVS), which quantifies and contextualizes a patient’s physical status based on continuous variables such as vital signs and deviations from age-dependent normative values. The vital signs, heart rate, oxygen saturation, respiratory rate, mean arterial blood pressure, and temperature were selected as input to the PVS pipeline. The method was applied to 70 pediatric patients in the intensive care unit (ICU), and its efficacy was evaluated by matching high values with septic events at different time points in patient care. Septic events included systemic inflammatory response syndrome (SIRS) and suspected or proven sepsis. The comparison of maximum PVS values between the presence and absence of a septic event showed significant differences (SIRS/No SIRS: p < 0.0001, η2 = 0.54; Suspected Sepsis/No Suspected Sepsis: p = 0.00047, η2 = 0.43; Proven Sepsis/No Proven Sepsis: p = 0.0055, η2 = 0.34). A further comparison between the most severe PVS in septic patients with the PVS at ICU discharge showed even higher effect sizes (SIRS: p < 0.0001, η2 = 0.8; Suspected Sepsis: p < 0.0001, η2 = 0.8; Proven Sepsis: p = 0.002, η2 = 0.84). The PVS is emerging as a data-driven tool with the potential to assess a patient’s vital status in the ICU objectively. Despite real-world data challenges and potential annotation biases, it shows promise for monitoring disease progression and treatment responses. Its adaptability to different disease markers and reliance on age-dependent reference values further broaden its application possibilities. Real-time implementation of PVS in personalized patient monitoring may be a promising way to improve critical care. However, PVS requires further research and external validation to realize its true potential.
The full text can be found here .
Funded by CAIMed
“Most discriminative stimuli for functional cell type clustering”
January 2024
ICLR 2024
Max F Burg, Thomas Zenkel, Michaela Vystrčilová, Jonathan Oesterle, Larissa Höfling, Konstantin Friedrich Willeke, Jan Lause, Sarah Müller, Paul G. Fahey, Zhiwei Ding, Kelli Restivo, Shashwat Sridhar, Tim Gollisch, Philipp Berens, Andreas S. Tolias, Thomas Euler, Matthias Bethge, Alexander S Ecker
CAIMed Groups:
Biomedical Image Recognition
Identifying cell types and understanding their functional properties is crucial for unraveling the mechanisms underlying perception and cognition. In the retina, functional types can be identified by carefully selected stimuli, but this requires expert domain knowledge and biases the procedure towards previously known cell types. In the visual cortex, it is still unknown what functional types exist and how to identify them. Thus, for unbiased identification of the functional cell types in retina and visual cortex, new approaches are needed. Here we propose an optimization-based clustering approach using deep predictive models to obtain functional clusters of neurons using Most Discriminative Stimuli (MDS). Our approach alternates between stimulus optimization with cluster reassignment akin to an expectation-maximization algorithm. The algorithm recovers functional clusters in mouse retina, marmoset retina and macaque visual area V4. This demonstrates that our approach can successfully find discriminative stimuli across species, stages of the visual system and recording techniques. The resulting most discriminative stimuli can be used to assign functional cell types fast and on the fly, without the need to train complex predictive models or show a large natural scene dataset, paving the way for experiments that were previously limited by experimental time. Crucially, MDS are interpretable: they visualize the distinctive stimulus patterns that most unambiguously identify a specific type of neuron.
The full text can be found here .
Funded by CAIMed
Preprints
“Auto-nnU-Net: Towards Automated Medical Image Segmentation”
22 May 2025 (submission date)
arxiv.org
CAIMed Groups:
AI & Decision
Human-Centered AI
Medical Image Segmentation (MIS) includes diverse tasks, from bone to organ segmentation, each with its own challenges in finding the best segmentation model. The state-of-the-art AutoML-related MIS-framework nnU-Net automates many aspects of model configuration but remains constrained by fixed hyperparameters and heuristic design choices. As a full-AutoML framework for MIS, we propose Auto-nnU-Net, a novel nnU-Net variant enabling hyperparameter optimization (HPO), neural architecture search (NAS), and hierarchical NAS (HNAS). Additionally, we propose Regularized PriorBand to balance model accuracy with the computational resources required for training, addressing the resource constraints often faced in real-world medical settings that limit the feasibility of extensive training procedures. We evaluate our approach across diverse MIS datasets from the well-established Medical Segmentation Decathlon, analyzing the impact of AutoML techniques on segmentation performance, computational efficiency, and model design choices. The results demonstrate that our AutoML approach substantially improves the segmentation performance of nnU-Net on 6 out of 10 datasets and is on par on the other datasets while maintaining practical resource requirements. Our code is available at this URL.
The full text can be found here .
Funded by CAIMed
„Artificial Intelligence in Pediatric Echocardiography: ExploringChallenges, Opportunities, and Clinical Applications withExplainable AI and Federated Learning“
27 March 2025 (submission date)
arxiv.org
CAIMed Groups:
Clinical Decision Support
Human-Centered AI
Pediatric heart diseases present a broad spectrum of congenital and acquired diseases. More complex congenital malformations require a differentiated and multimodal decision-making process, usually including echocardiography as a central imaging method. Artificial intelligence (AI) offers considerable promise for clinicians by facilitating automated interpretation of pediatric echocardiography data. However, adapting AI technologies for pediatric echocardiography analysis has challenges such as limited public data availability, data privacy, and AI model transparency. Recently, researchers have focused on disruptive technologies, such as federated learning (FL) and explainable AI (XAI), to improve automatic diagnostic and decision support workflows. This study offers a comprehensive overview of the limitations and opportunities of AI in pediatric echocardiography, emphasizing the synergistic workflow and role of XAI and FL, identifying research gaps, and exploring potential future developments. Additionally, three relevant clinical use cases demonstrate the functionality of XAI and FL with a focus on (i) view recognition, (ii) disease classification, (iii) segmentation of cardiac structures, and (iv) quantitative assessment of cardiac function.
The full text can be found here .
Funded by CAIMed