Index
Development of a Local LLM Agent System for Clinical Expert Support and Automation in MRI Planning for Radiation Therapy
RAG-Enhanced Low-Cost Vision-Language Models for Diabetic Retinopathy Classification and Automated Reporting
Diabetic Retinopathy (DR) affects over 160 million people globally, projected to reach 180 million by 2030, despite 90% of related blindness being preventable through early detection [1]. Current AI models achieve strong classification performance but lack interpretable clinical reports, limiting their adoption in low-resource settings. Although Vision-Language Models (VLMs) offer unified diagnosis and report generation, fundus captioning significantly underperforms compared with other imaging modalities [2,3], and state-of-the-art VLMs remain computationally expensive. Although Retrieval-Augmented Generation (RAG) has improved medical imaging accuracy [4], no prior study has integrated DR severity grading, lesion-aware reporting, and evidence retrieval within a low-cost, clinically deployable VLM.
Generalization of Self-Supervised Vision Models in Image Retrieval
This project investigates the generalization capabilities of prominent self-supervised vision models, such as DINOv2, CLIP, and MoCo, when applied to image retrieval tasks across diverse visual domains.
Overview
Self-supervised learning (SSL) models are increasingly important for creating robust visual features. However, their performance often degrades when transferring from standard natural image datasets (like ImageNet) to more specialized domains.
Our study systematically benchmarks these models on three distinct domains:
1. Natural Images (Baseline)
2. Scene-Centric Images (e.g., Places365)
3. Artistic Images (e.g., ArtPlaces)
We compare the models’ out-of-the-box generalization and analyze the impact of finetuning on their domain adaptation, providing crucial insights into the stability and robustness of the learned representations for practical retrieval applications.
Investigating the Influence of Different Motion Sensors for Detecting Parkinson’s Disease
Integrating Transformer Networks with Multi-Modal Learning for Document Layout Analysis
A Cascaded Encoder–Decoder Network for CT Image Restoration
Evaluate LORA finetuning for detection guided segmentation in CT images
Evaluation of LoRa tuning of grounded segmentation using MedSam. We will investigate if we can reduce the training parameters for optimal detection and segmentation performnce using SOTA methods and training paradigms on an Abdominal RSNA dataset.
Improving localization of VL models in CXRs
Patient-Specific Quality Assurance of Synthetic CT for MR-Only Radiotherapy
If you are interested in the thesis project, please send your application (CV, letter of motivation, current transcript
of records) with subject ’sCT PSQA – Thesis’ to: bernd-niklas.axer@extern.uk-erlangen.de
MSP_Thesis_Proposal_sCT_PSQA
Data-Driven Characterization and Modeling of the Radiotherapy Workflow
If you are interested in the project, please send your curriculum vitæ to: rafael.lobao@uk-erlangen.de
Having prior experience with Python, SQL, and statistics is advantageous.
Field of Study: Medical Engineering / Data Science
Thesis_Proposal__Workflow_2