Accurate and comprehensive interpretation of anterior segment diseases from slit-lamp and smartphone photographs remains a clinical challenge due to the limited specificity and structure of existing Artificial Intelligence tools. The purpose of this international, multicenter clinical trial is to developed and validated an agent-based framework that integrates vision-language models and large language models to enhance the diagnostic workflow of anterior segment diseases.
Study Type
OBSERVATIONAL
Enrollment
2,000
Multimodal Vision-language Model for Multi-task Diagnosis and Triage Suggestions of Ophthalmic Diseases Patients presenting with complaints of anterior segment diseases first complete a slit-lamp examination or take a mobile phone eye photograph. A multimodal vision-language model uses patient-related images (such as selfies and eye exam photos) to make an intelligent diagnosis. The diagnosis is kept private. The patient then seeks medical attention and undergoes a clinical examination by an experienced clinician. A second experienced clinician then reviews the clinical diagnosis. If the diagnosis agrees, it is considered the gold standard. If there is a discrepancy in the diagnosis, the consensus between the two clinicians is used as the gold standard.
Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University
Guangzhou, Guangdong, China
RECRUITINGDiagnostic accuracy of multimodal vision-language model.
For each patient, the diagnoses generated by the multimodal vision-language model and the clinical diagnosis provided by skilled clinicians were documented and compared. Consistency between the two diagnoses indicates the program's precision in clinical practice.
Time frame: from July 2025 to September 2025
This platform is for informational purposes only and does not constitute medical advice. Always consult a qualified healthcare professional.