Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Uncover This Tech Term: Large Vision-Language Models in Radiology
1
Zitationen
3
Autoren
2026
Jahr
Abstract
WHAT ARE LVLMs?Large multimodal models are typically transformer-based foundational models that can process and generate multiple types of data (modalities), including text, images, audio, and video [1,2].Large vision-language models (LVLMs) are a subset of large multimodal models that specifically focus on aligning and integrating visual and linguistic representations.Traditional artificial intelligence (AI) systems are trained to perform well-defined narrow tasks and have limited adaptability.By contrast, LVLMs generalize across diverse tasks and support flexible downstream applications without requiring task-specific retraining.
Ähnliche Arbeiten
Refinement and reassessment of the SERVQUAL scale.
1991 · 3.967 Zit.
Radiobiology for the Radiologist.
1974 · 3.502 Zit.
ACR Thyroid Imaging, Reporting and Data System (TI-RADS): White Paper of the ACR TI-RADS Committee
2017 · 2.431 Zit.
Accuracy of Physician Self-assessment Compared With Observed Measures of Competence
2006 · 2.325 Zit.
Technology as an Occasion for Structuring: Evidence from Observations of CT Scanners and the Social Order of Radiology Departments
1986 · 2.249 Zit.