Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Deployment-Oriented Benchmarking of Open-Source Large Language Models for Social Robots
0
Zitationen
2
Autoren
2026
Jahr
Abstract
Large language models (LLMs) offer new opportunities to enhance human–robot interaction by enabling humanoid robots to engage in natural, context-aware dialogue. However, deploying LLMs on social robots operating in real-time environments remains challenging due to latency constraints, limited onboard hardware, and privacy considerations. This paper introduces a deployment-oriented benchmarking framework for evaluating open-source LLMs that are feasible for on-device execution on humanoid robots. We implement and analyze ten lightweight LLMs (≤2 billion parameters), using the Pepper robot as a representative use case in CS1/CS2 laboratory courses where the robot functions as a teaching assistant. The models were evaluated using four normalized metrics: instruction-following accuracy, conversational clarity, response latency, and on-device feasibility. Results identify clear trade-offs within the lightweight tier, emphasizing models that best balance responsiveness with instructional quality. This work provides a reproducible methodology and practical deployment guidelines for integrating LLM-driven instructional capabilities into humanoid robots to support more autonomous, student-centered learning in introductory computer science education.
Ähnliche Arbeiten
ROS: an open-source Robot Operating System
2009 · 7.171 Zit.
An Experiment in Linguistic Synthesis with a Fuzzy Logic Controller
1999 · 5.632 Zit.
An experiment in linguistic synthesis with a fuzzy logic controller
1975 · 5.564 Zit.
IEMOCAP: interactive emotional dyadic motion capture database
2008 · 3.440 Zit.
A survey of socially interactive robots
2003 · 3.077 Zit.