Title:
Leveraging A Large Language Model for Movement Intention Recognition
Poster
Preview Converted Images may contain errors
Abstract
Current movement intention recognition systems rely on sensors such as electromyography sensors (EMG) and inertial measurement units (IMU) which rely on a change in movement to predict an intended movement. With recent advancements in large language models (LLMs) capable of analyzing visual data, there is potential to explore alternative approaches. This study investigates the feasibility of using LLMs for movement intention recognition by extracting frames from point-of-view videos of various movements, stitching these frames into a single image, and inputting the image into an LLM for interpretation. Experimental results show that the LLM achieved recognition accuracy between 70% and 85%, with average response times ranging from 5.5 to 7 seconds. Performance was notably influenced by the number of frames selected and the duration of the source video. While the approach demonstrates promise, the current speed and accuracy suggest that LLM-based video input systems may not yet be practical.
Authors
First Name |
Last Name |
Jared
|
Comeau
|
Leave a comment
Submission Details
Conference URC
Event Interdisciplinary Science and Engineering (ISE)
Department Electrical and Computer Engineering (ISE)
Group Electrical and Computer Engineering
Added April 20, 2025, 1:39 p.m.
Updated April 20, 2025, 1:39 p.m.
See More Department Presentations Here