Kinematic primitives in action similarity judgments: A human-centered computational modelShow others and affiliations
2023 (English)In: IEEE Transactions on Cognitive and Developmental Systems, ISSN 2379-8920, E-ISSN 2379-8939, Vol. 15, no 4, p. 1981-1992Article in journal (Refereed) Published
Abstract [en]
This article investigates the role that kinematic features play in human action similarity judgments. The results of three experiments with human participants are compared with the computational model that solves the same task. The chosen model has its roots in developmental robotics and performs action classification based on learned kinematic primitives. The comparative experimental results show that both model and human participants can reliably identify whether two actions are the same or not. Specifically, most of the given actions could be similarity judged based on very limited information from a single feature domain (velocity or spatial). Both velocity and spatial features were however necessary to reach a level of human performance on evaluated actions. The experimental results also show that human performance on an action identification task indicated that they clearly relied on kinematic information rather than on action semantics. The results show that both the model and human performance are highly accurate in an action similarity task based on kinematic-level features, which can provide an essential basis for classifying human actions.
Place, publisher, year, edition, pages
IEEE, 2023. Vol. 15, no 4, p. 1981-1992
Keywords [en]
Action similarity, action matching, biological motion, optical flow, point light display, kinematic primitives, computational model, comparative study
National Category
Psychology Computer Sciences Human Computer Interaction Computer Vision and Robotics (Autonomous Systems)
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:oru:diva-103866DOI: 10.1109/TCDS.2023.3240302ISI: 001126639000035Scopus ID: 2-s2.0-85148457281OAI: oai:DiVA.org:oru-103866DiVA, id: diva2:1732455
Funder
Knowledge Foundation, 2014022EU, Horizon 2020, 804388
Note
Funding agency:
AFOSR FA8655-20-1-7035
2023-01-312023-01-312024-06-24Bibliographically approved