SwePub
Sök i LIBRIS databas

  Extended search

WFRF:(Gutierrez Farewik Elena 1973 )
 

Search: WFRF:(Gutierrez Farewik Elena 1973 ) > A method of detecti...

A method of detecting human movement intentions in real environments

Liu, Yixing (author)
KTH,Farkostteknik och Solidmekanik
Wan, Zhao-Yuan (author)
KTH,Farkostteknik och Solidmekanik
Wang, Ruoli (author)
KTH,Farkostteknik och Solidmekanik
show more...
Gutierrez-Farewik, Elena, 1973- (author)
KTH,Farkostteknik och Solidmekanik
show less...
 (creator_code:org_t)
Institute of Electrical and Electronics Engineers (IEEE), 2023
2023
English.
In: 2023 international conference on rehabilitation robotics, ICORR. - : Institute of Electrical and Electronics Engineers (IEEE).
  • Conference paper (peer-reviewed)
Abstract Subject headings
Close  
  • Accurate and timely movement intention detection can facilitate exoskeleton control during transitions between different locomotion modes. Detecting movement intentions in real environments remains a challenge due to unavoidable environmental uncertainties. False movement intention detection may also induce risks of falling and general danger for exoskeleton users. To this end, in this study, we developed a method for detecting human movement intentions in real environments. The proposed method is capable of online self-correcting by implementing a decision fusion layer. Gaze data from an eye tracker and inertial measurement unit (IMU) signals were fused at the feature extraction level and used to predict movement intentions using 2 different methods. Images from the scene camera embedded on the eye tracker were used to identify terrains using a convolutional neural network. The decision fusion was made based on the predicted movement intentions and identified terrains. Four able-bodied participants wearing the eye tracker and 7 IMU sensors took part in the experiments to complete the tasks of level ground walking, ramp ascending, ramp descending, stairs ascending, and stair descending. The recorded experimental data were used to test the feasibility of the proposed method. An overall accuracy of 93.4% was achieved when both feature fusion and decision fusion were used. Fusing gaze data with IMU signals improved the prediction accuracy.

Subject headings

NATURVETENSKAP  -- Data- och informationsvetenskap -- Datorseende och robotik (hsv//swe)
NATURAL SCIENCES  -- Computer and Information Sciences -- Computer Vision and Robotics (hsv//eng)

Keyword

Robotic exoskeletons
movement intention prediction
eye tracker
wearable sensor

Publication and Content Type

ref (subject category)
kon (subject category)

To the university's database

Find more in SwePub

By the author/editor
Liu, Yixing
Wan, Zhao-Yuan
Wang, Ruoli
Gutierrez-Farewi ...
About the subject
NATURAL SCIENCES
NATURAL SCIENCES
and Computer and Inf ...
and Computer Vision ...
Articles in the publication
By the university
Royal Institute of Technology

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view