Human Motion Capture and Identification for Assistive Systems Design in Rehabilitation. Pubudu N. Pathirana
Чтение книги онлайн.

Читать онлайн книгу Human Motion Capture and Identification for Assistive Systems Design in Rehabilitation - Pubudu N. Pathirana страница 13

СКАЧАТЬ limbs, such as the angle of the elbow and knee. The second type was the angle between the orientation of a sensor and gravity. Moreover, in the telerehabilitation system developed by Luo et al. [218], angles of joints were utilised to encode the movements of the upper extremity. In this system, the angles of the shoulder and wrist were measured by two IMUs, as these two joints were modelled with three degrees of freedom on each joint while those of the elbow and fingers were measured by an optical linear encoder (OLE) and a glove made by multiple OLEs because these joints could be modelled with one degree of freedom. Additionally, Durfee et al. [102] introduced two bilateral electrogoniometers in a home telerehabilitation system for post-stroke patients. These bilateral electrogoniometers were attached to the wrist and hand of a subject, respectively. The angles of flexion and extension movements in the wrist and the first MCP joints were measured to represent the movement of the wrist and hand. Two potentiometers (refer to Figure 1.10a) were utilised to calculate the angles of joints (θ) as

      (1.4)theta equals 180 minus cosine Superscript negative 1 Baseline left-parenthesis StartFraction c squared plus upper L squared minus a squared Over 2 c upper L EndFraction right-parenthesis minus cosine Superscript negative 1 Baseline left-parenthesis StartFraction d squared plus upper L squared minus b squared Over 2 d upper L EndFraction right-parenthesis comma

      (1.5)upper L squared equals a squared plus c squared minus 2 a c cosine alpha period

Schematic illustration of the locations of five sensors worn by a subject.

      Apart from the above three examples, in some studies where motion trajectories of joints are captured, angle information is still derived for encoding human movements. For example, Adams et al. [15] developed a virtual reality system to assess the motor function of upper extremities in daily living. To encode the movement, they used the swing angle of the shoulder joint along the Y and Z axes, the twist angle of the shoulder, the angle of the elbow, their first and second derivatives, the bone length of the collarbone, upper arm and forearm, as well as the pose (position, yaw and pitch) of the vector along the collarbone to describe the movement of the upper body. Here the collarbone is a virtual bone connecting two shoulders. These parameters were utilised in an unscented Kalman filter as state, while the positions of the shoulders, elbows and wrists reading from a Kinect formed the observation. Another example is that of Wenbing et al. [378], who evaluated the feasibility of using a single Kinect with a series of rules to assess the quality of movements in rehabilitation. Five movements, including hip abduction, bowling, sit to stand, can turn and toe touch, were studied in this paper. For the first four movements, angles were used as encoders. For instance, the change of angle between left and right thighs (the vector from the hip centre to the left and right knee) was used to represent the angle of hip abduction, while the dot product of two vectors (from the hip centre to the left and right shoulders) was utilised to compute the angle encoding the movement of bowling. Additionally, Olesh et al. [267] proposed an automated approach to assess the impairment of upper limb movements caused by stroke. To encode the movement of the upper extremities, the angle of four joints, including shoulder flexion‐extension, shoulder abduction‐adduction, elbow flexion‐extension and wrist flexion‐extension, were calculated with the 3D positions of joints measured with Kinect.

      Though angles of joints, as well as their derivatives, are utilised widely in encoding human motions, trajectories of joints and their derivatives can also be observed in some rehabilitation and telerehabilitation applications.

      1.5.3 Summary and challenge

Photo depicts the pictures of animals.

      As a result, there remain challenges in developing formal descriptions and robust computational procedures for the automatic interpretation and representation of motions of patients. The majority of studies [92, 158] employed a variety of human motion encoders to recognise or decompose general movement, such as reaching, waving hands, jumping, walking and so on. Few of them investigated details in each general movement, for example, the even smaller atomic components included in these general movements that are of importance for syntactic and structural descriptions of human movements in detail, especially in a clinic and rehabilitation environment, where the details of movements of body parts require a form of motion language or, at least, syntax. A novel approach to encode human motion trajectories will be discussed in Chapter 3.

      In recent decades, with the advancements in telerehabilitation and associated motion capture technologies, an increasing number of research and development activities are focusing on the development of automated quantitative measures of patient performance in ADLs [136, 262]. Due to the important role played by the upper extremity in ADLs [99], an automated approach for measuring and assessing the ability of upper extremities to perform certain tasks is vital for telerehabilitation systems to deliver their full potential.

      1.6.1 Questionnaire‐based assessment scales

      In the past few decades, a number of approaches have been proposed for assessing upper extremities, the majority of which are questionnaire‐based. For musculoskeletal movement disorders of the extremities, most scales are generic. For instance, the self‐reported Musculoskeletal Function Assessment (MFA) instrument [226], Short Musculoskeletal Function Assessment (SMFA) questionnaire [344] and self‐administered СКАЧАТЬ