Musculoskeletal models enable movement scientists to examine muscle function by computing the mechanical work done by muscles during motor tasks. To estimate muscle work accurately requires a model that is physiologically plausible. Previous models of the human shoulder have coupled scapula movement to humeral movement. While coupled movement produces a stereotypical scapulohumeral rhythm, it cannot model shrugging or independent movement of the scapula and humerus. The artificial coupling of humeral elevation to scapular rotation permits muscles that cross the glenohumeral joint, such as the rotator-cuff muscles and deltoids, to do implausible work to elevate and rotate the scapula. In reality, the motion of the scapula is controlled by thoracoscapular muscles, yet the roles of these muscles in shoulder function remains unclear. To elucidate the roles of the thoracoscapular muscles, we developed a shoulder model with an accurate scapulothoracic joint and includes scapular muscles to drive its motion. We used the model to compute the work done by the thoracoscapular muscles during shrugging and arm elevation. We found that the bulk of the work done in upper-extremity tasks is performed by the largest muscles of the shoulder: trapezius, deltoids, pectoralis major, and serratus-anterior. Trapezius and serratus anterior prove to be important synergists in performing upward-rotation of the scapula. We show that the large thoracoscapular muscles do more work than glenohumeral muscles during arm-elevation tasks. The model, experimental data and simulation results are freely available on SimTK.org to enable anyone to explore our results and to perform further studies in OpenSim 4.0.
Emotional voice conversion aims at converting speech from one emotion state to another. This paper proposes to model timbre and prosody features using a deep bidirectional long shortterm memory (DBLSTM) for emotional voice conversion. A continuous wavelet transform (CWT) representation of fundamental frequency (F0) and energy contour are used for prosody modeling. Specifically, we use CWT to decompose F0 into a five-scale representation, and decompose energy contour into a ten-scale representation, where each feature scale corresponds to a temporal scale. Both spectrum and prosody (F0 and energy contour) features are simultaneously converted by a sequence to sequence conversion method with DBLSTM model, which captures both frame-wise and long-range relationship between source and target voice. The converted speech signals are evaluated both objectively and subjectively, which confirms the effectiveness of the proposed method.
Voice conversion (VC) aims to make one speaker (source) to sound like spoken by another speaker (target) without changing the language content. Most of the state-of-the-art voice conversion systems focus only on timbre conversion. However, the speaker identity is characterized by the source-related cues such as fundamental frequency and energy as well. In this work, we propose an exemplarbased sparse representation of timbre and prosody for voice conversion that does not necessitate separately timbre conversion and prosody conversions. The experiment results show that, in addition to the conversion of spectral features, the proper conversion of prosody features will improve the quality and speaker identity of the converted speech.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.