SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "WFRF:(Salvi Giampiero) srt2:(2005-2009)"

Sökning: WFRF:(Salvi Giampiero) > (2005-2009)

  • Resultat 1-10 av 16
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Agelfors, Eva, et al. (författare)
  • User evaluation of the SYNFACE talking head telephone
  • 2006
  • Ingår i: Computers Helping People With Special Needs, Proceedings. - Berlin, Heidelberg : Springer Berlin Heidelberg. - 3540360204 ; , s. 579-586
  • Konferensbidrag (refereegranskat)abstract
    • The talking-head telephone, Synface, is a lip-reading support for people with hearing-impairment. It has been tested by 49 users with varying degrees of hearing-impaired in UK and Sweden in lab and home environments. Synface was found to give support to the users, especially in perceiving numbers and addresses and an enjoyable way to communicate. A majority deemed Synface to be a useful product.
  •  
2.
  • Al Moubayed, Samer, et al. (författare)
  • Studies on Using the SynFace Talking Head for the Hearing Impaired
  • 2009
  • Ingår i: Proceedings of Fonetik'09. - Stockholm : Stockholm University. - 9789163348921 ; , s. 140-143
  • Konferensbidrag (övrigt vetenskapligt/konstnärligt)abstract
    • SynFace is a lip-synchronized talking agent which is optimized as a visual reading support for the hearing impaired. In this paper wepresent the large scale hearing impaired user studies carried out for three languages in the Hearing at Home project. The user tests focuson measuring the gain in Speech Reception Threshold in Noise and the effort scaling when using SynFace by hearing impaired people, where groups of hearing impaired subjects with different impairment levels from mild to severe and cochlear implants are tested. Preliminaryanalysis of the results does not show significant gain in SRT or in effort scaling. But looking at large cross-subject variability in both tests, it isclear that many subjects benefit from SynFace especially with speech with stereo babble.
  •  
3.
  • Al Moubayed, Samer, et al. (författare)
  • SynFace Phone Recognizer for Swedish Wideband and Narrowband Speech
  • 2008
  • Ingår i: Proceedings of The second Swedish Language Technology Conference (SLTC). - Stockholm, Sweden.. ; , s. 3-6
  • Konferensbidrag (övrigt vetenskapligt/konstnärligt)abstract
    • In this paper, we present new results and comparisons of the real-time lips synchronized talking head SynFace on different Swedish databases and bandwidth. The work involves training SynFace on narrow-band telephone speech from the Swedish SpeechDat, and on the narrow-band and wide-band Speecon corpus. Auditory perceptual tests are getting established for SynFace as an audio visual hearing support for the hearing-impaired. Preliminary results show high recognition accuracy compared to other languages.
  •  
4.
  • Al Moubayed, Samer, et al. (författare)
  • Virtual Speech Reading Support for Hard of Hearing in a Domestic Multi-Media Setting
  • 2009
  • Ingår i: INTERSPEECH 2009. - BAIXAS : ISCA-INST SPEECH COMMUNICATION ASSOC. ; , s. 1443-1446
  • Konferensbidrag (refereegranskat)abstract
    • In this paper we present recent results on the development of the SynFace lip synchronized talking head towards multilinguality, varying signal conditions and noise robustness in the Hearing at Home project. We then describe the large scale hearing impaired user studies carried out for three languages. The user tests focus on measuring the gain in Speech Reception Threshold in Noise when using SynFace, and on measuring the effort scaling when using SynFace by hearing impaired people. Preliminary analysis of the results does not show significant gain in SRT or in effort scaling. But looking at inter-subject variability, it is clear that many subjects benefit from SynFace especially with speech with stereo babble noise.
  •  
5.
  • Beskow, Jonas, et al. (författare)
  • Hearing at Home : Communication support in home environments for hearing impaired persons
  • 2008
  • Ingår i: INTERSPEECH 2008. - BAIXAS : ISCA-INST SPEECH COMMUNICATION ASSOC. - 9781615673780 ; , s. 2203-2206
  • Konferensbidrag (refereegranskat)abstract
    • The Hearing at Home (HaH) project focuses on the needs of hearing-impaired people in home environments. The project is researching and developing an innovative media-center solution for hearing support, with several integrated features that support perception of speech and audio, such as individual loudness amplification, noise reduction, audio classification and event detection, and the possibility to display an animated talking head providing real-time speechreading support. In this paper we provide a brief project overview and then describe some recent results related to the audio classifier and the talking head. As the talking head expects clean speech input, an audio classifier has been developed for the task of classifying audio signals as clean speech, speech in noise or other. The mean accuracy of the classifier was 82%. The talking head (based on technology from the SynFace project) has been adapted for German, and a small speech-in-noise intelligibility experiment was conducted where sentence recognition rates increased from 3% to 17% when the talking head was present.
  •  
6.
  • Beskow, Jonas, et al. (författare)
  • SynFace - Verbal and Non-verbal Face Animation from Audio
  • 2009
  • Ingår i: Auditory-Visual Speech Processing 2009, AVSP 2009. - Norwich, England : The International Society for Computers and Their Applications (ISCA).
  • Konferensbidrag (refereegranskat)abstract
    • We give an overview of SynFace, a speech-driven face animation system originally developed for the needs of hard-of-hearing users of the telephone. For the 2009 LIPS challenge, SynFace includes not only articulatory motion but also non-verbal motion of gaze, eyebrows and head, triggered by detection of acoustic correlates of prominence and cues for interaction control. In perceptual evaluations, both verbal and non-verbal movmements have been found to have positive impact on word recognition scores. 
  •  
7.
  • Krunic, Verica, et al. (författare)
  • Affordance based word-to-meaning association
  • 2009
  • Ingår i: ICRA. - : VDE Verlag GmbH. - 9781424427888 ; , s. 4138-4143
  • Konferensbidrag (refereegranskat)abstract
    • This paper presents a method to associate meanings to words in manipulation tasks. We base our model on an affordance network, i.e., a mapping between robot actions, robot perceptions and the perceived effects of these actions upon objects. We extend the affordance model to incorporate words. Using verbal descriptions of a task, the model uses temporal co-occurrence to create links between speech utterances and the involved objects, actions and effects. We show that the robot is able form useful word-to-meaning associations, even without considering grammatical structure in the learning process and in the presence of recognition errors. These word-to-meaning associations are embedded in the robot's own understanding of its actions. Thus they can be directly used to instruct the robot to perform tasks and also allow to incorporate context in the speech recognition task.
  •  
8.
  • Krunic, Verica, et al. (författare)
  • Associating word descriptions to learned manipulation task models
  • 2008
  • Ingår i: IEEE/RSJ International Conference on Intelligent RObots and Systems (IROS). - Nice, France.
  • Konferensbidrag (refereegranskat)abstract
    • This paper presents a method to associate meanings to words in manipulation tasks. We base our model on an affordance network, i.e., a mapping between robot actions, robot perceptions and the perceived effects of these actions upon objects. This knowledge is acquired by the robot in an unsupervised way by self-interaction with the environment. When a human user is involved in the process and describes a particular task, the robot can form associations between the (co-occurrence of) speech utterances and the involved objects, actions and effects. We extend the affordance model to incorporate a simple description of speech as a set of words. We show that, across many experiences, the robot is able form useful word-to-meaning associations, even without considering grammatical structure in the learning process and in the presence of recognition errors. Word-to-meaning associations are then used to instruct the robot to perform tasks and also allow to incorporate context in the speech recognition task.
  •  
9.
  • Lindblom, Björn, et al. (författare)
  • (Re)use of place features in voiced stop systems: : Role of phonetic constraints
  • 2008
  • Ingår i: Proceedings FONETIK 2008. - University of Gothenburg. - 9789197719605
  • Tidskriftsartikel (populärvet., debatt m.m.)abstract
    • Computational experiments focused on place of articulation in voiced stops were designed togenerate ‘optimal’ inventories of CV syllables from a larger set of ‘possible CV:s’ in the presenceof independently and numerically defined articulatory, perceptual and developmentalconstraints. Across vowel contexts the most salient places were retroflex, palatal and uvular.This was evident from acoustic measurements and perceptual data. Simulation results usingthe criterion of perceptual contrast alone failed to produce systems with the typologically widelyattested set [b] [d] [g], whereas using articulatory cost as the sole criterion produced inventoriesin which bilabial, dental/alveolar and velar onsets formed the core. Neither perceptualcontrast, nor articulatory cost, (nor the two combined), produced a consistent re-use ofplace features (‘phonemic coding’). Only systems constrained by ‘target learning’ exhibiteda strong recombination of place features.
  •  
10.
  • Salvi, Giampiero (författare)
  • Advances in regional accent clustering in Swedish
  • 2005
  • Ingår i: Proceedings of European Conference on Speech Communication and Technology (Eurospeech). ; , s. 2841-2844
  • Konferensbidrag (refereegranskat)abstract
    • The regional pronunciation variation in Swedish is analysed on a large database. Statistics over each phoneme and for each region of Sweden are computed using the EM algorithm in a hidden Markov model framework to overcome the difficulties of transcribing the whole set of data at the phonetic level. The model representations obtained this way are compared using a distance measure in the space spanned by the model parameters, and hierarchical clustering. The regional variants of each phoneme may group with those of any other phoneme, on the basis of their acoustic properties. The log likelihood of the data given the model is shown to display interesting properties regarding the choice of number of clusters, given a particular level of details. Discriminative analysis is used to find the parameters that most contribute to the separation between groups, adding an interpretative value to the discussion. Finally a number of examples are given on some of the phenomena that are revealed by examining the clustering tree.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-10 av 16

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy