hh.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Audio–visual person authentication using lip-motion from orientation maps
Halmstad University, School of Information Technology, Halmstad Embedded and Intelligent Systems Research (EIS).
Halmstad University, School of Information Technology, Halmstad Embedded and Intelligent Systems Research (EIS).ORCID iD: 0000-0002-4929-1262
2007 (English)In: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 28, no 11, p. 1368-1382Article in journal (Refereed) Published
Abstract [en]

This paper describes a new identity authentication technique by a synergetic use of lip-motion and speech. The lip-motion is defined as the distribution of apparent velocities in the movement of brightness patterns in an image and is estimated by computing the velocity components of the structure tensor by 1D processing, in 2D manifolds. Since the velocities are computed without extracting the speaker’s lip-contours, more robust visual features can be obtained in comparison to motion features extracted from lip-contours. The motion estimations are performed in a rectangular lip-region, which affords increased computational efficiency. A person authentication implementation based on lip-movements and speech is presented along with experiments exhibiting a recognition rate of 98%. Besides its value in authentication, the technique can be used naturally to evaluate the “liveness” of someone speaking as it can be used in text-prompted dialogue. The XM2VTS database was used for performance quantification as it is currently the largest publicly available database (≈300 persons) containing both lip-motion and speech. Comparisons with other techniques are presented.

Place, publisher, year, edition, pages
Amsterdam: North-Holland , 2007. Vol. 28, no 11, p. 1368-1382
Keywords [en]
Audio–visual recognition, Biometrics, Biometric recognition, Speaker verification, Speaker authentication, Person identification, Lip-movements, Motion, Structure tensor, Orientation, Optical flow, Hidden Markov model, Gaussian Markov model
National Category
Engineering and Technology
Identifiers
URN: urn:nbn:se:hh:diva-1335DOI: 10.1016/j.patrec.2007.02.017ISI: 000247807500013Scopus ID: 2-s2.0-34249752774Local ID: 2082/1714OAI: oai:DiVA.org:hh-1335DiVA, id: diva2:238553
Available from: 2008-04-16 Created: 2008-04-16 Last updated: 2018-03-23Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records BETA

Faraj, Maycel IsaacBigun, Josef

Search in DiVA

By author/editor
Faraj, Maycel IsaacBigun, Josef
By organisation
Halmstad Embedded and Intelligent Systems Research (EIS)
In the same journal
Pattern Recognition Letters
Engineering and Technology

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 159 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf