hh.sePublikationer
Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Audio–visual person authentication using lip-motion from orientation maps
Högskolan i Halmstad, Akademin för informationsteknologi, Halmstad Embedded and Intelligent Systems Research (EIS).
Högskolan i Halmstad, Akademin för informationsteknologi, Halmstad Embedded and Intelligent Systems Research (EIS).ORCID-id: 0000-0002-4929-1262
2007 (Engelska)Ingår i: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 28, nr 11, s. 1368-1382Artikel i tidskrift (Refereegranskat) Published
Abstract [en]

This paper describes a new identity authentication technique by a synergetic use of lip-motion and speech. The lip-motion is defined as the distribution of apparent velocities in the movement of brightness patterns in an image and is estimated by computing the velocity components of the structure tensor by 1D processing, in 2D manifolds. Since the velocities are computed without extracting the speaker’s lip-contours, more robust visual features can be obtained in comparison to motion features extracted from lip-contours. The motion estimations are performed in a rectangular lip-region, which affords increased computational efficiency. A person authentication implementation based on lip-movements and speech is presented along with experiments exhibiting a recognition rate of 98%. Besides its value in authentication, the technique can be used naturally to evaluate the “liveness” of someone speaking as it can be used in text-prompted dialogue. The XM2VTS database was used for performance quantification as it is currently the largest publicly available database (≈300 persons) containing both lip-motion and speech. Comparisons with other techniques are presented.

Ort, förlag, år, upplaga, sidor
Amsterdam: North-Holland , 2007. Vol. 28, nr 11, s. 1368-1382
Nyckelord [en]
Audio–visual recognition, Biometrics, Biometric recognition, Speaker verification, Speaker authentication, Person identification, Lip-movements, Motion, Structure tensor, Orientation, Optical flow, Hidden Markov model, Gaussian Markov model
Nationell ämneskategori
Teknik och teknologier
Identifikatorer
URN: urn:nbn:se:hh:diva-1335DOI: 10.1016/j.patrec.2007.02.017ISI: 000247807500013Scopus ID: 2-s2.0-34249752774Lokalt ID: 2082/1714OAI: oai:DiVA.org:hh-1335DiVA, id: diva2:238553
Tillgänglig från: 2008-04-16 Skapad: 2008-04-16 Senast uppdaterad: 2018-03-23Bibliografiskt granskad

Open Access i DiVA

Fulltext saknas i DiVA

Övriga länkar

Förlagets fulltextScopus

Personposter BETA

Faraj, Maycel IsaacBigun, Josef

Sök vidare i DiVA

Av författaren/redaktören
Faraj, Maycel IsaacBigun, Josef
Av organisationen
Halmstad Embedded and Intelligent Systems Research (EIS)
I samma tidskrift
Pattern Recognition Letters
Teknik och teknologier

Sök vidare utanför DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetricpoäng

doi
urn-nbn
Totalt: 174 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf