hh.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Audio-video synthesis methods for improving performance of biometric systems
Halmstad University, School of Information Technology, Halmstad Embedded and Intelligent Systems Research (EIS).
2007 (English)Licentiate thesis, comprehensive summary (Other academic)
Abstract [en]

System security is important for any automation. It is even more so in the case of biometric systems due to the sensitive nature of the data it uses for enrollment and authentication - the subjects physical or biological trait. The performance quantification of biometric systems, such as face tracking and recognition, highly depend on the database used for testing the systems. Systems trained and tested on realistic and represenative databases evidently perform better. In fact, the main reason for evaluating any system on test data is that these data sets represent problems that system might face in the real world. However, building biometric databases that represent the real world is an expensive task due to its high demand on the side of the participants. This becomes even more difficult and unrealistic if the data is to be collected in a natural environment such as supermarkets, offices, streets, etc.

This thesis presents a procedure to build a synthetic biometric database by damascening images from a studio recorded database with a realistic scenery. To this end, we developed an image segmenation procedure to spearate the background of a video recorded in studio conditions with the prupose to replace it with an arbitrary complex background. Furthermore, we present how several degradations such as affine transformation, imaging noise, and motion blur can be incorporated into the production of the new database to simulate natural recording environments. The system is applied to the entire XM2VTS database, which already consists of several terabytes of data, to produce the DXM2VTS - Damascened XM2VTS database.

Moreover, the thesis presents a method to segment a video sequence in the time domain based on its audio concept. The video is then reshuffled and used for testing resilience of text-prompted biometric systems against playback attacks. The playback is supported by pyramid based frame interpolation method to reduce discontinuities created at the digit boundaries in time.

Place, publisher, year, edition, pages
Gothenburg: Department of Signals and Systems, Chalmers University of Technology , 2007. , p. 31
Series
Technical report R, ISSN 1403-266X ; 2007:6
Keywords [en]
Biometrics, Audio-Video synthesis, Image segmentation, XM2VTS, DXM2VTS
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:hh:diva-1976Libris ID: 10417023Local ID: 2082/2371OAI: oai:DiVA.org:hh-1976DiVA, id: diva2:239194
Presentation
2007-04-24, R1107, Halmstad, 13:15
Supervisors
Available from: 2008-09-29 Created: 2008-09-29 Last updated: 2018-03-23Bibliographically approved
List of papers
1. Damascening video databases for evaluation of face tracking and recognition – The DXM2VTS database
Open this publication in new window or tab >>Damascening video databases for evaluation of face tracking and recognition – The DXM2VTS database
2007 (English)In: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 28, no 15, p. 2143-2156Article in journal (Refereed) Published
Abstract [en]

Performance quantification of biometric systems, such as face tracking and recognition highly depend on the database used for testing the systems. Systems trained and tested on realistic and representative databases evidently perform better. Actually, the main reason for evaluating any system on test data is that these data sets represent problems that systems might face in the real world. However, building biometric video databases with realistic background for testing is expensive especially due to its high demand of cooperation from the side of the participants. For example, XM2VTS database contain thousands of video recorded in a studio from 295 subjects. Recording these subjects repeatedly in public places such as supermarkets, offices, streets, etc., is not realistic. To this end, we present a procedure to separate the background of a video recorded in studio conditions with the purpose to replace it with an arbitrary complex background, e.g., outdoor scene containing motion, to measure performance, e.g., eye tracking. Furthermore, we present how an affine transformation and synthetic noise can be incorporated into the production of the new database to simulate natural noise, e.g. motion blur due to translation, zooming and rotation. The entire system is applied to the XM2VTS database, which already consists of several terabytes of data, to produce the DXM2VTS–Damascened XM2VTS database essentially without an increase in resource consumption, i.e., storage, bandwidth, and most importantly, the time of clients populating the database, and the time of the operators.

Place, publisher, year, edition, pages
Amsterdam: Elsevier, 2007
Keywords
Biometrics, XM2VTS, Damascened video, Face tracking, Face recognition, Performance quantification
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:hh:diva-1334 (URN)10.1016/j.patrec.2007.06.007 (DOI)000250377600026 ()2-s2.0-34548702005 (Scopus ID)2082/1713 (Local ID)2082/1713 (Archive number)2082/1713 (OAI)
Available from: 2008-04-16 Created: 2008-04-16 Last updated: 2018-03-23Bibliographically approved
2. Text Driven Face-Video Synthesis Using GMM and Spatial Correlation
Open this publication in new window or tab >>Text Driven Face-Video Synthesis Using GMM and Spatial Correlation
2007 (English)In: Image analysis: 15th Scandinavian Conference, SCIA 2007, Aalborg, Denmark, June 10-14, 2007 ; proceedings / [ed] Ersboll, B K, Pedersen, K S, Berlin: Springer Berlin/Heidelberg, 2007, p. 572-580Conference paper, Published paper (Refereed)
Abstract [en]

Liveness detection is increasingly planned to be incorporated into biometric systems to reduce the risk of spoofing and impersonation. Some of the techniques used include detection of motion of the head while posing/speaking, iris size in varying illumination, fingerprint sweat, text-prompted speech, speech-to-lip motion synchronization etc. In this paper, we propose to build a biometric signal to test attack resilience of biometric systems by creating a text-driven video synthesis of faces. We synthesize new realistic looking video sequences from real image sequences representing utterance of digits. We determine the image sequences for each digit by using a GMM based speech recognizer. Then, depending on system prompt (sequence of digits) our method regenerates a video signal to test attack resilience of a biometric system that asks for random digit utterances to prevent play-back of pre-recorded data representing both audio and images. The discontinuities in the new image sequence, created at the connection of each digit, are removed by using a frame prediction algorithm that makes use of the well known block matching algorithm. Other uses of our results include web-based video communication for electronic commerce and frame interpolation for low frame rate video.

Place, publisher, year, edition, pages
Berlin: Springer Berlin/Heidelberg, 2007
Series
Lecture Notes in Computer Science, ISSN 0302-9743 ; 4522
Keywords
Image analysis
National Category
Electrical Engineering, Electronic Engineering, Information Engineering Basic Medicine
Identifiers
urn:nbn:se:hh:diva-2130 (URN)000247364000058 ()2-s2.0-38049080023 (Scopus ID)2082/2525 (Local ID)978-3-540-73039-2 (ISBN)2082/2525 (Archive number)2082/2525 (OAI)
Conference
15th Scandinavian Conference on Image Analysis, Aalborg, Denmark, June 10-14, 2007
Available from: 2008-11-12 Created: 2008-11-12 Last updated: 2022-09-13Bibliographically approved

Open Access in DiVA

No full text in DiVA

Authority records

Teferi Lemma, Dereje

Search in DiVA

By author/editor
Teferi Lemma, Dereje
By organisation
Halmstad Embedded and Intelligent Systems Research (EIS)
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 302 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf