hh.sePublikasjoner
Endre søk
Link to record
Permanent link

Direct link
BETA
Alternativa namn
Publikasjoner (10 av 150) Visa alla publikasjoner
Alonso-Fernandez, F., Farrugia, R. A., Bigun, J., Fierrez, J. & Gonzalez-Sosa, E. (2019). A Survey of Super-Resolution in Iris Biometrics with Evaluation of Dictionary-Learning. IEEE Access, 7, 6519-6544
Åpne denne publikasjonen i ny fane eller vindu >>A Survey of Super-Resolution in Iris Biometrics with Evaluation of Dictionary-Learning
Vise andre…
2019 (engelsk)Inngår i: IEEE Access, E-ISSN 2169-3536, Vol. 7, s. 6519-6544Artikkel i tidsskrift (Fagfellevurdert) Published
Abstract [en]

The lack of resolution has a negative impact on the performance of image-based biometrics. While many generic super-resolution methods have been proposed to restore low-resolution images, they usually aim to enhance their visual appearance. However, an overall visual enhancement of biometric images does not necessarily correlate with a better recognition performance. Reconstruction approaches need thus to incorporate specific information from the target biometric modality to effectively improve recognition performance. This paper presents a comprehensive survey of iris super-resolution approaches proposed in the literature. We have also adapted an Eigen-patches reconstruction method based on PCA Eigentransformation of local image patches. The structure of the iris is exploited by building a patch-position dependent dictionary. In addition, image patches are restored separately, having their own reconstruction weights. This allows the solution to be locally optimized, helping to preserve local information. To evaluate the algorithm, we degraded high-resolution images from the CASIA Interval V3 database. Different restorations were considered, with 15 × 15 pixels being the smallest resolution evaluated. To the best of our knowledge, this is among the smallest resolutions employed in the literature. The experimental framework is complemented with six publicly available iris comparators, which were used to carry out biometric verification and identification experiments. Experimental results show that the proposed method significantly outperforms both bilinear and bicubic interpolation at very low-resolution. The performance of a number of comparators attain an impressive Equal Error Rate as low as 5%, and a Top-1 accuracy of 77-84% when considering iris images of only 15 × 15 pixels. These results clearly demonstrate the benefit of using trained super-resolution techniques to improve the quality of iris images prior to matching. © 2018, Emerald Publishing Limited.

sted, utgiver, år, opplag, sider
Piscataway, NJ: IEEE, 2019
Emneord
Iris hallucination, iris recognition, eigen-patch, super-resolution, PCA
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-38659 (URN)10.1109/ACCESS.2018.2889395 (DOI)000456549900001 ()2-s2.0-85059007584 (Scopus ID)
Forskningsfinansiär
Swedish Research Council, 2016-03497EU, FP7, Seventh Framework Programme, COST IC1106Knowledge Foundation, SIDUS-AIRKnowledge Foundation, CAISRVinnova, 2018-00472
Tilgjengelig fra: 2018-12-20 Laget: 2018-12-20 Sist oppdatert: 2020-01-31bibliografisk kontrollert
Hernandez-Diaz, K., Alonso-Fernandez, F. & Bigun, J. (2019). Cross Spectral Periocular Matching using ResNet Features. In: : . Paper presented at 12th IAPR International Conference on Biometrics, Crete, Greece, June 4-7, 2019.
Åpne denne publikasjonen i ny fane eller vindu >>Cross Spectral Periocular Matching using ResNet Features
2019 (engelsk)Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

Periocular recognition has gained attention in the last years thanks to its high discrimination capabilities in less constraint scenarios than other ocular modalities. In this paper we propose a method for periocular verification under different light spectra using CNN features with the particularity that the network has not been trained for this purpose. We use a ResNet-101 pretrained model for the ImageNet Large Scale Visual Recognition Challenge to extract features from the IIITD Multispectral Periocular Database. At each layer the features are compared using χ 2 distance and cosine similitude to carry on verification between images, achieving an improvement in the EER and accuracy at 1% FAR of up to 63.13% and 24.79% in comparison to previous works that employ the same database. In addition to this, we train a neural network to match the best CNN feature layer vector from each spectrum. With this procedure, we achieve improvements of up to 65% (EER) and 87% (accuracy at 1% FAR) in cross-spectral verification with respect to previous studies.

HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-40499 (URN)
Konferanse
12th IAPR International Conference on Biometrics, Crete, Greece, June 4-7, 2019
Forskningsfinansiär
Swedish Research Council, 2016-03497Knowledge Foundation, SIDUS-AIRKnowledge Foundation, CAISR
Tilgjengelig fra: 2019-09-04 Laget: 2019-09-04 Sist oppdatert: 2019-10-11
Hernandez-Diaz, K., Alonso-Fernandez, F. & Bigun, J. (2019). Cross-Spectral Biometric Recognition with Pretrained CNNs as Generic Feature Extractors. In: : . Paper presented at Swedish Symposium on Image Analysis, SSBA, Gothenburg, Sweden, March 19-20, 2019.
Åpne denne publikasjonen i ny fane eller vindu >>Cross-Spectral Biometric Recognition with Pretrained CNNs as Generic Feature Extractors
2019 (engelsk)Konferansepaper, Publicerat paper (Annet vitenskapelig)
Abstract [en]

Periocular recognition has gained attention in the last years thanks to its high discrimination capabilities in less constraint scenarios than face or iris. In this paper we propose a method for periocular verification under different light spectra using CNN features with the particularity that the network has not been trained for this purpose. We use a ResNet-101 pretrained model for the ImageNet Large Scale Visual Recognition Challenge to extract features from the IIITD Multispectral Periocular Database. At each layer the features are compared using χ 2 distance and cosine similitude to carry on verification between images, achieving an improvement in the EER and accuracy at 1% FAR of up to 63.13% and 24.79% in comparison to previous works that employ the same database. In addition to this, we train a neural network to match the best CNN feature layer vector from each spectrum. With this procedure, we achieve improvements of up to 65% (EER) and 87% (accuracy at 1% FAR) in cross-spectral verification with respect to previous studies.

HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-40625 (URN)
Konferanse
Swedish Symposium on Image Analysis, SSBA, Gothenburg, Sweden, March 19-20, 2019
Tilgjengelig fra: 2019-09-24 Laget: 2019-09-24 Sist oppdatert: 2019-12-16
Krish, R. P., Fierrez, J., Ramos, D., Alonso-Fernandez, F. & Bigun, J. (2019). Improving Automated Latent Fingerprint Identification Using Extended Minutia Types. Information Fusion, 50, 9-19
Åpne denne publikasjonen i ny fane eller vindu >>Improving Automated Latent Fingerprint Identification Using Extended Minutia Types
Vise andre…
2019 (engelsk)Inngår i: Information Fusion, ISSN 1566-2535, E-ISSN 1872-6305, Vol. 50, s. 9-19Artikkel i tidsskrift (Fagfellevurdert) Published
Abstract [en]

Latent fingerprints are usually processed with Automated Fingerprint Identification Systems (AFIS) by law enforcement agencies to narrow down possible suspects from a criminal database. AFIS do not commonly use all discriminatory features available in fingerprints but typically use only some types of features automatically extracted by a feature extraction algorithm. In this work, we explore ways to improve rank identification accuracies of AFIS when only a partial latent fingerprint is available. Towards solving this challenge, we propose a method that exploits extended fingerprint features (unusual/rare minutiae) not commonly considered in AFIS. This new method can be combined with any existing minutiae-based matcher. We first compute a similarity score based on least squares between latent and tenprint minutiae points, with rare minutiae features as reference points. Then the similarity score of the reference minutiae-based matcher at hand is modified based on a fitting error from the least square similarity stage. We use a realistic forensic fingerprint casework database in our experiments which contains rare minutiae features obtained from Guardia Civil, the Spanish law enforcement agency. Experiments are conducted using three minutiae-based matchers as a reference, namely: NIST-Bozorth3, VeriFinger-SDK and MCC-SDK. We report significant improvements in the rank identification accuracies when these minutiae matchers are augmented with our proposed algorithm based on rare minutiae features. © 2018 Elsevier B.V.

sted, utgiver, år, opplag, sider
Amsterdam: Elsevier, 2019
Emneord
Latent Fingerprints, Forensics, Extended Feature Sets, Rare minutiae features
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-38113 (URN)10.1016/j.inffus.2018.10.001 (DOI)000466056900002 ()2-s2.0-85054739072 (Scopus ID)
Prosjekter
BBfor2
Forskningsfinansiär
EU, FP7, Seventh Framework Programme, FP7-ITN-238803Knowledge Foundation, SIDUS-AIRKnowledge Foundation, CAISR
Merknad

R.K. was supported for the most part of this work by a Marie Curie Fellowship under project BBfor2 from European Commission (FP7-ITN-238803). This work has also been partially supported by Spanish Guardia Civil, and project CogniMetrics (TEC2015-70627-R) from Spanish MINECO/FEDER. The researchers from Halmstad University acknowledge funding from KK-SIDUS-AIR 485 project and the CAISR program in Sweden.

Tilgjengelig fra: 2018-10-08 Laget: 2018-10-08 Sist oppdatert: 2020-01-31bibliografisk kontrollert
Alonso-Fernandez, F., Farrugia, R. A., Fierrez, J. & Bigun, J. (2019). Super-Resolution for Selfie Biometrics: Introduction and Application to Face and Iris (1ed.). In: Ajita Rattani, Reza Derakhshani & Arun A. Ross (Ed.), Selfie Biometrics: Advances and Challenges (pp. 105-128). Cham: Springer
Åpne denne publikasjonen i ny fane eller vindu >>Super-Resolution for Selfie Biometrics: Introduction and Application to Face and Iris
2019 (engelsk)Inngår i: Selfie Biometrics: Advances and Challenges / [ed] Ajita Rattani, Reza Derakhshani & Arun A. Ross, Cham: Springer, 2019, 1, s. 105-128Kapittel i bok, del av antologi (Fagfellevurdert)
Abstract [en]

Biometric research is heading towards enabling more relaxed acquisition conditions. This has effects on the quality and resolution of acquired images, severly affecting the accuracy of recognition systems if not tackled appropriately. In this chapter, we give an overview of recent research in super-resolution reconstruction applied to biometrics, with a focus on face and iris images in the visible spectrum, two prevalent modalities in selfie biometrics. After an introduction to the generic topic of super-resolution, we investigate methods adapted to cater for the particularities of these two modalities. By experiments, we show the benefits of incorporating super-resolution to improve the quality of biometric images prior to recognition. © Springer Nature AG 2019

sted, utgiver, år, opplag, sider
Cham: Springer, 2019 Opplag: 1
Serie
Advances in Computer Vision and Pattern Recognition, ISSN 2191-6586, E-ISSN 2191-6594 ; 77
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-38508 (URN)10.1007/978-3-030-26972-2_5 (DOI)2-s2.0-85073171533 (Scopus ID)978-3-030-26971-5 (ISBN)978-3-030-26972-2 (ISBN)
Prosjekter
SIDUS-AIR
Forskningsfinansiär
Swedish Research CouncilVinnovaKnowledge Foundation
Merknad

Other funder: CogniMetrics (TEC2015-70627-R) from MINECO/FEDER

Tilgjengelig fra: 2018-12-06 Laget: 2018-12-06 Sist oppdatert: 2020-02-03bibliografisk kontrollert
Alonso-Fernandez, F., Bigun, J. & Englund, C. (2018). Expression Recognition Using the Periocular Region: A Feasibility Study. In: Gabriella Sanniti di Baja, Luigi Gallo, Kokou Yetongnon, Albert Dipanda, Modesto Castrillón-Santana & Richard Chbeir (Ed.), 2018 14th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS): . Paper presented at The 14th International Conference on Signal Image Technology & Internet Based Systems, SITIS 2018, Las Palmas de Gran Canaria, Spain, 26-29 November, 2018 (pp. 536-541). Los Alamitos: IEEE Computer Society
Åpne denne publikasjonen i ny fane eller vindu >>Expression Recognition Using the Periocular Region: A Feasibility Study
2018 (engelsk)Inngår i: 2018 14th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS) / [ed] Gabriella Sanniti di Baja, Luigi Gallo, Kokou Yetongnon, Albert Dipanda, Modesto Castrillón-Santana & Richard Chbeir, Los Alamitos: IEEE Computer Society, 2018, s. 536-541Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

This paper investigates the feasibility of using the periocular region for expression recognition. Most works have tried to solve this by analyzing the whole face. Periocular is the facial region in the immediate vicinity of the eye. It has the advantage of being available over a wide range of distances and under partial face occlusion, thus making it suitable for unconstrained or uncooperative scenarios. We evaluate five different image descriptors on a dataset of 1,574 images from 118 subjects. The experimental results show an average/overall accuracy of 67.0%/78.0% by fusion of several descriptors. While this accuracy is still behind that attained with full-face methods, it is noteworthy to mention that our initial approach employs only one frame to predict the expression, in contraposition to state of the art, exploiting several order more data comprising spatial-temporal data which is often not available.

sted, utgiver, år, opplag, sider
Los Alamitos: IEEE Computer Society, 2018
Emneord
Expression Recognition, Emotion Recognition, Periocular Analysis, Periocular Descriptor
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-38505 (URN)978-1-5386-9385-8 (ISBN)978-1-5386-9386-5 (ISBN)
Konferanse
The 14th International Conference on Signal Image Technology & Internet Based Systems, SITIS 2018, Las Palmas de Gran Canaria, Spain, 26-29 November, 2018
Prosjekter
SIDUS-AIR
Forskningsfinansiär
Swedish Research CouncilKnowledge Foundation
Merknad

Funding: Author F. A.-F. thanks the Swedish Research Council for funding his research. Authors acknowledge the CAISR program and the SIDUS-AIR project of the Swedish Knowledge Foundation.

Tilgjengelig fra: 2018-12-06 Laget: 2018-12-06 Sist oppdatert: 2019-05-16bibliografisk kontrollert
Alonso-Fernandez, F., Bigun, J. & Englund, C. (2018). Expression Recognition Using the Periocular Region: A Feasibility Study. In: DiBaja, G. S., Gallo, L., Yetongnon, K., Dipanda, A., CastrillonSantana, M., Chbeir, R. (Ed.), Proceedings. The 14th International Conference on Signal Image Technology & Internet Based Systems: SITIS 2018. Paper presented at 14th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS 2018), Las Palmas de Gran Canaria, Spain, November 26-29, 2018 (pp. 536-541). Institute of Electrical and Electronics Engineers (IEEE)
Åpne denne publikasjonen i ny fane eller vindu >>Expression Recognition Using the Periocular Region: A Feasibility Study
2018 (engelsk)Inngår i: Proceedings. The 14th International Conference on Signal Image Technology & Internet Based Systems: SITIS 2018 / [ed] DiBaja, G. S., Gallo, L., Yetongnon, K., Dipanda, A., CastrillonSantana, M., Chbeir, R., Institute of Electrical and Electronics Engineers (IEEE), 2018, s. 536-541Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

This paper investigates the feasibility of using the periocular region for expression recognition. Most works have tried to solve this by analyzing the whole face. Periocular is the facial region in the immediate vicinity of the eye. It has the advantage of being available over a wide range of distances and under partial face occlusion, thus making it suitable for unconstrained or uncooperative scenarios. We evaluate five different image descriptors on a dataset of 1,574 images from 118 subjects. The experimental results show an average/overall accuracy of 67.0/78.0% by fusion of several descriptors. While this accuracy is still behind that attained with full-face methods, it is noteworthy to mention that our initial approach employs only one frame to predict the expression, in contraposition to state of the art, exploiting several order more data comprising spatial-temporal data which is often not available. ©2018 IEEE

sted, utgiver, år, opplag, sider
Institute of Electrical and Electronics Engineers (IEEE), 2018
Emneord
Expression Recognition, Emotion Recognition, Periocular Analysis, Periocular Descriptor
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-41501 (URN)10.1109/SITIS.2018.00087 (DOI)000469258400076 ()2-s2.0-85065903319 (Scopus ID)978-1-5386-9385-8 (ISBN)978-1-5386-9386-5 (ISBN)
Konferanse
14th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS 2018), Las Palmas de Gran Canaria, Spain, November 26-29, 2018
Forskningsfinansiär
Swedish Research Council
Tilgjengelig fra: 2020-02-03 Laget: 2020-02-03 Sist oppdatert: 2020-02-03bibliografisk kontrollert
Hernandez-Diaz, K., Alonso-Fernandez, F. & Bigun, J. (2018). Periocular Recognition Using CNN Features Off-the-Shelf. In: 2018 International Conference of the Biometrics Special Interest Group (BIOSIG): . Paper presented at International Conference of the Biometrics Special Interest Group (BIOSIG), Darmstadt, Germany, Sept. 26-29, 2018. Piscataway, N.J.: IEEE
Åpne denne publikasjonen i ny fane eller vindu >>Periocular Recognition Using CNN Features Off-the-Shelf
2018 (engelsk)Inngår i: 2018 International Conference of the Biometrics Special Interest Group (BIOSIG), Piscataway, N.J.: IEEE, 2018Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

Periocular refers to the region around the eye, including sclera, eyelids, lashes, brows and skin. With a surprisingly high discrimination ability, it is the ocular modality requiring the least constrained acquisition. Here, we apply existing pre-trained architectures, proposed in the context of the ImageNet Large Scale Visual Recognition Challenge, to the task of periocular recognition. These have proven to be very successful for many other computer vision tasks apart from the detection and classification tasks for which they were designed. Experiments are done with a database of periocular images captured with a digital camera. We demonstrate that these offthe-shelf CNN features can effectively recognize individuals based on periocular images, despite being trained to classify generic objects. Compared against reference periocular features, they show an EER reduction of up to ~40%, with the fusion of CNN and traditional features providing additional improvements.

sted, utgiver, år, opplag, sider
Piscataway, N.J.: IEEE, 2018
Serie
2018 International Conference of the Biometrics Special Interest Group (BIOSIG), ISSN 1617-5468 ; 2018
Emneord
Periocular recognition, deep learning, biometrics, Convolutional Neural Network
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-37704 (URN)10.23919/BIOSIG.2018.8553348 (DOI)2-s2.0-85060015047 (Scopus ID)978-3-88579-676-3 (ISBN)
Konferanse
International Conference of the Biometrics Special Interest Group (BIOSIG), Darmstadt, Germany, Sept. 26-29, 2018
Prosjekter
SIDUS-AIR
Forskningsfinansiär
Knowledge Foundation, SIDUS-AIRSwedish Research Council, 2016-03497Vinnova, 2018-00472Knowledge Foundation, CAISR
Tilgjengelig fra: 2018-08-14 Laget: 2018-08-14 Sist oppdatert: 2020-02-03bibliografisk kontrollert
Ranftl, A., Alonso-Fernandez, F., Karlsson, S. & Bigun, J. (2017). A Real-Time AdaBoost Cascade Face Tracker Based on Likelihood Map and Optical Flow. IET Biometrics, 6(6), 468-477
Åpne denne publikasjonen i ny fane eller vindu >>A Real-Time AdaBoost Cascade Face Tracker Based on Likelihood Map and Optical Flow
2017 (engelsk)Inngår i: IET Biometrics, ISSN 2047-4938, E-ISSN 2047-4946, Vol. 6, nr 6, s. 468-477Artikkel i tidsskrift (Fagfellevurdert) Published
Abstract [en]

We present a novel face tracking approach where optical flow information is incorporated into a modified version of the Viola-Jones detection algorithm. In the original algorithm, detection is static, as information from previous frames is not considered; in addition, candidate windows have to pass all stages of the classification cascade, otherwise they are discarded as containing no face. In contrast, the proposed tracker preserves information about the number of classification stages passed by each window. Such information is used to build a likelihood map, which represents the probability of having a face located at that position. Tracking capabilities are provided by extrapolating the position of the likelihood map to the next frame by optical flow computation. The proposed algorithm works in real time on a standard laptop. The system is verified on the Boston Head Tracking Database, showing that the proposed algorithm outperforms the standard Viola-Jones detector in terms of detection rate and stability of the output bounding box, as well as including the capability to deal with occlusions. We also evaluate two recently published face detectors based on Convolutional Networks and Deformable Part Models, with our algorithm showing a comparable accuracy at a fraction of the computation time.

sted, utgiver, år, opplag, sider
Stevenage: The Institution of Engineering and Technology, 2017
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-33836 (URN)10.1049/iet-bmt.2016.0202 (DOI)000415218200012 ()2-s2.0-85034635497 (Scopus ID)
Prosjekter
SIDUS-AIR
Forskningsfinansiär
Swedish Research Council, 2012-4313Knowledge Foundation, CAISRKnowledge Foundation, SIDUS-AIR
Tilgjengelig fra: 2017-05-11 Laget: 2017-05-11 Sist oppdatert: 2020-02-03bibliografisk kontrollert
Alonso-Fernandez, F. & Bigun, J. (2017). An Overview of Periocular Biometrics. In: Christian Rathgeb & Christoph Busch (Ed.), Iris and Periocular Biometric Recognition: (pp. 29-53). London: The Institution of Engineering and Technology
Åpne denne publikasjonen i ny fane eller vindu >>An Overview of Periocular Biometrics
2017 (engelsk)Inngår i: Iris and Periocular Biometric Recognition / [ed] Christian Rathgeb & Christoph Busch, London: The Institution of Engineering and Technology , 2017, s. 29-53Kapittel i bok, del av antologi (Fagfellevurdert)
Abstract [en]

Periocular biometrics specifically refers to the externally visible skin region of the face that surrounds the eye socket. Its utility is specially pronounced when the iris or the face cannot be properly acquired, being the ocular modality requiring the least constrained acquisition process. It appears over a wide range of distances, even under partial face occlusion (close distance) or low resolution iris (long distance), making it very suitable for unconstrained or uncooperative scenarios. It also avoids the need of iris segmentation, an issue in difficult images. In such situation, identifying a suspect where only the periocular region is visible is one of the toughest real-world challenges in biometrics. The richness of the periocular region in terms of identity is so high that the whole face can even be reconstructed only from images of the periocular region. The technological shift to mobile devices has also resulted in many identity-sensitive applications becoming prevalent on these devices.

sted, utgiver, år, opplag, sider
London: The Institution of Engineering and Technology, 2017
Serie
IET security series ; 5
Emneord
face recognition, image reconstruction, unconstrained scenarios, eye socket, periocular biometrics, partial-face occlusion, face reconstruction, externally visible skin region, uncooperative scenarios, low-resolution iris
HSV kategori
Identifikatorer
urn:nbn:se:hh:diva-37705 (URN)978-1-78561-168-1 (ISBN)9781785611698 (ISBN)
Forskningsfinansiär
Knowledge Foundation, CAISRSwedish Research Council, 2016-03497EU, FP7, Seventh Framework Programme, COST Action IC1106
Tilgjengelig fra: 2018-08-14 Laget: 2018-08-14 Sist oppdatert: 2018-08-16bibliografisk kontrollert
Prosjekter
Läpprörelse, ansikt, och tal analys i synergi för människa-maskin gränssnitt [2008-03876_VR]; Högskolan i HalmstadSkala, orientering och belysning invariant kodning och avkodning -- En studie om invarianta visuella koder [2011-05819_VR]; Högskolan i HalmstadAnsiktsdetektering och robust igenkänning med avseende på bild deformationer [2012-04313_VR]; Högskolan i Halmstad
Organisasjoner
Identifikatorer
ORCID-id: ORCID iD iconorcid.org/0000-0002-4929-1262