-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathpublications.bib
More file actions
3737 lines (3564 loc) · 183 KB
/
publications.bib
File metadata and controls
3737 lines (3564 loc) · 183 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
%%%%%%%%%%%%
%% 2026 %%
%%%%%%%%%%%%
@article{acmhealth-2026,
author = {Queiroz Neto, Dilermando and Carlos, Anderson and Anjos,
Andr{\'{e}} and Berton, Lilian},
keywords = {Fairness, Foundation Models, World Health},
month = jan,
title = {Fair Foundation Models for Medical Image Analysis:
Challenges and Perspectives},
journal = {ACM Transactions on Computing for Healthcare},
year = {2026},
doi = {10.1145/3793542},
abstract = {Ensuring equitable Artificial Intelligence (AI) in
healthcare demands systems that make unbiased decisions across all
demographic groups, bridging technical innovation with ethical
principles. Foundation Models (FMs), trained on vast datasets
through self-supervised learning, enable efficient adaptation
across medical imaging tasks while reducing dependency on labeled
data. These models demonstrate potential for enhancing fairness,
though significant challenges remain in achieving consistent
performance across demographic groups. Our review indicates that
effective bias mitigation in FMs requires systematic interventions
throughout all stages of development. While previous approaches
focused primarily on model-level bias mitigation, our analysis
reveals that fairness in FMs requires integrated interventions
throughout the development pipeline, from data documentation to
deployment protocols. This comprehensive framework advances current
knowledge by demonstrating how systematic bias mitigation, combined
with policy engagement, can effectively address both technical and
institutional barriers to equitable AI in healthcare. The
development of equitable FMs represents a critical step toward
democratizing advanced healthcare technologies, particularly for
underserved populations and regions with limited medical
infrastructure and computational resources.},
}
%%%%%%%%%%%%
%% 2025 %%
%%%%%%%%%%%%
@ARTICLE{melba-2025,
author = {{\"{O}}zbulak, G{\"{o}}khan and Jimenez-del-Toro, Oscar
and Fatoretto, Ma{\'{\i}}ra and Berton, Lilian and Anjos, Andr{\'{e}}},
keywords = {machine learning, Medical Image Analysis,
multi-objective optimization, Multidimensional Fairness Evaluation,
Utility-Fairness Trade-off},
month = dec,
title = {A Multi-Objective Evaluation Framework for Analyzing
Utility-Fairness Trade-Offs in Machine Learning Systems},
journal = {The Journal of Machine Learning for Biomedical Imaging},
volume = {3},
year = {2025},
pages = {938-957},
doi = {10.59275/j.melba.2025-ab9a},
abstract = {The evaluation of fairness models in Machine Learning
involves complex challenges, such as defining appropriate metrics,
balancing trade-offs between utility and fairness, and there are
still gaps in this stage. This work presents a novel
multi-objective evaluation framework that enables the analysis of
utility-fairness trade-offs in Machine Learning systems. The
framework was developed using criteria from Multi-Objective
Optimization that collect comprehensive information regarding this
complex evaluation task. The assessment of multiple Machine
Learning systems is summarized, both quantitatively and
qualitatively, in a straightforward manner through a radar chart
and a measurement table encompassing various aspects such as
convergence, system capacity, and diversity. The framework’s
compact representation of performance facilitates the comparative
analysis of different Machine Learning strategies for
decision-makers, in real-world applications, with single or
multiple fairness requirements. In particular, this study focuses
on the medical imaging domain, where fairness considerations are
crucial due to the potential impact of biased diagnostic systems on
patient outcomes. The proposed framework enables a systematic
evaluation of multiple fairness constraints, helping to identify
and mitigate disparities among demographic groups while maintaining
diagnostic performance. The framework is model-agnostic and
flexible to be adapted to any kind of Machine Learning systems,
that is, black- or white-box, any kind and quantity of evaluation
metrics, including multidimensional fairness criteria. The
functionality and effectiveness of the proposed framework are shown
with different simulations, and an empirical study conducted on
three real-world medical imaging datasets with various Machine
Learning systems. Our evaluation framework is publicly available at
https://pypi.org/project/fairical.},
}
@article{cbm-2025,
author = {Amiot, Victor and Jimenez-del-Toro, Oscar and
Guex-Croisier, Yan and Ott, Muriel and Bogaciu, Teodora-Elena and
Banerjee, Shalini and Howell, Jeremy and Amstutz, Christoph and
Chiquet, Christophe and Bergin, Ciara and Meloni, Ilenia and
Tomasoni, Mattia and Hoogewoud, Florence and Anjos, Andr{\'{e}}},
keywords = {deep learning, fluorescein angiography, inter-grader
agreement, Optic disease grading, transformers, Uveitis, vasculitis},
month = jul,
title = {Automatic transformer-based grading of multiple retinal
inflammatory signs in uveitis on fluorescein angiography},
journal = {Computers in Biology and Medicine},
volume = {193},
number = {110327},
year = {2025},
issn = {0010-4825},
url = {https://github.com/JulesGoninRIO/uveitis_transformers},
doi = {10.1016/j.compbiomed.2025.110327},
abstract = {Grading fluorescein angiography (FA) for uveitis is
complex, often leading to the oversight of retinal inflammation
in clinical studies. This study aims to develop an automated
method for grading retinal inflammation.
Methods. Patients from Jules-Gonin Eye Hospital with active or
resolved uveitis who underwent FA between 2018 and 2021 were
included. FAs were acquired using a standardized protocol,
anonymized, and annotated following the Angiography Scoring for
Uveitis Working Group criteria, for four inflammatory signs of
the posterior pole. Intergrader agreement was assessed by four
independent graders. Four deep learning transformer models were
developed, and performance was evaluated using the Ordinal
Classification Index, accuracy, F1 scores, and Kappa scores.
Saliency analysis was employed to visualize model predictions.
Findings. A total of 543 patients (1042 eyes, 40987 images) were
included in the study. The models closely matched expert graders
in detecting vascular leakage (F1-score =
0{\textperiodcentered}87, 1-OCI = 0{\textperiodcentered}89),
capillary leakage (F1-score = 0{\textperiodcentered}86, 1-OCI =
0{\textperiodcentered}89), macular edema (F1-score =
0{\textperiodcentered}82, 1-OCI = 0{\textperiodcentered}86), and
optic disc hyperfluorescence (F1-score =
0{\textperiodcentered}72, 1-OCI = 0{\textperiodcentered}85).
Saliency analysis confirmed that the models focused on relevant
retinal structures. The mean intergrader agreement across all
inflammatory signs was F1-score = 0{\textperiodcentered}79 and
1-OCI = 0{\textperiodcentered}83.
Interpretation. We developed a vision transformer-based model for
the automatic grading of retinal inflammation in uveitis, utilizing
the largest dataset of FAs in uveitis to date. This approach
provides significant clinical benefits for the evaluation of
uveitis and paves the way for future advancements, including the
identification of novel biomarkers through the integration of
clinical data and other modalities.}
}
@article{asoc-2025,
author = {Fatoreto, Maira and {\"{O}}zbulak, G{\"{o}}khan and
Berton, Lilian and Anjos, Andr{\'{e}}},
month = sep,
title = {Optimizing Fairness and Utility in Healthcare Machine
Learning Models},
journal = {Applied Soft Computing},
volume = {181},
number = {113426},
year = {2025},
doi = {10.1016/j.asoc.2025.113426},
abstract = {Demographic fairness or equity is a crucial aspect of
machine learning models, particularly in critical domains such as
healthcare, where errors can have severe consequences. A fair model
avoids making distinctions between groups with different sensitive
or protected attributes. Although several metrics are available to
measure fairness and ensure equity between groups, proposed
solutions must uphold fairness without compromising the utility of
these models. Optimization problems that incorporate both fairness
and utility information can help find the best machine learning
model. Mathematical programming emerges as a valuable tool in this
context. One approach is to use optimization functions with
constraints, imposing a maximum difference between groups. In this
sense, considering multiple constraints that encompass various
sensitive attributes present in the dataset when adjusting the
models is essential to ensure intersectional fairness, minimize
hidden biases, and promote equitable decisions in diverse contexts.
In this work, we propose to constrain the minimization of the loss
function with multiple fairness-related metrics, ensuring that
fairness metrics do not exceed a maximum limit concerning the
impartiality of the decision boundary. We use metrics derived from
Pareto fronts, a method used in multi-objective optimization,
adapting it for single-objective optimization and incorporating
fairness characteristics into the constraints. The points observed
in this graph use different fairness thresholds. We compare our
proposed model with existing literature and demonstrate the
convergence of our model to logistic regression with simulated
data. Furthermore, we apply this strategy to health-related
datasets and other domains present in most fairness and
optimization articles. As a result, we found that, using the
proposed metrics, our model performs better, even with imbalanced
data concerning sensitive attributes and smaller datasets.}
}
@inproceedings{ijcb-2025,
author = {Vitek, Matej and Toma{\v s}evi{\'{c}}, Darian and Das,
Abhijit and Nathan, Sabari and {\"{O}}zbulak, G{\"{o}}khan and
Tataroğlu {\"{O}}zbulak, G{\"{o}}zde Ayşe and Calbimonte, Jean-Paul
and Anjos, Andr{\'{e}} and Hemant Bhatt, Hariohm and Dhirendra
Premani, Dhruv and Chaudhari, Jay and Wang, Caiyong and Jiang, Jian
and Zhang, Chi and Zhang, Qi and Iyappan Ganapathi, Iyyakutti and
Sadaf Ali, Syed and Velayudan, Divya and Assefa, Maregu and Werghi,
Naoufel and A Daniels, Zachary and John, Leeon and Vyas, Ritesh and
Nourmohammadi Khiarak, Jalil and Akbari Saeed, Taher and Nasehi,
Mahsa and Kianfar, Ali and Pashazadeh Panahi, Mobina and Sharma,
Geetanjali and Raj Panth, Pushp and Ramachandra, Raghavendra and
Nigam, Aditya and Pal, Umapada and Pedrini, Helio and Struc, Vitomir},
title = {Privacy-enhancing Sclera Segmentation Benchmarking
Competition: SSBC 2025},
booktitle = {International Joint Conference on Biometrics},
year = {2025},
publisher = {IEEE},
doi = {10.48550/arXiv.2508.10737},
abstract = {This paper presents a summary of the 2025 Sclera
Segmentation Benchmarking Competition (SSBC), which focused on the
development of privacy-preserving sclera-segmentation models
trained using synthetically generated ocular images. The goal of
the competition was to evaluate how well models trained on
synthetic data perform in comparison to those trained on real-world
datasets. The competition featured two tracks: one relying solely
on synthetic data for model development, and one combining/mixing
synthetic with (a limited amount of) real-world data. A total of
nine research groups submitted diverse segmentation models,
employing a variety of architectural designs, including
transformer-based solutions, lightweight models, and segmentation
networks guided by generative frameworks. Experiments were
conducted across three evaluation datasets containing both
synthetic and real-world images, collected under diverse
conditions. Results show that models trained entirely on synthetic
data can achieve competitive performance, particularly when
dedicated training strategies are employed, as evidenced by the top
performing models that achieved scores of over in the synthetic
data track. Moreover, performance gains in the mixed track were
often driven more by methodological choices rather than by the
inclusion of real data, highlighting the promise of synthetic data
for privacy-aware biometric development. The code and data for the
competition is available at: https://github.com/dariant/SSBC_2025.}
}
@inproceedings{aime-2025,
author = {Geissbuhler, Damien and Bornet, Alban and Marques,
Catarina and Anjos, Andr{\'{e}} and Pereira, S{\'{o}}nia and
Teodoro, Douglas},
keywords = {Hospital-Acquired Infections, Multidrug Resistance
Prediction, Sparse Clinical Data Integration, Spatiotemporal
Modeling, Temporal Graph Neural Networks},
month = 6,
title = {STM-GNN: Space-Time-and-Memory Graph Neural Networks for
Predicting Multi-Drug Resistance Risks in Dynamic Patient Networks},
booktitle = {International Conference on Artificial Intelligence in Medicine},
year = {2025},
location = {Pavia, Italy},
isbn = {978-3-031-95838-0},
doi = {10.1007/978-3-031-95838-0_16},
abstract = {Hospital-acquired infections (HAIs), particularly those
caused by multidrug-resistant (MDR) bacteria, pose significant
risks to vulnerable patients. Accurate predictive models are
important for assessing infection dynamics and informing infection
prediction and control (IPC) programmes. Graph-based methods,
including graph neural networks (GNNs), offer a powerful approach
to model complex relationships between patients and environments
but often struggle with data sparsity, irregularity, and
heterogeneity. We propose the space-time-and-memory (STM)-GNN, a
temporal GNN enhanced with recurrent connectivity designed to
capture spatiotemporal infection dynamics. STM-GNN effectively
integrates sparse, heterogeneous data combining network information
from patient-environment interactions and internal memory from
historical colonization and contact patterns. Using a unique IPC
dataset containing clinical and environmental colonization
information collected from a long-term healthcare unit, we show
that STM-GNN effectively addresses the challenges of limited and
irregular data in an MDR prediction task. Our model reaches 0.84
AUROC, and achieves the most balanced performance overall compared
to classic machine learning algorithms, as well as temporal GNN approaches.}
}
%%%%%%%%%%%%
%% 2024 %%
%%%%%%%%%%%%
@misc{ssrn-2024,
title = {Automatic Transformer-Based Grading of Multiple Retinal
Inflammatory Signs on Fluorescein Angiography},
url = {https://papers.ssrn.com/abstract=4960069},
doi = {10.2139/ssrn.4960069},
abstract = {Background: Grading fluorescein angiography ({FA}) in
the context of uveitis is complex, often leading to the oversight
of retinal inflammation in clinical studies. This study aims to
develop an automated method for grading retinal inflammation.},
number = {4960069},
author = {Amiot, Victor and Jimenez-del-Toro, Oscar and
Guex-Croisier, Yan and Ott, Muriel and Bogaciu, Teodora-Elena and
Banerjee, Shalini and Howell, Jeremy and Amstutz, Christoph and
Chiquet, Christophe and Bergin, Ciara and Meloni, Ilenia and
Tomasoni, Mattia and Hoogewoud, Florence and Anjos, André},
year = {2024},
month = 9,
day = 24,
keywords = {capillaropathy, Deep Learning, disease grading,
fluorescein angiography, inter-grader agreement, macular edema,
optic disc hyperfluorescence, ordinal classification index,
papillitis, retinal inflammation, transformers, Uveitis, vascular
leakage, vasculitis},
}
@inproceedings{miccai-2024,
author = {Queiroz Neto, Dilermando and Anjos, Andr{\'{e}} and Berton, Lilian},
keywords = {Fairness, Foundation Model, Medical Image},
month = 10,
title = {Using Backbone Foundation Model for Evaluating Fairness in
Chest Radiography Without Demographic Data},
booktitle = {Proceedings of the International Conference on Medical
Image Computing and Computer Assisted Intervention (MICCAI)},
year = {2024},
abstract = {Ensuring consistent performance across diverse
populations and incorporating fairness into machine learning models
are crucial for advancing medical image diagnostics and promoting
equitable healthcare. However, many databases do not provide
protected attributes or contain unbalanced representations of
demographic groups, complicating the evaluation of model
performance across different demographics and the application of
bias mitigation techniques that rely on these attributes. This
study aims to investigate the effectiveness of using the backbone
of Foundation Models as an embedding extractor for creating groups
that represent protected attributes, such as gender and age. We
propose utilizing these groups in different stages of bias
mitigation, including pre-processing, in-processing, and
evaluation. Using databases in and out-of-distribution scenarios,
it is possible to identify that the method can create groups that
represent gender in both databases and reduce in 4.44\% the
difference between the gender attribute in-distribution and 6.16\%
in out-of-distribution. However, the model lacks robustness in
handling age attributes, underscoring the need for more
fundamentally fair and robust Foundation models. These findings
suggest a role in promoting fairness assessment in scenarios where
we lack knowledge of attributes, contributing to the development of
more equitable medical diagnostics.},
pdf =
{https://publications.idiap.ch/attachments/papers/2024/QueirozNeto_CVPR_2024.pdf}
}
@inproceedings{eccv-2024,
author = {Queiroz Neto, Dilermando and Carlos, Anderson and
Fatoretto, Ma{\'{\i}}ra and Nakayama, Luis Filipe and Anjos,
Andr{\'{e}} and Berton, Lilian},
projects = {FAIRMI},
month = 10,
title = {Does Data-Efficient Generalization Exacerbate Bias in
Foundation Models?},
booktitle = {Proceedings of the 18th European Conference on
Computer Vision (ECCV)},
year = {2024},
abstract = {Foundation models have emerged as robust models with
label efficiency in diverse domains. In medical imaging, these
models contribute to the advancement of medical diagnoses due to
the difficulty in obtaining labeled data. However, it is unclear
whether using a large amount of unlabeled data, biased by the
presence of sensitive attributes during pre-training, influences
the fairness of the model. This research examines the bias in the
Foundation model (RetFound) when it is applied to fine-tune the
Brazilian Multilabel Ophthalmological Dataset (BRSET), which has a
different population than the pre-training dataset. The model
evaluation, in comparison with supervised learning, shows that the
Foundation Model has the potential to reduce the gap between the
maximum AUC and minimum AUC evaluations across gender and age
groups. However, in a data-efficient generalization, the model
increases the bias when the data amount decreases. These findings
suggest that when deploying a Foundation Model in real-life
scenarios with limited data, the possibility of fairness issues
should be considered.},
pdf =
{https://publications.idiap.ch/attachments/papers/2024/QueirozNeto_ECCV_2024.pdf}
}
@inproceedings{euvip-2024-2,
author = {Jimenez-del-Toro, Oscar and Aberle, Christoph and Schaer,
Roger and Bach, Michael and Flouris, Kyriakos and Konukoglu, Ender
and Stieltjes, Bram and Obmann, Markus M. and Anjos, Andr{\'{e}}
and M{\"{u}}ller, Henning and Depeursinge, Adrien},
month = 9,
title = {Comparing Stability and Discriminatory Power of
Hand-crafted Versus Deep Radiomics: A 3D-Printed Anthropomorphic
Phantom Study},
booktitle = {Proceedings of the 12th European Workshop on Visual
Information Processing},
year = {2024},
abstract = {Radiomics have the ability to comprehensively quantify
human tissue characteristics in medical imaging studies. However,
standard radiomic features are highly unstable due to their
sensitivity to scanner and reconstruction settings. We present an
evaluation framework for the extraction of 3D deep radiomics
features using a pre-trained neural network on real computed
tomography (CT) scans for tissue characterization. We compare both
the stability and discriminative power of the proposed 3D deep
learning radiomic features versus standard hand-crafted radiomic
features using 8 image acquisition protocols with a 3D-printed
anthropomorphic phantom containing 4 classes of liver lesions and
normal tissue. Even when the deep learning model was trained on an
external dataset and for a different tissue characterization task,
the resulting generic deep radiomics are at least twice more stable
on 8 CT parameter variations than any category of hand-crafted
features. Moreover, the 3D deep radiomics were also discriminative
for the tissue characterization between 4 classes of liver tissue
and lesions, with an average discriminative power of 93.5\%.},
pdf =
{https://publications.idiap.ch/attachments/papers/2024/Jimenez-del-Toro_EUVIP2024_2024.pdf}
}
@inproceedings{euvip-2024-1,
author = {G{\"{u}}ler, {\"{O}}zg{\"{u}}r and G{\"{u}}nther, Manuel
and Anjos, Andr{\'{e}}},
month = 9,
title = {Refining Tuberculosis Detection in CXR Imaging: Addressing
Bias in Deep Neural Networks via Interpretability},
booktitle = {Proceedings of the 12th European Workshop on Visual
Information Processing},
year = {2024},
abstract = {Automatic classification of active tuberculosis from
chest X-ray images has the potential to save lives, especially in
low- and mid-income countries where skilled human experts can be
scarce. Given the lack of available labeled data to train such
systems and the unbalanced nature of publicly available datasets,
we argue that the reliability of deep learning models is limited,
even if they can be shown to obtain perfect classification accuracy
on the test data. One way of evaluating the reliability of such
systems is to ensure that models use the same regions of input
images for predictions as medical experts would. In this paper, we
show that pre-training a deep neural network on a large-scale proxy
task, as well as using mixed objective optimization network (MOON),
a technique to balance different classes during pre-training and
fine-tuning, can improve the alignment of decision foundations
between models and experts, as compared to a model directly trained
on the target dataset. At the same time, these approaches keep
perfect classification accuracy according to the area under the
receiver operating characteristic curve (AUROC) on the test set,
and improve generalization on an independent, unseen dataset. For
the purpose of reproducibility, our source code is made available online.},
pdf =
{https://publications.idiap.ch/attachments/papers/2024/Guler_EUVIP24_2024.pdf}
}
@article{mvr-2024,
author = {Mautuit, Thibaud and Cunnac, Pierre and Truffer,
Fr{\'{e}}d{\'{e}}ric and Anjos, Andr{\'{e}} and Dufrane, Rebecca
and Ma{\^{\i}}tre, Gilbert and Geiser, Martial and Chiquet, Christophe},
month = 1,
title = {Absolute retinal blood flow in healthy eyes and in eyes
with retinal vein occlusion},
journal = {Microvascular Research},
volume = {152},
year = {2024},
issn = {0026-2862},
doi = {10.1016/j.mvr.2023.104648},
abstract = {Purpose: To measure non-invasively retinal venous blood flow
(RBF) in healthy subjects and patients with retinal venous occlusion (RVO).
Methods: The prototype named AO-LDV (Adaptive Optics Laser Doppler
Velocimeter), which combines a new absolute laser Doppler velocimeter with
an adaptive optics fundus camera (rtx1, Imagine Eyes{\textregistered},
Orsay, France), was studied for the measurement of absolute RBF as a
function of retinal vessel diameters and simultaneous measurement of red
blood cell velocity. RBF was measured in healthy subjects (n = 15) and
patients with retinal venous occlusion (RVO, n = 6). We also evaluated two
softwares for the measurement of retinal vessel diameters: software 1
(automatic vessel detection, profile analysis) and software 2 (based on the
use of deep neural networks for semantic segmentation of vessels, using a
M2u-Net architecture).
Results: Software 2 provided a higher rate of automatic retinal vessel
measurement (99.5 \% of 12,320 AO images) than software 1 (64.9 \%) and
wider measurements (75.5 ± 15.7 μm vs 70.9 ± 19.8 μm, p < 0.001). For
healthy subjects (n = 15), all the retinal veins in one eye were measured
to obtain the total RBF. In healthy subjects, the total RBF was 37.8 ± 6.8
μl/min. There was a significant linear correlation between retinal vessel
diameter and maximal velocity (slope = 0.1016; p < 0.001; r2 = 0.8597) and
a significant power curve correlation between retinal vessel diameter and
blood flow (3.63 × 10−5 × D2.54; p < 0.001; r2 = 0.7287). No significant
relationship was found between total RBF and systolic and diastolic blood
pressure, ocular perfusion pressure, heart rate, or hematocrit. For RVO
patients (n = 6), a significant decrease in RBF was noted in occluded veins
(3.51 ± 2.25 μl/min) compared with the contralateral healthy eye (11.07 ±
4.53 μl/min). For occluded vessels, the slope between diameter and velocity
was 0.0195 (p < 0.001; r2 = 0.6068) and the relation between diameter and
flow was Q = 9.91 × 10−6 × D2.41 (p < 0.01; r2 = 0.2526).
Conclusion: This AO-LDV prototype offers new opportunity to study
RBF in humans
and to evaluate treatment in retinal vein diseases.},
}
%%%%%%%%%%%%
%% 2023 %%
%%%%%%%%%%%%
@inproceedings{cbms-2023,
author = {Amiot, Victor AND Jimenez-del-Toro, Oscar AND Eyraud,
Pauline AND Guex-Crosier, Yan AND Bergin, Ciara AND Anjos, André
AND Hoogewoud, Florence AND Tomasoni, Mattia},
title = {Fully Automatic Grading of Retinal Vasculitis on
Fluorescein Angiography Time-lapse from Real-world Data in Clinical Settings},
booktitle={2023 IEEE 36th International Symposium on Computer-Based
Medical Systems (CBMS)},
year = {2023},
month = 6,
doi = {10.1109/CBMS58004.2023.00301},
abstract = {The objective of this study is to showcase a pipeline
able to perform fully automated grading of retinal inflammation
based on a standardised, clinically-validated grading scale. The
application of such scale has so far been hindered by the the
amount of time required to (manually) apply it in clinical
settings. Our dataset includes 3,205 fluorescein angiography images
from 148 patients and 242 eyes from the uveitis department of Jules
Gonin Eye Hospital. The data was automatically extracted from a
medical device, in hospital settings. Images were graded by a
medical expert. We focused specifically on one type of
inflammation, namely retinal vasculitis. Our pipeline comprises
both learning-based models (Pasa model with F1 score = 0.81, AUC =
0.86), and an intensity-based approach to serve as a baseline (F1
score = 0.57, AUC = 0.66). A recall of up to 0.833 computed in an
independent test set is comparable to the scores obtained by
available state-of-the-art approaches. Here we present the first
fully automated pipeline for the grading of retinal vasculitis from
raw medical images that is applicable to a real-world clinical data.},
}
@article{ijtld-2023,
title = {The rise of artificial intelligence reading of chest
X-rays for enhanced TB diagnosis and elimination},
doi = {10.5588/ijtld.22.0687},
abstract = {We provide an overview of the latest evidence on
computer-aided detection (CAD) software for automated
interpretation of chest radiographs (CXRs) for TB detection. CAD is
a useful tool that can assist in rapid and consistent CXR
interpretation for TB. CAD can achieve high sensitivity TB
detection among people seeking care with symptoms of TB and in
population-based screening, has accuracy on-par with human readers.
However, implementation challenges remain. Due to diagnostic
heterogeneity between settings and sub-populations, users need to
select threshold scores rather than use pre-specified ones, but
some sites may lack the resources and data to do so. Efficient
standardisation is further complicated by frequent updates and new
CAD versions, which also challenges implementation and comparison.
CAD has not been validated for TB diagnosis in children and its
accuracy for identifying non-TB abnormalities remains to be
evaluated. A number of economic and political issues also remain to
be addressed through regulation for CAD to avoid furthering health
inequities. Although CAD-based CXR analysis has proven remarkably
accurate for TB detection in adults, the above issues need to be
addressed to ensure that the technology meets the needs of
high-burden settings and vulnerable sub-populations.},
journal = {INT J TUBERC LUNG DIS},
volume = {27},
journaltitle = {International Journal of Tuberculosis and Lung Diseases},
author = {Geric, C. AND Qin, Z. Z. AND Denkinger, C. M. AND Kik, S.
V. AND Marais, B. AND Anjos, André AND David, P.-M. AND Khan, F. A.
AND Trajman, A.},
year = {2023},
month = 5,
keywords = {computer-aided detection; chest radiology; pulmonary
disease; tuberculosis; AI technology},
}
%%%%%%%%%%%%
%% 2022 %%
%%%%%%%%%%%%
@article{elsevier-csal-2022,
title = {Towards lifelong human assisted speaker diarization},
issn = {0885-2308},
doi = {10.1016/j.csl.2022.101437},
abstract = {This paper introduces the resources necessary to
develop and evaluate human assisted lifelong learning speaker
diarization systems. It describes the {ALLIES} corpus and
associated protocols, especially designed for diarization of a
collection audio recordings across time. This dataset is compared
to existing corpora and the performances of three baseline systems,
based on x-vectors, i-vectors and {VBxHMM}, are reported for
reference. Those systems are then extended to include an active
correction process that efficiently guides a human annotator to
improve the automatically generated hypotheses. An open-source
simulated human expert is provided to ensure reproducibility of the
human assisted correction process and its fair evaluation. An
exhaustive evaluation, of the human assisted correction shows the
high potential of this approach. The {ALLIES} corpus, a baseline
system including the active correction module and all evaluation
tools are made freely available to the scientific community.},
journal = {Computer Speech \& Language},
journaltitle = {Computer Speech \& Language},
author = {Shamsi, Meysam and Larcher, Anthony and Barrault, Loic
and Meignier, Sylvain and Prokopalo, Yevheni and Tahon, Marie and
Mehrish, Ambuj and Petitrenaud, Simon and Galibert, Olivier and
Gaist, Samuel and Anjos, André and Marcel, Sebastien and
Costa-jussà, Marta R.},
year = {2022},
month = 7,
date = {2022-07-27},
pdf = {https://www.idiap.ch/~aanjos/papers/elsevier-csal-2022.pdf},
keywords = {Evaluation, Human assisted learning, Lifelong learning,
Speaker diarization},
}
@inproceedings{union-2022,
author = {Raposo, Geoffrey and Trajman, Anete and Anjos, Andr{\'{e}}},
month = 11,
title = {Pulmonary Tuberculosis Screening from Radiological Signs
on Chest X-Ray Images Using Deep Models},
booktitle = {Union World Conference on Lung Health},
year = {2022},
addendum = {(Issued from master thesis supervision)},
date = {2022-11-01},
organization = {The Union},
abstract = {Background: The World Health Organization has recently
recommended the use of computer-aided detection (CAD) systems for
screening pulmonary tuberculosis (PT) in Chest X-Ray images.
Previous CAD models are based on direct image to probability
detection techniques - and do not generalize well (from training
to validation databases). We propose a method that overcomes
these limitations by using radiological signs as intermediary
proxies for PT detection.
Design/Methods: We developed a multi-class deep learning model,
mapping images to 14 radiological signs such as cavities,
infiltration, nodules, and fibrosis, using the National Institute
of Health (NIH) CXR14 dataset, which contains 112,120 images.
Using three public PTB datasets (Montgomery County - MC, Shenzen
- CH, and Indian - IN), summing up 955 images, we developed a
second model mapping F probabilities to PTB diagnosis (binary
labels). We evaluated this approach for its generalization
capabilities against direct models, learnt directly from PTB
training data or by transfer learning via cross-folding and
cross-database experiments. The area under the specificity vs.
sensitivity curve (AUC) considering all folds was used to
summarize the performance of each approach.
Results: The AUC for intra-dataset tests baseline direct
detection deep models achieved 0.95 (MC), 0.95 (CH) and 0.91
(IN), with up to 35\% performance drop on a cross-dataset
evaluation scenario. Our proposed approach achieved AUC of 0.97
(MC), 0.90 (CH), and 0.93 (IN), with at most 11\% performance
drop on a cross-dataset evaluation (Table/figures). In most
tests, the difference was less than 5\%.
Conclusions: A two-step CAD model based on radiological signs
offers an adequate base for the development of PT screening systems
and is more generalizable than a direct model. Unlike commercially
available CADS, our model is completely reproducible and available
open source at https://pypi.org/project/bob.med.tb/.}
}
@article{nsr-2022,
title = {State-of-the-art retinal vessel segmentation with
minimalistic models},
volume = {12},
rights = {2022 The Author(s)},
issn = {2045-2322},
url = {https://www.nature.com/articles/s41598-022-09675-y},
pdf = {https://www.nature.com/articles/s41598-022-09675-y.pdf},
doi = {10.1038/s41598-022-09675-y},
abstract = {The segmentation of retinal vasculature from eye fundus
images is a fundamental task in retinal image analysis. Over recent
years, increasingly complex approaches based on sophisticated
Convolutional Neural Network architectures have been pushing
performance on well-established benchmark datasets. In this paper,
we take a step back and analyze the real need of such complexity.
We first compile and review the performance of 20 different
techniques on some popular databases, and we demonstrate that a
minimalistic version of a standard U-Net with several orders of
magnitude less parameters, carefully trained and rigorously
evaluated, closely approximates the performance of current best
techniques. We then show that a cascaded extension (W-Net) reaches
outstanding performance on several popular datasets, still using
orders of magnitude less learnable weights than any previously
published work. Furthermore, we provide the most comprehensive
cross-dataset performance analysis to date, involving up to 10
different databases. Our analysis demonstrates that the retinal
vessel segmentation is far from solved when considering test images
that differ substantially from the training data, and that this
task represents an ideal scenario for the exploration of domain
adaptation techniques. In this context, we experiment with a simple
self-labeling strategy that enables moderate enhancement of
cross-dataset performance, indicating that there is still much room
for improvement in this area. Finally, we test our approach on
Artery/Vein and vessel segmentation from {OCTA} imaging problems,
where we again achieve results well-aligned with the
state-of-the-art, at a fraction of the model complexity available
in recent literature. Code to reproduce the results in this paper
is released.},
addendum = {(Issued from internship supervision)},
pages = {6174},
number = {1},
journal = {Nature Scientific Reports},
journaltitle = {Scientific Reports},
shortjournal = {Sci Rep},
author = {Galdran, Adrian and Anjos, André and Dolz, José and
Chakor, Hadi and Lombaert, Hervé and Ayed, Ismail Ben},
year = {2022},
month = 4,
date = {2022-04-13},
langid = {english},
note = {Number: 1
Publisher: Nature Publishing Group},
keywords = {Biomedical engineering, Computer science, Machine learning},
}
%%%%%%%%%%%%
%% 2021 %%
%%%%%%%%%%%%
@inproceedings{cbic-2021,
title = {Development of a lung segmentation algorithm for analog
imaged chest X-Ray: preliminary results},
url = {https://sbic.org.br/eventos/cbic_2021/cbic2021-123/},
pdf = {https://www.idiap.ch/~aanjos/papers/cbic-2021.pdf},
doi = {10.21528/CBIC2021-123},
shorttitle = {Development of a lung segmentation algorithm for
analog imaged chest X-Ray},
addendum = {(Issued from internship supervision)},
abstract = {Analog X-Ray radiography is still used in many
underdeveloped regions around the world. To allow these
populations to benefit from advances in automatic computer-aided
detection (CAD) systems, X-Ray films must be digitized.
Unfortunately, this procedure may introduce imaging artefacts,
which may severely impair the performance of such systems.
This work investigates the impact digitized images may cause to
deep neural networks trained for lung (semantic) segmentation on
digital x-ray samples. While three public datasets for lung
segmentation evaluation exist for digital samples, none are
available for digitized data. To this end, a U-Net-style
architecture was trained on publicly available data, and used to
predict lung segmentation on a newly annotated set of digitized images.
Using typical performance metrics such as the area under the
precision-recall curve (AUPRC), our results show that the model
is capable to identify lung regions at digital X-Rays with a high
intra-dataset (AUPRC: 0.99), and cross-dataset (AUPRC: 0.99)
efficiency on unseen test data. When challenged against
digitized data, the performance is substantially degraded (AUPRC: 0.90).
Our analysis also suggests that typical performance markers,
maximum F1 score and AUPRC, seems not to be informative to
characterize segmentation problems in test images. For this goal
pixels does not have independence due to natural connectivity of
lungs in images, this implies that a lung pixel tends to be
surrounded by other lung pixels.
This work is reproducible. Source code, evaluation protocols and
baseline results are available at: https://pypi.org/project/bob.ip.binseg/.},
eventtitle = {Congresso Brasileiro de Inteligência Computacional},
pages = {1--8},
booktitle = {Anais do 15. Congresso Brasileiro de Inteligência Computacional},
year = {2021},
month = 10,
publisher = {{SBIC}},
author = {Renzo, Matheus A. and Fernandez, Natália and Baceti,
André A. and Moura Junior, Natanael Nunes and Anjos, André},
}
%%%%%%%%%%%%
%% 2020 %%
%%%%%%%%%%%%
@misc{arxiv-2020,
title = {The Little W-Net That Could: State-of-the-Art Retinal
Vessel Segmentation with Minimalistic Models},
author = {Galdran, Adrian and Anjos, André and Dolz, José and
Chakor, Hadi and Lombaert, Hervé and Ayed, Ismail Ben},
year = {2020},
month = 9,
doi = {10.48550/arXiv.2009.01907},
eprinttype = {arxiv},
eprint = {2009.01907},
archivePrefix = {arXiv},
primaryClass = {cs.CV},
journaltitle = {{arXiv}:2009.01907 [cs, eess] (submitted to Nature
Scientific Reports)},
url = {https://arxiv.org/abs/2009.01907},
pdf = {https://arxiv.org/pdf/2009.01907},
abstract = {The segmentation of the retinal vasculature from eye
fundus images represents one of the most fundamental tasks in
retinal image analysis. Over recent years, increasingly complex
approaches based on sophisticated Convolutional Neural Network
architectures have been slowly pushing performance on
well-established benchmark datasets. In this paper, we take a step
back and analyze the real need of such complexity. Specifically, we
demonstrate that a minimalistic version of a standard UNet with
several orders of magnitude less parameters, carefully trained and
rigorously evaluated, closely approximates the performance of
current best techniques. In addition, we propose a simple
extension, dubbed W-Net, which reaches outstanding performance on
several popular datasets, still using orders of magnitude less
learnable weights than any previously published approach.
Furthermore, we provide the most comprehensive cross-dataset
performance analysis to date, involving up to 10 different
databases. Our analysis demonstrates that the retinal vessel
segmentation problem is far from solved when considering test
images that differ substantially from the training data, and that
this task represents an ideal scenario for the exploration of
domain adaptation techniques. In this context, we experiment with a
simple self-labeling strategy that allows us to moderately enhance
cross-dataset performance, indicating that there is still much room
for improvement in this area. Finally, we also test our approach on
the Artery/Vein segmentation problem, where we again achieve
results well-aligned with the state-of-the-art, at a fraction of
the model complexity in recent literature. All the code to
reproduce the results in this paper is released.},
}
@article{compbiomed-2020,
title = {Competitive neural layer-based method to identify people
with high risk for diabetic foot},
volume = {120},
url = {https://www.sciencedirect.com/science/article/pii/S0010482520301244},
pdf = {https://www.idiap.ch/~aanjos/papers/compbiomed-2020.pdf},
doi = {10.1016/j.compbiomed.2020.103744},
abstract = {Background and objective: To automatically identify
patients with diabetes mellitus (DM) who have high risk of
developing diabetic foot, via an unsupervised machine learning
technique. Methods: We collected a new database containing 54
known risk factors from 250 patients diagnosed with diabetes
mellitus. The database also contained a separate validation cohort
composed of 73 subjects, where the perceived risk was annotated by
expert nurses. A competitive neuron layer-based method was used to
automatically split training data into two risk groups. Results:
We found that one of the groups was composed of patients with
higher risk of developing diabetic foot. The dominant variables
that described group membership via our method agreed with the
findings from other studies, and indicated a greater risk for
developing such a condition. Our method was validated on the
available test data, reaching 71\% sensitivity, 100\% specificity,
and 90\% accuracy. Conclusions Unsupervised learning may be
deployed to screen patients with diabetes mellitus, pointing out
high-risk individuals who require priority follow-up in the
prevention of diabetic foot with very high accuracy. The proposed
method is automatic and does not require clinical examinations to
perform risk assessment, being solely based on the information of a
questionnaire answered by patients. Our study found that
discriminant variables for predicting risk group membership are
highly correlated with expert opinion.},
journal = {Computers in Biology and Medicine},
author = {Ferreira, Ana Cl\'audia Barbosa Hon\'orio and Ferreira,
Danton Diego and Oliveira, Henrique Ceretta and Resende, Igor
Carvalho de and Anjos, Andr\'e and Lopes, Maria Helena Baena de Moraes},
month = 5,
year = {2020},
keywords = {Artificial neural network, Diabetes mellitus, Diabetic foot},
}
%%%%%%%%%%%%
%% 2019 %%
%%%%%%%%%%%%
@misc{arxiv-2019,
title = {On the Evaluation and Real-World Usage Scenarios
of Deep Vessel Segmentation for Retinography},
author = {Tim Laibacher and Andr\'e Anjos},
addendum = {(Issued from intership supervision)},
year = {2019},
month = 9,
eprint = {1909.03856},
archivePrefix = {arXiv},
primaryClass = {cs.CV},
doi = {10.48550/arXiv.1909.03856},
url = {https://arxiv.org/abs/1909.03856},
pdf = {https://arxiv.org/pdf/1909.03856},
journaltitle = {{arXiv}:1909.03856 [cs] (submitted to IEEE
International Symposium on Biomedical Imaging 2021)},
abstract = {We identify and address three research gaps in the
field of vessel segmentation for retinography. The first focuses
on the task of inference on high-resolution fundus images for which
only a limited set of ground-truth data is publicly available.
Notably, we highlight that simple rescaling and padding or cropping
of lower resolution datasets is surprisingly effective. We further
explore the effectiveness of semi-supervised learning for better
domain adaptation in this context. Our results show competitive
performance on a set of common public retina datasets, using a
small and light-weight neural network. For HRF, the only very
high-resolution dataset currently available, we reach comparable,
if not superior, state-of-the-art performance by solely relying on
training images from lower-resolution datasets. The second topic
we address concerns the lack of standardisation in evaluation
metrics. We investigate the variability of the F1-score on the
existing datasets and report results for recently published
architectures. Our evaluation show that most reported results are
actually comparable to each other in performance. Finally, we
address the issue of reproducibility, by open-sourcing the complete
framework used to produce results shown here.},
}
@patent{3dfv-patent-2019,
author = {Sonna Momo, Lambert and Cerqueira Torres, Luciano and
Marcel, S\'ebastien and Anjos, Andr\'e and Liebling, Michael and
Shajkofci, Adrian and Amoos, Serge and Woeffray, Alain and Sierro,
Alexandre and Roduit, Pierre and Ferrez, Pierre and Bonvin, Lucas},
title = {Method and Device for Biometric Vascular Recognition
and/or Identification},
year = {2019},
month = 8,
day = 8,
number = {WO/2019/150254},
type = {Patent},
filing_num = {PCT/IB2019/050708},
yearfiled = {2019},
monthfiled = 1,
dayfiled = 29,
pat_refs = {P\&TS SA (AG, LTD.); Av. J.-J. Rousseau 4 P.O. Box 2848
2001 Neuchâtel, CH},
abstract = {The invention concerns a method and a biometric
acquisition device for biometric vascular recognition and/or
identification. The method comprising a step of capturing a
plurality of veins images (116, 117, 118) of supposed subcutaneous
veins (21) of a same inspecting portion (20) of a presented entity
(2) from various converging orientations (113, 114, 115). The
method further comprises a step of determine if said entity is a
spoof based on estimated likelihood that said supposed subcutaneous
veins within said plurality of veins images (116, 117, 118) are
likely projections of solid veins (120).},
url = {https://patentscope.wipo.int/search/en/detail.jsf?docId=WO2019150254}
}
@article{tifs-2019-2,
author = {George, Anjith and Mostaani, Zohreh and Geissenbuhler,
David and Nikisins, Olegs and Anjos, Andr{\'{e}} and Marcel, S{\'{e}}bastien},
title = {Biometric Face Presentation Attack Detection with
Multi-Channel Convolutional Neural Network},
journal = {IEEE Transactions on Information Forensics and Security},
month = 5,
year = {2019},
doi = {10.1109/TIFS.2019.2916652},
pdf = {https://www.idiap.ch/~aanjos/papers/tifs-2019-2.pdf},
abstract = {Face recognition is a mainstream biometric
authentication method. However, vulnerability to presentation
attacks (a.k.a spoofing) limits its usability in unsupervised
applications. Even though there are many methods available for
tackling presentation attacks (PA), most of them fail to detect
sophisticated attacks such as silicone masks.
As the quality of presentation attack instruments improves over
time, achieving reliable PA detection with visual spectra alone
remains very challenging. We argue that analysis in multiple
channels might help to address this issue. In this context, we
propose a multi-channel Convolutional Neural Network based
approach for presentation attack detection (PAD).
We also introduce the new Wide Multi-Channel presentation Attack
(WMCA) database for face PAD which contains a wide variety of 2D
and 3D presentation attacks for both impersonation and obfuscation
attacks. Data from different channels such as color, depth,
near-infrared and thermal are available to advance the research in
face PAD. The proposed method was compared with feature-based
approaches and found to outperform the baselines achieving an ACER
of 0.3\% on the introduced dataset. The database and the software
to reproduce the results are made available publicly.},
}
@incollection{hopad-2019,
title = {Recent Advances in Face Presentation Attack Detection},
author = {Bhattacharjee, Sushil and Mohammadi, Amir and Anjos,
Andr{\'{e}} and Marcel, S{\'{e}}bastien},
editor = "Marcel, S{\'{e}}bastien AND Nixon, Mark AND Fierrez,
Julian AND Evans, Nicholas",
edition = "2nd edition (in press)",
booktitle = "Handbook of Biometric Anti-Spoofing",
publisher = "Springer-Verlag",
year = "2019",
month = 1,
pages = "207--228",
isbn = "ISBN 978-3-319-92627-8",
doi = "10.1007/978-3-319-92627-8_10",
abstract = "The undeniable convenience of face-recognition (FR)
based biomet- rics has made it an attractive tool for access
control in various applications, from immigration-control to remote
banking. Widespread adoption of face biometrics, however, depends
on the how secure such systems are perceived to be. One particular
vulnerability of FR systems comes from presentation attacks (PA),
where a subject A attempts to impersonate another subject B, by
presenting, for example, a photograph of B to the biometric sensor
(i.e., the camera). PAs are the most likely forms of attacks on
face biometric systems, as the camera is the only component of the
biometric system that is exposed to the outside world. Robust
presentation attack detection (PAD) methods are necessary to
construct secure FR based access control systems. The first edition
of the Handbook of Biometric Anti-spoofing included two chapters on
face-PAD. In this chapter we present the significant advances in
face-PAD research since the publication of the first edition of
this book. In addition to PAD methods designed to work with color
images, we also discuss advances in face-PAD methods using other
imaging modalities, namely, near-infrared (NIR) and thermal
imaging. This chapter also presents a number of recently published
datasets for face-PAD experiments.",
}
@incollection{hopad-2019-2,
title = "Evaluation Methodologies for Biometric Presentation Attack
Detection",
author = {Chingovska, Ivana and Mohammadi, Amir and Anjos,
Andr{\'{e}} and Marcel, S{\'{e}}bastien},
editor = "Marcel, S{\'{e}}bastien AND Nixon, Mark AND Fierrez,
Julian AND Evans, Nicholas",
addendum = {(Issued from Ph.D co-supervision)},
edition = "2nd edition (in press)",
booktitle = "Handbook of Biometric Anti-Spoofing",
publisher = "Springer-Verlag",
year = "2019",
month = 1,
pages = "457--480",
isbn = "ISBN 978-3-319-92627-8",
doi = "10.1007/978-3-319-92627-8_20",
abstract = "Presentation attack detection (PAD, also known as