Constrained Design of Deep Iris Networks
Constrained Design of Deep Iris Networks
fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
IEEE TRANSACTIONS ON IMAGE PROCESSING 1
Abstract—Despite the promise of recent deep neural networks to provide more accurate and efficient iris recognition compared to
traditional techniques, there are vital properties of the classic IrisCode which are almost unable to be achieved with current deep iris
networks: the compactness of model and the small number of computing operations (FLOPs). This paper casts the iris network design
process as a constrained optimization problem which takes model size and computation into account as learning criteria. On one hand,
this allows us to fully automate the network design process to search for the optimal iris network architecture with the highest
recognition accuracy confined to the computation and model compactness constraints. On the other hand, it allows us to investigate
the optimality of the classic IrisCode and recent deep iris networks. It also enables us to learn an optimal iris network and demonstrate
state-of-the-art performance with less computation and memory requirements.
Index Terms—Iris Recognition, Deep Learning, Iris Network Design, Constrained Deep Network Design.
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
2 IEEE TRANSACTIONS ON IMAGE PROCESSING
stored and processed on a high-computing platform, then aiming to improve recognition performance of the iris bio-
there doesn’t need to be restrictions on model compactness metric system. There are two lines of deep iris networks in
or matching speed, and the design process can target the the literature: softmax-based and pairwise-based.
highest accuracy. In contrast, if the iris recognition system The softmax-based networks train images on a fixed
is performed on embedded or mobile systems with limited number of classes and seek to match a test image with those
resources, model compactness and matching speed have to classes. The typical examples of these are DeepIrisNet [11],
be prioritized. [12] and Off-the-shelf Iris CNNs [23]. DeepIrisNets stack
There are two main objectives of this paper: (1) Develop classic layers to create three networks versions. DeepIrisNet-
an algorithm to search for the optimal deep iris network that A has 8 convolutional layers, 4 pooling layers, 3 fully
satisfies both the pre-defined constraints in computation connected layers and 2 dropout layers. DeepIrisNet-B swaps
and model compactness and achieves the highest possible a number of late convolutional layers with Inception layers
performance; and (2) Use this algorithm to investigate the [35]. DeepIrisNet-2 adds spatial transformer layers [17] to
optimality of the classic IrisCode and recent iris networks. predict the coefficients of affine transformations within the
To achieve these objectives, we re-visit existing NAS ap- iris images. Training on these big networks requires large
proaches, and re-interpret them with additional constraints datasets, hence Off-the-shelf Iris CNNs apply transfer learn-
to model the design as a constrained optimization problem. ing. State of the art Off-the-shelf CNNs (AlexNet, VGG,
Solving this constrained optimization problem allows us to Inception, ResNet, DenseNet) pre-trained on the large-scale
automatically discover the optimal network (the network ImageNet dataset [31] are further fine-tuned to achieve
which achieves the highest achievable accuracy within the state of the art iris recognition accuracy [23]. The main
constraints of the computation and model compactness). requirement of the softmax-based networks is the test image
In addition, applying this algorithm with two constraints has to belong to one of classes on the training set, which
similar to those of the existing approaches, we are able means the networks will have to be re-trained whenever a
to: (i) understand the optimality of the classic handcrafted new class is added.
IrisCode and recent deep iris networks; and (ii) optimize The pairwise-based networks deal with this drawback
an existing network architecture to achieve the highest by learning networks to measure the similarity or dissimi-
accuracy under the same computational and memory cost. larity between two images, without knowing their classes.
Our major contributions can be highlighted as follows: The typical examples of these are DeepIris [20] and FeatNet
• Our work is the first effort to consider automatic ar- [38]. DeepIris applies a pairwise filter on the two input iris
chitecture search for deep iris recognition networks. images, followed by a stack of classic convolutional, pooling
• We re-interpret the search for network architecture and fully connected layers. DeepIris outputs a similarity
and parameters as a constrained optimization with score between two input images [20]. FeatNet uses a fully
design constraints related to the compactness of the convolutional architecture by replacing all fully connected
model along with the matching speed. layers with 1 × 1 convolutional layers to retain spatial corre-
• Our approach results in the full automation of deep spondence between two input images. The classic pairwise
iris network design, in both feature learning and loss [33] is extended to incorporate the translation and mask,
architecture learning. leading to an accurate similarity estimation between the
• We discover an optimal deep network for iris recog- input couple [38].
nition achieving state-of-the-art performance with All these deep iris networks have their architectures
less computation and memory required compared to handcrafted, which means potential redundant connections,
existing deep iris networks. filters and layers, and no optimal accuracy guarantee. Learn-
• We provide a way to understand the optimality of ing both a feature representation and an optimal architecture
the classic IrisCode approach. is challenging considering the huge number of potential
architectures and their parameters. This challenge will be
The remainder of the paper is organized as follows: Sec- addressed in this paper.
tion 2 discusses recent deep iris recognition networks along
with feature learning and architecture learning approaches;
Section 3 presents how we model the network design as a bi- 2.2 Neural Architecture Search
level constrained optimization problem; Section 4 illustrates There does not exist any work in network architecture
our experimental results; and the paper is concluded in search in the iris recognition setting. In general deep learn-
Section 5. ing, modern architecture search approaches usually rely
on Evolution Theory and Reinforcement Learning (RL) to
2 R ELATED WORK design searching policies. For example, Real et al. brought
across ideas from the natural evolution process to gradually
This section reviews the existing attempts in the literature
update the architecture of the network to achieve higher
on applying deep learning approaches for iris recognition
accuracies [29], [30]. At each iteration, a number of archi-
and architecture learning via neural architecture search.
tectures, called the population, are investigated. The best
architecture is mutated (i.e. randomly add/remove layers)
2.1 Deep iris recognition networks to generate a new child architecture which is then added
A number of deep networks have been proposed to take into the population. Subsequently, the worst architecture
advantage of their automatic feature engineering capability [30] or the oldest architecture [29] is discarded to generate a
to automatically learn feature representation for iris images, new population. This algorithm is considered as a discrete
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
NGUYEN et al.: CONSTRAINED DESIGN OF DEEP IRIS NETWORKS 3
Fig. 1. Re-interpreting layer connections. Traditionally, two layers, i and j , are connected by one operation selected from the operation set, O =
{o1 , ..., oN }. This means activating only operation, ok , in the operation set with a coefficient 1, while disabling all others with coefficients of 0.
(i,j)
The coefficients only take a discrete value {0, 1}. We relax this to allow the coefficients αk to take a continuous value between [0, 1]. This
re-interpretation enables us to make the architecture search space continuous.
optimization process. In contrast, Zoph et al. trained a RNN node is computed as a summary of network operations
controller to iteratively sample candidate architectures, and applied on its predecessors,
trains them to convergence to measure their performance on X
the desired task. The controller then uses the performance xj = o(i,j) (xi ), (1)
i<j
as a guiding signal to find more promising architectures
[39], [40]. Parameter sharing can be forced on all child where o (i,j)
is the candidate network operation between
models to improve the search speed at a slight cost of node i and node j . The candidate operation o(i,j) is defined
performance [24]. Both evolution and RL approaches are ex- by two properties: its operation type, which is one of the
tremely computationally expensive despite their remarkable types in the operation set O; and the parameters of the
performance. For example, the network architecture search chosen operation type. The operation set can include a
for CIFAR-10 needs 1800 GPU days of RL [40] or 3150 GPU convolutional operator, conv ; a pooling operator, pool; a
days of evolution [29]. skip connection operator, Identity ; and a no connection
One of our main motivations arises from a recent body indicator, zero. Each convolutional operator is followed
of works on gradient-based architecture search [14], [19] by a batch norm operation by default. The size of the
where the problem is encoded as a bi-level optimization convolutional kernel can vary, e.g. 3 × 3 and 3 × 5. There
problem with one level to optimize based on the architecture are two types of convolution: traditional convolution and
and the other level to optimize based on the weights of its dilated version [37]. The Identity operation will function
the chosen architecture. This strategy is able to discover as a skip connection similar to the ResNet architecture [15],
high-performing architectures achieving high classification which adds the original signal from the input feature map
accuracy in only tens of GPU hours. This strategy is highly to the output feature map. The Zero operation will model
desired in the iris recognition scenario to search for the the lack of connection between two feature maps.
highest-accuracy architecture. However, it is not directly With this notation, the network design is interpreted
applicable to the iris recognition scenario because in contrast as two tasks: (1) searching for a set of network operations
to the natural image recognition scenario, the approaches {o(i,j) : j : 1..L and i : 1..j }; and (2) searching for the
in the iris recognition scenario need to take into account operation weights (a.k.a parameter values) to achieve the
design constraints as discussed earlier to be applicable in highest performance network in the iris recognition setting.
real-world applications. In this paper, we introduce a new For each network, we consider two constraints:
design procedure to address these challenges.
It is also worth noting that two related tasks: hyperpa- • Compactness of the model: number of parameters:
rameter optimization and network simplification/pruning P , which is directly related to memory required: MP ,
are simpler tasks considering the dimension of the parame- • Computation: number of FLOPS: K .
ters compared with the whole architecture. In addition, a Our aim is to design a network that achieves the highest
recent body of work on resources-aware network design accuracy possible conditioned on these constraints.
such as MorphNet [13] and NetAdapt [36] have considered
resource constraints in the design process. However, these 3.1 Re-interpreting layer connections
techniques work on simplifying a pre-trained model to
match the resource constraints, which is much simpler than Considering o(i,j) is discrete, as shown in the literature,
an architecture search from scratch as being addressed in searching in the discrete space is extremely computational
this paper. heavy and may result in missing the optimal point [14],
[22]. We employ one adjustment to make the search space
(i,j)
continuous by assigning a coefficient αo for each candi-
3 D ESIGNING CONSTRAINTS date operation in the operation set O. Other than activating
A network architecture, α, is defined as a directed acyclic a single operation while all others are disabled between
graph consisting of an ordered sequence of L nodes [19]. two nodes (i, j), we activate all candidate operations in the
Each node, xj , in the graph represents a layer in the network operation set but only one operation is strongly encouraged
architecture. The first node is the input node, which is the with a high coefficient value while others are strongly
input image. The final node is the output. Each intermediate discouraged with small coefficient values. We call this a
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
4 IEEE TRANSACTIONS ON IMAGE PROCESSING
Fig. 2. One search architecture example: with connection bundles, the architecture search becomes an optimization task to find the best set
(i,j)
{αo }, i, j = 1..L, i < j to minimize the loss.
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
NGUYEN et al.: CONSTRAINED DESIGN OF DEEP IRIS NETWORKS 5
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
6 IEEE TRANSACTIONS ON IMAGE PROCESSING
TABLE 1
Statistics of three datasets, ND-CrossSensor-2013, CASIA-Iris-Thousand and UBIRIS.v2 in this research.
one outer circular boundary. These two circles are detected recognition [20], [38]. Hence we also split the data
by an integro-differential operator [5] as, into subject-disjoint training and testing subsets.
∂ I(x, y)
I
In summary, there are two data splitting schemes in this
maxr,x0 ,y0 Gσ (r) ∗ ds , (7)
∂r r,x0 ,y0 2πr work: (1) sample-disjoint splitting and (2) subject-disjoint
splitting. Depending on the task to be performed, our exper-
where I(x, y) denotes the input image and and Gσ refers
iments will apply one of these two schemes. For the sample-
to a Gaussian with a standard deviation σ . This integro-
disjoint scheme, we split images of each subject into 70% of
differential operator detects circles by iteratively searching
the images for training, 10% of the images for validation
the circular arc ds with a radius r centered at the location
and 20% of the images for testing. For the subject-disjoint
(x0 , y0 ). A mask is also added to eliminate the occlusion
scheme, we split 70% of the subjects for training, 10% of the
impact of eyelids and eyelash.
subjects for validation and 20% of the subjects for testing.
We subsequently normalize the segmented iris region
Intra-dataset performance We perform the intra-dataset
to a fixed size rectangle by a rubber-sheet model [5] by
experiment on the ND-CrossSensor-Iris-2013 dataset, not
re-mapping the segmented iris image IS (x, y), from the
the other two datasets, due to its large size suitable for train-
raw Cartesian coordinates (x, y) to the dimensionless polar
ing. The training subset is to train a network architecture to
coordinates (r, θ) as,
find the best weights. The validation subset is used to find
x(r, θ) = (1 − r)xp (θ) + rxs (θ), (8) the best network architecture. The testing subset is used to
y(r, θ) = (1 − r)yp (θ) + rys (θ), (9) report the intra-dataset performance.
The classic IrisCode is used as a handcrafted baseline.
where r and θ are the radius and the angle in the range Our implementation achieves a F RR = 3.76% at F AR =
of [0, 1] and [0, 2π] respectively. The normalization step 0.1% and an EER = 1.75% on the ND-CrossSensor-2013
helps to reduce the rotations of the eye (e.g., due to the dataset, which is comparable to the state of the art imple-
head movement), to simple translation during matching. mentation [38].
We choose an open-source software, USIT v2.2, from the Cross-dataset performance The best network learned is
University of Salzburg [28] for the pre-processing phase and further investigated for generalization capability through
generate normalized images with a fixed size of 64 × 512 training in one dataset and testing on others. The best
pixels. Some manual corrections are subsequently made by network discovered in the ND-CrossSensor-Iris-2013 dataset
removing a small portion of wrongly segmented images. is tested on the other two datasets, CASIA-Iris-Thousand
and UBIRIS.v2 to understand its generalizability. We do not
4.2 Performance metrics perform network search on the CASIA and UBIRIS datasets
To report the performance, we rely on False Rejection Rate as their small number of images will restrict the search
(FRR) and Equal Error Rate (EER). In this work, FRRs at space.
False Acceptance Rate (FAR) = 0.1% are experimented and
reported due to its popular adoption in the field. 4.3 Case 1: Handcrafted - IrisCode
State-of-the-art iris networks employ two types of losses:
Firstly, we are interested to see how well the deep networks
(1) cross-entropy loss [11], [23] and (2) pairwise loss [20],
perform if they have to limit their computation akin to the
[38].
one in the classic handcrafted IrisCode [6]. We impose one
• Cross-entropy loss: the important property of the computation constraint, i.e. the maximum number of FLOPs,
cross-entropy loss is using the same identities in the to be akin to the one in the classic IrisCode, i.e. 0.9M ,
training and testing datasets, which means sample- and investigate the best accuracy a deep network could
disjoint but not subject-disjoint. This is shown in the achieve compared to the accuracy of the IrisCode. We run
softmax classifier of [11] and the SVM classifier of the constrained design algorithm to find the best network
[23]. Hence to be comparable with the state of the art, architecture yielding the highest accuracy or smallest EER
we first divide our dataset into the sample-disjoint conditioned on the IrisCode computation.
but not subject-disjoint training and testing subsets. Operation set We apply popular operations, which are
• Pairwise loss: the pairwise loss measures the sim- widely used in the existing deep iris networks, in the oper-
ilarity or dissimilarity between two input images, ation set O: 3 × 3 and 3 × 5 convolution, 2 × 2 max pooling,
deciding whether they are from the same class or 2 × 2 average pooling, Identity and Zero. All operations
not. This loss allows us to have unseen subjects, i.e. are of stride 1.
not present in the training phase, in the test phase. Loss choice We apply the most popular softmax clas-
This loss has been shown to be effective for iris sifier with a classification cross-entropy loss. A majority of
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
NGUYEN et al.: CONSTRAINED DESIGN OF DEEP IRIS NETWORKS 7
Fig. 5. The architecture that achieves the same accuracy level with the handcrafted IrisCode. Notice it achieves the same performance level with 8
times additional computation.
Fig. 6. The optimal network architecture that is discovered by our constrained search algorithm.
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
8 IEEE TRANSACTIONS ON IMAGE PROCESSING
Algorithm
Inputs Outputs
hyperparameters
- O = {3 × 3, 3 × 5 conv ;
- α∗
K = KResN et18 2 × 2 max pool; 2 × 2
- w∗
P = PResN et18 avg pool; Identity ; Zero}
- EER∗
- Loss = cross-entropy
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
NGUYEN et al.: CONSTRAINED DESIGN OF DEEP IRIS NETWORKS 9
TABLE 7 R EFERENCES
Statistical significance of the performance.
[1] B. Bhanu and A. Kumar. Deep Learning for Biometrics. Springer,
2017.
Methods Avg EER [2] S. Boyd and L. Vandenberghe. Convex Optimization. Stanford
FeatNet [38] 1.79 × 10−2 University, 2004.
Ours 1.32 × 10−2 [3] S. Bubeck. Convex optimization: Algorithms and complexity.
p-value 2.61 × 10−3 Foundations and Trends in Machine Learning, 8(3):231–357, 2015.
[4] Chinese Academy of Sciences Institute of Automation. CASIA iris
image database, http://biometrics.idealtest.org/, Aug 2017.
TABLE 8 [5] J. Daugman. How iris recognition works? IEEE Transactions on
Cross-dataset performance of the network discovered. Circuits and Systems for Video Technology, 14:21 – 30, 2004.
[6] J. Daugman. New methods in iris recognition. IEEE Transactions
on Systems, Man and Cybernetics, 37:1167 – 75, 2007.
CASIA-Iris-1K UBIRISv2 [7] J. Daugman. Information theory and the iriscode. IEEE Transac-
FRR EER FRR EER tions on Information Forensics and Security, 11(2):400–409, Feb 2016.
IrisCode 5.53% 3.46% 14.31% 8.33% [8] J. Daugman. Information theory and the iriscode. IEEE Transac-
DeepIris [20] 4.25% 2.17% 13.20% 7.12% tions on Information Forensics and Security, 11:400–409, Feb 2016.
FeatNet [38] 3.98% 1.93% 13.93% 6.69% [9] J. Daugman and C. Downing. Searching for doppelgangers:
Ours 3.07% 1.54% 11.12% 5.98% assessing the universality of the iriscode impostors distribution.
IET Biometrics, 5:65–75, 2016.
[10] L. Franceschi, P. Frasconi, S. Salzo, R. Grazzi, and M. Pontil.
Bilevel programming for hyperparameter optimization and meta-
4.6 Generalizability of the model learning. In International Conference on Machine Learning (ICML),
2018.
Finally, we want to understand the generalization capability [11] A. Gangwar and A. Joshi. Deepirisnet: Deep iris representation
of the network architecture discovered by performing a with applications in iris recognition and cross-sensor iris recog-
nition. In IEEE International Conference on Image Processing (ICIP),
cross-dataset experiment on two smaller-size datasets, CA- pages 2301–2305, Sep 2016.
SIA [4] and UBIRIS [26]. While the CASIA dataset captured [12] A. K. Gangwar, A. Joshi, P. Joshi, and R. Raghavendra. Deepiris-
the NIR iris images using a different camera, the UBIRIS net2: Learning deep-iriscodes from scratch for segmentation-
robust visible wavelength and near infrared iris recognition. CoRR,
dataset captured the iris images with a visible camera. abs/1902.05390, 2019.
This demonstrates the wide range of imaging conditions to [13] Z. Gordon, E. Eban, O. Nachum, B. Chen, T.-J. Yang, and E. Choi.
test the generalizability. Two datasets CASIA and UBIRIS Morphnet: Fast & simple resource-constrained structure learning
of deep networks. In IEEE International Conference on Computer
are split into 20% for training and 80% for testing. Three Vision and Pattern Recognition (CVPR), 2018.
networks, i.e.: ours, DeepIris and FeatNet, trained as per [14] W. Grathwohl, E. Creager, S. K. S. Ghasemipour, and R. Zemel.
Section 4.5 are further fine-tuned using the training subset Gradient-based optimization of neural network architecture. In
International Conference on Learning Representation (ICLR), 2018.
and tested in the testing subset of the two datasets for the [15] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for
cross-dataset performance investigation. image recognition. In IEEE Conference on Computer Vision and
The performance is presented in Table 8. The network Pattern Recognition (CVPR), pages 770–778, Jun 2016.
[16] K. P. Hollingsworth, K. W. Bowyer, and P. J. Flynn. Improved
discovered by our architecture search algorithm outper- iris recognition through fusion of hamming distance and fragile
forms the state-of-the-art approaches in both CASIA and bit distance. IEEE Transactions on Pattern Analysis and Machine
UBIRIS datasets, illustrating a high level of generalization Intelligence (PAMI), 33(12):2465–2476, Dec 2011.
[17] M. Jaderberg, K. Simonyan, A. Zisserman, and k. kavukcuoglu.
across different sensors, different imaging distances and Spatial transformer networks. In C. Cortes, N. D. Lawrence, D. D.
different levels of subject cooperation. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural
Information Processing Systems (NIPS), pages 2017–2025. 2015.
[18] A. W. K. Kong, D. Zhang, and M. S. Kamel. An analysis of iriscode.
IEEE Transactions on Image Processing, 19(2):522–532, Feb 2010.
5 C ONCLUSIONS [19] H. Liu, K. Simonyan, and Y. Yang. Darts: Differentiable architec-
ture search. arXiv:1806.09055, 2018.
This paper proposes an algorithm to design a deep iris [20] N. Liu, M. Zhang, H. Li, Z. Sun, and T. Tan. Deepiris: Learning
recognition network with attention to computation and pairwise filter bank for heterogeneous iris verification. Pattern
memory constraints. By modeling the design process as a Recognition Letters, 82:154 – 161, 2016.
[21] M. Lopez, J. Daugman, and E. Canto. Hardware-software co-
bi-level constrained optimization approach, our algorithm design of an iris recognition algorithm. IET Information Security,
is able to search for the optimal network which achieves the 5:60–68, Mar 2011.
[22] R. Luo, F. Tian, T. Qin, E. Chen, and T.-Y. Liu. Neural architecture
best possible performance conditioned on the pre-defined optimization. In Advances on Neural Information Processing Systems
computation and model compactness constraints. This al- (NIPS), 2018.
gorithm enables us to investigate the effectiveness of the [23] K. Nguyen, C. Fookes, A. Ross, and S. Sridharan. Iris recognition
with off-the-shelf cnn features: A deep learning perspective. IEEE
classic handcrafted IrisCode compared with deep network
Access, 6:18848 – 18855, 2017.
counterparts. It also enables us to further improve the [24] H. Pham, M. Guan, B. Zoph, Q. Le, and J. Dean. Efficient
existing deep iris recognition networks to achieve similar neural architecture search via parameter sharing. In International
or better accuracy with the same level of computation Conference on Machine Learning (ICML), 2018.
[25] P. Phillips, W. Scruggs, A. O’Toole, P. Flynn, K. Bowyer, C. Schott,
and memory cost. The design algorithm also discovers an and M. Sharpe. Frvt 2006 and ice 2006 large-scale experimental re-
optimal network with competitive performance with less sults. IEEE Transactions on Pattern Analysis and Machine Intelligence,
computation and memory required than the state-of-the- 32(5):831–846, May 2010.
[26] H. Proenca, S. Filipe, R. Santos, J. Oliveira, and L. Alexandre. The
art approaches, in both intra-dataset and cross-dataset ex- UBIRIS.v2: A database of visible wavelength images captured on-
periments. More importantly, this algorithm simultaneously the-move and at-a-distance. IEEE Transactions on Pattern Analysis
achieves both automatic feature engineering and network and Machine Intelligence, 32(8):1529–1535, Aug 2010.
[27] H. Proena and J. C. Neves. Irina: Iris recognition (even) in
architecture engineering, opening us to full automation in inaccurately segmented data. In IEEE Conference on Computer
deep iris recognition network design. Vision and Pattern Recognition (CVPR), pages 6747–6756, Jul 2017.
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2020.2999211, IEEE
Transactions on Image Processing
10 IEEE TRANSACTIONS ON IMAGE PROCESSING
1057-7149 (c) 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Carleton University. Downloaded on June 28,2020 at 03:21:42 UTC from IEEE Xplore. Restrictions apply.