Computational Vision and Medical Image Processing
Computational Vision and Medical Image Processing
3D Vision;
Computational Bio- Imaging and Visualization;
Computational Vision and Medical Image Processing. VipIMAGE 2015 will be useful
to academics, researchers and professionals in Biomechanics, Biomedical Engineering,
Computational Vision (image processing and analysis), Computer Sciences, Computational
Mechanics, Signal Processing, Medicine and Rehabilitation.
Sponsored by
Editors
João Manuel R.S. Tavares & R.M. Natal Jorge
Faculdade de Engenharia, Universidade do Porto, Porto, Portugal
All rights reserved. No part of this publication or the information contained herein may be reproduced,
stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical,
by photocopying, recording or otherwise, without written prior permission from the publisher.
Although all care is taken to ensure integrity and the quality of this publication and the information
herein, no responsibility is assumed by the publishers nor the author for any damage to the property or
persons as a result of operation or use of this publication and/or the information contained herein.
Table of contents
Acknowledgements ix
Preface xi
Invited lectures xiii
Scientific committee xv
Invited lectures
An adaptive non-rigid image registration technique using hierarchical B-splines 3
A. Pawar, Y. Zhang, X. Wei, Y. Jia, T. Rabczuk, C.L. Chan & C. Anitescu
Medical image segmentation using Object Shape Models: A critical review on recent trends,
and alternative directions 9
A.X. Falcão, T.V. Spina, S.B. Martins & R. Phellan
Primal-dual method for continuous max-flow approaches 17
K. Wei, X.-C. Tai, T.F. Chan & S. Leung
Image restoration using variational approaches: Some recent advances 25
A. Lanza, S. Morigi & F. Sgallari
Virtual and augmented medical imaging environments: Applications to simulation, training,
surgical planning and interventional guidance 33
C.A. Linte
Contributed papers
The fast method of creating High Dynamic Range image for fluorescent microscopy
applications 41
A. Bal
Automatic cheek detection in digital images 49
M. Frackiewicz, H. Palus & K. Radlak
A variational model for image fusion with simultaneous cartoon and texture
decomposition 57
M. Dodangeh, I.N. Figueiredo & G. Gonçalves
Image contrast enhancement using split Bregman method 63
S.Gh. Bardeji, I.N. Figueiredo & E. Sousa
Extraction of spectral drowsy component from the resting electroencephalographic signal
for quick, objective and direct testing of sleepiness in absolute terms 69
A.A. Putilov, O.G. Donskaya & E.G. Verevkin
The illusion of a blackbody at the human ear and the human temperature measurement 75
A. Cardoso
Video-based Turkish Sign Language recognition systems 81
M. Aktaş & E.B. Sonmez
vi
vii
viii
Acknowledgements
The editors and the Conference co-chairs acknowledge the support towards the publication of the Book
of Proceedings and the organization of the V ECCOMAS Thematic Conference VipIMAGE to the
following organizations:
– Universidade do Porto (UP)
– Faculdade de Engenharia da Universidade do Porto (FEUP)
– Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial (INEGI)
– European Community on Computational Methods in Applied Sciences (ECCOMAS)
– International Association for Computational Mechanics (IACM)
– Fundação para a Ciência e a Tecnologia (FCT)
– Associação Portuguesa de Mecânica Teórica Aplicada e Computacional (APMTAC)
ix
Preface
This book contains invited lectures and full papers presented at VipIMAGE 2015—V ECCOMAS
Thematic Conference on Computational Vision and Medical Image Processing, which was held in Ten-
erife, Canary Islands, Spain, during the period 19–21 October 2015. The event had 6 invited lectures,
and 50 contributed presentations originated from 19 countries: Algeria, Brazil, China, Czech Republic,
Egypt, France, Germany, Hungary, Italy, Mexico, Norway, Poland, Portugal, Romania, Russia, Spain,
The United Kingdom, The United States of America and Turkey.
Computational methodologies of signal processing and analyses have been commonly used in our
society. For instances, full automatic or semi-automatic Computational Vision systems have been increas-
ing used in surveillance tasks, traffic analysis, recognition process, inspection purposes, human-machine
interfaces, 3D vision, deformation analysis and aided medical procedures.
One of the notable aspects of the Computational Vision domain is the inter- and multi-disciplinarily.
Actually, principles and methodologies of other sciences, such as Informatics, Mathematics, Statistics,
Psychology, Mechanics and Physics, are regularly embraced into this domain. One of the key motives
that contributes for the continually effort done in this field of the human knowledge is the high number of
applications that can be easily found in Medicine. For instance, computational algorithms can be applied
on medical images for shape reconstruction, motion and deformation analysis, tissue characterization or
computer-assisted diagnosis and therapy.
The main objective of these ECCOMAS Thematic Conferences on Computational Vision and Medi-
cal Image Processing, initiated in 2007, is to promote a comprehensive forum for discussion on the recent
advances in the related fields in order to identify potential collaboration between researchers of differ-
ent sciences. Henceforth, VipIMAGE 2015 brought together researchers representing fields related to
Biomechanics, Biomedical Engineering, Computational Vision, Computer Graphics, Computer Sciences,
Computational Mechanics, Electrical Engineering, Mathematics, Statistics, Medical Imaging, Medicine
and Rehabilitation.
The expertises spanned a broad range of techniques for Signal Processing and Analysis, Image Acqui-
sition, Image Processing and Analysis, Data Interpolation, Registration, Acquisition and Compression,
Image Segmentation, Tracking and Analysis of Motion and Deformation, 3D Vision, Computer Simula-
tion, Medical Imaging, Computer Aided Diagnosis, Surgery, Therapy and Treatment, Computational
Bio-imaging and Visualization, Telemedicine, Virtual and Enhanced Reality, Satellite imagery, Software
Development and Applications.
The conference co-chairs would like to take this opportunity to express gratitude for the support given
by The International European Community on Computational Methods in Applied Sciences and The
Portuguese Association of Theoretical, Applied and Computational Mechanics, and thank to all spon-
sors, to all members of the Scientific Committee, to all Invited Lecturers, to all Session-Chairs and to all
Authors for submitting and sharing their knowledge.
xi
Invited lectures
During VipIMAGE 2015, Invited lectures were delivered by 6 Expertises from 4 countries:
− Alexandre Xavier Falcão, Universidade de Campinas, Brazil
− Cristian A. Linte, Mayo Clinic, USA
− Fiorella Sgallari, University of Bologna, Italy
− Yongjie Zhang, Carnegie Mellon University, USA
− Xiaochuan Pan, The University of Chicago, USA
− Xue-cheng Tai, University of Bergen, Norway
xiii
Scientific committee
All works submitted to VipIMAGE 2015 were evaluated by an International Scientific Committee
composed by 116 expert researchers from recognized institutions of 26 countries:
xv
xvi
xvii
Yue Jia, Timon Rabczuk, Chiu Ling Chan & Cosmin Anitescu
Institute of Structural Mechanics, Bauhaus-Universität Weimar, Weimar, Germany
secutive levels, l and l + 1, in the following steps: the total number of basis functions at level-0. For
l , , ..., lmax − 1,
1. At a particular refinement level l, a subset of
B-spline basis functions satisfying the refine- 1. Compute the absolute difference in the intensi-
ment criterion are identified, which form the set ties, I d , between I1l f ( x )) and I 2 ( ) , at the cell
φ l . The remaining B-spline basis functions are coordinates of the hierarchical B-spline con-
set as active (φal ). trol grid at level-l. Here I1l f ( x )) is the evolved
2. The children B-spline basis functions of all the image obtained from the previous level-(l −1). It
basis functions φ l are set as active (φal+1 ). is the initial source image if l = 0. Let Dmean be
3. The process ends by collecting all the active the average difference in the intensities between
B-spline basis functions at levels l and l + 1, the two images.
in order to get the hierarchical B-splines basis 2. Loop over the active B-spline basis functions
l +1.
functions, φhb φ lj , j 1 2 Nb, where Nb is the total number
of active B-spline basis functions at level-l. Loop
Thus we can obtain, over supp(φ lj ), and find the average of the I d val-
l +1
ues in supp(φ lj ), denoted as D j . If D j Dmean ,
φhb φal ∪ φal +1. (7) then we refine the particular B-spline according
SSD
4
RS (%) CPU Time (in seconds) ( initial = 2.2943 × 10 ) Control points used
Levels
(Iterations) HBS UBS HBS UBS HBS UBS HBS UBS
Level 1 (40) 91.36 91.00 74.06 70.84 1,982.28 2,064.87 1,089 1,089
Level 2 (40) 97.02 96.65 155.25 303.60 682.76 768.80 2,328 3,969
Level 3 (20) 97.96 97.80 333.67 393.27 (at 6th 468.62 505.55 6,254 15,129
iteration)
Total time 562.98 767.71
(seconds)
Table 2. Medical image (Brain MRI example, image size = 320 × 320): Comparison of the convergence of the solution
and the time taken using our method (HBS) with uniform B-spline grid refinement (UBS).
SSD
3
RS (%) CPU time (in seconds) i ii Control points used
Levels
(Iterations) HBS UBS HBS UBS HBS UBS HBS UBS
Level 1 (10) 54.05 54.05 81.81 37.40 1,170.30 1,170.30 1,849 1,849
Level 2 (20) 75.41 73.73 208.60 326.61 626.25 669.12 3,997 6,889
Level 3 (20) 79.07 79.51 770.10 2,247.60 532.95 521.93 11,089 26,569
Total time 1,060.51 2,611.61
(seconds)
transformation is then evaluated on the finer levels to capture the features of the images correctly.
to further improve the accuracy. Table 1 shows the Also, the time step is reduced as compared to syn-
comparison of the time taken, the value of RS and thetic images in order to gradually and smoothly
the number of control points used for each refine- deform the image. In this example, we use a uni-
ment level of the grid. From Table 1, it can be seen form control grid consisting of 40 × 40 elements
that the total time taken to achieve the desired and 43 × 43 control points at the first level. The
accuracy using the adaptive refinement scheme threshold parameter ρ, is set as 1.5. The number
is less than that of the uniform refinement strat- of control points in the second and third refine-
egy. Although the initialization time taken to set ment levels are 3,997 and 11,089 respectively. The
up the hierarchical data structure in adaptive grid time step for the entire registration process is set
refinement is more than that taken for the uni- to be 0.00025. The solution reached the desired
form grid refinement, during the computation accuracy at the end of the third refinement level.
of the deformed image, the time taken by the For the first level, the total number of iterations
adaptive refinement method is much less. This equals to 10, whereas for the second and third
is because the number of control points used to level they were increased to 20. The values of RS
compute the spatial transformation of the deform- at the end of each level are 54.05%, 75.41% and
ing image is much lower than the uniformly refined 79.07%. The final images obtained at each refine-
grid. This makes the adaptive registration process ment level are shown in Figures 2 (g–i). The dif-
much more efficient. ference between the resulting and target images
at the end of each level are shown in Figures 2
(j–l). Table 2 shows the comparison of the time
3.2 Medical image example (Brain MRI)
taken, the value of RS and the number of control
The registration model is tested on a pair of med- points used for each refinement level of the grid.
ical images as shown in Figures 2 (a–b). Unlike From Table 2, it can be seen that the total time
synthetic images, medical images have more taken to achieve the desired accuracy using adap-
complex feature information. Thus we use more tive refinement scheme was less than that of the
control points at the first refinement level so as uniform refinement strategy.
ACKNOWLEDGEMENTS
REFERENCES
Alexandre Xavier Falcão, Thiago Vallin Spina & Samuel Botter Martins*
Institute of Computing, University of Campinas, Campinas, Brazil
Renzo Phellan
Eldorado Research Institute, Campinas, Brazil
ABSTRACT: Segmentation is important to define the spatial extension of body anatomic structures
(objects) in medical images for quantitative analysis. In this context, it is desirable to eliminate (at least
minimize) user interaction. This aim is feasible by combining object delineation algorithms with Object
Shape Models (OSMs). While the former can better capture the actual shape of the object in the image,
the latter provides shape constraints to assist its location and delineation. We review two important
classes of OSMs for medical image segmentation: Statistical (SOSMs) and Fuzzy (FOSMs). SOSMs
rely on the image mapping onto a reference coordinate system, which indicates the probability of each
voxel to be in the object (a probabilistic atlas built from a set of training images and their segmentation
masks). Imperfect mappings due to shape and texture variations asks for object delineation algorithms,
but the methods usually assume that the atlas is at the best position for delineation. Multiple atlases per
object can mitigate the problem and a recent trend is to use each training mask as an individual atlas.
By mapping them onto the coordinate system of a new image, object delineation can be accomplished
by label fusion. However, the processing time for deformable registration is critical to make SOSMs suit-
able for large scale studies. FOSMs appear as a recent alternative to avoid reference systems (deformable
registration) by translating the training masks to a common reference point for model construction.
This relaxes the shape constraints, but asks for a more effective object delineation algorithm and some
efficient approach for object’s location. One of the solutions, named optimum object search, translates
the model inside an estimated search region in the image while a criterion function guides translation and
determines the best delineated object among candidates. This makes segmentation with FOSMs consid-
erably faster than with SOSMs, but SOSMs that adopt the optimum object search can be more effective
and with less atlases per object. We then discuss the pros and cons of the recent FOSM and SOSM
approaches by providing alternative directions, which also include the user to correct segmentation errors
and improve the models.
10
11
12
4 INTERACTIVE CORRECTION
OF OSM SEGMENTATION
13
14
15
Ke Wei
Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong
Xue-Cheng Tai
Department of Mathematics, University of Bergen, Bergen, Norway
Tony F. Chan
Office of the President, Hong Kong University of Science and Technology, Hong Kong
Shingyu Leung
Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong
ABSTRACT: We review the continuous max-flow approaches for the variational image segmentation
models with piecewise constant representations. The review is conducted by exploring the primal-dual
relationships between the continuous min-cut and max-flow problems. In addition, we introduce the
parameter free primal-dual method for solving those max-flow problems. Empirical results show that the
primal-dual method is competitive to the augmented Lagrangian method.
17
n n
where a1i … ami is a binary representation of i, denoted
u ∈{
min
}
∑ ∫Ω fi i (u ) ∑ ∫Ω | ∇ i ( u)) | dx
d , (3) by i [a1i ami . The interpretation of this model as
i =1 i =1
a min-cut problem and its corresponding max-flow
model was given in (Bae, Lellmann, & Tai 2013,
This model was first introduced in (Lie, Bae & Tai 2015, Liu, Tai, Leung, & Huang 2014).
Lysaker, & Tai 2005, Lie, Lysaker, & Tai 2006b).
Augmented Lagrangian was used in (Lie, Lysaker, &
Tai 2005, Lie, Lysaker, & Tai 2006b) to solve it. In 2 CONTINUOUS MAX-FLOW
(Bae & Tai 2009), discrete graph cut was used to APPROACHES
solve this model which is very much related to the
graph based method in (Ishikawa 2003, Darbon & Three piecewise constant representations and the cor-
Sigelle 2006). The interpretation of this model as a responding formulations for labelling problems were
min-cut problem and its corresponding max-flow introduced in the last section. All the three formula-
model was given in (Bae, Yuan, Tai, & Boykov tions are non-convex because either the constraints
2010, Bae, Yuan, Tai, & Boykov 2014a). This prob- are non-convex or both the objective functional
lem can also be transformed into a binary mini- and the constraints are non-convex. A number of
zation problem. If we replace the regulariser by approaches have appeared in the literature recently
R ⎛⎝{ i }in=1 ⎞⎠ = ∫ ∇u dx and introduce showing that these models can be interpreted as
Ω continuous min-cut problems. There exists a corre-
sponding continuous max-flow model for each of
⎧1 if u( x ) > i these three min-cut problems. Exploring the connec-
λi ( x ) = ⎨ , i 1, …, n − 1
⎩0 if u( x ) ≤ i tion between these continuous min-cut and max-flow
problems, convex global minimization method could
be derived for these non-convex problems.
and λ0 ( ) = 1, λn ( x ) = 0 . Then model (1) can also Since the pioneer work of Chan et al. (Chan,
be written as a binary minimization problem (Liu, Esedoglu, & Nikolova 2006), one of the important
Tai, Leung, & Huang 2014) research topics in image segmentation is to find
good convex relaxations for the non-convex mini-
mization problems arising from the variational
1
In the sequel, we will omit the notation x when there is methods and then design numerical algorithms for
no confusion from the context. the corresponding convex relaxations. In a series of
18
min ∫ ( u )Cs dx
d ∫Ω uCt dx + ∫Ω ∇u ddx 2.2.1 Binary value representation
u∈[ ,
] Ω Following the convex relaxation technique for the
= min max ∫ ∫Ω upt dx + ∫Ω u div qdx two-phase Chan-Vese model, a natural convex
u∈[ , ] ps Cs Ω
pt ≤Ct
relaxation for (2) is
| |≤α n
= min max ∫ min ∑ ∫ ui fi dx + ∑ ∫Ω ∇ui dx.
u∈R ps ≤Cs Ω
∫Ω upt dx + ∫Ω u div qdx u S
i =1
Ω
i =1
(9)
p ≤Ct
|q|≤α
We can form a dual problem of (9) by introducing
= max min ∫ ∫Ω upt dx + ∫Ω u div qdx another n variables pi i = 1, , n which corresponds
ps ≤Cs u∈R Ω
p ≤Ct to the data fidelity term and an auxiliary variable ps:
| |≤α
n n
= max min ∫
ps ≤Cs u∈R Ω
∫Ω u ( − + ) dx ( 1
min
n )∈S
∑ ∫Ω ui fi dx + ∑ ∫Ω ∇ui dx
i =11 i 1
p ≤Ct ⎛ n ⎞ n
| |≤α
⎧ p ≤ Cs , pt ≤ Ct , | | ≤ α
= min max ∫
ui ≥ 0 ps
⎜1 −
Ω⎜
⎝
∑ ⎟
ui ⎟ ps dx ∑ ∫Ω ui fi dx
i =1 ⎠
= max ∫ ps dx
d s. t. ⎨ s i =1
⎩div q − ps + pt = 0,
Ω n
+ ∑ ∫Ω ∇ui dx
(8) i =1
19
20
The maximisation problems (8), (10), (12) and (13) where F ( p q γ ) is (linear) functional of p, q
share the following general form and γ . Then following the general set-up of the
primal-dual method in (Chambolle & Pock 2011,
⎧ p Cp Esser, Zhang, & Chan 2010), we can solve (16) as
⎪ follows (Alg. 2):
max F ( p ) s. t. ⎨q Cq , (14)
p
⎪⎩Lq + Dp = 0,
∂F ( p ,q l ,γ l )
• pl pl + τ lp ,
∂pp
where ∂F ( pl ,q ,γ l )
• ql q l + τ lq ,
∂q
• F ( p ) is a (linear) functional of p,
∂F ( pl ,q l +1 ,γ )
• L and D are linear mappings, • γl γ l − βl ,
∂γ
• C p and Cq are convex sets.
where τ lp ,τ lq and β l are adaptive stepsizes. The
The augmented Lagrangian formulation of (14)
primal-dual method is parameter free compared
is
with the augmented Lagrangian method in Sec. 3.1
c 2 as there is no proximal regularizer in Alg. 2. More-
L( γ ) = F ( p ) + γ , Lq + Dp − Lq + Dp . over, we can choose the stepsizes automatically to
2 (15) accelerate the algorithm, for example see Table 1.
In general, τ li (i p q ) and β l should be inverse
Then with a triple of intial points ( 0 0 γ 0 ), to each other, which are proportional to l and 1/l
we can solve (14) by the augmented Lagrangian respectively.
method as follows (Alg. 1):
• p
l +1
arg max p C p L( p, q l ,γ l ), 4 NUMERICAL EXPERIMENTS
• q
l +1
q Cq
( p , q, γ ),
l l
This section compares the aforementioned two
• γ l
γ − c ( Lq Dp
l
D ). algorithms on the image segmentation problems.
The tests are conducted on a Mac Pro laptop with
When there are no explicit solutions to the first Intel Core i5 CPUs @ 2.4 GHz and executed from
two subproblems, they can be solved approximately Matlab. For the discretized version of the gradient,
by several projected ascent iterations (Yuan, Bae, divergence and TV operators, we refer the readers
Tai, & Boykov 2014, Bae & Tai 2015, Bae, Yuan, to (Zhu 2008, Ch. 1.3).
Tai, & Boykov 2014b, Bae, Lellmann, & Tai 2013,
Yuan, Bae, Tai, & Boykov 2010). The penalization
parameter c is chosen by trial-and-error in numeri-
cal tests.
τ lp τ lq βl
21
Table 2. Computational results of the image segmentation for the two-phase (cameraman) and multi-phase (3 col-
our) image segmentations by the primal-dual and the augmented Lagrangian methods. The algorithm terminates if el
defined in (17) is less than ε.
Cameraman 3 color
ε =1 −3 ε =1 −5 ε =1 −3 ε =1 −5
Iter Time (s) Iter Time (s) Iter Time (s) Iter Time (s)
In our tests, the data term has the form We test the primal-dual and the augmented
fi | u − ci |β , where u is the input image and ci Lagrangian methods on the standard cameraman
is the average intensity of the region i which is image (see the left of Fig. 1) for the two-phase
assumed to be fixed. Apparently, fi is convex image segmentation problem and on the standard
when β ≥ 1 and non-convex when β < 1. Here β noisy 3 color image (see the right of Fig. 1) for the
is set to 1. multi-phase image segmentation problem. Due to
In order to compare the convergence rate of the the limited space, we only conduct the tests for the
two tested algorithms, we first (numerically) com- variational model with the integer value represen-
pute the optimal primal energy, denoted by E ∗ , by tation in Sec. 2.2.2. In Alg. 1, c is selected to be 0.3,
running either of the algorithms 10000 iterations. and the adaptive stepsize selections in Alg. 2 are
Then we evaluate the progress of the algorithm by listed in Table 1.
computing the relative error of the primal energy The segmentation results are plotted in Figure 2.
defined as By comparing the plots before and after thresh-
olding, we can see that both the primal-dual and
| El − E∗ | the augmented Lagrangian methods can achieve
el = , (17)
E∗ very good performance since they all converge to
a result which takes the values 0 and 1 almost eve-
where E l is the primal energy in the lth iteration. rywhere. The computational results are presented
22
23
24
ABSTRACT: In image formation, the observed images are usually blurred by optical instruments and/
or transfer medium and contaminated by noise, which makes image restoration a classical inverse prob-
lem. A general principle for dealing with the intrinsic numerical instability of most inverse problems is that
of regularization. Among the regularization approaches the most common is Tikhonov regularization.
Another very popular choice in the literature for regularization is based on the Total Variation (TV) semi-
norm. Since then, in order to provide more reliable and efficient models, several variational formulations
have been investigated which involve two terms, the data-fidelity term and the regularization term, with
the goal to capture characteristic and essential image features. In this review, we will focus on variational
methods for the restoration of images corrupted by additive noise which we assume to be sampled from a
known a-priori distribution. In particular, we will discuss recent advanced proposals for the regularization
and fidelity terms in the optimization problem. Numerical results illustrate the efficacy of the different
models.
25
u. If we let ∇ i ::=
= (Gx ,i u,G
,G y,i u )T , with Gx,i, Gy,i rep-
resenting the ith rows of the x and y-directional where Λ = diag( λ1 λd ) is an d 2 d 2 diago-
finite difference operators Gx, Gy, respectively, then nal matrix with λi representing the regularization
the regularization term R( ) is defined by the parameter for the ith pixel, Aα (ui ) = || ∇α i ui ||2 ,
TV-semi-norm where α i represents the fractional order of differ-
entiation for the ith pixel, and
d2
|| u ||TV = || A( ) ||1: ∑ ( x ,i )2 ( yi )2 .
∇α i = (Gxα,ii ,,G
::= G yα,ii u )T (4)
i =1
26
where τ is a scalar parameter controlling the vari- the upper and lower limits of the whiteness set,
ance of the residue image Ku − g . respectively.
27
∗ || u ||
SNR( ) g10 dB, (10)
|| u∗ u ||
2
where u∗ ∈ R d is the computed
p estimate of the
2
uncorrupted image u ∈R d . This quantity pro-
vides a quantitative measure of the quality of the
restored image: a high SNR value indicates that u∗
is an accurate approximation of u .
The corrupted images taken as input by the
algorithms have been obtained as follows. First,
the original d d image u is blurred by a Gaussian
kernel characterized by two parameters and
Figure 2. Top row: step image (left), realization of and generated through the MATLAB com-
Gaussian (center) and uniform (right) noise distribu- mand ( ).
tions; bottom row: associated histograms. The larger the , the stronger the blur will be.
28
29
30
31
Cristian A. Linte
Department of Biomedical Engineering, Chester F. Carlson Center for Imaging Science,
Rochester Institute of Technology, Rochester, NY, USA
ABSTRACT: While effective therapy cannot be delivered without the appropriate equipment and
instrumentation that enables access to the internal organs and tissues under minimally invasive condi-
tions, real-time visualization and navigation information is critical to guide the instruments to the correct
targets and manipulate the diseased tissues in absence of direct vision. Over the past decades, thanks to
the advances in medical image acquisition, visualization and display, surgical tracking and image comput-
ing infrastructure, a wide variety of technology has emerged that facilitates diagnosis, procedure planning,
intra-operative guidance and treatment monitoring while providing safer and less invasive approaches for
therapy delivery. Nevertheless, successful exploitation, use and integration of the multi-modality images
and signals available from both pre- and intra-operative sources require the development and validation
of engineering tools that allow users to extract critical information for decision making during planning,
visualization, guidance or training. This article serves as a quick snapshot of various applications of
biomedical imaging and modeling, their integration into platforms for planning and guidance of inter-
ventions, and their potential to develop new paradigms for virtual anatomy visualization for simulation,
teaching and training applications.
33
34
35
36
37
38
A. Bal
Institute of Automatic Control, Silesian University of Technology, Gliwice, Poland
ABSTRACT: In this work simple and effective method for creating of High Dynamic Range (HDR)
images for use in the fluorescence microscopy applications is presented. For creating of the HDR image
the Camera Response Function (CRF) is used, so the pixel values of the HDR image, in general, are pro-
portional to the signal which have been reached the camera sensor (i.e. irradiance of sensor). Therefore
the obtained HDR image can be used not only for qualitative analysis of the specimen, but also for its
quantitative analysis. The proposed method has been tested on real data.
41
Ii ( i 1) α i Ii ( i ) τ + Δi , (3)
where
ti −1
Figure 2. Profiles A–B of images 1(a)–(c) i.e. for αi = (4)
DAPI channel: (a) cell image with marked line segment
ti
AB (red) for which the profiles have been calculated,
(b) image profiles in normalized range of pixel values. and
1
2 DESCRIPTION OF THE METHOD
Δi = ∑ Ii −1( x, y).
|| Ci −1 || pi −1∈Ci −1
(5)
In the proposed method the HDR image I is The C( j ) is an inner contour of region ωj and
assembled from two or more LDR images I i ∈ I , it is defined as follows
where I = { 1, 2 , , i , , imax } . The index i repre-
sents exposure time ti of the image Ii, and images Cj { j( y) p j ( ) j ∈ (ω j , m )}, (6)
I i ∈ I are sorted in the descending order of its
exposure time i.e. ti ti +1. The images Ii and Ij are where ∈( j , ) is a result of morphological erosion
differ only in the exposure times ti and tj used for of region ωj with structuring element m, which in
their acquisition. this case is defined as 3 × 3 array of 1.
42
43
ψ Ii ( x y)
NRLDRi ( x y ) = 20 log10
SNR , (8)
ψ I i ( x y ) + ηr 2 + ηd 2
SNR
NRHDR ( x y ) SNRLDR
S y ) |Ii ( x y ) = f
Ri ( x y) I i ( x y )) , (9)
44
Iτ ( , y ) − I ( , y )
Δ τ ( , y) = , (11)
I ( , y)
where
Figure 8. Maps of SNR for LDR and HDR image in
FITC channel: (a) LDR image acquired with exposure
1
time t = 2.5 s, (b) HDR image; (c) difference ΔSNR for
images (a) and (b).
I( y) ∑ Iτ (
|| T || τ ∈T
y ) ⋅ 100%, (12)
45
imax
T ( x y) ∑ ti ( x y), (14)
i =1
ACKNOWLEDGEMENTS
46
47
ABSTRACT: This paper deals with some problems of cheek detection in colour images. Authors
reviewed the literature in this field and proposed a new effective method for cheek detection. A character-
istic feature of the proposed method is the use of 13 essential face landmarks generated by a known face
landmark detector. In the paper four methods from the literature and one’s own method are described
in detail. All these methods have been tested experimentally using a suitable face image database PUT.
Preliminary tests showed the advantages of our method. Additionally we investigated the impact of head
movements on the location detected cheeks. A small number of false cheek pixels lying outside the face
contour and a small colour error caused by movements of the head predetermines the proposed method
for further research and applications.
1 INTRODUCTION (Su et al. 2013) both cheeks are detected after the
previous using the Viola-Jones algorithm for eye
Several applications of face image processing may and mouth detection. Based on the colours of
require detecting facial regions such the mouth, cheek pixels the haemoglobin and melanin com-
eyes, eyebrows, nose, cheeks and chin. Automatic ponents are separated, making it easier to assess
cheek detection is an important operation in this the health of a person. Sometimes the colours of
field. The cheek colour well represents the face the face in selected locations (five blocks each of
colour, because the cheeks are not covered by hair 32 × 32 pixels) are determined and on the basis
as sometimes the forehead or the chin and not of these colours such liver diseases as hepatitis
generate highlights as the nose area. Each cheek are diagnosed (Liu and Guo 2007). Two of these
is a consistent patch of skin. Among the typical blocks are located on the surface of both cheeks
cheek detection applications are medical diagnos- through finding another facial features. Some
tics, human face beautification as well as coding of other application of this approach is described in
videophone sequences. the paper (Kim et al. 2008), that presents an idea to
Person’s emotional state generates colour use the colours of the forehead centre and the lips
changes in its facial skin (Ramirez et al. 2014). for cardiovascular disease detection.
In this work the colour changes were considered An original application of the cheek detection is
and measured in three regions of the face, i.e. the used to beautify the human face in photographic por-
forehead and both cheeks. These regions of inter- traits (Lai et al. 2013). Beautification is performed
est limited to small squares were determined on the by the total removal or the magnitude reduction of
basis of data from the facial feature tracker. Good unwanted skin details: scars, pores, blemishes, wrin-
results are obtained from the combination of the kles, etc. The developed skin changing system can,
forehead and the left cheek. Automatic cheek on the one of the early processing stages, automati-
detection may be particularly useful in medical cally extract the cheeks by using the active shape
diagnostics. Human health evaluation based on modelling and the watershed algorithm.
the colour of his/her face is used in traditional Automatic cheek detection was also applied in
oriental medicine for thousands of years (Yang the field of model-based coding of videophone
et al. 2012, Seo et al. 2014, Zhao et al. 2014). The sequences (Hess and Martinez 2004). The human
result of the latest research is a strong statistical face in such sequence can be segmented into few
relationship between perceptually uniform colour facial regions such as e.g. cheeks. The cheek borders
space CIELAB components in specific regions of were detected based on the well-known SUSAN
the face and pathological patterns used in orien- algorithm (Smith and Brady 1997) and additional
tal medicine. In the paper (Yang et al. 2012) both knowledge-based rules.
cheek skin boxes are detected after the previous The main goal of this study is to compare the
eye and mouth detection. Similarly, in the work automatic cheek detection methods existing in the
49
50
3 EXPERIMENTAL TESTS
51
52
53
4 CONCLUSION
REFERENCES
Table 2. Average colour errors of cheek regions.
Campadelli, P., G. Lipori, & R. Lanzarotti (2007).
Meth. 1 Meth. 2 Meth. 3 Meth. 4 Meth. 5 Automatic facial feature extraction for face recognition.
INTECH Open Access Publisher.
Person 1 6.23 6.23 3.52 4.12 4.12 Cristinacce, D. & T.F. Cootes (2006). Feature detection
Person 2 10.58 15.14 6.42 4.41 4.30 and tracking with constrained local models. In Proc.
Person 3 7.18 12.14 5.82 3.98 4.02 BMVC, pp. 95.1–95.10. doi:10.5244/C.20.95.
Person 4 5.75 8.89 4.01 3.29 2.95 Dalal, N. & B. Triggs (2005). Histograms of oriented
gradients for human detection. In IEEE Computer
AVG 7.44 10.60 4.94 3.95 3.85
Society Conference on Computer Vision and Pattern
Recognition, Volume 1, pp. 886–893.
Fairchild, M. (1998). Color appearance models. Addison-
calculated an average colour error ΔE E76 using a Wesley.
Gross, R., I. Matthews, J. Cohn, T. Kanade, & S. Baker
standard distance formula in the perceptually uni-
(2010). Multi-pie. Image and Vision Computing 28(5),
form CIELAB colour space (Fairchild 1998): 807–813.
Hess, M. & G. Martinez (2004). Facial feature extraction
Δ 76 = ( L0 − Li )2 + (a − ai )2 + (b0 − bi )2 (6) based on the smallest univalue segment assimilating
nucleus (susan) algorithm. In Proceedings of Picture
Coding Symposium, Volume 1, pp. 261–266.
where L0 , a0 , b0 are the components of average Kasinski, A., A. Florek, & A. Schmidt (2008). The PUT
cheek colour for central positioned image and face database. Image Processing and Communications
Li , ai , bi —components of average cheek colour for 13(3–4), 59–64.
54
55
Gil Gonçalves
INESC—Coimbra, Portugal
ABSTRACT: Image fusion is a technique that merges the information of multiple images, represent-
ing the same scene, to produce a single image that should gather the major and meaningful information
contained in the different images. On the other hand, cartoon+texture image decomposition is another
image processing technique, that decomposes the image into the sum of a cartoon image, containing the
major geometric information, i.e., piece-wise smooth regions, and a textural image, containing the small
details and oscillating patterns. We propose a model to perform the fusion of multiple images, relying on
gradient information, that provides as well a cartoon and texture decomposition of the fused image. The
model is formulated as a variational minimization problem and solved with the split Bregman method.
The suitability of the proposed model is illustrated with some earth observation images and also medical
images.
57
58
+ d2 − b2k ⋅ n z k
s (v k +1 + w k , λα ),
λ1 + λ2 ck ck − ek
4
∇(u k +1 + v k +1 ) g,
where n is the unit outward normal to ∂Ω and w k
w −z
k k
v k +1.
“⋅” denotes the inner pproduct in R 2 . End
− The formulas for d1k +1 and d2k +1 are explicit using Output—Fused image u k v k , where u k is the
the shrinkage operators, so at each pixel (i, j) in Ω cartoon part and v k the texture part.
d1ki j
⎝
(
⎛ ( k
∇ u 1
+ v k ) g + b1k )i j , λ1 ⎞⎟⎠ , 4 EXPERIMENTS
1
⎛
d2ki+1j = shri
r nk ⎜ ∇
⎝
( k+
+ b2k )i , j , λ1 ⎞⎟⎠ , In this section we apply the proposed tech-
2 nique to three different types of images. In the
59
Figure 1. Top left: Multispectral (R, G, B) scene of Sydney, Australia. Top middle: Panchromatic image. Bottom left:
Cartoon. Bottom middle: Fused Image. Right: Texture.
60
E AG × 103
61
REFERENCES
Table 3. The computational time in seconds for the
three figures. Columns 2–4 correspond to the execution Bertalmío, M., V. Caselles, E. Provenzi, & A. Rizzi
time of the proposed method (fusion and simultaneous (2007). Perceptual color correction through vari-
cartoon+texture decomposition of the fused image) of ational techniques. Image Processing, IEEE Transac-
the red, green and blue bands, respectively. tions on 16(4), 1058–1072.
Bregman, L.M. (1967). The relaxation method of finding
Red Green Blue
the common point of convex sets and its application
to the solution of problems in convex programming.
Figure 1 0.3276 0.3120 0.3744
USSR Computational Mathematics and Mathematical
Figure 2 1.7940 1.6224 1.7472 Physics 7(3), 200–217.
Figure 3 0.2184 0.2652 0.1560 Bresson, X., S. Esedoglu, P. Vandergheynst, J.-P. Thiran, &
S. Osher (2007). Fast global minimization of the active
contour/snake model. Journal of Mathematical Imag-
ing and vision 28(2), 151–167.
All the numbers reported in Tables 1 and 2 are Fang, F., F. Li, G. Zhang, & C. Shen (2013). A varia-
averages of the three channels when it is applica- tional method for multisource remote-sensing image
ble. The displayed results confirm that the average fusion. International Journal of Remote Sensing 34(7),
gradient AG and entropy E for the fused images 2470–2486.
are improved when compared to AG and E of the Goldstein, T. & S. Osher (2009). The Split Bregman
method for L1-regularized problems. SIAM Journal
input (see Table 1). Also, from Table 2, one can
on Imaging Sciences 2(2), 323–343.
see that PSNR and SNR of the cartoon and fused Gonzalez, R.C. & R.E. Woods (2006). Digital Image
images are similar. Processing (3rd Edition). Upper Saddle River, NJ,
To evaluate the computational efficiency of USA: Prentice-Hall, Inc.
our model, we computed the execution time. Huynh-Thu, Q. & M. Ghanbari (2008, June). Scope of
Table 3 shows that the proposed method is quite validity of PSNR in image/video quality assessment.
fast. Electronics Letters 44(13), 800–801.
Liu, W., J. Huang, & Y. Zhao (2006). Image fusion based
on PCA and undecimated discrete wavelet transform.
In I. King, J. Wang, L.-W. Chan, and D. Wang (Eds.),
5 CONCLUSIONS
Neural Information Processing, Volume 4233 of Lec-
ture Notes in Computer Science, pp. 481–488. Springer
We have presented a variational model for image Berlin Heidelberg.
fusion with a simultaneous cartoon and texture Piella, G. (2009). Image fusion for enhanced visualiza-
decomposition of the fused image. The main moti- tion: a variational approach. International Journal of
vation for performing these two operations at the Computer Vision 83(1), 1–11.
62
ABSTRACT: In this paper we propose a variational method for image contrast enhancement, by keep-
ing the image details and correcting the non-uniform illumination. It is a minimization problem, where the
objective functional consists of two different fitting terms: a L1 term that matches the gradients of the input
and reconstructed images, for preserving the image details, and a L2 term that measures the misfit between
the reconstructed image and the mean value of the input image, for reducing the image variance and thus
correcting the illumination. For solving this minimization problem we apply the split Bregman method,
which is an efficient and fast iterative method suitable for this type of non-differentiable and convex mini-
mization problem, involving a L1 term. Some experimental results show the effectiveness of the method.
{∫ }
*This work was partially supported by the research
project PTDC/MATNAN/0593/2012 of FCT (the Por- min
u Ω
u f ddx
x ∫Ω ( ) dx
{ },
tuguese national funding agency for science, research and
technology) and by CMUC (the Centre for Mathematics = min || ∇u − ∇f ||2L2 ( ) + λ || u − u ||2L2 ( ) (1)
of the University of Coimbra—UID/MAT/00324/2013), u
funded by the Portuguese Government through FCT/
MEC and co-funded by the European Regional Develop- where f R is the original (grayscale) input
ment Fund through the Partnership Agreement PT2020. image, Ω ⊂ R 2 represents the image pixel domain,
63
}
misfit between ∇u and ∇f and we also replace the
α
mean value u by f . This yields the following vari- + || d (∇ ∇f ) ||2L2 ( )
. (3)
ational model 2
{∫ }
Then, the split Bregman method consists in
min
u Ω
u f dx
d ∫Ω ( f ) dx solving the following sequence of problems for
k = 0, 1, 2, …
u {
= min || ∇u − ∇f ||L1 ( )
λ
+ || u − f ||2L2 (
2 ) } (2)
⎧ k
⎪(
k +1
{
) = arg min || d ||L ( )
λ
|| u − f ||2L2 ( )
d ,u 2
}
⎪
1
where L ( ) is the space of absolutely integrable ⎨ α
+|| d − (∇u ∇ff ) − b k ||2L2 ( ,
functions in Ω and || . ||L1 ( ) denotes the L1-norm. The ⎪ 2 )
⎪b k
reason for replacing the L2-norm by the L1-norm is ⎩ b k + ∇uu k f − d k +1.
related to the fact that the L2-norm of the gradient
tends to smear image discontinuities, as opposed to (4)
the L1-norm that tends to preserve the discontinui- k
where the new vector b results from the Bregman
ties, which in image processing corresponds to sharp iteration (Bregman 1967), that is a strategy for
edges. In addition to this advantage, the presence of enforcing the constraint d = ∇u − ∇ff , using a fixed
the L1 fitting term in (2) permits the use of fast and penalty parameter α. This strategy is an alternative
effective algorithms for computing its solution. In to the conventional continuation technique to solve
effect, problem (2) belongs to the general class of (3) with an increasing sequence of penalty param-
L1-regularized problems of the form eters α1 < α 2 < … < α n tending to ∞, for accurately
enforcing the constraint. The minimization prob-
min{|| φ ( ) ||L1 ( )
H ( )} lem in (4) is solved by iteratively minimizing with
u
respect u and d, alternatively, which means the fol-
lowing two steps are performed.
where both || φ ( ) ||L1 ( ) and H(u) are convex func-
tions. This kind of models can be efficiently solved Step 1—Minimization with respect to u (with d
with the split Bregman method of (Goldstein & fixed)
Osher 2009). This is an appropriate algorithm for
solving non-differentiable convex minimization ⎧λ α
u k +1 arg min ⎨ || u f ||2L2 + || d k
problems, involving L1 or TV (total variation) u ⎩2 2
terms. We refer to (Fang, Li, Zhang, & Shen 2013,
Getreuer 2012b, Goldstein, Bresson, & Osher 2010,
−(∇u ∇f ) − b k ||2L2 ( ) , } (5)
Yin, Osher, Goldfarb,& Darbon 2008) for a few
examples of different applications of the method. for which the optimality condition (derived from
In our case φ ( ) = ∇u − ∇ff and H (u ) = the Euler-Lagrange equation) is, in Ω
λ || u f ||L2 ( ), and in the next section we apply
split Bregman method to solve (2). (λ )u k + = λ f α div( k
∇f k
),
64
1 4 APPLICATIONS
uik +j 1 [αU i j + λ fi j − α vik j ]
λ + 4α
Some results obtained with our proposed model
U = (U i j ) := (uik +1j + uik+1, j + ik, +j 11 + uik, j +1 )
are shown in this section. All the experiments
vk = ( i , j ) ::=
k
( (d k + ∇f − b k ))i , j . were implemented with the software MATLAB®
R2014a (The Mathworks, Inc.)
Here we use finite differences for approximating Figure 1 shows the contrast enhancement with
the derivatives in the gradient ∇ and divergence div our method for a standard test image (a scalar
operators, respectively.
p In particular the discretiza- image with 512 × 343 pixels), downloaded from
tion used for vik j is obtained by applying
y backward the IPOL archive (http://www.ipol.im/). As this
finite differences for divd k and divb k and centered figure demonstrates, the details are kept and the
finite differences for Δf
Δ dark regions become more visible in the enhanced
image. In addition, and as expected, when λ
vik j dik j − dik − dik j −1 − 2bik j bik + bik j −1 increases (λ is the parameter associated with the
j j
fitting term intended to reduce the non-uniform
+ fi k 1 j + fi k j fikj −1 + fikj +1 fikj . illumination) the result tends to the mean value of
the input image.
The Neumann boundary condition is implicitly Figure 2 depicts the results of our method
imposed in ∂Ω, the boundary of the rectangular applied to a medical (RGB—red, green, blue)
pixel domain, by using backward finite differences, image (with 536 × 536 pixels), acquired with the
in the right and top sides, and forward finite differ- wireless capsule Pillcam Colon 2 of Given Imag-
ences in the left and bottom sides. ing. It displays a colonic polyp (the reddish region
Step 2—Minimization with respect to d (with u
fixed)
k +1
{
rg min || d ||L1 (
d )
α
2
∇u k +1 ∇f )
|| d − (∇u
− b k ||2L2 ( ) }.
This problem can be explicitly solved using
shrinkage operation (known as well as soft thresh-
olding) at each pixel (i, j)
⎛ 1⎞
dik j = shrink (∇u k +1 − ∇f
∇f b k )i j , ⎟
⎝ α⎠
where for z ,γ R
65
in the top left subfigure) exhibiting strong texture. Figure 3. First column: Original (RGB) retinal fun-
We applied the algorithm independently to each dus image and results with (λ = 1, α = 5) in 2nd row and
channel. The original medical image (the top left (λ = 5, α = 5) in 3rd row. Second column: Grayscale ver-
subfigure) has a non-uniform illumination, with sion and results with (λ = 1, α = 5) in 2nd row and (λ = 5,
low contrast in some regions, that is corrected and α = 5) in 3rd row.
enhanced with the proposed method. The influ-
ence of the model parameters (λ and α) is also
illustrated in these results. Increasing λ results in
an averaged image, tending to the mean value of
the input image, and by increasing α the contrast
enhancement is enforced.
In Figure 3 we can see the results for another medi-
cal image (with size 536 × 536 pixels). It is an inhomo-
geneous illuminated retinal fundus image, provided
by the company Retmarker (http://www.retmarker.
com/), and obtained from a patient screened accord-
ing to the Diabetic Retinopathy Screening Program
of Portugal. We have processed with our method
the grayscale version (second column) as well as
each color channel separately (first column). Again
these results show the good contrast enhancement
improvement achieved with our method.
Figure 4 shows the result of the proposed meth-
odology for a MRI image, downloaded from the
IPOL archive (http://www.ipol.im/), and also a
comparison to the result obtained with the online
demo of (Morel, Petro, & Sbert 2014). We remark Figure 4. Top left: Original MRI and input for the pro-
posed method. Top right: Original MRI with 2% satu-
that our method operates directly on the input
ration and input for the model solved with the screened
image (exhibiting intensity inhomogeneities) Poisson equation (Morel, Petro, & Sbert 2014). Bottom
whereas the displayed result with the technique of left: Result for the proposed method with (λ = 1, α = 10).
(Morel, Petro, & Sbert 2014) involve pre and post Bottom right: Result applying a simplest color balance
processing steps that consist on the application of with 2% saturation to the solution obtained with the
a simplest color balance technique (Limare, Lisani, screened Poisson equation and with λ = 0.05.
66
5 CONCLUSIONS REFERENCES
In this paper we propose an inverse variational Aubert, G. & P. Kornprobst (2006). Mathematical prob-
model for image contrast enhancement. It is a modi- lems in image processing: partial differential equations
fication of the screened Poisson equation for image and the calculus of variations, Volume 147. Springer
contrast enhancement as proposed by (Morel, Science & Business Media.
Bregman, L.M. (1967). The relaxation method of finding
Petro, & Sbert 2014). The difference with respect the common point of convex sets and its application
to this latter model is in the term that matches the to the solution of problems in convex programming.
gradients of the input and reconstructed images, USSR Computational Mathematics and Mathematical
that is now measured in the L1-norm (instead of Physics 7(3), 200–217.
the L2-norm as in (Morel, Petro, & Sbert 2014)), Fang, F., F. Li, G. Zhang, & C. Shen (2013). A varia-
because the former has the advantage of better tional method for multisource remote-sensing image
preserving edges in image processing applications. fusion. International Journal of Remote Sensing 34(7),
The presence of this L1-norm term enables the 2470–2486.
proposed model to be solved by the split Bregman Getreuer, P. (2012a). Automatic color enhancement (ace)
and its fast implementation. Image Processing On Line
technique, which is an efficient procedure that has (2012).
the advantage of not requiring regularization, Getreuer, P. (2012b). Rudin-Osher-Fatemi total variation
continuation, or the enforcement of inequality denoising using split Bregman. Image Processing On
constraints. Line 10.
Applications of the proposed method to dif- Goldstein, T., X. Bresson, & S. Osher (2010). Geometric
ferent types of images show its good perform- applications of the split bregman method:
ance. We also remark that our methodology does segmentation and surface reconstruction. Journal of
not include pre and post processing steps as it Scientific Computing 45(1–3), 272–293.
happens in the methodology presented in (Morel, Goldstein, T. & S. Osher (2009). The Split Bregman
method for L1-regularized problems. SIAM Journal
Petro, & Sbert 2014), where the best reconstructed on Imaging Sciences 2(2), 323–343.
images have been obtained with pre and post Gonzalez, R.C. & R.E. Woods (2008). Digital Image
processing steps, involving the application of a Processing (3rd Edition). NJ, USA: Prentice-Hall, Inc.
simplest color balance technique (Limare, Lisani, Limare, N., J.-L. Lisani, J.-M. Morel, A.B. Petro, &
Morel, Petro, & Sbert 2011) applied before and C. Sbert (2011). Simplest color balance. Image Process-
after solving the model with the screened Poisson ing On Line 1.
equation. Morel, J.-M., A.-B. Petro, & C. Sbert (2014). Screened
The model we propose involves some parameters Poisson equation for image contrast enhancement.
that are tuned and fixed manually. In the future Image Processing On Line 4, 16–29.
Yin, W., S. Osher, D. Goldfarb, & J. Darbon (2008).
an automatic or self-adapting method for choos- Bregman iterative algorithms for L1-minimization
ing these parameters will be studied. Moreover an with applications to compressed sensing. SIAM
additional analysis concerning the comparison of Journal on Imaging Sciences 1(1), 143–168.
the results obtained with the present method with
67
ABSTRACT: In spite of the extraordinary progress that has been made by sleep and behavioral sciences
in the last decades, the assessment of sleepiness is one of the remaining challenges. The lack of practical
tools for objective sleepiness testing has become one of the critical barriers to reducing the threats of
increased sleepiness to public health, productivity, and safety. The first aim of the present report was to
examine whether objective sleepiness testing can be facilitated by the quantitative analysis of the resting
Electroencephalographic (EEG) signal recorded at the first minute following eyes closing. The second aim
was to examine applicability of a new spectral EEG index for differentiation of 4 alertness-sleepiness sub-
states (alert, neither, dozy, and sleepy) in participants of two sleep deprivation experiments. In the first
(exploration) study of applicability of the new index, sleepiness of 15 young adults was self-scored every
second hour simultaneously with recording of the resting EEG in the course of 43–61-hour wakefulness.
In another (confirmation) study, one-minute eyes closed EEG recordings were obtained with a 3-hour
interval during 24-hour wakefulness of 130 adolescents and adults. The sleepiness testing was based on
calculation of Spectral Drowsy Component Score (SDCS). To compute SDCS, 16 (1–16 Hz) single-Hz
powers of the EEG spectra were obtained for the first one-minute interval of eyes closed EEG signal, log-
transformed, ipsatized, weighted, and summed. The 16 weights were the correlation coefficients between
the time course of self-scored sleepiness and the time courses of single-Hz powers. The results indicate that
SDCS provided reliable criteria for distinguishing between four sub-states of alertness-sleepiness. These
criteria performed even better than the traditional criteria based on latency to sleep onset. Moreover,
the occipital SDCS changed their sign exactly at the boundary between neither and dozy sub-states thus
providing yes-or-no criterion of onset of sleepiness state as opposed to the preceding alert state. We con-
cluded that SDCS might be recommended for quick (app. one-minute), direct and objective testing of
sleepiness in absolute terms.
69
70
71
72
73
74
António Cardoso
Materiais e Produtos, CATIM—Centro de Apoio Tecnológico à Indústria Metalomecânica, Porto, Portugal
ABSTRACT: The insertion of probes in the human ear or any other technique involving contact ther-
mometers, or the use of the inner canthus for temperature measurement, using infrared thermal imagers,
with the purpose of obtaining a reference to determine the core temperature in humans, can be compro-
mised by a wide number of factors that may lead to faulty results.
In this paper, an innovative approach based on the use of a thermographic camera and on the char-
acteristics of the external auditory meatus is proposed. This new approach enables to solve the problems
associated with the previously proposed methods, adding new possibilities namely alowing the precise
measurement of the core temperature in humans in a quick and simple way, without body contact, open-
ing the possibility of setting self-screening stations, working without operators in fully automatic mode.
75
76
77
78
79
80
ABSTRACT: Automatic recognition of sign language will be extremely beneficial for hearing impaired
people, by allowing them to easily communicate with the outside world. Sign language is a non-verbal and
visual language, which evolved detached from the spoken language and it is based on the simultaneous
contribution of both temporal and spatial configurations. Even if it is spread all over the world, sign lan-
guage is not a universal language and every country needs to build its own automatic system. This paper
challenges the issue of the development of automatic video-based Turkish Sign Language (TSL) recogni-
tion systems; it presents two computer vision-based automatic systems for TSL recognition and compares
their performances. Since we are using a newly released TSL database, this paper has also the side-effect
to divulge the database to the research community.
81
82
4.1 Reading and preprocessing Figure 2. A sample gray frame before (a) and after
(b) moving the useless background information.
We used all the 1700 videos of section 1, the aver-
age number of frames per video is 43, their initial
size is 640 × 480. Due to out of memory problems, successive frames is to be able to determine the area
it has been necessary to extract and memorize only where the motion of hands is significant because,
the relevant information already during the read- if the values of some pixels change from one frame
ing step. to another, these form a lighter area within the dif-
First of all, RGB frames were converted into ference of images, IDi.
gray frames, after that, a naive cropping method Subsequently, the difference image is converted
was developed and applied on the frames of each into a binary image by a threshold function; the
video. The aim of this cropping is to eliminate value of the threshold is database depended and it
useless background information, and keep only has been fixed, manually, to 50. Figure 3 shows the
the player’s body figure; this results in frames of thresholded image of differences.
size 361 × 427. The input and output frames of Finally, all thresholded images of differences are
naive cropping function can be seen in Figure 2. summed up to calculate a spatial-temporal feature,
named by Abbas et al. the accumulated motion
4.2 Feature extraction image:
Every two consecutive frames in a video produce n −1
an ‘image of differences’, which is calculated using Avi ∑ i 1 Ith i
83
c1 c2 c3 c4 c5 c6
c1 1 0 0 0 0 0
c2 0 4 0 0 0 0
c3 0 0 3 1 0 0
c4 0 0 1 3 0 0
c5 0 0 0 0 5 0
c6 0 0 0 0 0 3
4.3 Classification
Table 4. Accuracy table using spatial-
We used the K-fold cross validation method to temporal features.
separate all samples into training and test sets;
we fixed K = 3, and we divided the data set into Fold 1 Fold 2 Fold 3
three nearly equal subsets. Since the algorithm uses
random selection to assign samples to folders, the Trial 1 84.30 82.54 84.63
whole process is repeated for 5 times and the given Trial 2 82.54 80.78 83.04
performance is the average of 3 × 5 = 15 trials. Trial 3 79.72 85.89 80.57
We used the Euclidean distance to built a Trial 4 83,95 84.10 81.48
Distancei,j matrix of dimensions M × N, where M Trial 5 82.54 80.92 78.48
is the number of test samples and N is the number
of training samples:
Table 5. Accuracy table using frequency
Distancei j T ri − TrainVectorj )2
(TestVector domain features.
84
85
ABSTRACT: We propose a new variational approach for the restoration of images simultaneously cor-
rupted by blur and impulse noise. Promising results have been obtained by the 1-TV variational model
which contains a Total Variation (TV) regularization term and a nonsmooth convex 1-norm data-fidelity
term. We introduce a sparsity-promoting nonconvex data-fidelity term which performs better but makes
the problem more difficult to handle. The main contribution of this paper is to develop a numerical algo-
rithm based on the Alternating Direction Method of Multipliers (ADMM) strategy and on the use of
proximal operators. Preliminary experiments are presented which strongly indicate that using nonconvex
versus convex fidelities holds the potential for more accurate restorations at a modest computational
extra-cost.
87
d2
In the following, we denote the sets of non-negative
and positive real numbers as R + : { ∈ R : ≥ } { ∗ ∗ ∗
}← in ∑ { (r
g min (ri a)) || i || }
u ,t ,r
and R∗+ : { ∈ R : > }, respectively. To induce i =1 (3)
sparsity of the residual Ku − g according to the s.t.: t = Du, r = Ku
K − g,
proposed restoration model (2), we consider 2 2 2
parameterized nonconvex penalty functions where D : ( Dh ;DDv ) ∈ R 2 d d , Dh Dv ∈ R d are lin-
φ ( ) : R → R, with parameter a ∈ R∗+, satisfying ear operators representing the discretization of the
the following assumptions: first-order horizontal and vertical partial derivatives,
φ( ) 1 | | ⎛ 1 + 2 | x |⎞ π
a log(1 + a | x |) 1 + | x |//2 atan −
⎝ 3 ⎠ 6
a 3 /2
φ( ) 0 0 0
φ ′( ±
) ±1 ±1 ±1
±
φ ″( ) −a −a −a
lim φ ( ) +∞ 2
x→±∞ a 2π 3
a 9
88
89
⎧
if | zik | ≤ 1/β r y∗ + ηφ
φ ′ ( y∗ ; a ) | b | = 0 , 0 < y∗ |b|. (25)
k +1 ⎪0
ri = ⎨ , (18)
⎪sign( z k ) r∗ if | zik | > 1/β r
⎩ i Moreover, in case that | b | > η , the Newton-
Raphson iterative scheme defined by:
where r∗ denotes the (unique) solution of the fol-
lowing constrained nonlinear equation: y0 b
y kφ y k ;;aa φ ′( y k ;aa ) b/η
1∗ y k +1 = , (26)
r + φ ′(r∗; a ) − | zik | = 0 , 0 < r∗ < | zik | . (19) φ ″( y k ;aa ) + 1 / η
βr
According to (26), the solution r∗ of (19) can be converges with quadratic rate to the unique solu-
obtained as the limit point of the following (quad- tion y∗ of the constrained nonlinear equation
ratically convergent) Newton-Raphson iteration: in (25).
Proof First, we notice that for any penalty
function φ satisfying assumptions A1)–A6), the
r 0 zik
proximity map proxηφ in (23) has the following
r kφ r k ;;a − φ ′(r k ;a ) + zik β r properties:
r k +1 = . (20)
φ ″(r k ;a ) + β r
proxηφ ( ) 0 proxηφ ( ) = − proxηφ ( ) ∀ b ∈ R , (27)
For what concerns computational efficiency of
(18)–(20), first we notice that iterations (20) has to which follow immediately from the definitions in
be carried out only for the subset of indexes i for (21) and (23). Since according to (27) proxηφ is a
which the quantity | k | is above the threshold odd function, it follows that:
in (18). Then, the quantity zik β r in the numera-
tor of (20) is constant during iterations and can proxηφ ( ) sign( ) proxηφ (| |)) b ∈R . (28)
be precomputed. Last but not the least, very few
iterations (20) are sufficient to compute accurate
estimates of r∗ , as it will be demonstrated in the Therefore, in the rest of the proof we focus on
experimental section. the case b ∈ R∗+ and then apply (28) to derive the
Proposition 1. Let φ : R → R be a function sat- expression of proxηφ for b ∈R .
isfying assumptions A1)–A6) and a,η ∈ R∗+, b ∈R For any a,η ∈ R∗+ satisfying condition (22) and
be given constants. Then, the function f : R → R any b ∈ R∗+ , the function f in (21) is strictly convex
defined as and, in particular, has the following properties:
90
91
92
93
R.P. Tornow
University of Erlangen–Nuremberg, Department of Ophthalmology, Erlangen, Germany
ABSTRACT: Automatic and precise segmentation of retinal blood vessels can help in computer aided
assessment of various retinal diseases, especially diseases related to cardiovascular system or glaucoma.
Hence, an accurate detection of retinal vascular structures is one of the most addressed topic in the field
of retinal image processing today. Most of the methods are designed directly for utilization with a special
dataset only or have problems to segment blurry and noisy images. In this study, we showed that com-
bination of three standard approaches—matched filtering, Hessian-based approach, and morphological
processing, together with support vector machine classification technique, can be satisfactorily used as a
universal segmentation approach in retinal images acquired using entirely different devices.
95
MFR
FRk ,ϕ I ∗ hk ,ϕ . (1)
OMFFR
k ,ϕ
{MFR
M R }. k (2)
96
Bhat ( I ) = (I
(I S
SE ) I , (5)
97
4.2 Segmentation of static fundus images ground truth segmentations are available, quanti-
tative evaluation of the proposed method has been
In the second step, the proposed method was tested
performed via computation of SE, SP, and ACC
on static fundus images in HRF database—dataset
parameters (Table 2). Values of other authors are
A (Fig. 4). The HRF database contains three sets
also presented to compare the proposed approach
of fundus images: one of healthy retinas, one of
with recent state-of-the-art methods, which were
glaucomatous, and one of Diabetic Retinopathy
evaluated on HRF dataset. It is shown the proposed
(DR) retinas. Each set is composed of 15 color
method is well comparable with other methods.
fundus images with corresponding gold standard
Next, the proposed approach was tested on
segmentation made by a human observer. Since
static patient fundus image dataset—dataset B.
(Fig. 5). The dataset contains 45 images of patients
Table 1. Classification performance. with various diseases—DR, glaucoma, and mac-
ular degeneration. In this case, the method was
Evaluation parameters Average results evaluated only qualitatively, since no gold standard
segmentation is available. The original method was
SE = TP/(TP + FN) 96.86 ± 1.66% developed especially for images of HRF database.
SP = TN/(TN + FP) 98.76 ± 0.89% However, the results revealed the utilization for
ACC = (TP + TN)/ 97.78 ± 0.81% other dataset is possible.
(TP + FN + TN + FP)
98
99
100
Vladimir A. Krylov
Department of Electrical, Electronic and Telecommunication Engineering and Naval Architecture,
University of Genoa, Genoa, Italy
ABSTRACT: In this paper we address the problem of extracting curvilinear structures from images
in the presence of Poisson-distributed noise that is typically found in low-quality, low-contrast data, as
well as in numerous medical imaging modalities. A contrast-robust phase congruency estimation method
is proposed such that the optimal range of scales is selected in an adaptive manner for each location in
the image. The adaptive regime is driven by a statistically well-principled variance test statistic which is
based upon a local measure of the dispersion index. Experimental validation confirms that the adap-
tive scheme delivers superior line extraction results on mammographic imagery when compared to other
recent attempts including the non-adaptive phase congruency.
101
102
103
104
105
106
J.F. Gomes
Institute of Biology and Institute of Computing, University of Campinas, Campinas, Brazil
C.T.N. Suzuki
Immunocamp Science and Technology LTDA, Brazil
ABSTRACT: Intestinal parasitic infections can cause serious health problems, especially in children
and immunodeficient adults. In order to make the diagnosis of intestinal parasites fast and effective, we
have developed an automated system based on optical microscopy image analysis. This work presents a
deep learning approach to discover more effective parasite image features from a small training set. We
also discuss how to prepare the training set in order to cope with object scale and pose variations. By
using random kernels, our approach considerably simplifies the learning task of a suitable convolutional
network architecture for feature extraction. The results demonstrate significant accuracy gains in classifi-
cation of the 15 most common species of human intestinal parasites in Brazil with respect to our previous
state-of-the-art solution.
107
3 CONVOLUTIONAL NETWORK
108
3.2 Activation
The activation
operation simply creates an image
Kˆ = ( J ) (a sparse code since we use zero mean
and unit norm kernels) by K i ( p ) = max( Ji ( p ), 0 ),
where p DJ are pixels of the image, and
i { , , , n} are the image bands.
Mi ( p)
Oi ( p ) = (1)
∑ j = ∑ ∀q∈C ( p ) M j (q )M j (q )
n
109
Balanced Unbalanced
The experiments used a dataset with 16,437 objects
(segmented images with their corresponding class Protozoa 0.8288 ± 0.0150 0.8171 ± 0.0190
as confirmed by experts), including examples from Eggs 0.8994 ± 0.0212 0.8310 ± 0.0358
the 15 most common species of human intestinal Larvae 0.8735 ± 0.0778 0.8688 ± 0.0240
parasites in Brazil and impurities.
110
111
112
D. Veiga
Enermeter, Parque Industrial de Celeirós, Lugar de Gaião, Braga, Portugal
University of Minho, Centro Algoritmi, Azurém, Guimarães, Portugal
N. Martins
Centro de Computação Gráfica, Azurém, Guimarães, Portugal
Enermeter, Parque Industrial de Celeirós, Lugar de Gaião, Braga, Portugal
J. Monteiro
University of Minho, Centro Algoritmi, Azurém, Guimarães, Portugal
ABSTRACT: The appearance of microaneurysms is the first clinical sign in Diabetic retinopathy. These
lesions appear as small red dots in color fundus images. Their detection in a preliminary phase could
prevent disease progression which leads to blindness. This paper addresses the problem of automatic
microaneurysms detection in color fundus images. Laws texture-energy measures were selected as feature
extractor and descriptor. The suitability of these measures were investigated by testing several kernel sizes
followed by support vector machine classification.
Tests were made on the LaTIM (laboratoire de Traitement de l’Information Médicale) database and
results are evaluated in terms of the Free-Response Receiver-Operating Characteristic (FROC) curve. The
best performance was achieved with 31 × 31 kernel size.
To our knowledge, this is the first time that Laws local texture-energy transform is used to segment
microaneurysms. Results are promising and a next step to analyse each microaneurysm candidate is needed.
113
114
with,
KernelSize × 3
w= − 1, (2)
2
115
116
ACKNOWLEDGMENTS
117
118
ABSTRACT: Digital Breast Tomosynthesis (DBT) is a new technique that takes several projection
images with different angles and combines them to reconstruct 3D image using approximately the same
dose of two mammograms. Currently, handling noise is a major problem in DBT. Total Variation (TV)
based minimization algorithms have achieved great success due to its virtue of preserving edges while
reducing image noise. Because TV denoising depends on noise model, we tested an algorithm for Poisson
noise in real DBT images.
For unfiltered images we have a Contrast to Noise Ratio (CNR), a spatial resolution and TV of 1.110,
0.483 mm and 2.319E+08, respectively. After filter application, the CNR improves up to 5.3%, spatial
resolution value doesn’t suffer degradation (variations up to 0.4%) and TV decrease up to 40.4%.
We conclude that application of TV denoising to DBT images is a promising technique and allows a
wide field of research in the future.
119
FWHM of a Gaussian curve fitted to the Point Here Δx+, Δx−, Δy+ and Δy− denote forward and
Spread Function of one microcalcification was backward one-sided differences in x and y direc-
considered as an indicator of spatial resolution. tions, respectively; and Δx0 and Δy0 indicate central
CNR and FWHM values were calculated before difference in x and y directions, respectively. This
and after filter application for the two scenarios. formulation is valid only for interior points, and
Boundary conditions were defined as in Ref Rudin
et al. 1992. To solve the problem with these bound-
2.3 Formulation of TV minimization algorithm ary conditions for u, a routine was implemented
With this TV minimization algorithm, starting and integrated in the image reconstruction soft-
from a constrained minimization problem, it is ware implemented in IDL.
obtained an unconstrained problem that is com- Δt should be a small number that forces the
posed by a fidelity term corresponding to the algorithm to make small jumps. Its optimal value
assumed noise model and a regularization term depends on the function. If it is too large, the algo-
that is the TV function. rithm will diverge and if it is too small, it will take
Our unconstrained TV minimization problem, a long time to converge.
based on ROF model but for Poisson noise, can be λ is the Lagrange multiplier, also called as regu-
formulated as Equation 2 (Le et al. 2007, Sawatzky larization or fitting parameter since it controls
et al. 2009): image regularization, between removing noise and
preserving edges.
In a first phase, results for Δt ranging between
min { (u)) ( l u )}
f ln (2)
u 0.05 and 0.25 segs, with a fixed λ = 60 were ana-
lyzed (from previous studies we know that this is an
where f and u are the original and denoised image, acceptable value for this level of noise. See (Mota
respectively, and TV(u) is the TV of denoised et al., 2015)). It is considered a good Δt the one that
image. TV values were determined according to is effective in minimizing the TV and that can achieve
Equation 3, considering the selected 2D slice. convergence with a reasonable number of iterations.
In the next stage, taking into account Δt values cho-
m n sen, different λ values were compared. Finally, with
TV (u ) ∑∑ ( x ui , j )
2
+( y ui j )
2
(3) the most appropriate Δt and λ, filtered images were
i =1 j =1 quantitatively and quantitatively analyzed.
120
4 DISCUSSION
Figure 3. CNR values for phantom unfiltered and fil-
We presented here improved results that can be opti-
tered reconstructed images plotted as a function of itera-
tion number, for λ = 60 and Δt = 0.25 segs. mized in the sense of achieve an objective correla-
tion between the parameters and image noise level.
The combination Δ = 0.25 segs and λ = 60 works well
TV minimization. The results obtained for differ- and can achieve convergence in iteration number 10.
ent λ values, taking into account Δt = 0.25 are pre- λ = 10 doesn’t ensure the fidelity term has enough
sented in Figure 2. weight to keep the fundamental information of the
For the image quantitative and qualitative anal- initial image and λ between 30 and 50 diverge. In
ysis, λ = 60 was the value chosen and the results this way those values were automatically excluded.
obtained for CNR and FWHM are shown in From λ = 60 to 120 we choose λ = 60 because it
Figure 3 and Figure 4. For a visual inspection images allows the better CNR in 10 iterations.
of a mass and a cluster of microcalcifications of the We can see the a decrease in image TV of 40.4%
phantom are presented in Figure 5 and Figure 6. as well as the fact that this method improves CNR
121
122
A. El-Rafei
Engineering Physics and Mathematics Department, Faculty of Engineering, Ain Shams University, Cairo, Egypt
S. El-Ramly
Electronics and Communication Engineering Department, Faculty of Engineering, Ain Shams University,
Cairo, Egypt
G. Michelson
Department of Ophthalmology, Friedrich-Alexander University, Erlangen-Nuremberg, Germany
J. Hornegger
Pattern Recognition Laboratory, Department of Computer Science, Friedrich-Alexander University,
Erlangen-Nuremberg, Germany
ABSTRACT: Retinal Image Quality Assessment (RIQA) is an essential preliminarily step in Automatic
Retinal Screening Systems (ARSS) to avoid misdiagnosis of retinal disease. In this work, a no-reference
wavelet based RIQA algorithm suitable for real-time ARSS is introduced. Retinal images are classified
into good and bad quality images using two proposed metrics. These measures are extracted from the
wavelet decomposition subbands of the image to reflect sharpness and contrast. Wavelet level and regional
analyses are performed to determine the most relevant levels and retinal region to be used in the RIQA.
The algorithm is tested on two publicly available datasets with different resolutions. Area under the ROC
curve (AUC) of 1 and 0.985 are obtained for the low and high resolution datasets respectively with run-
time ranging between 0.1 and 2 seconds depending on image resolution. Comparison with other RIQA
algorithms shows superiority of the proposed algorithm in both performance and time.
123
124
125
Table 1. Area under the receiver operating characteristic Table 3. Comparison of the proposed algorithm with
curve (AUC) for the DRIMDB database. existing RIQA techniques.
L1 L2 L3 L4 L5 Proposed Fasih Thomas Davis
Algorithm 2014 2013 2009
Global 0.973 0.987 0.998 0.981 0.876
Local 0.963 0.995 0.998 0.998 0.895 AUC 0.985 0.796 0. 883 0.852
126
127
128
129
ABSTRACT: Distinguishing vessels in a retinal image is a solid first step towards a computer-aided
diagnosis of several medical conditions. Mathematical morphology provides simple yet powerful trans-
formations to process images. Extended with fuzzy techniques, it achieves the ability to handle the uncer-
tainty present in medical imaging, both in the images and in the expert diagnosis. Within this paradigm,
we propose a method that leverages the geometric and photometric features that characterize vessels as
such. It is tested with two databases, DRIVE and STARE. The first promising results outperform some
of the state of the art methods with which it was compared, positioning it among the most competitive
techniques.
131
132
3 EXPERIMENTATION RESULTS
133
⎧ y λ ), if ( x y ) [ λ ]2 ,
λ
TMT ( x y ) = ⎨max( x
⎩max( x y ), otherwise .
134
We remark that only around 10%–15% are “ves- Table 1. Comparison of the average of the F1 Score and
sel” pixels, so the trivial estimation that considers Accuracy of all possible images in both databases, where
each pixel as “non-vessel” would obtain an Accu- both the results and the ground truth were available.
racy of 0.85–0.90.
DRIVE STARE
A statistical test is needed to compare two meth-
ods with each other. Given n images, each method F1 Acc F1 Acc
Mi provides a series of results ri (ri( ) , …, ri( n ) ) ,
which can be interpreted as samples of a probabil- 2nd expert 0.788 0.964 0.740 0.952
ity distribution. We use the Wilcoxon signed-rank Martínez [10] 0.697 0.950 – –
test to contrast whether the mean of two paired Niemeijer [11] 0.750 0.960 – –
distributions differ or not. It is robust against non Hoover [8] – – 0.651 0.946
normal data, which is our case. The significance Chanwim. [3] 0.670 0.942 0.638 0.935
level is set to α = 0.05. Staal [13] 0.764 0.962 – –
For each database, the method has been executed Soares [12] 0.748 0.960 0.698 0.957
and tested against the ground truth provided. In Our method 0.733 0.958 0.689 0.960
the DRIVE database, the 20 test images have been
considered for experimenting purposes, using the
other 20 images as training samples when needed.
The 20 manually segmented images of the STARE of our method. However, with the STARE data-
database are considered, using the strategy leave- base, which presents images with more noise and
one-out1 when training is required. The mean of lower contrast, our method obtains extremely good
the metrics of all samples is showed in Table 1. segmentations, with a F1 Score only improved—
Along the 20 test images of the DRIVE data- without statistical significance—by Soares et al. It
base, the method provides a mean F1 Score of also yields the best Accuracy result, which is signifi-
0.733 and an Accuracy of 0.958. The 20 STARE catively better than all the other methods except for
images yield an average F1 Score of 0.689 and an the results by Soares et al. Surprisingly, the Accu-
Accuracy of 0.960. racy result is even better than that of the second
As we can see in Table 1, the metrics of our expert segmentation—being the only method pro-
method are in all cases similar to the best ones. viding a better result than the second expert. This
With respect to the DRIVE database, our method is due to the fact that our method and the second
is statistically better than the one presented by ground truth differ from the first manual segmenta-
Chanwimaluang, although the method presented by tion in different ways: the F1 Score of the latter is
Staal et al. has significative better results than those clearly above the one of the former.
The methods by Chanwimaluangand Guoliang
[3] and Soares [12] were executed using the source
code published by their authors. Hoover et al.
1
Leave-one-out consists on computing each estimation [8] published the segmentations of their method
by training a new classifier with the remaining samples. with the STARE database, and Staal et al. did so
135
136
ABSTRACT: The distinction between a transmural and subendocardial infarction is important for
both therapeutic and prognostic reasons. To investigate the possibility to differentiate these two cases,
a numerical model of a LV phantom with uniform wall and wall with inclusion simulating subendocar-
dial infarction was created. Using meshes of these phantoms synthetic ultrasonic data was generated
using the Field II software. Block matching based estimation of displacements was carried out using
the 2-dimensional cross correlation coefficient. Eulerian strain tensor and principal strain components
were obtained using displacement gradient tensors. The components of the principal strain, particularly
the first component, show strong differences in the affected wall as compared with the homogenous wall.
The methodology has the potential for identification of non-transmural changes in the left ventricle wall.
137
138
Figure 2. Maps of the first component of the principal strain, LAX view, data without scan conversion. FE simula-
tion reference data—homogenous model (top-left), model with insertion (top-right). Estimation results—homogenous
model (bottom-left), model with insertion (bottom-right). The effect of the non-transmural insertion can be seen in
the (top-right) and (bottom-right) parts. Abscissa is scaled in scan line number, ordinate in RF sample number, strain
nondimensionalised [%].
139
Figure 3. Maps of the second component of the principal strain, LAX view, data without scan conversion.
Organization as in Figure 2.
140
REFERENCES
141
142
N. Martins
Centro de Computação Gráfica, Azurém, Guimarães, Portugal
Enermeter, Parque Industrial de Celeirós, Lugar de Gaião, Braga, Portugal
N. Alves
Centro de Computação Gráfica, Azurém, Guimarães, Portugal
M. Delgado
Dr Campos Costa, Imagiologia Clínica, Rua de Avis, Porto, Portugal
ABSTRACT: Breast Cancer is one of the most common types of cancer in Women. Mammography
is widely used as a screening method because it can detect pathologies in an early stage. There are some
Computer Aided Diagnosis (CAD) systems, but their performance can be improved with new and bet-
ter image processing algorithms. In this paper, we present a new microcalcification detection method for
Full Field Digital Mammograms (FFDM), based on the wavelet transform, Expectation Maximization
with Markov Random Fields bootstrap and a final SVM classifier. Tests were made in a public database
(InBreast) and results shown that our approach achieved better False Positive (FP) rates and sensitivities
compared to the literature. Features were also presented and the problems with the train addressed.
143
{A s+ }W ( As )
Dsh+ Dsv+ Dsd+ = DWT (1)
{As } ⎛ h v d⎞
⎝ As , Ds , Ds , Ds ⎠ (2) Figure 1. Microcalcifications enhancement result.
144
145
ACKNOWLEDGMENTS
REFERENCES
Figure 4. FROC of the proposed method. A. Bria, N.K. (2014). Learning from unbalanced data:
A cascade-based approach for detecting clustered micro-
calcifications. Medical Image Analysis 18, 241–252.
Table 2. Operation points of some published algorithms. Afsaneh Jalalian, S.B.M. (2013). Computer-aided detec-
tion/diagnosis of breast cancer in mammography
Method Sensitivity (%) FP and ultrasound: a review computer-aided detection/
diagnosis of breast cancer in mammography and
Fixed threshold 34.5 164 ultrasound: a review. ELSEVIER—Clinical Imaging.
Outlier detection 45.8 60 Domingues, I., C.J. (2011). Inbreast: Toward a full-field
Mathematical morphology 40.3 225 digital mammographic database inbreast: Toward a
Bayesian surprise 1 60.3 108 full-field digital mammographic database. Technical
Report: Acad Radiology.
146
147
N. Strisciuglio
Department of Computer Engineering and Electrical Engineering and Applied Mathematics, University of Salerno,
Salerno, Italy
Johann Bernoulli Institute for Mathematics and Computer Science, University of Groningen, Groningen, The Netherlands
M. Vento
Department of Computer Engineering and Electrical Engineering and Applied Mathematics, University of Salerno,
Salerno, Italy
G. Azzopardi
Department of Intelligent Computer Systems, University of Malta, Msida, Malta
Johann Bernoulli Institute for Mathematics and Computer Science, University of Groningen, Groningen, The Netherlands
N. Petkov
Johann Bernoulli Institute for Mathematics and Computer Science, University of Groningen, Groningen, The Netherlands
ABSTRACT: Retinal imaging has gained particular popularity as it provides an opportunity to diag-
nose various medical pathologies in a non-invasive way. One of the basic and very important steps in
the analysis of such images is the delineation of the vessel tree from the background. Such segmentation
facilitates the investigation of the morphological characteristics of the vessel tree and the analysis of any
lesions in the background, which are both indicators for various pathologies. We propose a novel method
called B-COSFIRE for the delineation of the vessel tree. It is based on the classic COSFIRE approach,
which is a trainable nonlinear filtering method. RE filter to be configured by the automatic analysis of any
given vessel-like pattern. The responses of a B-COSFIRE filter is achieved by combining the responses of
difference-of-Gaussians filters whose areas of support are determined in an automatic configuration step.
We configure two types of B-COSFIRE filters, one that responds selectively along vessels and another that
is selective to vessel endings. The segmentation of the vessel tree is achieved by summing up the response
maps of both types of filters followed by thresholding. We demonstrate high effectiveness of the proposed
approach by performing experiments on four public data sets, namely DRIVE, STARE, CHASE_DB1 and
HRF. The delineation approach that we propose also has lower time complexity than existing methods.
150
Figure 4. Examples of retinal fundus images (first column) from the four concerned data sets (DRIVE, STARE,
CHASE DB1 and HRF) together with the manually segmented ground truth (second column). The output of the
delineation algorithm (third column) is thresholded in order to obtain the final segmented vessel tree (fourth column).
151
152
153
DRIVE STARE
Unsupervised B-COSFIRE 0.7655 0.9704 0.9614 0.9442 0.7716 0.9701 0.9563 0.9497
(Mendonca and 0.7344 0.9764 – 0.9463 0.6996 0.9730 – 0.9479
Campilho 2006)
(Martinez-Pérez et al. 2007) 0.7246 0.9655 – 0.9344 0.7506 0.9569 – 0.9410
(Al-Rawi et al. 2007) – – 0.9435 0.9535 – – 0.9467 0.9090
(Ricci and Perfetti 2007) – – 0.9558 0.9563 – – 0.9602 0.9584
(Lam et al. 2010) – – 0.9614 0.9472 – – 0.9739 0.9567
Supervised (Niemeijer et al. 2004) – – 0.9294 0.9416 – – – –
(Staal et al. 2004) – – 0.9520 0.9441 – – 0.9614 0.9516
(Soares et al. 2006) 0.7332 0.9782 0.9614 0.9466 0.7207 0.9747 0.9671 0.9480
(Ricci and Perfetti 2007) – – 0.9633 0.9595 – – 0.9680 0.9646
(Marin et al. 2011) 0.7067 0.9801 0.9588 0.9452 0.6944 0.9819 0.9769 0.9526
(Fraz et al. 2012) 0.7406 0.9807 0.9747 0.9480 0.7548 0.9763 0.9768 0.9534
CHASE_DB1 HRF
Unsup. B-COSFIRE 0.7585 0.9587 0.9487 0.9387 0.7511 0.9745 0.9510 0.9644
(Odstrcilik et al. 2013) – – – – 0.7740 0.9669 0.9493 0.9668
Sup. (Fraz et al. 2012) 0.7224 0.9711 0.9712 0.9469 – – – –
154
155
H. Rieiro
Universidade de Vigo, Vigo, Spain
Barrow Neurological Institute, Phoenix, AZ, USA
157
Figure 1. Input image and result of applying the crease- Figure 2. Example of keypoint localization in a two-
ness transformation. The creaseness transformation has photon microscopy section. Grey dots over imposed on
a two-fold effect, blurring the image and removing most blood vessels mark the detected keypoints, which tend to
of the background noise. fall in the points of vessel branching.
158
159
160
161
ABSTRACT: Most of biomedical images are monochromatic and thus their assessment may be sig-
nificantly improved by means of the colorization techniques. Such colorized versions of images not only
benefit from more pleasing appearance, but also some fine details are easily noticeable, what may substan-
tially improve the diagnosis. In this article we present a novel method of colorization, which is based on
a combination of distance transformation and modified bilateral filtering approach. Tunable parameters
are introduced to allow for precise adjustment of final colorization outcomes. Our proposed algorithm
enables a simple and efficient region highlighting and interactive segmentation of biomedical images. It
may be also employed for the educational and presentation purposes. We show several examples of appli-
cations of our algorithm using various biomedical images obtained by ultrasonography, transmission
electron microscopy, computer tomography, magnetic resonance and infrared imaging.
163
2 ALGORITHM
N2
I0 ∑ I k wk (2)
k =1
164
⎧ ⎛ ( I I )2 ⎞ ⎫
M0 a ⎨M k exp ⎜ − 0 2k ⎟ ⎬ , (4)
k , ,N 2
⎩ ⎝ 2σ ⎠⎭
165
166
167
5 SUMMARY
ACKNOWLEDGMENT
168
169
J.C.L. Stutz
Universidade do Estado do Rio de Janeiro—Instituto Politécnico, Nova Friburgo, Rio de Janeiro, Brasil
Universidade Estácio de Sá, Nova Friburgo, Rio de Janeiro, Brasil
ABSTRACT: The use of computed microtomography (μCT) has revolutionized many areas of research,
such as non-invasive, fast and high precision techniques, which allows immediate visualization of internal
structures without any risk to the object of study. Its usefulness is undeniable for medicinal purposes, is also
important for purposes of zoological research and paleontological. In this work, we used the μCT to inves-
tigate the internal structure of bones from mammalian and poultry. We studied the bones of rats and quail.
Through of microtomography images, we observed that the bones of the poultry have a bony structure in
the form of a trellis that is not present in bones of mammals. These trellises bony is an evolutionary adapta-
tion that allowed the bones of the birds they become longer and lighter maintaining its strength. It was also
observed that the percentage of the trabecular area in poultry is almost half of that observed in mammals.
The results obtained validate the use of μCT as a technique that allows the study of bone structures in small
samples. Enabling explore the morphological differences between the bones of those species.
171
172
4 CONCLUSION
REFERENCES
173
ABSTRACT: In this work we propose a technique for the modellisation of biological structures for its
applications in Finite Elements Methods (FEM). The key of this method consists on deforming the mesh
of a generic structure in order to obtain a mesh of a specific one. The input data are both the mesh of the
generic structure and the data collected by Computerised Tomography (CT) scan, Magnetic Resonance
Imaging (MRI), etc... This technique involves genetic algorithms for adjusting the nodes of the input
mesh to the collected points, Radial Basis Functions (RBF) to deform the input mesh, and optimisation
of the final mesh.
1 INTRODUCTION
175
Figure 2. Skeleton deformation stages. a) Model input mesh; b) Patient input mesh; c) Superposed skeletons of model
(blue) and patient (red) meshes; d) Deformed model (blue) skeleton to fit the patient (red) skeleton; e) Deformed model
surface mesh.
176
3 RESULTS
177
Table 2. Evolution of the Hausdorff distance Table 3. Evolution of the final mesh quality after opti-
as the number of handle points increases. misation and untangling. All case finished with 0 tangled
elements.
Mesh n. handle P. Hausdorff dist.
n. tangled ele. Mean Q. Mean Q.
Case 1 67 0.3833
Case 2 1094 0.3725 Mesh Pre Pre Post
Case 3 1453 0.3179
Case 4 1557 0.3112 Case 1 571 0.568 0.756
Case 2 595 0.566 0.755
Case 3 637 0.557 0.755
Case 4 624 0.559 0.755
4 DISCUSSION
178
179
Helio Pedrini
Institute of Computing, University of Campinas, Campinas-SP, Brazil
ABSTRACT: Several medical imaging techniques have allowed the acquisition of volumetric images
for non-invasive diagnosis purposes, such as Computerized Tomography, Magnetic Resonance Imaging,
X-Ray Microtomography, Positron Emission Tomography and Ultrasonography. A substantial increase
in both data size and image resolution has been possible due to recent technological advances. Interactive
exploration of volumetric data through visualization approaches can be used to enhance internal struc-
tures of interest. This work describes and evaluates a novel projection technique for volume rendering
based on gradient information and viewing direction to generate values of color and opacity from the
data values. Experiments are conducted on several data sets to demonstrate the effectiveness of the pro-
posed technique compared to other projection approaches.
181
Cˆ i = β i Cˆ i −1 + (1 − β i αˆ i −1)α iCi
(4)
3 METHODOLOGY αˆ i = βi αˆ i −1 + (1 − βi αˆ i −1)α i
This work proposes a new projection technique where β i δ i and δ i = max( I i I maxi , 0 ) , such
for volumetric rendering, which is based on gradi- that I maxi is the current maximum intensity value
ent information and viewing direction to produce along the viewing ray and I i is the intensity from
values of color and opacity from the data values. the segment i along the viewing ray.
182
Figure 1. Results for Manix data set for different projection techniques.
183
Figure 2. Results for Chest data set for different projection techniques.
184
tissue covering the jaw bone, which is not visible in on gradient information and viewing direction to
the image generated by Contour MIP. generate a final image.
In Figure 2, it is possible to observe that only Experiments are conducted on several data sets
Contour MIP (Fig. 2e) and Contour MIDA and the results obtained through the proposed method
(Fig. 2f) are able to show the pulmonary bronchi- are compared with other projection approaches.
oles and the grooves on the base of the left lung.
The main difference between both is the depth and
ACKNOWLEDGEMENTS
volumetric information. The depth information
makes it possible to perceive that the bones of the
The authors are grateful to FAPESP, CNPq and
ribs are behind the lungs, not in front of them.
CAPES for the financial support.
In Figure 3, only Contour MIP (Fig. 2e) and
Contour MIDA (Fig. 2f) are able to show the colon
in the images. However, the Contour MIDA allowed REFERENCES
to differentiate the bends of the colon, as well as to
show the internal parts of the pelvis bone. Bruckner, S. & M.E. Gröller (2009, June). Instant Vol-
ume Visualization using Maximum Intensity Differ-
ence Accumulation. Computer Graphics Forum 28(3),
775–782.
5 CONCLUSIONS Csébfalvi, B., L. Mroz, H. Hauser, A. König, & M.E. Gröller
(2001). Fast Visualization of Object Contours by
This work proposed and evaluated a novel projec- Non-Photorealistic Volume Rendering. In Computer
tion technique for medical volume rendering based Graphics Forum, Volume 20, pp. 452–460.
185
186
Sami Dhahbi
Research Team on Intelligent Systems in Imaging and Artificial Vision (SIIVA)—RIADI Laboratory,
Ariana, Tunisia
Michal Kruk
The Faculty of Applied Informatics and Mathematics, Warsaw University of Life Sciences, Warsaw, Poland
Walid Barhoumi
Research Team on Intelligent Systems in Imaging and Artificial Vision (SIIVA)—RIADI Laboratory,
Ariana, Tunisia
Grzegorz Wieczorek
The Faculty of Applied Informatics and Mathematics, Warsaw University of Life Sciences, Warsaw, Poland
Ezzeddine Zagrouba
Research Team on Intelligent Systems in Imaging and Artificial Vision (SIIVA)—RIADI Laboratory,
Ariana, Tunisia
ABSTRACT: The paper presents improved method for breast cancer diagnosis. The previous method
has been presented in Dhahbi et al. 2015. The suggested improvements regards applying more diagnostic
features such as wavelet packet decomposition, Hilbert matrix, fractal texture features, etc. Moreover we
investigated several classifiers such as Random Forest, Support Vector Machine and Decision Tree. In this
paper larger database (number of images/trials) has been used and reached better accuracy.
187
2 DATABASES
188
189
log (N )
d = lim
ε →∞ log( )
190
191
M.E. Fantacci
Dipartimento di Fisica, Universita’ degli Studi di Pisa, Italy
Istituto Nazionale di Fisica Nucleare Sezione di Pisa, Italy
A. Retico
Istituto Nazionale di Fisica Nucleare Sezione di Pisa, Italy
E. Lopez Torres
CEADEN, Havana, Cuba
Istituto Nazionale di Fisica Nucleare Sezione di Torino, Italy
ABSTRACT: M5L, a Web-based fully automated Computer-Aided Detection (CAD) system for the
automated detection of lung nodules in thoracic Computed Tomography (CT), is based on a multi-thread
analysis with two independent CAD subsystems, the lung Channeler Ant Model (lungCAM) and the
Voxel-Based Neural Analysis (VBNA), and on the combination of their results. The M5L performance,
extensively validated on 1043 CT scans from 3 independent data-sets, including the full LIDC/IDRI data-
base, is homogeneous across the databases: the sensitivity is about 0.8 at 6–8 False Positive findings per
scan, despite the different annotation criteria and acquisition and reconstruction conditions. In order
to make CAD algorithms and their results available to users without requiring the installation of CAD
software or new hardware for CAD computations, the project has proposed a Cloud SaaS (Software as a
Service) approach composed by three main building blocks: a front-end web which handles the workflow,
image upload, CAD results notification and direct on-line annotation of the exam by the radiologist; the
OpenNebula-based cloud Iaas (Infrastructure as a Service) batch farm allocates virtual computing and
storage resources; the M5L CAD provides the nodule detection functionality.
193
194
195
196
197
198
199
Giuseppe Palestra
Department of Computer Science, University of Bari, Italy
Liliana Ruta
Department of Developmental Neuroscience, Stella Maris Scientific Institute, Calambrone, Pisa, Italy
ABSTRACT: Assistive technology is a generic system that is used to increase, help or improve the func-
tional capabilities of people with disability. Recently, its employment has generated innovative solutions
also in the field of Autism Spectrum Disorder (ASD), where it is extremely challenging to obtain feed-
back or to extract meaningful data. In this work, a study about the possibility to understand the visual
exploration in children with ASD is presented. In order to obtain an automatic evaluation, an algorithm
for free gaze estimation is employed. The proposed gaze estimation method can work without constrains
nor using additional hardware, IR light sources or other intrusive methods. Furthermore, no initial cali-
bration is required. These relaxations of the constraints makes the technique particularly suitable to be
used in the critical context of autism, where the child is certainly not inclined to employ invasive devices.
In particular, the technique is used in a scenario where a closet containing specific toys, that are neatly
disposed from the therapist, is opened to the child. After a brief environment exploration, the child will
freely choose the desired toy that will be subsequently used during therapy. The video acquisition have
been accomplished by a Microsoft Kinect sensor hidden into the closet in order to obtain both RGB and
depth images, that can be processed by the estimation algorithm, therefore computing gaze tracking by
intersection with data coming from the well-known initial disposition of toys. The system has been tested
with children with ASD, allowing to understand their choices and preferences, letting to optimize the toy
disposition for cognitive-behavioural therapy.
201
202
203
Fixation Count 4 3 6
Sequence 5-2-5-7 2-5-6 5-2-3-5-2-1
First Fixation Cell 5 2 5
Selected Toy 7 4 1
Figure 3. An example of toys disposition into the
Most Viewed toy 2 6 1
closet.
204
4 DISCUSSION
205
206
207
Szabolcs Urbán
Institute of Informatics, University of Szeged, Szeged, Hungary
László Ruskó
GE Healthcare, Szeged, Hungary
Antal Nagy
Institute of Informatics, University of Szeged, Szeged, Hungary
ABSTRACT: Tumor contouring is a challenging task for physicians especially when functional as well
as anatomical images are incorporated. This paper presents a semi-automated technique to solve this
problem. The proposed method is a variant of the level-set segmentation where the initial region is based
on the functional image and the speed function combines both images. The presented approach allows
the user to balance between the information of two input images. The method was evaluated on registered
head and neck PET-MRI image pairs using manually defined tumor contours as reference. The algorithm
was tested for various types of tumors using different weights to combine the functional and the anatomi-
cal information. The best results showed good correlation with the reference (3% volume difference and
80% DICE similarity in average).
209
The goal of this work was to develop an interactive In FDG-PET images high uptake values show high
tool that can combine the information of registered glucose metabolism (high intensities), which can be
PET and MRI images to facilitate tumor contour- normal inside some regions (e.g. brain, muscles),
ing in the head and neck region which contains but it usually reflects pathology (tumor, inflam-
various-shaped tumors and the mentioned images mation) in the area of the neck. In the input PET
can contain complementary/contradictoire infor- images the intensities are represented by Standard-
mation. The main idea was to extend a standard ized Uptake Value (SUV). Inside the user-defined
segmentation technique to incorporate functional ROI the maximal (SUV Vmax ) value is calculated and
and anatomical images in a user-friendly way. Our its 40% is used for thresholding the high-uptake
primary goal was to develop a MR-PET tumor region. The largest 3D-connected component of
contouring tool the result of which is consistent the detected region, which can be considered as the
with both input images and the user’s expectations. active part of the tumor, is the basis of the starting
This goal was primarily motivated by discussions region. Due to the lower spatial resolution of the
with physicians working at the field of radiother- functional image, the border of this region cannot
apy planning. They are challenged with defining be always considered as part of the tumor in the
the target volume (i.e. the tumor) when both func- anatomical image (that has much higher spatial
tional and anatomical images are given because resolution). Furthermore the input images were
they have to switch back and forth between the globally registered, which cannot guarantee the
input images when they manually contour the dif- local accuracy of the registration at each tumor.
ferent parts of the tumor. This way they include all To take these observation into account, the initial
regions which may belong to the tumor based on region is defined by the skeleton of the connected
any of the input images. According to our experi- region. The applied 3D skeletonization algorithm
ments with head-and-neck cases fully automated (Lee, Kashyap, & Chu 1994) guarantees the con-
approach is hardly achievable, while the presented nectedness of the result. The generated skeleton
semi-automated technique can facilitate this com- is a thin structure, therefore dilation is applied to
plex task. The proposed method consists of two that. The dilation incorporates the thresholded
main steps. First, an initial region is extracted from uptake region, so that the final skeleton is always
the user-defined Region Of Interest (ROI) based located inside that. The proposed initial region rep-
on the functional image. Then, 3D fastmarching resents the shape of the functionally active region
approach is applied to generate a set of contours, and it is not sensitive for small local registration
from which the user can select one (referred as level) inaccuracy. An example for initial region is shown
that fits both images as well as the user’s expec- in Figure 1a.
tations. The contour propagation is driven by a In order to combine the functional and the ana-
hybrid speed function that incorporates both PET tomical information of the input images a hybrid
and the MRI image intensities such that the weight speed function is defined in the following way.
of the PET (referred as λ) and MRI (referred as First, the intensity range of both PET and MRI
1 − λ ) images can be specified by the user. This way images is normalized to the [0,1] interval, using
the user can balance between the PET and MRI the min-max normalization technique. Since the
information according to the characteristics of the intensity maximum of an MRI images can be
tumor to be contoured. Figure 1 demonstrates the associated with some voxels affected by local imag-
main steps of the algorithm. ing artifacts, we use the largest intensity that is
The segmentation requires a starting region. This higher than 99% of the voxels of the MRI image.
region is the basis of the contour propagation, so it This step is substantial because the input images
must be fully located inside the tumor. In order to involve different information, generated by differ-
get such a region, only the functional image is used. ent acquisition techniques. The functional (PET)
210
1 1
F =λ⋅ I ffn − β ffn + ( − λ)⋅ I an − β aan (1) with the restriction that the speed function can
− −
1+ e
α ffn
1+ e α aan
never change sign. It can defined as follows:
211
212
Figure 4. The method integrated into the MITK framework: the generated red contour inside the defined ROI using
λ = 3 and level = 11.
Cases 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
DSC (%) 74 71 80 83 81 75 83 86 79 78 82 81 81 87 88
λ (%) 60 30 20 30 30 70 30 20 20 50 70 20 30 30 20
Level 96 16 6 16 30 46 23 11 14 26 16 11 13 11 21
Volume (ml) 149.0 8.0 1.9 3.4 23.9 55.0 41.6 3.4 2.2 17.0 1.8 11.7 8.5 8.9 2.8
AVD (%) 25.5 3.5 1.4 0.1 1.4 1.5 1.3 1.3 1.0 1.0 1.7 1.9 1.0 2.9 0.3
213
214
ABSTRACT: In this paper an application of confidence map for bones surface detection in ultrasound
images for Rheumatoid Arthritis is presented. The bone feature descriptor is based on confidence maps,
which are directly related with the likelihood of ultrasound transmission and allows for highly accurate local-
ization of the bone surface. The quality of the confidence map has been significantly improved by applying
the Trimmed Non-Local Means filtering. The results were evaluated on the publicly available MEDUSA
database that contains manually annotated markers of bones surface. This research is a preliminary step in
the development of fully automatic system for the assessment of degree of synovitis inflamation.
215
216
217
218
Figure 8. Some failure examples of the bone detection and corresponding confidence and bone localizer maps.
219
220
S. Osowski
Warsaw University of Technology, Military University of Technology, Warsaw, Poland
R. Koktysz
Military Institute of Medicine, Warsaw, Poland
ABSTRACT: The paper presents the improved system of recognition of the Fuhrman grading in the
analysis of clear-cell renal carcinoma. The previous system was described in (Kruk et al. 2014). The pro-
posed improvement is based on the new segmentation method which gives better results. A redundant
wavelet transformation is used for denoising the image and enhancing the edges in multiple levels of
resolution. The image gradient is estimated using the wavelet transformation. The watershed transform
is then applied to the obtained gradient image, thanks to this the segmented regions that do not satisfy
specific criteria are removed.
221
222
223
Class 1 2 3 4
224
REFERENCES
225
A. Ferreira
ICVS/3B’s—PT Government Associate Laboratory, Braga/Guimarães, Portugal
P. Morais
ICVS/3B’s—PT Government Associate Laboratory, Braga/Guimarães, Portugal
INEGI, Faculdade de Engenharia, Universidade do Porto, Porto, Portugal
S. Queirós
ICVS/3B’s—PT Government Associate Laboratory, Braga/Guimarães, Portugal
Algoritmi Center, School of Engineering, University of Minho, Guimarães, Portugal
F. Veloso
ICVS/3B’s—PT Government Associate Laboratory, Braga/Guimarães, Portugal
N.F. Rodrigues
ICVS/3B’s—PT Government Associate Laboratory, Braga/Guimarães, Portugal
DIGARC, Polytechnic Institute of Cávado and Ave, Barcelos, Portugal
J. Correira-Pinto
ICVS/3B’s—PT Government Associate Laboratory, Braga/Guimarães, Portugal
J.L. Vilaça
ICVS/3B’s—PT Government Associate Laboratory, Braga/Guimarães, Portugal
DIGARC, Polytechnic Institute of Cãvado and Ave, Barcelos, Portugal
ABSTRACT: Pectus Excavatum (PE) is the most common congenital chest wall deformity, affecting
1 in 400 live births. This deformity is commonly corrected using the minimally invasive Nuss procedure,
where a bar is positioned under the sternum. Although recent procedure advances based on patient-
specific prosthesis were proposed, correct bar placement is still challenging.
In this work, we propose a novel augmented reality system to guide the surgeon during PE bar place-
ment. This system combines a 3D sensor with a projector to superimpose the thoracic ribs cage on the
chest wall of the patient, thus indicating the optimal insertion and bar placement points.
This system was validated in three different scenarios: 1) simulated chest surface models; 2) 3D printed
phantom; and 3) 3D commercial thoracic phantom. An error of 3.93 ± 3.44 mm, and 3.08 ± 1.57 mm were
obtained in the first and second experiments, respectively. In the final experiment, visual assessment of
the result proved that a high similarity was obtained between the projected model and the real ribs cage
position.
Overall, the proposed system showed high feasibility with low error, proving that 3D projection of the
ribs on the patient’s chest wall may facilitate PE bar insertion and ultimately provide useful information
to guide Nuss procedure.
227
228
a robust multi-dimensional features that describe Consequently, the ICP algorithm is used to refine
the local geometry around a point for 3D point the alignment.
clouds datasets. After estimation of the FPFH for The ICP is a popular algorithm employed to
the Kinect and CT point clouds, the SAC-IA pro- register two sets of point clouds X, P, by itera-
ceed as follow: tively matching points in one set to the clos-
est points in another set, consequently refining
i. Select sample points from one 3D model while
the transformation between the two sets in
making sure that their pairwise distances are
order to minimize the distance between both
greater than a user-defined minimum distance;
point clouds. Then, given two corresponding
ii. For each of the sample points, find a list of
points whose histograms are similar to the
point sets: {x , ,xxn } P = { p pn }, the
ICP algorithm computes the translation t and
sample points’ histogram. From these, select
rotation R that minimizes the sum of the squared
one randomly which will be considered as sam-
error:
ple points’ correspondence;
iii. Compute the rigid transformation defined by N
the sample points and their correspondences, 1 p
∑ xi
2
and compute an error metric for the point E (R t ) = R i
Rp t , (1)
N p i =1
cloud that computes the quality of the trans-
formation (Rusu et al. 2009).
where xi and pi are the corresponding points (Besl
These three steps are repeated, and the transfor- and McKay 1992). To compute the transformation
mation that presents the best error metric is stored between X and P, the algorithm iterate the follow-
and used to roughly align the two point clouds. ing steps:
229
Figure 2. Mean distance error of the 3D registration using targets with different SNR of white Gaussian noise: (a)
25 dB, 30 dB and 35 dB.
230
231
232
ABSTRACT: The market share of mobile devices has spectacularly increased in a few years. The novel
characteristics of this kind of devices: intuitive interfaces and everywhere/everytime access, has promoted
its widespread use in many different scenarios. In addition, medical trainees and specialist are familiar-
ized with them. On the other hand, wearable devices are a very promising new line of products even for
medical applications. In recent years, many computational techniques have been proposed to help the
specialist in the fracture reduction process. However, due to its intrinsic complexity and its high techno-
logical requirements, this research line is still open. The increasing computational power and the birth of
new technology favor the chances of automating more tasks reducing risks and surgery time exposition.
Nevertheless, the use of mobile devices is not widespread in this specific medical context. In this paper,
we briefly introduce the possibilities of mobile and wearable devices in a fracture reduction process from
a global perspective. Different areas of research have to be taken into account to get a practical overview
of the problem and the connections among them. Specifically the focus is placed on the visualization of
medical data, the interaction, collaborative aspects and augmented reality possibilities.
Keywords: mobile devices; bone fracture reduction; 3D medical images; visualization; computer sup-
ported cooperative work; user interaction; augmented reality
233
234
3 VISUALIZATION OF 3D MEDICAL
IMAGES
235
236
237
238
E. Binaghi
Dipartimento di Scienze Teoriche e Applicate, Università degli Studi dell’Insubria, Varese, Italy
S. Balbi
Dipartimento di Biotecnologie e Scienze della Vita, Università degli Studi dell’Insubria, Varese, Italy
D. Lattanzi
Scuola di Specializzazione in Neurochirurgia, Ospedale di Circolo e Fondazione Macchi, Varese, Italy
ABSTRACT: Peritumoral brain edema is a frequently encountered phenomenon that strongly contrib-
utes to neurological signs and symptoms. This paper presents a fully automated procedure based on
Support Vector Machine for the accurate and efficient edema delineation in brain MRI imagery. The
segmentation procedure is conceived fully automatic and specifically oriented to non-volumetric data to
facilitate the insertion in clinical practice. The major efforts in building a robust classifier were concen-
trated in the extraction and selection of a set of significant features from multispectral brain MR images.
Preliminary experimental results obtained by processing in-house collected data, are encouraging creat-
ing the premises for an effective use of the automated segmentation procedure in brain tumor clinical
studies. To see if the proposed supervised method can be considered an alternative to other contemporary
approaches, the results obtained were compared with those obtained by an unsupervised region-based
segmentation procedure based on Graph Cut.
239
240
241
242
243
Guanghua Zhang
Chongqing University of China, Chongqing, China
Graphics and Imaging Laboratory, University of Girona, Girona, Spain
Angel Romero
Faculty of Infirmary, University of Girona, Girona, Spain
ABSTRACT: One of the key parameters that has to be considered in wound treatment is the size of the
wound. Usually, this is monitored by an expert at regular intervals. To carry out this process, image-based
approaches can be applied. Generally, these approaches start applying a segmentation technique to isolate
the wound from the rest of the image. Due to the importance of segmentation, our aim is to evaluate
the performance of four segmentation techniques (grabcut, snakes, region growing and Support Vector
Machine (SVM)) on pressure ulcer images. In our study, we have considered twenty images captured fol-
lowing normal clinical practice and under non-controlled conditions which makes the processing more
difficult. The images have been manually segmented by two experts and the obtained results have been
compared with the automatic segmentation obtained with the different methods. From the comparison, we
can conclude that both snakes and grabcut methods clearly behave better than region growing and SVM.
245
Figure 1. From left to right the user interaction and the obtained segmentation with snakes, region growing, grab cut,
SVM and SVM-neigh techniques.
246
In our implementation, the method computes the The third columns of Figure 1 and 2 show two
standard deviation of the image intensity on a win- examples of selected regions by the user and the
dow centered in the seed point. Then, the neigh- obtained results.
bour pixels are included if their intensity is in the
range [ k k ], where I is the intensity of
2.4 Support Vector Machine (SVM)
the seed point, k is a parameter of the method, and
σ is the standard deviation computed in the initial SVM method is a clustering-based technique that
step. Thus, the method has two parameters: the classifies each pixel using the Support Vector
size of the window where the standard deviation is Machine algorithm (Cortes & Vapnik 1995) from
computed and k. We have used the a training set, which is given by the user. For each
implementation provided by the opencv library pixel, a vector of features is obtained and used by
(Bradski 2000). In this method, the user interac- the SVM algorithm to perform the classification.
tion with the image simply consists in the selection In our implementation, the RGB color values and
of the seed point and the adjustment of k and the the distance to the lesion center (given by the user)
window size to find out the best result. e set σ are used as features. In order to do not have fea-
between 10 and 50. The second columns of Figure tures with different range of values, the distance
1 and 2 show two examples of the seed placement to the lesion center has been divided by the image
and the obtained results. diagonal and multiplied by 255. In this way, every
feature takes values in the range [0,255]. In our
framework, we have used the SVM() implemen-
2.3 Grabcut
tation provided by the opencv library (Bradski
The grabcut method (Rother, Kolmogorov, & 2000). The parameters of this implementation are:
Blake 2004) is a graph-based technique and, in par- the four selected features corresponding to R, G,
ticular, based on graph cuts. To start the process, and B values and the pixel to the center of selected
the user defines a bounding box around the object foreground distance. We have also implemented
to be segmented. Then, the method estimates the the SVM method adding local information. For
color distributions of the target object and the each pixel, we have considered the R, G, and B val-
background using a Gaussian mixture model. This ues from the top, bottom, left and right neighbour
model is used to construct a Markov random field pixels obtaining a feature vector with 12 values.
over the pixel labels, with an energy function that We denote this method SVM-neigh. The user
prefers connected regions with the same label, and interaction for SVM method consists in the defi-
running a graph cut based optimization to infer nition of some rectangular regions corresponding
their values. Since this estimate is more accurate to either background or foreground wound pixels.
than the one taken from the bounding box, this The fourth and fifth columns of Figure 1 and 2
two-step procedure is repeated until convergence. show two examples of selected regions by the user
To correct the estimates the user can point out and the obtained results with the SVM and SVM-
misclassified regions and rerun the optimization. neigh, respectively.
The method also corrects the results to preserve
edges. In our framework, we have used the grabCut
method from the opencv library (Bradski 2000). 3 EXPERIMENTAL SETUP
This implementation does not require any param-
eter. The user interaction for this method consists This study was done in collaboration with clini-
in drawing a square contour containing the wound cians from the Faculty of Infirmary from the
to get an initial segmentation and also selecting University of Girona. For the experiments we
the background and foreground with some points. used 20 images of 2304 × 3072) resolution stored
247
Figure 4. Masks of manual segmentation and the automatic ones are compared and points are labelled as true and
false positive or true and false negative.
248
Method P R F-score
can define some parameters that can be optimally automate wound measurement from images. Wound
set, for most of cases, after about 10 iterations. segmentation is a key step of these methods. Due to
The total time to obtain the final result is about the importance of segmentation in these approaches
50 seconds. The grabcut algorithm also achieve we have evaluated four segmentation techniques and
very good results, but the interaction takes more compared with manual segmentation. From our
time than snakes, since the method is much more tests we observed that snakes and grabcut methods
computationally demanding. The final result is obtain the best results, which are slightly less accu-
achieved in this case in about 2 minutes. Despite rate than manual edition. As future work, we will
the poor results of the region growing, its interac- improve the user interface to easily interact with the
tion is very fast and a final result can be obtained method parameters, we will integrate the methods
in 30 seconds. SVM and SVM-neigh are the most into an e-learning tool and we will investigate new
computationally demanding methods, but the image features to improve the SVM method results.
manual interaction is not very high. In this case,
the total time is about 2 minutes.
In Figure 4, we show three different cases that ACKNOWLEDGMENTS
represent different wound shapes: a regular one,
an irregular one with concavities, and a lengthened This work was supported by the Spanish Govern-
one, respectively. From these results we can observe ment (Grant No. TIN2013-47276-C6-1-R) and by
that the the snakes method performs better with the Catalan Government (Grant No. 2014-SGR-
regular shapes. Snakes method assumes a smooth 1232). Guanghua Zhang received financial sup-
contour and, when the wound has an irregular or port from the European Commission (Erasmus
lengthened shape, the method does not achieve Mundus project Techno, ref. 204323-1-2011-1-FR-
correct results. Region growing method has a bad EMA21).
performance in all the cases which is due to irregu-
larities on the lesion colour but not on the shape.
Grabcut method has a good performance in all the REFERENCES
cases and irregular shapes do not affect the results.
Moreover, the interaction is almost the same, inde- Ahmad-Fauzia, M., I. Khansab, K. Catignanic,
pendently of the wound shape. If we compare SVM G. Gordillo, K. Send, & M. Gurcan (2015). Computerized
and SVM-neigh, we can observe that in all the segmentation and measurement of chronic wound
cases considering neighbour information leads to images. Computers in Biology and Medicine, 74–85.
Bilgin, M. & U.Y. Gunes (2014). A comparison of 3
better results. Although SVM and SVM-neigh are wound measurement techniques. J Wound Ostomy
not affected on the contour shape themselves, in Continence Nurs 40(6), 590–593.
the case of lengthened shapes, the pixels of wound Bradski, G. (2000). The OpenCV Library. Dr. Dobb’s
extremes are misclassified. This is due to the fact Journal of Software Tools.
that the distance to the central point of manual Burns, M. J. Enderle, E. Rosow, & Q. Zhu (2002). Devel-
interaction is considered as a SVM feature. opment of a wound assessment system for quantita-
tive chronic wound monitoring. In Proc. IEEE Annual
North-east Bioeng. Conference.
5 CONCLUSIONS Cortes, C. & V. Vapnik (1995). Support-vector networks.
Machine Learning 20(3), 273–297.
Gethin, G. (2006). The importance of continuous wound
The standard practice for the treatment of wounds measuring. Wounds 2(2), 60–68.
includes monitoring the size of the wound at regu- Gethin, G. & S. Cowman (2006). Wound measurement
lar intervals. Different methods based on image- comparing the use of acetate tracings and visitrak dig-
processing techniques have been proposed to ital planimetry. J Clinical Nursing 15, 422–27.
249
250
S.M. Shontz
The University of Kansas, Kansas, USA
J.S. Tahara
BSMP Student, The University of Kansas, Kansas, USA
BSMP Student, The University of Sorocaba, Sorocaba, Brazil
D.O. McLaurin
CD-adapco
D.J.L. Colbry
Michigan State University, Michigan, USA
B. Parikh
The University of Kansas, Kansas, USA
ABSTRACT: Image segmentation is a challenging problem due to sensor-based noise, artifacts caused
by motion, and partial obstruction of objects. Although numerous algorithms for image segmentation
have been invented, there is no single algorithm which can be used to successfully segment all images.
Similarly, there is no fully automated image segmentation tool despite recent progress in this direction.
In this talk, we present a machine learning tool for automated segmentation of a set of images. Our
technique first uses an existing image segmentation algorithm to semi-automatically segment the first
image in the data set. It then proceeds by using our multiobjective optimization technique to match the
segmented shapes in the initial image to their most similar shapes in the second image in the data set.
Optimal shapes are determined based upon factors such as color, geometry, motion, and texture. This
process is then repeated in order to segment the remaining images in the data set. Machine learning is used
throughout the optimization procedure in order to determine the most appropriate values for the weights
of the relevant terms in the objective function for a set of images. We present results obtained from testing
our algorithms on various sets of biological and medical images. In addition, we present results on the
scalability of our method.
251
f w1, w2 , w3 ) = w1(hue
( huenew huessegmented
egmented )
2
+ w2 ( saturationnew
− satura
r tionsegmented )2
+ w3 (valuenew
− valuesegmented )2 .
3 EXPERIMENTAL RESULTS
252
REFERENCES
253
Irina Grigorescu
Computer Science and Information Technology Department, University Politehnica Bucharest, Bucharest, Romania
Roxana Radu
LOTUS Hospital SRL, Romania
ABSTRACT: This paper proposes a method to statistically characterize the proprieties of cervical cells
extracted from samples labeled as normal and abnormal such as the ones affected by pathological modi-
fications or subjected to various processes like electroporation. The characterization is based on fractal
descriptors applied on 3D images obtained employing digital holographic microscopy technique. Fractal
descriptions are using the measures of the dilated volume and surface of the holographic phase images of
the cells. The method is a promising tool to discriminate among the normal and abnormal cervical cells
by exploiting the whole image of the cell not only limited zones or random cross sections, with potential
to be automatically implemented.
255
3 HOLOGRAM AQUISITION
AND RECONSTRUCTION
Figure 1. Typical normal cervical cell (a) and abnormal Figure 2. Hologram of a cervical cell (left) and the
cervical cell (b) in a marked smear on glass plate (Nikon corresponding phase image of the reconstructed object
Ti-U microscope, 40× objective). (right) for a) normal cell, b) abnormal cell.
256
4 FRACTAL DESCRIPTORS
257
log N (r )
d ∝ − lim . (5)
r→0 log r
where the radii are computed according to the In the present work the descriptors (8) and (8′)
Euclidean metric are used together following a version of the method
used by Da Silva Oliveira et al. (Da Silva Oliveira
rijk i2 + j2 k2 , (7) et al. 2014):
( ) ⎫⎪⎬.
domly positioning of the first sphere followed by
⎧⎪ log N Σ s rijk
a neighborhood strategy. For the following posi-
D∑ = ⎨ (8)
log rijk tions concentric spheres around this first position
⎩⎪ ⎭⎪ are considered at increasing distances such as no
overlapping are allowed between the candidate
Similarly, the descriptor DV for volume case is element and the all other already placed spheres.
the number of disjoint balls filling the volume Vr If the candidate does not overlap with any of the
(see Fig. 5): previously placed spheres, then it will be kept and
the procedure moves forward with one step. The
DV = ⎨
(
⎧⎪ log NV s rijk ) ⎫⎪⎬. (8′)
algorithm stops when all possible candidates from
log rijk a given distance are outside the cell’s boundary.
⎪⎩ ⎪⎭ For every radius of the structural element the
258
259
260
Emeric Veron-Tocquet
Laboratoire de Traitement de l’Information Médical, LaTIM-INSERM U1101, France
Université de Bretagne Occidentale, France
Valérie Burdin
Laboratoire de Traitement de l’Information Médical, LaTIM-INSERM U1101, France
Institut Mines Telecom-Telecom Bretagne, France
Joel Savéan
Laboratoire de Traitement de l’Information Médical, LaTIM-INSERM U1101, France
Université de Bretagne Occidentale, France
ABSTRACT: Medical rehabilitation therapies depend on the accurate motion of bone structures,
which is often impossible to acquire without using invasive markers. Current clinical studies that evaluate
measurements of motion use skin markers even though its drawbacks due to skin motion are well-known.
Recent developments in motion analysis use markerless Time-of-Flight (ToF) cameras that can acquire
three dimensional data. This study was focused on integrating two synchronized ToF cameras to evaluate
knee flexion movement and compare the findings with standard Vicon system data. A rigid structure was
developed and the best relative location of ToF cameras on this structure relative to the patient was
analyzed. In the experimental process, the rigid structure was placed in the motion analysis lab and
standard skin markers were placed on the lower leg of a volunteer. The volunteer performed knee flexion
movements while lying on a bed and knee flexion data was simultaneously acquired by both the Vicon
system as well as the integrated ToF cameras during nine camera placement combinations. The RMS
error between the two measurement methods ranged from 5.61° to 9.64° depending on camera locations.
261
262
α { ; ; } d θ ={ ; ; }.
263
4 DISCUSSION
5 CONCLUSIONS
This study successfully implemented an integrated
ToF system to determine knee flexion angle and This markerless ToF camera system provides
in doing so illustrated the feasibility of using mul- valuable insights and feasibility in its use as a
tiple ToF camera system as a viable, quick, and rehabilitative diagnostic tool. The best RMS
264
265
A. Evangelista
Ospedale San Giovanni Calibita Fatebenefratelli—Isola Tiberina, Roma, Italy
267
268
269
Figure 6. Primary Strains (PS) measured by 3DSTE along the cardiac cycle in one representative healthy subject
(left panel) from our series of 45 subjects and in one patient with primary pressure overload (right panel) from our
series of 22 HCM patients at endocardial level: it is evident that at end-systolic homologous time (around 400 ms)
the patient presents greatly reduced PS in both LV and LA levels, indicating a reduced capacity to contrast intra–
chamber pressures, a possible consequence of reduced elasticity due to intrinsic heart muscle pathology typical of
HCM. Interestingly enough, in comparison to the representative healthy subject of the left panel, the reduced elastic-
ity indexed by greatly lower PS of the endocardium is dimensionally similar at LV and LA levels in the representative
HCM patient shown in the right panel.
effective for subendocardial layers, which dilat- values which correspond to the circumferential
ing reduce the circumferential shortening induced PSL are smaller.
by muscle contraction. Better is the capacity of
the elastic response of the endocardial surface, REFERENCES
smaller is the dilation induced by blood pressure.
This concept is well represented by endocardial [1] Piras, P., Evangelista, A., Gabriele, S., Nardinocchi, P.,
Primary Strain (PS) lines and values, which rep- Teresi, L., Torromeo, C., Schiariti, M., Varano, V.,
resent at a point of the endocardial surface the Puddu, P.E., 2014. 4D-analysis of Left Ventricu-
directions where the smallest are to be expected. lar Heart Cycle Using Procrustes Motion Analysis.
In this sense, we interpreted the primary strain val- PLoS ONE 9(1): e86896. (doi:10.1371/journal.
ues measured by 3DSTE along the cardiac cycle pone.0086896).
as shown in Figure 6. our representative healthy [2] Evangelista, A., Gabriele, S., Nardinocchi, P., Piras, P.,
subject shows a PS value which denotes a larger Puddu, PE., Teresi, L., Torromeo, C., Varano, V.,
2014. A comparative analysis of the strain-line pat-
shortening than the corresponding value shown by tern in the human left ventricle: experiments vs
the representative HCM patient. It means that the modeling. Computer Methods in Biomechanics and
capacity to contrast blood pressure is reduced in Biomedical Engineering/Imaging & Visualization.
patients with with primary pressure overload due Published online: Jun 23, 2014. (doi:10.1080/216811
to intrinsic heart muscle pathology, hence the PS 63.2014.927741).
270
271
ABSTRACT: In this paper, we proposed a numerical left ventricle model with simplified conical
geometry. The numerical simulation of the model deformation is carried out using the Yeoh constitu-
tive model, hexahedral elements, and nonlinear geometry (Abaqus CAE, Simulia). To validate the model
utility, the displacements and strains of some representative points are compared with clinical data. The
proposed computational LV model can simulate the LV wall deformations to some extent. Although, it
cannot fully mimic the strain distribution and absolute peak strain values obtained in clinical practice, a
physical model approximating the LV shape with a cone and deformed in a dedicated measurement stand
will be a profitable source of reliable data for the validation of myocardial strain imaging modalities.
273
274
275
of apical segment. The opposite trend may be The twist along the model wall for the nodes
observed for the basal segment. In the mid seg- located at the outer (epi) and inner (endo) surface
ment, axial strain component for the epi- and myo- of the conical part is presented in Figure 6.
cardial layers are almost identical and higher than A twist gradient may be observed across the model
for the endocardial layer. wall with twist values for the endocardial layer higher
276
Table 2. Maximal absolute strain and twist values obtained in clinical practice.
Clinical data*
Strain Wall
component [%] segment Epi Mio Endo
277
278
ABSTRACT: Minimally Invasive Surgery is a surgery technique that provides evident advantages for
the patients, but also some difficulties for the surgeons. In medicine, the Augmented Reality (AR) technol-
ogy allows surgeons to have a sort of “X-ray” vision of the patient’s body and can help them during the
surgical procedures. In this paper we present an application of Augmented Reality that could be used as
support for a more accurate preoperative surgical planning and also for an image-guided surgery. The AR
can support the surgeon during the needle insertion for the Radiofrequency Ablation of the liver tumours
in order to guide the needle and to have an accurate placement of the surgical instrument within the
lesion. The augmented visualization can avoid as much as possible to destroy healthy cells of the liver.
279
280
281
282
ABSTRACT: In this paper the virtual simulation of an anthropomorphic robot forefinger, analyzed and
synthesized as a four-link mechanism open chain is proposed. The metacarpal bone is considered fixed;
the proximal, medial and distal phalanges are links in movement of said mechanism. The five degrees of
freedom of the system are reproduced without interference between the links at the corresponding kin-
ematic pairs and using a unique angular parameter to generate the different geometries. The simulation
is parameterized using a spreadsheet and graphic soft ware. After describing the advantages of virtual
simulation their results are presented, which show the possibility of studying the movement of the system
in different circumstances without higher costs. Additionally, application possibilities are shown in the
care of diseases of the fingers.
1 INTRODUCTION
2 NOTATION
283
4 DEGREES OF FREEDOM
5 SYSTEM COMPONENTS
284
7 PRACTICAL APPLICATION
F 2 r2 zD L2 − L3 senθ3 (5)
m3 = (1)
2rF xD L3 cos θ3 (6)
285
8 ADVANTAGES OF VIRTUAL
SIMULATION
9 ANALYSIS OF RESULTS
286
10 CONCLUSIONS
11 FUTURE WORK
ACKNOWLEDGEMENTS
REFERENCES
Figure 16. Swan Neck deformity.
[1] Neil A. Davidoff, Andris Freivalds (August 4,
1993). “A graphic model of the human hand using
An iteratively refined design was necessary to CATIA”. Department of Industrial and Management
make possible the above limits, without interfer- Systems Engineering, Pennsylvania State University,
ence between the links converging on different kin- University Park, PA 16802, USA.
ematic pairs. The limits obtained are higher than [2] Jung Won Kang, Hyun Seok Hong, Bong Sung Kim,
average limits of a person. In Figure 14, detail of Myung Jin Chung, “Work Assistive Mobile Robots
Assisting the Disabled in a Real Manufacturing
Figure 9 in the kinematic pair between the proxi-
Environment”, International Journal of Assistive
mal and medial phalanges, it can be seen that a Robotics and Mechtronics, Volume 8 Number 3,
larger positive flexion angle of 135 degrees may September 2007.
cause interference between these bones. [3] J.K. Salisbury, M.T. Mason, “Robot Hands and the
Deeper knowledge of the behaviour from Mechanics of Manipulation”, MIT Press, Cambridge,
the point of view of the kinematic joints of the MA, 1985.
287
ABSTRACT: This paper deals with modelling the tissue blood perfusion and focuses on the identifica-
tion of the model parameters using the patient-specific data obtained using the standard CT, or MRI
investigations. The research is aimed to develop a complex model which would assist in planning the liver
surgery. Recently we proposed a multicompartment model of the blood perfusion in liver which serves as
the feed back for simulations of dynamic CT investigation (Rohan, Jonášová, & Lukeš 2014). The flow
can be characterized at several scales for which different models are used. Flow in larger branching vessels
is described using a simple 1D model based of the Bernoulli equation with correction terms respecting
the pressure losses due to the dissipation. This model is coupled through point sources/sinks with a 3D
model describing multicompartment flows at the lower hierarchies of the perfusion trees penetrating to
the parenchyma. The compartments in the liver tissue are associated with segments which confine the
flow to subdomains within the organ, and hierarchies which reflect the flow complexity on branching
vascular trees of the portal and hepatic veins. For simulations of the CT perfusion test, a model of the
dynamic transport of the contrast fluid in the compartments was developed; the time-space distribution
of the so-called tissue density can be computed and compared with the measured data obtained form the
CT. To find suitable values of the perfusion model parameters, we formulate an optimization problem
where the objective function expresses the difference between the standard CT images and the corre-
sponding perfusion maps of the contrast agent concentration. This is computed by solving the perfusion
problem to obtain a steady distribution of the blood pressure at all the compartments, and the transport
problem describing the contrast fluid saturation. The optimization problem can be solved using a suitable
gradient-based method. For this the sensitivity analysis formulae were derived using the adjoint system
method. A simplified identification problem was implemented and solved numerically to show viability
of the proposed approach.
289
290
L(α ( p, ) ( λ ,ψ ))
) Φ(α ) 0 ∫
− ∂α aαi ⎛⎝ pi , S iψ i ⎞⎠ dt. (8)
⎛ ⎞
+ ∑ aαi ⎛⎝ pi , λ i ⎞⎠ ∑ bαij ⎛⎝ pi p j , λ i ⎞⎠ − g i ( λ i )⎟
i ⎝ j i ⎠ 4 SIMPLIFIED PROBLEM
T ⎛ ∂S i i i ⎞ i
+∑∫ ∫0 ⎜ ∂t + ∇ ⋅ ( S ) ψ dtdx
d We consider a simplified problem which was
Ωi ⎝ ⎠ implemented and solved to examine some poten-
i
tial difficulties arising in this kind of inverse prob-
T
+ ∑ ∑ ∫ bαij ⎛⎝ pi − p j , Z ij ( )ψ i ⎞⎠ dt lems. The liver perfusion is described by a reduced
0
i j ≠i multicompartment model which incorporates the
T vascular tree of the portal vein only, Figure 1(a).
−∑∫ ∫Ω (Sin f+ + S
i i i
f−i )ψ i dxdt , Upon decomposing the tree in 2 compartments,
0 i
i the hierarchical flow is represented by three com-
partments which are defined in terms of two per-
where λ i and ψ i are the Lagrange multipliers; meabilities, 1, K 2 and two inter-compartment
λ i ∈V
V0i and ψ j = 0 at t T . perfusion coefficients, G12 and G23. The 3rd com-
partment is degenerate, i.e. the pressure p3 = 0 is
3.2 Optimality conditions prescribed. The flow is described by only y two pres-
sure fields, p ( p1, p2 ), such that p j j
j = 1, 2.
The optimality conditions associated with feasibil- Thus, the state problem is given by the following
ity of the state problem read: 2 equations:
δS δS δp δp=0, (7)
∫Ω K ∇p ⋅ ∇q + ∫ΩG2 ( )q1 = ∫Ω F 1q1,
1 1 1 1 1
− 2
291
ϕ (α ) → min
p = p(α ) solves ( ), (10)
292
= −2∫ G21 (
Ω
( 1
− 2
) − J )G21q1,
∫Ω K ∫Ω (G1 ( ) + G32 ) q2
2
∇ 2
⋅ ∇q 2 2 2
− 1 2
(
= 2∫ G21 ( p1 − p2 ) J G21q 2 ,
Ω
) (12)
∂αtotk φ , p) = 2∫ (α k G k ( pi − p j ) − J k )G k ( pi − p j )
Ωk
+∫ G k ( p − p )( λ − λ ) ,
1 2 1 2
Ωk
5 CONCLUSIONS
293
294
Insaf Setitra
Research Center on Scientific and Technical Information Cerist, Algeria
University of Science and Technology Houari Boumediene, Algeria
Slimane Larabi
University of Science and Technology Houari Boumediene, Algeria
ABSTRACT: Scale Invariant Feature Transform (SIFT) since its first apparition in 2004 has been (and
still is) extensively used in computer vision to classify and match objects in RGB and grey level images
and videos. However, since the descriptor used in SIFT approach is based on gradient magnitude and
orientation, it has always been considered as texture feature and received less interest when treating binary
images. In this work we investigate the power of discrimination of SIFT applied to binary images. A theo-
retical and experimental studies show that SIFT can still describe shapes and can be used to distinguish
objects of several classes.
295
LoG ( x y )=− 4 ⎢1 − e 2σ
⎥ (1) defined above, it is considered potential maxima
πσ
σ ⎣ 2σ 2 ⎦ and its coordinates are saved.
Once potential maxima over all scales defined, a
where x and y are respectively row and column coor- further refinement is performed in order to eliminate
dinates in the shape to be convolved and σ is the weak edges. This is done by suppressing from each
standard deviation of the Gaussian distribution. convoluted image Li + 1 potential maxima which are
Using the Laplacian of Gaussian besides being an less than an input threshold (measure of stability).
accurate edge detector due to its precision in detect- Maxima detected in this step are returned as
ing zeros-crossings, has many other advantages. output of this phase and are called keypoints. Each
First, the Laplacian of Gaussian is the second deriv- maxima is defined by its row coordinate r, column
ative of an image convolved with a Gaussian which coordinate c and its radius rad which is computed
prevents from noise in the image. Secondly, as the as follows:
filter is pre-calculated in advance, processing time is
reduced. Convolving the image several times allows
to define the zeros-crossing over different scales. rad σ k i −1 × 2 (4)
Number of convolutions c is computed as
follows: Changes in scale allows to see the image across
different resolutions. The next step is to describe
keypoints detected in this step.
gσ ⎤
⎡ log c − log
c=⎢ ⎥ (2)
⎢ log k ⎥
2.2 Keypoint description
where ⎡ is the ceiling function which maps a real A binary image can be defined as a matrix which
number to its smallest following integer, σ and σc can have only two possible values typically relative
are respectively initial and final standard deviation to black and white. Depending on the image rep-
of the Laplacian of Gaussian Filter and are given resentation, those two values can be either 0 and
as input parameters. log is the natural logarithm of 255 or 0 and 1. In a sake of simplicity we use the
base 2 and k is the scale step. second representation in this paper.
At each of the c convolutions applied to I, σ is Description of keypoints localized in the first
updated as follows: step is done by first, computing the gradient mag-
nitude and direction of the original image I around
σ i = σ × k i −1 (3) its keypoints and quantifying orientations weighted
by the magnitude to form the final descriptor. We
where 1 ≤ i ≤ c is a counter ranging from 1 to c. The describe the whole process in what follows.
initial image I is convoluted with the new σ using First gradient of the original image I:
the formula 1. ( x ( ∂I//∂ ) I y = ( ∂I /∂y )) is computed. Hence,
Once the initial image I is convoluted c times, Gradient of an image can have either −1, 0 or
local maxima are defined and their coordinates 1 value in both directions.
(row, column indices and radius) are returned. The Gradient magnitude and Gradient orienta-
maxima is defined as follows: tion are computed. As gradient orientation is
Let Li, Li + 1 and Li + 2 be three adjacent images based on the gradient magnitude, 8 Gradient
computed as follow: orientations can then take up to eight values:
− 3π , − π , − π , π , 0, 3π , π , π .
LoG ( x y; 4 2 4 4 2 4
Li i ) ∗ I, Li = LoG
L G (x y i +1 ) I
Once orientations of the gradient for the whole
Li LoG ( x y; i +2 ) ∗ I image computed, the second step is to get for each
keypoint its window computed as follows:
The local maxima is first computed on the neigh-
borhood of Li + 1 which is the center of the current xo ⎡⎢ max( , r rad
add ) ⎤⎥ , yo ⎡⎢ max( , c − rad ) ⎤⎥
3 − tuples of convoluted images (Li, Li + 1, Li + 2).
More precisely, each pixel p(x, y) in Li + 1 is com- xe = ⎡⎢ min
i (r + rad , n ) ⎤⎥ ye = ⎡⎢ min(c + radd , m ) ⎤⎥
pared to its 8 neighbors. If p is either minimum or
maximum, it is considered as a potential maxima where xo, yo, xe, ye are respectively the (row, col-
of the inner scale and its coordinates are stored. umn) of the beginning respectively end of the key-
Then, p is compared to its 18 neighbors in a 3×3 point window.
windows in Li and Li + 2 centered at (x, y). r, c and rad are row, column coordinates and
Comparisons are done the same way as in the radius of the keypoint. n and m are the two dimen-
inner scale i.e. if the central pixel in Li + 1 is either sion sizes of the image I.
296
297
4.2 Training and classification using bag of words where l is the dimension of the feature vector equal
to number of clusters of the K − means algorithm.
After having Bag of Features descriptor of each
of training and test images, the distance between
Bag of Features of test images and training images 5 EXPERIMENTAL RESULTS
is computed, for each test image, the minimal dis-
tance is chosen, and label of the nearest training 5.1 Shape retrieval accuracy
of the dataset ETH-80
Retrieval experiments are done on the dataset
ETH-80 [9]. The dataset comprises 8 categories
each of which has 400 images. We choose ran-
domly 10 images of each category which results
in 80 images in the dataset prepared for retrieval.
Then, we chose randomly an image from each
category for retrieval. We use the bullseye score
for retrieval accuracy measured as the percent
of correct retrievals in the top 20 ranks. Table 1
shows retrieval accuracy of retrieval. The first
column represents the query shape, the second
column represents shapes retrieved in the order
of similarity i.e. similar shapes start in the top left
of the column and are ordered line wise. Names of
shapes are the same as in the original database1.
The third column stands for the processing time.
The table shows that best rates are for shapes with
few curvatures whereas weak rates are observed
for shapes with high curvatures. This is because
keypoints in curvatures are repetitive in many
shapes. However, processing time for shapes with
few curvatures is higher. Besides, shapes with few
curvatures are most often matched to shapes with
few curvatures, examples include: apple, pear and
tomato. However, some shapes of high curvatures
are confused with shapes with few curvatures.
The reason is that, many keypoints in a shape
can be matched to only one keypoint in another
shape and distance is minimal. The opposite is
also true. One solution could be to match a key-
point in a query shape to only one keypoint of a
shape in the dataset. We let this improvement to
a future work.
Figure 1. Example of Bag of Features representation
of an image I (a) the image I to be represented by the
visual vocabulary, (b) Visual Vocabulary represented by
1
its centers, (c) Bag of Features of the image I. http://people.csail.mit.edu/jjl/libpmk/samples/eth.html
298
299
300
ABSTRACT: In most image processing applications, the raw data acquired by an image sensor passes
through a demosaicing algorithm in order to obtain a full color image. The demosaicing step can take time
and resources that are of great importance when developing real-time applications. Most of the times, this
is done with single purpose of rendering the images into a viewable format for humans. However, most of
nowadays sensors allow the retrieval of images in raw format and processing the data as is. In this paper
we present a study on the direct usage of raw Bayer data for real-time color object detection, in the sce-
nario of autonomous robot soccer. The experimental results that we provide prove that the efficiency of
the chosen object detection algorithm is not lowered and that several improvements of the autonomous
robotic vision system used in this study can be noted: the bandwidth needed to transmit the data between
the digital camera and the PC is reduced, allowing, in some models, to increase the maximum frame rate.
Moreover, a decrease in the delay between the perception and action of the autonomous robot was meas-
ured, when using raw image data for processing the environment.
301
1
http://robotica.ua.pt/CAMBADA
2
http://sweet.ua.pt/an/uavision/
302
3 EXPERIMENTAL RESULTS
303
304
305
306
ABSTRACT: Nowadays there are many computer vision algorithms dedicated to solve the problem of
object detection, from many different perspectives. Many of these algorithms take a considerable process-
ing time even for low resolution images. Research areas that are more concerned about real-time capabili-
ties of a system, such as the case of robotics for example, are focusing on adapting the existing algorithms
or developing new ones that cope with time constraints. Color object detection algorithms are often used
in time-constrained application since the color of an object is an important clue for its detection. In
addition, color segmentation algorithms have a smaller computational cost than generic object detection
algorithms. Most of the color object detection algorithms include a step of image scanning in search of
pixels of the color of interest. In this paper we present a study on the influence of an image scanning pat-
tern while searching for pixels of a certain color versus the image resolution in the results of three color
object detection algorithms. We present results comparing different scanning approaches versus different
resolution images for color object detection algorithms based on blob formation, contours detection and
region growing clustering. Our experiments show that using search lines for color segmentation, followed
by blob formation provides faster and more accurate results.
307
308
Mean Max Mean Max Mean Max Blobs Contours Region growing
1024 × 1024 8.72 10 17.16 35 8.86 21 1024 × 1024 100% 68.5% 73.1%
512 × 512 2.9 7 5.16 16 2.03 7 512 × 512 95.7% 63.7% 69%
256 × 256 1 5 1.02 3 1 1 256 × 256 85.7% 24.6% 53.9%
309
310
vision system in the hardware architecture of the results are even less reliable for balls found at 9 m
robot. When the ball is aligned with any of these or more from the robot. In this case, only 10% of
bars, the ball is not visible. the balls are correctly detected.
For images of 256 × 256 pixels, the contours The region growing clustering algorithm delivers
detection algorithms provides quite poor results. slightly better results, compared to the contours
When the ball is found at 6 m or more from the detection algorithm but it is still far from achieving
robot, it is only detected in 16% of the cases. The the detection rate of the blob detection algorithm.
311
REFERENCES
Figure 8. Ball detection using the blob algorithm at
12 m. The detected ball is marked with a magenta circle. Akin, H.L., T. Mericli, E. Ozukur, C. Kavaklioglu, &
B. Gokce (2014). Cerberus Team Description. RoboCup
2014, Joao Pessoa, Brazil.
Bradski, G. & A. Kaehler (2008, September). Learning
These results prove that working with images at OpenCV (First ed.). O’Reilly.
full resolution can provide optimal results in a Canny, J.F. (1986, November). A computational ap-
small amount of time when using an appropriate proach to edge detection. IEEE Transactions on Pat-
scanning pattern of the image. For applications tern Analysis and Machine Intelligence 8 (6).
where detection accuracy and total lack of false Garcia, D., E. Carbajal, C. Quintana, E. Torres,
positives is required, the algorithm based on radial I. Gonzalez, C. Bustamante, & L. Garrido (2012).
scanlines and blob formation is a suitable solution Borregos Team Description. RoboCup 2012, Mexico
since its detection rate is 100% and it can be used City, Mexico.
with sensors working at 100 fps, if the hardware Gonzalez, R.C. & R.E. Woods (2001). Digital Image
Processing (2 ed.). Addison-Wesley Longman Pub-
allows it. A smaller resolution can lead to an even lishing Co.
smaller processing time, for applications where fast Neves, A., J. Azevedo, N.L.B. Cunha, J. Silva, F. Santos,
performance is more crucial than having 100% G. Corrente, D.A. Martins, N. Figueiredo, A. Pereira,
correct detection. L. Almeida, L.S. Lopes, & P. Pedreiras (In Vladan
In Figure 8 we present an image with the ball Papic (Ed.), Robot Soccer, 2010). CAMBADA soc-
correctly detected by the blob algorithm at a dis- cer team: from robot architecture to multiagent
tance of 12 m from the robot. coordination, Chapter 2. I-Tech Education and Pub-
The results that have been presented in this lishing, Vienna, Austria.
paper have been obtained on a computer with an Neves, A.J.R., D.A. Martins, & A.J. Pinho (2008, April).
A hybrid vision system for soccer robots using
Intel Core i5-3340M CPU @ 2.70 GHz 4 proces- radial search lines. In Proc. of the 8th Conference
sor, which is the main processing unit of the soccer on Autonomous Robot Systems and Competitions,
robots used in the experiments. Portuguese Robotics Open—ROBOT-ICA’2008,
Aveiro, Portugal, pp. 51–55.
Neves, A.J.R., A.J. Pinho, D.A. Martins, & B. Cunha
5 CONCLUSIONS (2011, mar). An efficient omni directional vision
system for soccer robots: from calibration to object
In this paper we have presented a study about the detection. Mechatronics 21 (2), 399–410.
influence of image scanning in the pipeline of a Rofer, T., T. Laue, C. Graf, T. Kastner, A. Fabisch, &
C. Thedieck (2011). B-Human Team Description.
color object detection digital vision system. Using RoboCup 2011, Istanbul, Turkey.
the real-time scenario of robotic soccer and the Trifan, A., A. Neves, N. Lau, & B. Cunha (2012). A
digital vision system of an autonomous robot, we modular real-time vision system for humanoid robots.
presented results for three different algorithms for In Proceedings of IS&T/SPIE Electronic Imaging
object detection, using different scanning patterns 2012.
and different image resolutions. The results Trifan, A., A.J.R. Neves, B. Cunha, & J.L. Azevedo
presented show that when reducing the image (2014). Uavision: a modular time-constrained library
resolution, object detection algorithms tend to lower for soccer robots. In Proc. of RoboCup 2014.
their performance, due to loss of information.
312
H. Erdmann
Department of Research and Technology, Boolabs, São Paulo, Brazil
F.T. Ito
Department of Computer Science, Universidade Federal de Sao Carlos, Brazil
D. Takabayashi
Department of Research and Technology, Boolabs, São Paulo, Brazil
ABSTRACT: The amount of images used to train a classifier has a great impact in the algorithm’s per-
formance. In the domain of e-commerce, most of the dataset is formed by studio images and we demon-
strate in this work that by multiplying the amount of images with different transformations can provide a
significant boost on the overall performance of the classifier. Moreover, we list the challenges encountered
in such task and present the improvements obtained with a classifier trained with additional images that
were synthetically generated by applying several transformations on the original dataset.
313
314
315
316
317
318
319
ABSTRACT: The generation of surface models of bone fragments is useful for many medical applica-
tions. On the one hand, bone fragment models can be visualized in a more effective way. On the other
hand, simulation and analysis techniques usually need surface models as input. Simulation and analysis
processes require surface models to have certain properties and features that depend on the methods used
for these processes. Therefore, the suitability of the method utilized to generate the surface will depend on
the simulation or analysis performed. In this paper, the suitability of two algorithms that generate bone
fragment surface models from point clouds is tested. A proven 2D region growing based method enables
the segmentation of bone fragments from CT stacks. Then a new approach to extract contour points and
estimated normals is used in a pre-process step. The experiments performed allow extracting the benefits
and drawbacks of each method applied to different bones and fracture types. Moreover, the quality of
the obtained results is analyzed focusing on the special features of the data obtained from CT stacks rep-
resenting bone fractures. To the best of our knowledge, nowadays studied methods are not widespread in
the generation of surfaces to represent bone fragments.
Keywords: bone fragments; surface reconstruction; 3D medical images; bone fracture reduction;
segmentation; contour extraction; poisson; ball-pivoting
321
Figure 1. Point clouds extracted from CT scans. These points are used as input for the surface reconstruction algo-
rithms. A—Spiral fracture of distal fibula. B—Simple fracture of patella. C—Comminuted fracture of radius.
322
Table 1. Number of vertices and faces generated for each of the datasets.
323
Figure 2. Surfaces generated using the Ball-pivoting algorithm. A—Spiral fracture of distal fibula. B—Simple frac-
ture of patella. C—Comminuted fracture of radius.
Figure 3. Surfaces generated using the Poisson algorithm. A—Spiral fracture of distal fibula. B—Simple fracture of
patella. C—Comminuted fracture of radius.
324
325
326
ABSTRACT: Patch based reconstruction methods have been and still are one of the top performing
approaches to 3D reconstruction to date. Their local approach to refining the position and orientation
of a patch, free of global minimisation and independent of surface smoothness, make patch based meth-
ods extremely powerful in recovering fine grained detail of an objects surface. However, patch based
approaches still fail to faithfully reconstruct textureless or highly occluded surface regions; thus, though
performing well under lab conditions, they deteriorate in industrial or real world situations. They are also
computationally expensive. Current patch based methods generate point clouds with holes in texturesless
or occluded regions that require expensive energy minimisation techniques to fill and interpolate a high
fidelity reconstruction. Such shortcomings hinder the adaptation of the methods for industrial applica-
tions where object surfaces are often highly textureless and the speed of reconstruction is an important
factor. This paper presents on-going work towards a multi-resolution approach to address the problems,
utilising particle swarm optimisation to reconstruct high fidelity geometry, and increasing robustness to
textureless features through an adapted approach to the normalised cross correlation. The work also aims
to speed up the reconstruction using advances in GPU technologies and remove the need for costly feature
based initialisation and expansion. Through the combination of these enhancements, it is the intention of
this work to create denser patch clouds even in textureless regions within a reasonable time. Initial results
show the potential of such an approach to construct denser point clouds with a comparable accuracy to
that of the current top-performing algorithms.
327
328
329
330
331
lower but not higher resolutions, where this region To improve the performance of the NCC, a photo-
becomes steadily more uniform as the resolution detail consistency measure is defined, weighting
increases. the NCC based on the patches texture and view-
Once the patch cloud is stably initialised at the ing angle. To increase the point clouds density
lowest level the image resolution can be raised a PSO patch refinement scheme is run per pixel
and the results propagated upwards. PSO is run allowing for only approximate initialisation and
at the higher resolution using the previous resolu- removing the need for feature based matching
tions optimised result as a velocity factor added to and expansion. Finally a multi-resolution scheme
Equation 5. This allows for reconstruction of sur- using the wavelet transform is proposed, propagat-
face details not expressed in the lower resolution ing refined patch results for textureless surfaces up
images. Further to this we can decide not to run through the resolutions, where consistency meas-
PSO refinement for a higher resolution pixel if the ures start to fail. The experimental results show
detail coefficients of the wavelet transform reflect a the improved patch clouds and the potential for
non-textured patch. In these cases, rather than run more accurate MVS reconstruction in textureless
PSO, we smooth the patch parameters to obtain a regions through propagation of data from lower
consistent patch cloud. Figure 2 shows promise that resolutions.
with correct propagation the refinement of patches
in textureless regions can be improved significantly.
REFERENCES
5 CONCLUSIONS Bratton, D. & J. Kennedy (2007). Defining a standard for
particle swarm optimization. In Swarm Intelligence
This paper has presented on-going work towards Symposium, 2007. SIS 2007. IEEE, pp. 120–127.
a multi-resolution patch based approach to MVS. IEEE.
332
333
ABSTRACT: In-situ measurements of tools allow inspecting a running metal forming process. This
enables a fast adaption of the process and helps reducing costs. However, the requirements on the
measurement systems are high, as the measurement duration needs to be low and the available space
inside the machines is limited. Commercially available measurement systems are unable to operate in
these conditions. An application in metal forming is presented followed by the introduction of a novel
design for an optical measurement system based on the fringe projection profilometry technique. The new
design employs flexible fiber bundles to transport the images of the fringe projection system to a compact
sensor head which is based around gradient-index lenses. The pattern generation unit and camera are
housed in a base unit, which can be placed outside of the machine. While this new approach allows
flexible positioning of the sensor head, it creates challenges for the pattern evaluation algorithms and the
optical design of the system. The laser-based optical setup will be presented followed by measurements
of the tool of the aforementioned forming process. To evaluate the applicability of the new measurement
system, the determined tool parameters will be compared to reference measurements obtained by a tactile
Hommel T8000 system. In combination with a positioning device, which is currently being developed, the
new measurement system will be used to obtain wear information of the running forming process in order
to control the process quality loop in the future.
335
336
337
Height left 3.38 mm 3.38 mm 3.42 mm 3.36 mm 3.42 mm 3.33 mm 0.00 mm 0.03 mm
Height right 3.38 mm 3.38 mm 3.40 mm 3.38 mm 3.45 mm 3.38 mm −0.05 mm 0.00 mm
Angle left 110.0° 110.0° 110.5° 109.6° 110.1° 109.9° 0.4° −0.3°
Angle right 110.0° 110.0° 110.6° 110.8° 109.9° 109.8° 0.7° 1.0°
338
339
ABSTRACT: This paper deals with methods suitable for movable robotics platform navigation. We can
divide these methods into suitability for indoor navigation, outdoor navigation or its combination. At
present, the navigation methods of robotics platforms, suitable for outdoor spaces, are based on GNSS
(Global Navigation Satellite System). Accurate position estimation is a big advantage for this type of
navigation. Unfortunately this type of navigation is not suitable for indoor spaces and industrial areas,
mainly because of the weak signal and low position accuracy. An alternative navigation system is neces-
sary for these solutions—system independent on satellite navigation. Our research is focused on movable
robotics platform navigation using advanced image processing methods. Descriptions of these sophisti-
cated methods are presented in this paper.
Keywords: monocular camera vision; image processing; visual odometry; robot navigation; surface
detection
341
z (∑dx
d ∑dy
d ∑ dx d )
d ∑ dy (5)
3 EXPERIMENTAL RESULT
342
343
Parameter Value
VFOV 60°
HFOV 35, 98°
H 17.5 cm
L 44 cm
α 21.68°
Figure 6. Relationship between angle β and focal Width × height 320 × 240 pix
length f.
344
is a detected previous position of the template. At The research was supported by the Internal Grant
the top of this image, estimated differences between Agency of University of Pardubice, the project
black and white centres are shown in centimetres. No. SGSFEI 2015003.
At first, this distance is calculated in pixel distance
and later, it is transformed to real coordinates.
REFERENCES
6 CONCLUSION Bay, H. et al, 2008, Speeded-Up Robust Features
(SURF), In: LEONARDIS, Computer Vision and
In this paper a modern navigation method based Image Understanding: 9th European Conference on
on an advanced image processing method called Computer Vision, 2008, s. 346–359.
345
346
K. Nurzynska
Institute of Informatics, Silesian University of Technology, Poland
B. Smolka
Institute of Automatic Control, Silesian University of Technology, Poland
ABSTRACT: Many affective disorders can influence the capacity of emotion expression and
understanding. For instance, in the case of depression, the frequency and intensity of smiling is diminish-
ing when the disease proceeds. Analysing these changes by a psychologist could bring information about
the effectiveness of the prescribed medical treatment. Therefore, this work describes a system which auto-
matically analyses video recordings from sessions with patients and determines the length and intensity
of smile.
The proposed system exploits a standard database to train the classifier, which is later used to dis-
tinguish between the smiling and neutral facial expression depicted in the video frames. The obtained
information is presented graphically, indicating the class membership and temporal smile intensity. The
performed experiments revealed that the classification accuracy exceeds 85%. Moreover, the case studies
show, that the variations in smile magnitude enable better insight into the smile analysis.
347
348
349
350
351
352
353
3D Vision;
Computational Bio- Imaging and Visualization;
Computational Vision and Medical Image Processing. VipIMAGE 2015 will be useful
to academics, researchers and professionals in Biomechanics, Biomedical Engineering,
Computational Vision (image processing and analysis), Computer Sciences, Computational
Mechanics, Signal Processing, Medicine and Rehabilitation.
Sponsored by