2020-Lin - Deep Hough-transform Line Priors
2020-Lin - Deep Hough-transform Line Priors
1 Introduction
Line segment detection is a classic Computer Vision task, with applications such
as road-line detection for autonomous driving [17,22,30,36], wireframe detection
for design in architecture [18,54,55], horizon line detection for assisted flying
[12,32,39], image vectorization [41,56]. Such problems are currently solved by
state-of-the-art line detection methods [18,54,51] by relying on deep learning
models powered by huge, annotated, datasets.
Training deep networks demands large datasets [2,35], which are expensive
to annotate. The amount of needed training data can be significantly reduced
by adding prior knowledge to deep networks [5,19,21]. Priors encode inductive
solution biases: e.g. for image classification, objects can appear at any loca-
tion and size in the input image. The convolution operation adds a translation-
equivariance prior [21,43], and multi-scale filters add a scale-invariance prior
[37,40]. Such priors offer a strong reduction in the amount of required data: built-
in knowledge no longer has to be learned from data. Here, we study straight line
detection which allows us to exploit the line equation.
In this work we add geometric line priors into deep networks for improved
data efficiency by relying on the Hough transform. The Hough transform has a
long and successful history for line detection [10,20,26]. It parameterizes lines
in terms of two geometric terms: an offset and an angle, describing the line
2 Y. Lin, S.L. Pintea, and J.C. van Gemert
Ground truth Learned local features Added line priors Line predictions
Fig. 1. We add prior knowledge to deep networks for data efficient line detection. We
learn local deep features, which are combined with a global inductive line priors, using
the Hough transform. Adding prior knowledge saves valuable training data.
equation in polar coordinates. This gives a global representation for every line in
the image. As shown in figure 1, global information is essential to correctly locate
lines, when the initial detections are noisy. In this work we do not exclusively
rely on prior knowledge as in the classical approach [6,7,31,44] nor do we learn
everything in deep architectures [18,51,54]. Instead, we take the best of both: we
combine learned global shape priors with local learned appearance.
This paper makes the following contributions: (1) we add global geometric
line priors through Hough transform into deep networks; (2) we improve data
efficiency of deep line detection models; (3) we propose a well-founded man-
ner of adding the Hough transform into an end-to-end trainable deep network,
with convolutions performed in the Hough domain over the space of all pos-
sible image-line parameterizations; (4) we experimentally show improved data
efficiency and a reduction in parameters on two popular line segment detection
datasets, Wireframe (ShanghaiTech) [18] and York Urban [8].
2 Related work
Image Gradients. Lines are edges, therefore substantial work has focused on
line segment detection using local image gradients followed by pixel grouping
strategies such a region growing [31,44], connected components [6], probabilistic
graphical models [7]. Instead of knowledge-based approach for detecting local
line features, we use deep networks to learn local appearance-based features,
which we combine with a global Hough transform prior.
Hough transform. The Hough transform is the most popular algorithm for
image line detection where the offset-angle line parameterization was first used
in 1972 [10]. Given its simplicity and effectiveness, subsequent line-detection
work followed this approach [11,20,49], by focusing on analyzing peaks in Hough
space. To overcome the sensitivity to noise, previous work proposed statistical
analysis of Hough space [50], and segment-set selection based on hypothesis
testing [45]. Similarly, a probabilistic Hough transform for line detection, followed
by Markov Chain modelling of candidate lines is proposed in [1], while [26]
creates a progressive probabilistic Hough transform, which is both faster and
Deep Hough-Transform Line Priors 3
Fig. 2. HT-IHT block: The input featuremap, F, coming from the previous convolu-
tional layer, learns local edge information, and is combined on a residual branch with
line candidates, detected in global Hough space. The input featuremap of 128×128×256
is transformed channel-wise to the Hough domain through the HT layer into multiple
183 × 60 maps. The result is filtered with 1D channel-wise convolutions. Two subse-
quent 1D convolutions are added for merging and reducing the channels. The output
is converted back to the image domain by the IHT layer. The two branches are con-
catenated together. Convolutional layers are shown in blue, and in red the HT and
IHT layers. Our proposed HT-IHT block can be used in any architecture.
0 0
0
10 10
20
20 20
40
30 60 30
40 80 40
50 100 50
60 120 60
0 10 20 30 40 50 0 10 20 30 40 50
0 20 40 60 80 100 120
(a) Input line (b) Line HT (c) Line IHT (d) Mask B(x0 , y 0 )
Fig. 3. (a) A line together with its (ρ, θ) parameterization. (b) The Hough transform
(HT ) of the line. (c) The inverse Hough transform (IHT ) of the Hough map. (d) The
binary mask B, mapping the pixel location (x0 , y 0 ) highlighted in blue in (c) to its
corresponding set of bins in the Hough domain.
The traditional Hough transform [10,26] uses binary input where featuremaps
are real valued. Instead of binarizing the featuremaps, we define the Hough trans-
form similar to the Radon transform [3]. Therefore for a certain (ρ, θ) bin, our
Hough transform accumulates the featuremap activations F of the corresponding
pixels residing on that image direction:
X
HT (ρ, θ) = Fρ,θ (x(i), y(i)), (2)
i
where the relation between the pixel (x(i), y(i)) and bin (ρ, θ) is given in equa-
tion (1), and Fρ,θ (x(i), y(i)) is the featuremap value of the pixel indexed by i
along the (ρ, θ) line in the image. The HT is computed channel-wise, but for
simplicity, we ignore the channel dimension here. Figure 3.(b) shows the Hough
transform map for the input line in figure 3.(a), where we highlight in red the
bin corresponding to the line.
Note that in equation (2), there is a correspondence between the pixel (x(i), y(i))
and the bin (ρ, θ). We store this correspondence in a binary matrix, so we do
not need to recompute it. For each featuremap pixel, we remember in which HT
bins it votes, and generate a binary mask B of size: [W, H, Nρ , Nθ ] where [W, H]
is the size of the input featuremap F, and [Nρ , Nθ ] is the size of the HT map.
Thus, in practice when performing the Hough transform, we multiply the input
feature map F with B, channel-wise:
HT = FB. (3)
For gradient stability, we additionally normalize the HT by the width of the
input featuremap.
We transform to the Hough domain for each featuremap channel by looping
over all input pixels, F, rather than only the pixels along a certain line, and we
consider a range of discrete line parameters, (ρ, θ) where the pixels can vote. The
(ρ, θ) pair is mapped into Hough bins by uniformly sampling 60 angles in the
range [0, π] and 183 offsets in the range [0, d], where d is the image diagonal, and
the computed offsets from θ are assigned to the closest sampled offset values.
6 Y. Lin, S.L. Pintea, and J.C. van Gemert
Fig. 4. Local filters in the Hough domain correspond to global structure in the image
domain. (a) An input line in orange. (b) The line becomes a point in Hough domain.
(c) A local [−1, 0, 1]| filter in Hough domain. (d) The inverse of the local Hough filter
corresponds to a global line filter in the image domain.
In the backward pass, ∂F∂HT(x,y) , we use equation (4) without the normalization
over the number of angles, Nθ .
Similar to the forward Hough transform pass, we store the correspondence
between the pixels in the input featuremap (x, y) and the Hough transform bins
(ρ, θ), in the binary matrix, B. We implement the inverse Hough transform as a
matrix multiplication of B with the learned HT map, for each channel:
1
IHT = B HT . (5)
Nθ
Figure 3.(c) shows the IHT of the Hough transform map in figure 3.(b), while
figure 3.(d) shows the binary mask B for the pixel (x0 , y 0 ) highlighted in blue in
figure 3.(c), mapping it to its corresponding set of bins in the Hough map.
Fig. 5. Noisy local features aggregated globally by learning filters in the Hough do-
main. (a) Input featuremap with noisy discontinuous lines. (b) The output of the HT
layer using 183 offsets and 60 angles. (c) The result after filtering in the Hough domain.
The Hough map contains only the bins corresponding to lines. (d) The output of IHT
layer which receives as input the filtered Hough map. The lines are now clearly visible.
over the offsets, ρ, as the Hough transform is also computed channel-wise over
the input featuremaps. In Figure 5 we show an example; note that the input
featuremap lines are noisy and discontinuous and after applying 1D convolutions
in Hough space the informative bins are kept and when transformed back to the
image domain by the IHT contains clean lines.
Inspired by the Radon literature [23,29,46] we initialize the channel-wise
filters, f , with sign-inverted Laplacians by using the second order derivative of
a 1D Gaussian with randomly sampled scale, σ:
init ∂ 2 g(ρ, σ)
f (ρ) = − , (6)
∂ρ2
where g(ρ, σ) is a 1D Gaussian kernel. We normalize each filter to have unit
L1 norm and clip it to match the predefined spatial support. We, subsequently,
add two more 1D convolutional layers for reducing and merging the channels of
the Hough transform map. This lowers the computations needed in the inverse
Hough transform. Our block is visualized in Figure 2.
4 Experiments
We conduct experiments on three datasets: a controlled Line-Circle dataset, the
Wireframe (ShanghaiTech) [18] dataset and the York Urban [8] dataset. We
evaluate the added value of global Hough priors, convolutions in the Hough
domain, and data efficiency. We provide our source code online1 .
Fig. 6. Exp 1: Results in AP (average precision) and image examples of the Line-
Circle dataset. Using local+global information detects not only the direction of the
lines, as the global-only does, but also their extent.
Experimental setup. We evaluate our HT-IHT block design, specifically, the ef-
fect of convolutions in the Hough domain on a subset of the Wireframe (Shang-
haiTech) dataset [18]. The Wireframe dataset contains 5,462 images. We sample
from the training set 1,000 images for training, and 256 images for validation,
and use the official test split. As in [55], we resize all images to 512 × 512 px.
The goal is predicting pixels along line segments, where we report AP using the
same evaluation setup as in Exp 1, and we optimize a binary cross entropy loss.
We use a ResNet [16] backbone architecture, containing 2 convolutional layers
with ReLU, followed by 2 residual blocks, and another convolutional layer with a
sigmoid activation. The evaluation is done on predictions of 128×128 px, and the
ground truth are binary images with line segments. We insert our HT-IHT block
after every residual block. All layers are initialized with the He initialization [15].
We test the effect of convolutions in the Hough domain by considering in
our HT-IHT block: (0) not using any convolutions, (1) using a 1D convolution
over the offsets, (2) a channel-wise 1D convolution initialized with sign-inverted
Laplacian filters, (3) our complete HT-IHT block containing Laplacian-initialized
1D convolution and two additional 1D convolutions for merging and reducing
the channels, and (4) using three standard 3 × 3 convolutions.
Experimental analysis. Table 1 shows that using convolutions in the Hough do-
main is beneficial. The channel-wise Laplacian-initialized convolution is more ef-
fective than the standard 1D convolution using the He initialization [15]. Adding
extra convolutions for merging and reducing the channels gives a small improve-
ment in AP, however we use these for practical reasons rather than improved
performance. When comparing option (3) with (4), we see clearly the added
value of performing 1D convolutions over the offsets instead of using standard
3 × 3 convolutions. This experiment confirms that our choices, inspired from the
Radon transform practices, are indeed effective for line detection.
10 Y. Lin, S.L. Pintea, and J.C. van Gemert
6 Structural AP on Wireframe subsets 6 Junction mAP on Wireframe subsets PR curves for structural AP10
Precision
0.6
0.5
0 61.5 58.1 53.7 46.1 38.1
0 59.2 56.3 53.1 48.4 44.4
0.4
0.3
2 2 0.2
0.1 Dotted lines: L-CNN(9.7M)
Solid lines: HT-CNN(9.3M)
4100% 50% 25% 10% 5% 4100% 50% 25% 10% 5% 0.0
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Percentage of training samples Percentage of training samples Recall
6 Structural AP on Wireframe subsets 5 Junction mAP on Wireframe subsets PR curves for structural AP10
HT-HAWP(10.5M) HT-HAWP(10.5M) 5%
4 0.9
5 HT-HAWP(6.5M)
HAWP(10.3M)
HT-HAWP(6.5M)
HAWP(10.3M) 0.8
10%
25%
4 HAWP(6.5M) 3 HAWP(6.5M)
0.7 50%
100%
Precision
3 2 0.6
0.5
2 1 0.4
1 0 59.9 55.4 50.8 46.9 42.7
0.3
0.2
0 65.2 59.8 52.9 47.0 40.8
1 0.1 Dotted lines: HAWP(6.5M)
Solid lines: HT-HAWP(6.5M)
1100% 50% 25% 10% 5% 2100% 50% 25% 10% 5% 0.0
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Percentage of training samples Percentage of training samples Recall
(a) Structural-AP10 (b)Junction-mAP (c) PR for structural-AP10
Input image LCNN (100%) HT-LCNN (100%) LCNN (10%) HT-LCNN (10%)
Precision
0.6
0.5 0.4
0.4 0.3
0.3 LSD LCNN
Linelet HT-LCNN 0.2
0.2 MCMLSD HAWP
0.1 AFM HT-HAWP 0.1
WF-Parser
0.0 0.0
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
Recall Recall
(a) Precision-recall on Wireframe (ShanghaiTech) (b) Precision-recall on York Urban
Fig. 9. Exp 3.(b): Comparing our HT-LCNN and HT-HAWP with seven existing
methods using precision-recall scores on the Wireframe (ShanghaiTech) and York Ur-
ban datasets. Traditional knowledge-based methods are outperformed by deep learning
methods. Among the learning-based methods, our proposed HT-LCNN and HT-HAWP
achieve state-of-the-art performance, even in the full-data regime.
Precision
0.5 LCNN (9.7M) f=0.6 0.5 f=0.6
0.4 HT-LCNN (9.3M) f=0.5 0.4 f=0.5
0.3 HAWP (10.3M) f=0.5 0.3 f=0.5
f=0.4 f=0.4
0.2 HT-HAWP (10.5M) f=0.3 0.2 f=0.3
f=0.2 f=0.2
0.1 PPGNet (68.5M) 0.1
0.00.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.00.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Recall Recall
(a) AP on Wireframe (ShanghaiTech) (b) AP on York Urban
Fig. 10. Exp 3.(b): Comparing PPGNet[53] with (HT-)LCNN and (HT-)HAWP
on the Wireframe (ShanghaiTech) and York Urban datasets. PPGNet shows better
performance on the York Urban dataset, especially in high-recall region, while being
slightly less precise on the Wireframe dataset when compared to our HT-LCNN and
HT-HAWP methods. We show between brackets the number of parameters.
5 Conclusion
References
1. Almazan, E.J., Tal, R., Qian, Y., Elder, J.H.: Mcmlsd: A dynamic programming
approach to line segment detection. In: Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition. pp. 2031–2039 (2017)
2. Barbu, A., Mayo, D., Alverio, J., Luo, W., Wang, C., Gutfreund, D., Tenenbaum,
J., Katz, B.: Objectnet: A large-scale bias-controlled dataset for pushing the lim-
its of object recognition models. In: Advances in Neural Information Processing
Systems. pp. 9448–9458 (2019)
3. Beatty, J.: The Radon Transform and the Mathematics of Medical Imaging. Honors
thesis, Digital Commons @ Colby (2012)
4. Beltrametti, M.C., Campi, C., Massone, A.M., Torrente, M.L.: Geometry of the
hough transforms with applications to synthetic data. CoRR (2019)
5. Bruna, J., Mallat, S.: Invariant scattering convolution networks. IEEE transactions
on pattern analysis and machine intelligence 35(8), 1872–1886 (2013)
6. Burns, J.B., Hanson, A.R., Riseman, E.M.: Extracting straight lines. IEEE trans-
actions on pattern analysis and machine intelligence (4), 425–455 (1986)
7. Cho, N.G., Yuille, A., Lee, S.W.: A novel linelet-based representation for line seg-
ment detection. IEEE Transactions on Pattern Analysis and Machine Intelligence
40(5), 1195–1208 (2017)
8. Denis, P., Elder, J.H., Estrada, F.J.: Efficient edge-based methods for estimating
manhattan frames in urban imagery. In: European Conference on Computer Vision.
pp. 197–210. Springer (2008)
9. Do, M.N., Vetterli, M.: The finite ridgelet transform for image representation. IEEE
Transactions on image Processing 12(1), 16–28 (2003)
10. Duda, R.O., Hart, P.E.: Use of the hough transformation to detect lines and curves
in pictures. Communications of the ACM 15(1), 11–15 (1972)
11. Furukawa, Y., Shinagawa, Y.: Accurate and robust line segment extraction by
analyzing distribution around peaks in hough space. Computer Vision and Image
Understanding 92(1), 1–25 (2003)
12. Gershikov, E., Libe, T., Kosolapov, S.: Horizon line detection in marine images:
which method to choose? International Journal on Advances in Intelligent Systems
6(1) (2013)
13. Guerreiro, R.F., Aguiar, P.M.: Connectivity-enforcing hough transform for the ro-
bust extraction of line segments. IEEE Transactions on Image Processing 21(12),
4819–4829 (2012)
14. He, J., Ma, J.: Radon inversion via deep learning. In: Medical Imaging (2018)
15. He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: Surpassing human-
level performance on imagenet classification. In: Proceedings of the IEEE interna-
tional conference on computer vision. pp. 1026–1034 (2015)
16. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In:
Proceedings of the IEEE conference on computer vision and pattern recognition.
pp. 770–778 (2016)
17. Hillel, A.B., Lerner, R., Levi, D., Raz, G.: Recent progress in road and lane detec-
tion: a survey. Machine vision and applications 25(3), 727–745 (2014)
18. Huang, K., Wang, Y., Zhou, Z., Ding, T., Gao, S., Ma, Y.: Learning to parse
wireframes in images of man-made environments. In: Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition. pp. 626–635 (2018)
19. Jacobsen, J.H., van Gemert, J., Lou, Z., Smeulders, A.W.: Structured receptive
fields in cnns. In: Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition. pp. 2610–2619 (2016)
16 Y. Lin, S.L. Pintea, and J.C. van Gemert
20. Kamat-Sadekar, V., Ganesan, S.: Complete description of multiple line segments
using the hough transform. Image and Vision Computing 16(9-10), 597–613 (1998)
21. Kayhan, O.S., van Gemert, J.C.: On translation invariance in cnns: Convolutional
layers can exploit absolute spatial location. In: Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition. pp. 14274–14285 (2020)
22. Lee, S., Kim, J., Shin Yoon, J., Shin, S., Bailo, O., Kim, N., Lee, T.H., Seok Hong,
H., Han, S.H., So Kweon, I.: Vpgnet: Vanishing point guided network for lane and
road marking detection and recognition. In: Proceedings of the IEEE international
conference on computer vision. pp. 1947–1955 (2017)
23. Magnusson, M.: Linogram and Other Direct Fourier Methods for Tomographic
Reconstruction. Linköping studies in science and technology: Dissertations, De-
partment of Mechanical Engineering, Linköping University (1993)
24. Maire, M., Arbelaez, P., Fowlkes, C., Malik, J.: Using contours to detect and lo-
calize junctions in natural images. In: 2008 IEEE Conference on Computer Vision
and Pattern Recognition. pp. 1–8. IEEE (2008)
25. Martin, D.R., Fowlkes, C.C., Malik, J.: Learning to detect natural image bound-
aries using local brightness, color, and texture cues. IEEE transactions on pattern
analysis and machine intelligence 26(5), 530–549 (2004)
26. Matas, J., Galambos, C., Kittler, J.: Robust detection of lines using the progressive
probabilistic hough transform. Computer Vision and Image Understanding 78(1),
119–137 (2000)
27. Min, J., Lee, J., Ponce, J., Cho, M.: Hyperpixel flow: Semantic correspondence with
multi-layer neural features. In: Proceedings of the IEEE International Conference
on Computer Vision. pp. 3395–3404 (2019)
28. Nguyen, V.N., Jenssen, R., Roverso, D.: Ls-net: Fast single-shot line-segment de-
tector. CoRR (2019)
29. Nikolaev, D.P., Karpenko, S.M., Nikolaev, I.P., Nikolayev, P.P.: Hough transform:
underestimated tool in the computer vision field. In: Proceedings of the 22th Eu-
ropean Conference on Modelling and Simulation. vol. 238, p. 246 (2008)
30. Niu, J., Lu, J., Xu, M., Lv, P., Zhao, X.: Robust lane detection using two-stage
feature extraction with curve fitting. Pattern Recognition 59, 225–233 (2016)
31. Pătrăucean, V., Gurdjos, P., Von Gioi, R.G.: A parameterless line segment and
elliptical arc detector with enhanced ellipse fitting. In: European Conference on
Computer Vision. pp. 572–585 (2012)
32. Porzi, L., Rota Bulò, S., Ricci, E.: A deeply-supervised deconvolutional network for
horizon line detection. In: Proceedings of the 24th ACM international conference
on Multimedia. pp. 137–141 (2016)
33. Qi, C.R., Litany, O., He, K., Guibas, L.J.: Deep hough voting for 3d object de-
tection in point clouds. In: Proceedings of the IEEE International Conference on
Computer Vision. pp. 9277–9286 (2019)
34. Rim, D.: Exact and fast inversion of the approximate discrete radon transform
from partial data. Applied Mathematics Letters 102, 106159 (2020)
35. Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z.,
Karpathy, A., Khosla, A., Bernstein, M., Berg, A.C., Fei-Fei, L.: ImageNet Large
Scale Visual Recognition Challenge. International Journal of Computer Vision
(IJCV) 115(3), 211–252 (2015). https://doi.org/10.1007/s11263-015-0816-y
36. Satzoda, R.K., Trivedi, M.M.: Efficient lane and vehicle detection with integrated
synergies (elvis). In: 2014 IEEE Conference on Computer Vision and Pattern
Recognition Workshops. pp. 708–713 (2014)
37. Shelhamer, E., Wang, D., Darrell, T.: Blurring the line between structure and
learning to optimize and adapt receptive fields. CoRR (2019)
Deep Hough-Transform Line Priors 17
38. Sheshkus, A., Ingacheva, A., Arlazarov, V., Nikolaev, D.: Houghnet: neural network
architecture for vanishing points detection. International Conference on Document
Analysis and Recognition (ICDAR) (2019)
39. Simon, G., Fond, A., Berger, M.O.: A-contrario horizon-first vanishing point detec-
tion using second-order grouping laws. In: Proceedings of the European Conference
on Computer Vision (ECCV). pp. 318–333 (2018)
40. Sosnovik, I., Szmaja, M., Smeulders, A.: Scale-equivariant steerable networks. In-
ternational Conference on Learning Representations (2020)
41. Sun, J., Liang, L., Wen, F., Shum, H.Y.: Image vectorization using optimized
gradient meshes. ACM Transactions on Graphics (TOG) 26(3), 11–es (2007)
42. Toft, P.: The Radon Transform: Theory and Implementation. Department of Math-
ematical Modelling, Section for Digital Signal Processing, Technical University of
Denmark (1996)
43. Urban, G., Geras, K.J., Kahou, S.E., Aslan, O., Wang, S., Caruana, R., Mohamed,
A., Philipose, M., Richardson, M.: Do deep convolutional nets really need to be
deep and convolutional? International Conference on Learning Representations
(2016)
44. Von Gioi, R.G., Jakubowicz, J., Morel, J.M., Randall, G.: Lsd: A fast line segment
detector with a false detection control. IEEE Transactions on Pattern Analysis and
Machine Intelligence 32(4), 722–732 (2008)
45. Von Gioi, R.G., Jakubowicz, J., Morel, J.M., Randall, G.: On straight line segment
detection. Journal of Mathematical Imaging and Vision 32(3), 313 (2008)
46. Wei, H., Bing, W., Yue, Z.: X-linenet: Detecting aircraft in remote sensing images
by a pair of intersecting line segments. CoRR (2019)
47. Wei, Q., Feng, D., Zheng, W.: Funnel transform for straight line detection. CoRR
(2019)
48. Workman, S., Zhai, M., Jacobs, N.: Horizon lines in the wild. British Machine
Vision Conference (2016)
49. Xu, Z., Shin, B.S., Klette, R.: Accurate and robust line segment extraction using
minimum entropy with hough transform. IEEE Transactions on Image Processing
24(3), 813–822 (2014)
50. Xu, Z., Shin, B.S., Klette, R.: A statistical method for line segment detection.
Computer Vision and Image Understanding 138, 61–73 (2015)
51. Xue, N., Bai, S., Wang, F., Xia, G.S., Wu, T., Zhang, L.: Learning attraction
field representation for robust line segment detection. In: The IEEE Conference on
Computer Vision and Pattern Recognition (June 2019)
52. Xue, N., Wu, T., Bai, S., Wang, F., Xia, G.S., Zhang, L., Torr, P.H.: Holistically-
attracted wireframe parsing. In: Conference on Computer Vision and Pattern
Recognition (2020)
53. Zhang, Z., Li, Z., Bi, N., Zheng, J., Wang, J., Huang, K., Luo, W., Xu, Y., Gao,
S.: Ppgnet: Learning point-pair graph for line segment detection. In: Conference
on Computer Vision and Pattern Recognition (2019)
54. Zhou, Y., Qi, H., Ma, Y.: End-to-end wireframe parsing. In: Proceedings of the
IEEE International Conference on Computer Vision. pp. 962–971 (2019)
55. Zhou, Y., Qi, H., Zhai, Y., Sun, Q., Chen, Z., Wei, L.Y., Ma, Y.: Learning to
reconstruct 3d manhattan wireframes from a single image. In: Proceedings of the
IEEE International Conference on Computer Vision. pp. 7698–7707 (2019)
56. Zou, J.J., Yan, H.: Cartoon image vectorization based on shape subdivision. In:
Proceedings. Computer Graphics International 2001. pp. 225–231 (2001)