YL

Authored

6 records found

Deep Vanishing Point Detection

Geometric priors make dataset variations vanish

Deep learning has improved vanishing point detection in images. Yet, deep networks require expensive annotated datasets trained on costly hardware and do not generalize to even slightly different domains, and minor problem variants. Here, we address these issues by injecting deep ...
A number of computer vision deep regression approaches report improved results when adding a classification loss to the regression loss. Here, we explore why this is useful in practice and when it is beneficial. To do so, we start from precisely controlled dataset variations and ...
Current work on lane detection relies on large manually annotated datasets. We reduce the dependency on annotations by leveraging massive cheaply available unlabelled data. We propose a novel loss function exploiting geometric knowledge of lanes in Hough space, where a lane can b ...
The humanly constructed world is well-organized in space. A prominent feature of this artificial world is the presence of repetitive structures and coherent patterns, such as lines, junctions, wireframes of a building, and footprints of a city. These structures and patterns facil ...
Classical work on line segment detection is knowledge-based; it uses carefully designed geometric priors using either image gradients, pixel groupings, or Hough transform variants. Instead, current deep learning methods do away with all prior knowledge and replace priors by train ...
Transformers can generate predictions in two approaches: 1. auto-regressively by conditioning each sequence element on the previous ones, or 2. directly produce an output sequences in parallel. While research has mostly explored upon this difference on sequential tasks in NLP, we ...

Contributed

7 records found

Although the pixel-wise labelling approaches have been exploited in depth and achieve good results in segmentation tasks, the grouped pixels are not ideal output for many end-users. In this paper, we propose a vertex-voting-based approach that can directly extract the polygon rep ...
Neural Radiance Fields (NeRFs) have demonstrated remarkable capabilities in photo-realistic 3D reconstruction. NeRFs often take as input posed images where the camera poses come from either off-the-shelf S extit{f}M or online optimization together with NeRFs. However, we find tha ...
Convolutional Neural Networks (CNNs) have made significant strides in the field of image processing over the last decade. Different approaches have been taken and improvements have been suggested. This paper looks at a newer novelty to neural networks for image counting, which is ...
Convolutional Neural Networks are particularly vulnerable to attacks that manipulate their data, which are usually called adversarial attacks. In this paper, a method of filtering images using the Fast Fourier Transform is explored, along with its potential to be used as a defens ...
Wheat is among the most important grains worldwide. For the assessment of wheat fields, image detection of spikes atop the plant containing grain is used. Previous work in deep learning for precision agriculture employs the already established object detectors, Faster R-CNN and Y ...
Structure-from-Motion (SfM) and Neural Radiance Fields (NeRFs) have significantly advanced 3D reconstruction in multi-view scenarios. Despite their success in handling non-repetitive, texture-rich scenes, applying such techniques to real-world scenarios with texture-less and repe ...
2D to 3D transfer learning has proven to be an effective method to transfer strong 2D representations into a 3D network. Recent advancements show that casting semantic segmentation of outdoor LiDAR point clouds as a 2D problem, such as through range projection, enables the proces ...