<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Publications &#8211; Jan-Nico Zaech</title>
	<atom:link href="/category/publications/feed/" rel="self" type="application/rss+xml" />
	<link>/</link>
	<description></description>
	<lastBuildDate>Tue, 11 Jun 2024 08:26:17 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.5.4</generator>
	<item>
		<title>Probabilistic Sampling of Balanced K-Means using Adiabatic Quantum Computing</title>
		<link>/probabilistic-sampling-of-balanced-k-means-using-adiabatic-quantum-computing/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Sat, 01 Jun 2024 00:00:06 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Computer Vision]]></category>
		<category><![CDATA[Quantum Computing]]></category>
		<guid isPermaLink="false">/?p=256</guid>

					<description><![CDATA[Jan-Nico Zaech, Martin Danelljan, Tolga Birdal, Luc Van Gool IEEE Conference on Computer Vision and Pattern Recognition 2024 (CVPR) Abstract Adiabatic quantum computing (AQC) is a promising approach for discrete and often NP-hard optimization problems. Current AQCs allow to implement problems of research interest, which has sparked the development of quantum representations for many computer [&#8230;]]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image size-full is-resized"><img fetchpriority="high" decoding="async" width="1071" height="315" src="/wp-content/uploads/2023/11/teaser.png" alt="" class="wp-image-258" style="width:1242px;height:auto" srcset="/wp-content/uploads/2023/11/teaser.png 1071w, /wp-content/uploads/2023/11/teaser-300x88.png 300w, /wp-content/uploads/2023/11/teaser-1024x301.png 1024w, /wp-content/uploads/2023/11/teaser-768x226.png 768w" sizes="(max-width: 1071px) 100vw, 1071px" /></figure>



<p>Jan-Nico Zaech, Martin Danelljan, Tolga Birdal, Luc Van Gool</p>



<p>IEEE Conference on Computer Vision and Pattern Recognition 2024 (CVPR)</p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Adiabatic quantum computing (AQC) is a promising approach for discrete and often NP-hard optimization problems. Current AQCs allow to implement problems of research interest, which has sparked the development of quantum representations for many computer vision tasks. Despite requiring multiple measurements from the noisy AQC, current approaches only utilize the best measurement, discarding information contained in the remaining ones. In this work, we explore the potential of using this information for probabilistic balanced k-means clustering. Instead of discarding non-optimal solutions, we propose to use them to compute calibrated posterior probabilities with little additional compute cost. This allows us to identify ambiguous solutions and data points, which we demonstrate on a D-Wave AQC on synthetic tasks and real visual data.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Optimizing Long-Term Player Tracking and Identification in NAO Robot Soccer by fusing Game-state and External Video</title>
		<link>/optimizing-long-term-player-tracking-and-identification-in-nao-robot-soccer-by-fusing-game-state-and-external-video/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Fri, 02 Jun 2023 15:45:51 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[RoboCup]]></category>
		<category><![CDATA[Tracking]]></category>
		<guid isPermaLink="false">/?p=83</guid>

					<description><![CDATA[A collaborative sensing approach for multi object tracking of robots.]]></description>
										<content:encoded><![CDATA[
<div class="wp-block-columns is-layout-flex wp-container-core-columns-is-layout-1 wp-block-columns-is-layout-flex">
<div class="wp-block-column is-layout-flow wp-block-column-is-layout-flow" style="flex-basis:53%">
<figure class="wp-block-image size-large"><img decoding="async" width="1024" height="505" src="/wp-content/uploads/2023/06/method-1-1024x505.jpg" alt="" class="wp-image-84" srcset="/wp-content/uploads/2023/06/method-1-1024x505.jpg 1024w, /wp-content/uploads/2023/06/method-1-300x148.jpg 300w, /wp-content/uploads/2023/06/method-1-768x379.jpg 768w, /wp-content/uploads/2023/06/method-1-1536x757.jpg 1536w, /wp-content/uploads/2023/06/method-1.jpg 1588w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>
</div>



<div class="wp-block-column is-vertically-aligned-center is-layout-flow wp-block-column-is-layout-flow">
<figure class="wp-block-image size-large"><img decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/field-edited-1024x576.jpg" alt="" class="wp-image-41" srcset="/wp-content/uploads/2023/06/field-edited-1024x576.jpg 1024w, /wp-content/uploads/2023/06/field-edited-300x169.jpg 300w, /wp-content/uploads/2023/06/field-edited-768x432.jpg 768w, /wp-content/uploads/2023/06/field-edited.jpg 1307w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>
</div>
</div>



<p>Giuliano Albanese*, Arka Mitra*, Jan-Nico Zaech*, Yupeng Zhao*, Ajad Chhatkuli, and Luc Van Gool</p>



<p>International Conference on Robotics and Automation Workshops, ICRA 2023 (<a href="https://coperception.github.io/index.html">CoPerception: Collaborative Perception and Learning</a>)</p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Monitoring a fleet of robots requires stable long-term tracking with re-identification, which is yet an unsolved challenge in many scenarios. One application of this is the analysis of autonomous robotic soccer games at RoboCup. Tracking these games requires the handling of identically looking players, strong occlusions, and non-professional video recordings, but also offers state information estimated by the robots. In order to make effective use of the information coming from the robot sensors, we propose a robust tracking and identification<br>pipeline. It fuses external non-calibrated camera data with the robots’ internal states using quadratic optimization for tracklet matching. The approach is validated using game recordings from previous RoboCup World Cups.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Adiabatic Quantum Computing for Multi Object Tracking</title>
		<link>/adiabatic-quantum-computing-for-multi-object-tracking/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Sun, 19 Jun 2022 00:00:18 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Quantum Computing]]></category>
		<category><![CDATA[Tracking]]></category>
		<guid isPermaLink="false">/?p=74</guid>

					<description><![CDATA[A Multi-Object Tracking algorithm that can be solved with Adiabatic Quantum Computing]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image size-large is-resized"><img loading="lazy" decoding="async" src="/wp-content/uploads/2023/06/teaser-1-1024x576.jpg" alt="" class="wp-image-75" width="1239" height="697" srcset="/wp-content/uploads/2023/06/teaser-1-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-1-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-1-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-1-1536x863.jpg 1536w, /wp-content/uploads/2023/06/teaser-1.jpg 1612w" sizes="(max-width: 1239px) 100vw, 1239px" /></figure>



<p>Jan-Nico Zaech, Alexander Liniger, Martin Danelljan, Dengxin Dai, Luc Van Gool</p>



<p><em>Conference on Computer Vision and Pattern Recognition, CVPR 2022</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Multi-Object Tracking (MOT) is most often approached in the tracking-by-detection paradigm, where object detections are associated through time. The association step naturally leads to discrete optimization problems. As these optimization problems are often NP-hard, they can only be solved exactly for small instances on current hardware. Adiabatic quantum computing (AQC) offers a solution for this, as it has the potential to provide a considerable speedup on a range of NP-hard optimization problems in the near future. However, current MOT formulations are unsuitable for quantum computing due to their scaling properties. In this work, we therefore propose the first MOT formulation designed to be solved with AQC. We employ an Ising model that represents the quantum mechanical system implemented on the AQC. We show that our approach is competitive compared with state-of-the-art optimization-based approaches, even when using of-the-shelf integer programming solvers. Finally, we demonstrate that our MOT problem is already solvable on the current generation of real quantum computers for small examples, and analyze the properties of the measured solutions.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Learnable Online Graph Representations for 3D Multi-Object Tracking</title>
		<link>/learnable-online-graph-representations-for-3d-multi-object-tracking/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Mon, 23 May 2022 16:41:28 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<category><![CDATA[Tracking]]></category>
		<guid isPermaLink="false">/?p=109</guid>

					<description><![CDATA[An online 3D Multi-Object Tracking method based on graph neural networks.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="575" src="/wp-content/uploads/2023/06/teaser-3-1024x575.jpg" alt="" class="wp-image-111" srcset="/wp-content/uploads/2023/06/teaser-3-1024x575.jpg 1024w, /wp-content/uploads/2023/06/teaser-3-300x168.jpg 300w, /wp-content/uploads/2023/06/teaser-3-768x431.jpg 768w, /wp-content/uploads/2023/06/teaser-3-1536x862.jpg 1536w, /wp-content/uploads/2023/06/teaser-3.jpg 1916w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Jan-Nico Zaech, Dengxin Dai, Alexander Liniger, Martin Danelljan, Luc Van Gool</p>



<p><em>International Conference on Robotics and Automation Workshops, ICRA 2022</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Tracking of objects in 3D is a fundamental task in computer vision that finds use in a wide range of applications<br>such as autonomous driving, robotics or augmented reality. Most recent approaches for 3D multi object tracking (MOT) from LIDAR use object dynamics together with a set of handcrafted features to match detections of objects. However, manually designing such features and heuristics is cumbersome and often leads to suboptimal performance. In this work, we instead strive towards a unified and learning based approach to the 3D MOT problem. We design a graph structure to jointly process detection and track states in an online manner. To this end, we employ a Neural Message Passing network for data association that is fully trainable. Our approach provides a natural way for track initialization and handling of false positive detections, while significantly improving track stability. We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Unsupervised robust domain adaptation without source data</title>
		<link>/unsupervised-robust-domain-adaptation-without-source-data/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Tue, 04 Jan 2022 20:39:34 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Domain Adaptation]]></category>
		<guid isPermaLink="false">/?p=123</guid>

					<description><![CDATA[A method that keeps a network robust against adversarial images during source-free domain adaptation.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-4-1024x576.jpg" alt="" class="wp-image-124" srcset="/wp-content/uploads/2023/06/teaser-4-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-4-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-4-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-4-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-4.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Peshal Agarwal, Danda Pani Paudel, Jan-Nico Zaech, Luc Van Gool</p>



<p><em>Winter Conference on Applications of Computer Vision, WACV 2022</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>We study the problem of robust domain adaptation in the context of  unavailable target labels and source data. The considered robustness is  against adversarial perturbations. This paper aims at answering the  question of finding the right strategy to make the target model robust  and accurate in the setting of unsupervised domain adaptation without  source data. The major findings of this paper are:(i) robust source  models can be transferred robustly to the target;(ii) robust domain  adaptation can greatly benefit from non-robust pseudo-labels and the  pair-wise contrastive loss. The proposed method of using non-robust  pseudo-labels performs surprisingly well on both clean and adversarial  samples, for the task of image classification. We show a consistent  performance improvement of over 10% in accuracy against the tested  baselines on four benchmark datasets.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Decoder fusion RNN: Context and interaction aware decoders for trajectory prediction</title>
		<link>/decoder-fusion-rnn-context-and-interaction-aware-decoders-for-trajectory-prediction/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Mon, 27 Sep 2021 20:47:00 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Action prediction]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<guid isPermaLink="false">/?p=129</guid>

					<description><![CDATA[A multi-headed attention based method for vehicle trajectory prediction using map data encoded on a graph.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-5-1024x576.jpg" alt="" class="wp-image-130" srcset="/wp-content/uploads/2023/06/teaser-5-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-5-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-5-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-5-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-5.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Edoardo Mello Rella, Jan-Nico Zaech, Alexander Liniger, Luc Van Gool</p>



<p><em>International Conference on Intelligent Robots and Systems, IROS 2021</em></p>



<h2 class="wp-block-heading">Abstract</h2>



<p>Forecasting the future behavior of all traffic agents in the vicinity is a key task to achieve safe and reliable autonomous driving systems. It is a challenging problem as agents adjust their behavior depending on their intentions, the others’ actions, and the road layout. In this paper, we propose Decoder Fusion RNN (DF-RNN), a recurrent, attention-based approach for motion forecasting. Our network is composed of a recurrent behavior encoder, an inter-agent multi-headed attention module, and a context-aware decoder. We design a map encoder that embeds polyline segments, combines them to create a graph structure, and merges their relevant parts with<br>the agents’ embeddings. We fuse the encoded map information with further inter-agent interactions only inside the decoder and propose to use explicit training as a method to effectively utilize the information available. We demonstrate the efficacy of our method by testing it on the Argoverse motion forecasting dataset and show its state-of-the-art performance on the public benchmark.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Action sequence predictions of vehicles in urban environments using map and social context</title>
		<link>/action-sequence-predictions-of-vehicles-in-urban-environments-using-map-and-social-context/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Tue, 27 Oct 2020 21:05:00 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Action prediction]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<guid isPermaLink="false">/?p=135</guid>

					<description><![CDATA[Views the traffic agent trajectory prediction task form a classification perspective and proposes a method to automatically annotate trajectory data by using graph-based maps.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-6-1024x576.jpg" alt="" class="wp-image-137" srcset="/wp-content/uploads/2023/06/teaser-6-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-6-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-6-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-6-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-6.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Jan-Nico Zaech, Dengxin Dai, Alexander Liniger, Luc Van Gool</p>



<p><em>International Conference on Intelligent Robots and Systems, IROS 2020</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>This work studies the problem of predicting the sequence of future actions for surrounding vehicles in real-world driving scenarios. To this aim, we make three main contributions. The first contribution is an automatic method to convert the trajectories recorded in real-world driving scenarios to action sequences with the help of HD maps. The method enables automatic dataset creation for this task from large-scale driving data. Our second contribution lies in applying the method to the well-known traffic agent tracking and prediction dataset Argoverse, resulting in 228,000 action sequences. Additionally, 2,245 action sequences were manually annotated for testing. The third contribution is to propose a novel action sequence prediction method by integrating past positions and velocities of the traffic agents, map information and social context into a single end-to-end trainable neural network. Our experiments prove the merit of the data creation method and the value of the created dataset – prediction performance improves consistently with the size of the dataset and shows that our action prediction method outperforms comparing models.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Texture Underfitting for Domain Adaptation</title>
		<link>/texture-underfitting-for-domain-adaptation/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Mon, 28 Oct 2019 19:05:45 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Domain Adaptation]]></category>
		<guid isPermaLink="false">/?p=150</guid>

					<description><![CDATA[A method to use image structure in domain adaptation implemented as a two stage training process.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-8-1024x576.jpg" alt="" class="wp-image-151" srcset="/wp-content/uploads/2023/06/teaser-8-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-8-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-8-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-8-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-8.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Jan-Nico Zaech, Dengxin Dai, Martin Hahner, Luc Van Gool</p>



<p><em>Inteligent Transportation Systems Conference (IEEE), ITSC 2019</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Comprehensive semantic segmentation is one of the key components for robust scene understanding and a<br>requirement to enable autonomous driving. Driven by large scale datasets, convolutional neural networks show impressive results on this task. However, a segmentation algorithm generalizing to various scenes and conditions would require an enormously diverse dataset, making the labour intensive data acquisition and labeling process prohibitively expensive. Under the assumption of structural similarities between segmentation<br>maps, domain adaptation promises to resolve this challenge by transferring knowledge from existing, potentially simulated datasets to new environments where no supervision exists. While the performance of this approach is contingent on the concept that neural networks learn a high level understanding of scene structure, recent work suggests that neural networks are biased towards overfitting to texture instead of learning structural and shape information. Considering the ideas underlying semantic segmentation, we employ random image stylization to augment the training dataset and propose a training procedure that facilitates texture underfitting to improve the performance of domain adaptation. In experiments with supervised as well as unsupervised methods for the task of synthetic-to-real domain adaptation, we show that our approach outperforms conventional training methods.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Learning to avoid poor images: Towards task-aware C-arm cone-beam CT trajectories</title>
		<link>/learning-to-avoid-poor-images-towards-task-aware-c-arm-cone-beam-ct-trajectories/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Sun, 13 Oct 2019 08:48:48 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<category><![CDATA[Medical Imaging]]></category>
		<guid isPermaLink="false">/?p=146</guid>

					<description><![CDATA[A robotic CBCT system that that predicts an acquisition trajectory optimized online during a scan.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-7-1024x576.jpg" alt="" class="wp-image-147" srcset="/wp-content/uploads/2023/06/teaser-7-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-7-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-7-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-7-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-7.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Jan-Nico Zaech, Cong Gao, Bastian Bier, Russell Taylor, Andreas Maier, Nassir Navab, Mathias Unberath</p>



<p><em>Medical Image Computing and Computer Assisted Intervention, MICCAI 2019</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Metal artifacts in computed tomography (CT) arise from a mismatch between physics of image formation and idealized assumptions during tomographic reconstruction. These artifacts are particularly strong around metal implants, inhibiting widespread adoption of 3D cone-beam CT (CBCT) despite clear opportunity for intra-operative verification of implant positioning, e. g. in spinal fusion surgery. On synthetic and real data, we demonstrate that much of the artifact can be avoided by acquiring better data for reconstruction in a task-aware and patient-specific manner, and describe the first step towards the envisioned task-aware CBCT protocol. The traditional short-scan CBCT trajectory is planar, with little room for scene-specific adjustment. We extend this trajectory by autonomously adjusting out-of-plane angulation. This enables C-arm source trajectories that are scene-specific in that they avoid acquiring ”poor images”, characterized by beam hardening, photon starvation, and noise. The recommendation of ideal out-of-plane angulation is performed on-the-fly using a deep convolutional neural network that regresses a detectability-rank derived from imaging physics.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
