<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Autonomous Systems &#8211; Jan-Nico Zaech</title>
	<atom:link href="/tag/autonomous-systems/feed/" rel="self" type="application/rss+xml" />
	<link>/</link>
	<description></description>
	<lastBuildDate>Thu, 13 Jul 2023 17:13:18 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.5.4</generator>
	<item>
		<title>Learnable Online Graph Representations for 3D Multi-Object Tracking</title>
		<link>/learnable-online-graph-representations-for-3d-multi-object-tracking/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Mon, 23 May 2022 16:41:28 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<category><![CDATA[Tracking]]></category>
		<guid isPermaLink="false">/?p=109</guid>

					<description><![CDATA[An online 3D Multi-Object Tracking method based on graph neural networks.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img fetchpriority="high" decoding="async" width="1024" height="575" src="/wp-content/uploads/2023/06/teaser-3-1024x575.jpg" alt="" class="wp-image-111" srcset="/wp-content/uploads/2023/06/teaser-3-1024x575.jpg 1024w, /wp-content/uploads/2023/06/teaser-3-300x168.jpg 300w, /wp-content/uploads/2023/06/teaser-3-768x431.jpg 768w, /wp-content/uploads/2023/06/teaser-3-1536x862.jpg 1536w, /wp-content/uploads/2023/06/teaser-3.jpg 1916w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Jan-Nico Zaech, Dengxin Dai, Alexander Liniger, Martin Danelljan, Luc Van Gool</p>



<p><em>International Conference on Robotics and Automation Workshops, ICRA 2022</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Tracking of objects in 3D is a fundamental task in computer vision that finds use in a wide range of applications<br>such as autonomous driving, robotics or augmented reality. Most recent approaches for 3D multi object tracking (MOT) from LIDAR use object dynamics together with a set of handcrafted features to match detections of objects. However, manually designing such features and heuristics is cumbersome and often leads to suboptimal performance. In this work, we instead strive towards a unified and learning based approach to the 3D MOT problem. We design a graph structure to jointly process detection and track states in an online manner. To this end, we employ a Neural Message Passing network for data association that is fully trainable. Our approach provides a natural way for track initialization and handling of false positive detections, while significantly improving track stability. We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Decoder fusion RNN: Context and interaction aware decoders for trajectory prediction</title>
		<link>/decoder-fusion-rnn-context-and-interaction-aware-decoders-for-trajectory-prediction/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Mon, 27 Sep 2021 20:47:00 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Action prediction]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<guid isPermaLink="false">/?p=129</guid>

					<description><![CDATA[A multi-headed attention based method for vehicle trajectory prediction using map data encoded on a graph.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-5-1024x576.jpg" alt="" class="wp-image-130" srcset="/wp-content/uploads/2023/06/teaser-5-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-5-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-5-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-5-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-5.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Edoardo Mello Rella, Jan-Nico Zaech, Alexander Liniger, Luc Van Gool</p>



<p><em>International Conference on Intelligent Robots and Systems, IROS 2021</em></p>



<h2 class="wp-block-heading">Abstract</h2>



<p>Forecasting the future behavior of all traffic agents in the vicinity is a key task to achieve safe and reliable autonomous driving systems. It is a challenging problem as agents adjust their behavior depending on their intentions, the others’ actions, and the road layout. In this paper, we propose Decoder Fusion RNN (DF-RNN), a recurrent, attention-based approach for motion forecasting. Our network is composed of a recurrent behavior encoder, an inter-agent multi-headed attention module, and a context-aware decoder. We design a map encoder that embeds polyline segments, combines them to create a graph structure, and merges their relevant parts with<br>the agents’ embeddings. We fuse the encoded map information with further inter-agent interactions only inside the decoder and propose to use explicit training as a method to effectively utilize the information available. We demonstrate the efficacy of our method by testing it on the Argoverse motion forecasting dataset and show its state-of-the-art performance on the public benchmark.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Action sequence predictions of vehicles in urban environments using map and social context</title>
		<link>/action-sequence-predictions-of-vehicles-in-urban-environments-using-map-and-social-context/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Tue, 27 Oct 2020 21:05:00 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Action prediction]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<guid isPermaLink="false">/?p=135</guid>

					<description><![CDATA[Views the traffic agent trajectory prediction task form a classification perspective and proposes a method to automatically annotate trajectory data by using graph-based maps.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-6-1024x576.jpg" alt="" class="wp-image-137" srcset="/wp-content/uploads/2023/06/teaser-6-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-6-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-6-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-6-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-6.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Jan-Nico Zaech, Dengxin Dai, Alexander Liniger, Luc Van Gool</p>



<p><em>International Conference on Intelligent Robots and Systems, IROS 2020</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>This work studies the problem of predicting the sequence of future actions for surrounding vehicles in real-world driving scenarios. To this aim, we make three main contributions. The first contribution is an automatic method to convert the trajectories recorded in real-world driving scenarios to action sequences with the help of HD maps. The method enables automatic dataset creation for this task from large-scale driving data. Our second contribution lies in applying the method to the well-known traffic agent tracking and prediction dataset Argoverse, resulting in 228,000 action sequences. Additionally, 2,245 action sequences were manually annotated for testing. The third contribution is to propose a novel action sequence prediction method by integrating past positions and velocities of the traffic agents, map information and social context into a single end-to-end trainable neural network. Our experiments prove the merit of the data creation method and the value of the created dataset – prediction performance improves consistently with the size of the dataset and shows that our action prediction method outperforms comparing models.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Learning to avoid poor images: Towards task-aware C-arm cone-beam CT trajectories</title>
		<link>/learning-to-avoid-poor-images-towards-task-aware-c-arm-cone-beam-ct-trajectories/</link>
		
		<dc:creator><![CDATA[zaech]]></dc:creator>
		<pubDate>Sun, 13 Oct 2019 08:48:48 +0000</pubDate>
				<category><![CDATA[Publications]]></category>
		<category><![CDATA[Autonomous Systems]]></category>
		<category><![CDATA[Medical Imaging]]></category>
		<guid isPermaLink="false">/?p=146</guid>

					<description><![CDATA[A robotic CBCT system that that predicts an acquisition trajectory optimized online during a scan.]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="576" src="/wp-content/uploads/2023/06/teaser-7-1024x576.jpg" alt="" class="wp-image-147" srcset="/wp-content/uploads/2023/06/teaser-7-1024x576.jpg 1024w, /wp-content/uploads/2023/06/teaser-7-300x169.jpg 300w, /wp-content/uploads/2023/06/teaser-7-768x432.jpg 768w, /wp-content/uploads/2023/06/teaser-7-1536x864.jpg 1536w, /wp-content/uploads/2023/06/teaser-7.jpg 1920w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Jan-Nico Zaech, Cong Gao, Bastian Bier, Russell Taylor, Andreas Maier, Nassir Navab, Mathias Unberath</p>



<p><em>Medical Image Computing and Computer Assisted Intervention, MICCAI 2019</em></p>



<h3 class="wp-block-heading">Abstract</h3>



<p>Metal artifacts in computed tomography (CT) arise from a mismatch between physics of image formation and idealized assumptions during tomographic reconstruction. These artifacts are particularly strong around metal implants, inhibiting widespread adoption of 3D cone-beam CT (CBCT) despite clear opportunity for intra-operative verification of implant positioning, e. g. in spinal fusion surgery. On synthetic and real data, we demonstrate that much of the artifact can be avoided by acquiring better data for reconstruction in a task-aware and patient-specific manner, and describe the first step towards the envisioned task-aware CBCT protocol. The traditional short-scan CBCT trajectory is planar, with little room for scene-specific adjustment. We extend this trajectory by autonomously adjusting out-of-plane angulation. This enables C-arm source trajectories that are scene-specific in that they avoid acquiring ”poor images”, characterized by beam hardening, photon starvation, and noise. The recommendation of ideal out-of-plane angulation is performed on-the-fly using a deep convolutional neural network that regresses a detectability-rank derived from imaging physics.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
