-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathpublication_list.bib
345 lines (298 loc) · 28.5 KB
/
publication_list.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
---
---
@STRING{CVPR = {Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR)}}
@STRING{ECCV = {Proc. of the European Conf. on Computer Vision (ECCV)}}
@STRING{ICCV = {Proc. of the IEEE International Conf. on Computer Vision (ICCV)}}
@STRING{THREEDV = {Proc. of the International Conf. on 3D Vision (3DV)}}
@STRING{NEURIPS = {Advances in Neural Information Processing Systems (NeurIPS)}}
@STRING{ARXIV = {arXiv.org}}
@article{hashemzadeh2021offline,
title={Offline-Online Reinforcement Learning: Extending Batch and Online RL},
author={Hashemzadeh, Maryam and Chung, Wesley and White, Martha},
year={2021}
}
@article{hashemzadeh2020language,
title={From Language to Language-ish: How Brain-Like is an LSTM's Representation of Nonsensical Language Stimuli?},
author={Hashemzadeh, Maryam and Kaufeld, Greta and White, Martha and Martin, Andrea E and Fyshe, Alona},
journal={arXiv preprint arXiv:2010.07435},
year={2020}
}
@article{cortese2021value,
title={Value signals guide abstraction during learning},
author={Cortese, Aurelio and Yamamoto, Asuka and Hashemzadeh, Maryam and Sepulveda, Pradyumna and Kawato, Mitsuo and De Martino, Benedetto},
journal={Elife},
volume={10},
pages={e68943},
year={2021},
publisher={eLife Sciences Publications Limited}
}
@article{hashemzadeh2020clustering,
title={Clustering subspace generalization to obtain faster reinforcement learning},
author={Hashemzadeh, Maryam and Hosseini, Reshad and Ahmadabadi, Majid Nili},
journal={Evolving Systems},
volume={11},
number={1},
pages={89--103},
year={2020},
publisher={Springer}
}
@article{hashemzadeh2018exploiting,
title={Exploiting generalization in the subspaces for faster model-based reinforcement learning},
author={Hashemzadeh, Maryam and Hosseini, Reshad and Ahmadabadi, Majid Nili},
journal={IEEE transactions on neural networks and learning systems},
volume={30},
number={6},
pages={1635--1650},
year={2018},
publisher={IEEE}
}
@inproceedings{Raj2023ARXIV,
author = {Raj, Amit and Kaza, Srinivas and Poole, Ben and Niemeyer, Michael and Ruiz, Nataniel and Mildenhall, Ben and Zada, Shiran and Aberman, Kfir and Rubinstein, Michael and Barron, Jonathan and Li, Yuanzhen and Jampani, Varun},
title = {DreamBooth3D: Subject-Driven Text-to-3D Generation},
year = {2023},
html = {https://dreambooth3d.github.io},
pdf = {https://arxiv.org/pdf/2303.13508.pdf},
img = {assets/img/publications/dreambooth3d.jpg},
booktitle = ARXIV,
}
@inproceedings{Tsalicoglou2023ARXIV,
author = {Tsalicoglou, Christina and Manhardt, Fabian and Tonioni, Alessio and Niemeyer, Michael and Tombari, Federico},
title = {TextMesh: Generation of Realistic 3D Meshes From Text Prompts},
year = {2023},
html = {https://fabi92.github.io/textmesh/},
pdf = {https://arxiv.org/pdf/2304.12439.pdf},
img = {assets/img/publications/textmesh.jpg},
booktitle = ARXIV,
}
@inproceedings{Matsuki2023ARXIV,
author = {Matsuki, Hidenobu and Tateno, Keisuke and Niemeyer, Michael and Tombari, Federico},
title = {NEWTON: Neural View-Centric Mapping for On-the-Fly Large-Scale SLAM},
year = {2023},
html = {https://arxiv.org/abs/2303.13654},
pdf = {https://arxiv.org/pdf/2303.13654.pdf},
img = {assets/img/publications/newton.jpg},
booktitle = ARXIV,
}
@inproceedings{Rakotosaona2023ARXIV,
author = {Rakotosaona, Marie-Julie and Manhardt, Fabian and Martin Arroyo, Diego and Niemeyer, Michael and Kundu, Abhijit and Tombari, Federico},
title = {NeRFMeshing: Distilling Neural Radiance Fields into Geometrically-Accurate 3D Meshes},
year = {2023},
html = {https://arxiv.org/abs/2303.09431},
pdf = {https://arxiv.org/pdf/2303.09431.pdf},
img = {assets/img/publications/nerfmeshing.jpg},
booktitle = ARXIV,
}
@misc{Yu2022SDFStudio,
author = {Yu, Zehao and Chen, Anpei and Antic, Bozidar and Peng, Songyou Peng and Bhattacharyya, Apratim
and Niemeyer, Michael and Tang, Siyu and Sattler, Torsten and Geiger, Andreas},
title = {SDFStudio: A Unified Framework for Surface Reconstruction},
year = {2022},
html = {https://autonomousvision.github.io/sdfstudio/},
img = {assets/img/publications/sdfstudio.jpg},
code = {https://github.com/autonomousvision/sdfstudio},
booktitle = {Open Source Project}
}
@inproceedings{Yu2022NEURIPS,
author = {Yu, Zehao and Peng, Songyou and Niemeyer, Michael and Sattler, Torsten and Geiger, Andreas},
title = {MonoSDF: Exploring Monocular Geometric Cues for Neural Implicit Surface Reconstruction},
booktitle = NEURIPS,
year = {2022},
html = {https://niujinshuchong.github.io/monosdf/},
pdf = {https://arxiv.org/pdf/2206.00665.pdf},
supp = {https://arxiv.org/pdf/2206.00665.pdf},
code = {https://github.com/autonomousvision/monosdf},
img = {assets/img/publications/monosdf.jpg},
}
@inproceedings{Schwarz2022NEURIPS,
author = {Schwarz, Katja and Sauer, Axel and Niemeyer, Michael and Liao, Yiyi and Geiger, Andreas},
title = {VoxGRAF: Fast 3D-Aware Image Synthesis with Sparse Voxel Grids},
booktitle = NEURIPS,
year = {2022},
pdf = {https://arxiv.org/pdf/2206.07695.pdf},
supp = {https://arxiv.org/pdf/2206.07695.pdf},
html = {https://katjaschwarz.github.io/voxgraf/},
code = {https://github.com/autonomousvision/voxgraf},
img = {assets/img/publications/voxgraf.jpg},
}
@inproceedings{Niemeyer2022CVPR,
author = {Michael Niemeyer and Jonathan T. Barron and Ben Mildenhall and Mehdi S. M. Sajjadi and Andreas Geiger and Noha Radwan},
title = {RegNeRF: Regularizing Neural Radiance Fields for View Synthesis from Sparse Inputs},
booktitle = CVPR,
year = {2022},
abstract = {Neural Radiance Fields (NeRF) have emerged as a powerful representation for the task of novel view synthesis due to their simplicity and state-of-the-art performance. Though NeRF can produce photorealistic renderings of unseen viewpoints when many input views are available, its performance drops significantly when this number is reduced. We observe that the majority of artifacts in sparse input scenarios are caused by errors in the estimated scene geometry, and by divergent behavior at the start of training. We address this by regularizing the geometry and appearance of patches rendered from unobserved viewpoints, and annealing the ray sampling space during training. We additionally use a normalizing flow model to regularize the color of unobserved viewpoints. Our model outperforms not only other methods that optimize over a single scene, but in many cases also conditional models that are extensively pre-trained on large multi-view datasets.},
img = {assets/img/publications/regnerf.jpg},
html = {https://m-niemeyer.github.io/regnerf},
pdf = {https://drive.google.com/file/d/1S_NnmhypZjyMfwqcHg-YbWSSYNWdqqlo/view?usp=sharing},
supp = {https://drive.google.com/file/d/15ip8Fvfxp6rNRfBnbJEnFCjIJeFMH4CE/view?usp=sharing},
video = {https://youtu.be/QyyyvA4-Kwc},
code = {https://github.com/google-research/google-research/tree/master/regnerf},
selected = {true},
award = {Oral Presentation},
poster ={https://drive.google.com/file/d/1kYknB2Ap3I3avstmPxAa9IiW8m85AZEF/view?usp=sharing},
}
@INPROCEEDINGS{Peng2021NEURIPS,
author = {Songyou Peng and Chiyu Max Jiang and Yiyi Liao and Michael Niemeyer and Marc Pollefeys and Andreas Geiger},
title = {Shape As Points: A Differentiable Poisson Solver},
booktitle = NEURIPS,
abstract = {In recent years, neural implicit representations gained popularity in 3D reconstruction due to their expressiveness and flexibility. However, the implicit nature of neural implicit representations results in slow inference times and requires careful initialization. In this paper, we revisit the classic yet ubiquitous point cloud representation and introduce a differentiable point-to-mesh layer using a differentiable formulation of Poisson Surface Reconstruction (PSR) which allows for a GPU-accelerated fast solution of the indicator function given an oriented point cloud. The differentiable PSR layer allows us to efficiently and differentiably bridge the explicit 3D point representation with the 3D mesh via the implicit indicator field, enabling end-to-end optimization of surface reconstruction metrics such as Chamfer distance. This duality between points and meshes hence allows us to represent shapes as oriented point clouds, which are explicit, lightweight and expressive. Compared to neural implicit representations, our Shape-As-Points (SAP) model is more interpretable, lightweight, and accelerates inference time by one order of magnitude. Compared to other explicit representations such as points, patches, and meshes, SAP produces topology-agnostic, watertight manifold surfaces. We demonstrate the effectiveness of SAP on the task of surface reconstruction from unoriented point clouds and learning-based reconstruction.},
year = {2021},
img = {assets/img/publications/sap.jpg},
award = {Oral Presentation},
html = {https://pengsongyou.github.io/sap},
code = {https://github.com/autonomousvision/shape_as_points},
video = {https://youtu.be/FL8LMk_qWb4},
poster = {https://pengsongyou.github.io/media/sap/sap_poster.pdf},
pdf = {https://arxiv.org/abs/2106.03452},
}
@InProceedings{Schwarz2020NEURIPS,
author = {Katja Schwarz and Yiyi Liao and Michael Niemeyer and Andreas Geiger},
title = {GRAF: Generative Radiance Fields for 3D-Aware Image Synthesis},
booktitle = NEURIPS,
year = {2020},
abstract = {While 2D generative adversarial networks have enabled high-resolution image synthesis, they largely lack an understanding of the 3D world and the image formation process. Thus, they do not provide precise control over camera viewpoint or object pose. To address this problem, several recent approaches leverage intermediate voxel-based representations in combination with differentiable rendering. However, existing methods either produce low image resolution or fall short in disentangling camera and scene properties, eg, the object identity may vary with the viewpoint. In this paper, we propose a generative model for radiance fields which have recently proven successful for novel view synthesis of a single scene. In contrast to voxel-based representations, radiance fields are not confined to a coarse discretization of the 3D space, yet allow for disentangling camera and scene properties while degrading gracefully in the presence of reconstruction ambiguity. By introducing a multi-scale patch-based discriminator, we demonstrate synthesis of high-resolution images while training our model from unposed 2D images alone. We systematically analyze our approach on several challenging synthetic and real-world datasets. Our experiments reveal that radiance fields are a powerful representation for generative image synthesis, leading to 3D consistent models that render with high fidelity.},
cvlibs_month = {12},
code = {https://github.com/autonomousvision/graf},
cvlibs_youtube = {akQf7WaCOHo},
groups = {cvlibs},
img = {assets/img/publications/graf.jpg},
pdf = {http://www.cvlibs.net/publications/Schwarz2020NEURIPS.pdf},
supp = {http://www.cvlibs.net/publications/Schwarz2020NEURIPS_supplementary.pdf},
video = {http://www.youtube.com/watch?v=akQf7WaCOHo&vq=hd1080&autoplay=1},
blog = {https://autonomousvision.github.io/graf/},
selected = {true},
html = {https://ps.is.mpg.de/publications/schwarz2020neurips},
}
@InProceedings{Niemeyer2021CVPR,
author = {Michael Niemeyer and Andreas Geiger},
title = {GIRAFFE: Representing Scenes as Compositional Generative Neural Feature Fields},
booktitle = CVPR,
year = {2021},
abstract = {Deep generative models allow for photorealistic image synthesis at high resolutions. But for many applications, this is not enough: content creation also needs to be controllable. While several recent works investigate how to disentangle underlying factors of variation in the data, most of them operate in 2D and hence ignore that our world is three-dimensional. Further, only few works consider the compositional nature of scenes. Our key hypothesis is that incorporating a compositional 3D scene representation into the generative model leads to more controllable image synthesis. Representing scenes as compositional generative neural feature fields allows us to disentangle one or multiple objects from the background as well as individual objects' shapes and appearances while learning from unstructured and unposed image collections without any additional supervision. Combining this scene representation with a neural rendering pipeline yields a fast and realistic image synthesis model. As evidenced by our experiments, our model is able to disentangle individual objects and allows for translating and rotating them in the scene as well as changing the camera pose.},
html = {https://m-niemeyer.github.io/project-pages/giraffe/index.html},
img = {assets/img/publications/giraffe.jpg},
pdf = {http://www.cvlibs.net/publications/Niemeyer2021CVPR.pdf},
supp = {http://www.cvlibs.net/publications/Niemeyer2021CVPR_supplementary.pdf},
poster = {http://www.cvlibs.net/publications/Niemeyer2021CVPR_poster.pdf},
video = {http://www.youtube.com/watch?v=fIaDXC-qRSg&vq=hd1080&autoplay=1},
code = {https://github.com/autonomousvision/giraffe},
award = {Oral Presentation, Best Paper Award}
}
@InProceedings{Niemeyer2021THREEDV,
author = {Michael Niemeyer and Andreas Geiger},
title = {CAMPARI: Camera-Aware Decomposed Generative Neural Radiance Fields},
booktitle = THREEDV,
year = {2021},
abstract = {Tremendous progress in deep generative models has led to photorealistic image synthesis. While achieving compelling results, most approaches operate in the two-dimensional image domain, ignoring the three-dimensional nature of our world. Several recent works therefore propose generative models which are 3D-aware, i.e., scenes are modeled in 3D and then rendered differentiably to the image plane. This leads to impressive 3D consistency, but incorporating such a bias comes at a price: the camera needs to be modeled as well. Current approaches assume fixed intrinsics and a predefined prior over camera pose ranges. As a result, parameter tuning is typically required for real-world data, and results degrade if the data distribution is not matched. Our key hypothesis is that learning a camera generator jointly with the image generator leads to a more principled approach to 3D-aware image synthesis. Further, we propose to decompose the scene into a background and foreground model, leading to more efficient and disentangled scene representations. While training from raw, unposed image collections, we learn a 3D- and camera-aware generative model which faithfully recovers not only the image but also the camera data distribution. At test time, our model generates images with explicit control over the camera as well as the shape and appearance of the scene.},
img = {assets/img/publications/campari.jpg},
pdf = {https://arxiv.org/pdf/2103.17269.pdf},
code = {https://github.com/autonomousvision/campari},
html = {https://github.com/autonomousvision/campari},
video = {http://www.youtube.com/watch?v=rrIIEc2qYjM},
poster ={https://www.cvlibs.net/publications/Niemeyer2021THREEDV_poster.pdf},
supp = {https://www.cvlibs.net/publications/Niemeyer2021THREEDV_supplementary.pdf},
}
@InProceedings{Oechsle2020THREEDV,
author = {Michael Oechsle and Michael Niemeyer and Christian Reiser and Lars Mescheder and Thilo Strauss and Andreas Geiger},
title = {Learning Implicit Surface Light Fields},
booktitle = THREEDV,
year = {2020},
abstract = {Implicit representations of 3D objects have recently achieved impressive results on learning-based 3D reconstruction tasks. While existing works use simple texture models to represent object appearance, photo-realistic image synthesis requires reasoning about the complex interplay of light, geometry and surface properties. In this work, we propose a novel implicit representation for capturing the visual appearance of an object in terms of its surface light field. In contrast to existing representations, our implicit model represents surface light fields in a continuous fashion and independent of the geometry. Moreover, we condition the surface light field with respect to the location and color of a small light source. Compared to traditional surface light field models, this allows us to manipulate the light source and relight the object using environment maps. We further demonstrate the capabilities of our model to predict the visual appearance of an unseen object from a single real RGB image and corresponding 3D shape information. As evidenced by our experiments, our model is able to infer rich visual appearance including shadows and specular reflections. Finally, we show that the proposed representation can be embedded into a variational auto-encoder for generating novel appearances that conform to the specified illumination conditions.},
cvlibs_month = {11},
code = {https://github.com/autonomousvision/cslf},
groups = {cvlibs},
img = {assets/img/publications/cslf.jpg},
pdf = {http://www.cvlibs.net/publications/Oechsle2020THREEDV.pdf},
supp = {http://www.cvlibs.net/publications/Oechsle2020THREEDV_supplementary.pdf},
poster = {http://www.cvlibs.net/publications/Oechsle2020THREEDV_poster.pdf},
html = {https://arxiv.org/abs/2003.12406},
}
@InProceedings{Peng2020ECCV,
author = {Songyou Peng and Michael Niemeyer and Lars Mescheder and Marc Pollefeys and Andreas Geiger},
title = {Convolutional Occupancy Networks},
booktitle = ECCV,
year = {2020},
abstract = {Recently, implicit neural representations have gained popularity for learning-based 3D reconstruction. While demonstrating promising results, most implicit approaches are limited to comparably simple geometry of single objects and do not scale to more complicated or large-scale scenes. The key limiting factor of implicit methods is their simple fully-connected network architecture which does not allow for integrating local information in the observations or incorporating inductive biases such as translational equivariance. In this paper, we propose Convolutional Occupancy Networks, a more flexible implicit representation for detailed reconstruction of objects and 3D scenes. By combining convolutional encoders with implicit occupancy decoders, our model incorporates inductive biases, enabling structured reasoning in 3D space. We investigate the effectiveness of the proposed representation by reconstructing complex geometry from noisy point clouds and low-resolution voxel representations. We empirically find that our method enables the fine-grained implicit 3D reconstruction of single objects, scales to large indoor scenes, and generalizes well from synthetic to real data.},
blog = {https://autonomousvision.github.io/convolutional-occupancy-networks/},
cvlibs_month = {08},
code = {https://github.com/autonomousvision/convolutional_occupancy_networks},
cvlibs_tags = {spotlight},
cvlibs_youtube = {EmauovgrDSM},
groups = {cvlibs},
img = {assets/img/publications/conv_onet.jpg},
pdf = {http://www.cvlibs.net/publications/Peng2020ECCV.pdf},
supp = {http://www.cvlibs.net/publications/Peng2020ECCV_supplementary.pdf},
video = {http://www.youtube.com/watch?v=EmauovgrDSM&vq=hd1080&autoplay=1},
award = {Spotlight Presentation},
html = {https://pengsongyou.github.io/conv_onet},
}
@InProceedings{Niemeyer2020CVPR,
author = {Michael Niemeyer and Lars Mescheder and Michael Oechsle and Andreas Geiger},
title = {Differentiable Volumetric Rendering: Learning Implicit 3D Representations without 3D Supervision},
booktitle = CVPR,
year = {2020},
abstract = {Learning-based 3D reconstruction methods have shown impressive results. However, most methods require 3D supervision which is often hard to obtain for real-world datasets. Recently, several works have proposed differentiable rendering techniques to train reconstruction models from RGB images. Unfortunately, these approaches are currently restricted to voxel- and mesh-based representations, suffering from discretization or low resolution. In this work, we propose a differentiable rendering formulation for implicit shape and texture representations. Implicit representations have recently gained popularity as they represent shape and texture continuously. Our key insight is that depth gradients can be derived analytically using the concept of implicit differentiation. This allows us to learn implicit shape and texture representations directly from RGB images. We experimentally show that our single-view reconstructions rival those learned with full 3D supervision. Moreover, we find that our method can be used for multi-view 3D reconstruction, directly resulting in watertight meshes.},
blog = {https://autonomousvision.github.io/differentiable-volumetric-rendering/},
cvlibs_month = {06},
code = {https://github.com/autonomousvision/differentiable_volumetric_rendering},
pdf = {http://www.cvlibs.net/publications/Niemeyer2020CVPR.pdf},
supp = {http://www.cvlibs.net/publications/Niemeyer2020CVPR_supplementary.pdf},
video = {https://www.youtube.com/watch?v=U_jIN3qWVEw},
img = {assets/img/publications/dvr.jpg},
poster = {http://www.cvlibs.net/publications/Niemeyer2020CVPR_poster.pdf},
html = {https://avg.is.mpg.de/publications/niemeyer2020cvpr},
}
@InProceedings{Niemeyer2019ICCV,
author = {Michael Niemeyer and Lars Mescheder and Michael Oechsle and Andreas Geiger},
title = {Occupancy Flow: 4D Reconstruction by Learning Particle Dynamics},
booktitle = ICCV,
year = {2019},
abstract = {Deep learning based 3D reconstruction techniques have recently achieved impressive results. However, while state-of-the-art methods are able to output complex 3D geometry, it is not clear how to extend these results to time-varying topologies. Approaches treating each time step individually lack continuity and exhibit slow inference, while traditional 4D reconstruction methods often utilize a template model or discretize the 4D space at fixed resolution. In this work, we present Occupancy Flow, a novel spatio-temporal representation of time-varying 3D geometry with implicit correspondences. Towards this goal, we learn a temporally and spatially continuous vector field which assigns a motion vector to every point in space and time. In order to perform dense 4D reconstruction from images or sparse point clouds, we combine our method with a continuous 3D representation. Implicitly, our model yields correspondences over time, thus enabling fast inference while providing a sound physical description of the temporal dynamics. We show that our method can be used for interpolation and reconstruction tasks, and demonstrate the accuracy of the learned correspondences. We believe that Occupancy Flow is a promising new 4D representation which will be useful for a variety of spatio-temporal reconstruction tasks.},
blog = {https://autonomousvision.github.io/occupancy-flow/},
cvlibs_month = {10},
code = {https://github.com/autonomousvision/occupancy_flow},
cvlibs_youtube = {c0yOugTgrWc},
groups = {cvlibs},
img = {assets/img/publications/oflow.jpg},
pdf = {http://www.cvlibs.net/publications/Niemeyer2019ICCV.pdf},
supp = {http://www.cvlibs.net/publications/Niemeyer2019ICCV_supplementary.pdf},
poster = {http://www.cvlibs.net/publications/Niemeyer2019ICCV_poster.pdf},
video = {http://www.youtube.com/watch?v=c0yOugTgrWc&vq=hd1080&autoplay=1},
html = {https://avg.is.mpg.de/publications/niemeyer2019iccv},
}
@InProceedings{Oechsle2019ICCV,
author = {Michael Oechsle and Lars Mescheder and Michael Niemeyer and Thilo Strauss and Andreas Geiger},
title = {Texture Fields: Learning Texture Representations in Function Space},
booktitle = ICCV,
year = {2019},
abstract = {In recent years, substantial progress has been achieved in learning-based reconstruction of 3D objects. At the same time, generative models were proposed that can generate highly realistic images. However, despite this success in these closely related tasks, texture reconstruction of 3D objects has received little attention from the research community and state-of-the-art methods are either limited to comparably low resolution or constrained experimental setups. A major reason for these limitations is that common representations of texture are inefficient or hard to interface for modern deep learning techniques. In this paper, we propose Texture Fields, a novel texture representation which is based on regressing a continuous 3D function parameterized with a neural network. Our approach circumvents limiting factors like shape discretization and parameterization, as the proposed texture representation is independent of the shape representation of the 3D object. We show that Texture Fields are able to represent high frequency texture and naturally blend with modern deep learning techniques. Experimentally, we find that Texture Fields compare favorably to state-of-the-art methods for conditional texture reconstruction of 3D objects and enable learning of probabilistic generative models for texturing unseen 3D models. We believe that Texture Fields will become an important building block for the next generation of generative 3D models.
},
blog = {https://autonomousvision.github.io/texture-fields/},
cvlibs_month = {10},
code = {https://github.com/autonomousvision/texture_fields},
cvlibs_tags = {oral},
cvlibs_youtube = {y8XHkl3vtpI},
groups = {cvlibs},
img = {assets/img/publications/tfield.jpg},
pdf = {http://www.cvlibs.net/publications/Oechsle2019ICCV.pdf},
supp = {http://www.cvlibs.net/publications/Oechsle2019ICCV_supplementary.pdf},
video = {http://www.youtube.com/watch?v=y8XHkl3vtpI&vq=hd1080&autoplay=1},
poster = {http://www.cvlibs.net/publications/Oechsle2019ICCV_poster.pdf},
html = {https://avg.is.mpg.de/publications/oechsle2019iccv},
award = {Oral Presentation},
}
@InProceedings{Mescheder2019CVPR,
author = {Lars Mescheder and Michael Oechsle and Michael Niemeyer and Sebastian Nowozin and Andreas Geiger},
title = {Occupancy Networks: Learning 3D Reconstruction in Function Space},
booktitle = CVPR,
year = {2019},
abstract = {With the advent of deep neural networks, learning-based approaches for 3D~reconstruction have gained popularity. However, unlike for images, in 3D there is no canonical representation which is both computationally and memory efficient yet allows for representing high-resolution geometry of arbitrary topology. Many of the state-of-the-art learning-based 3D~reconstruction approaches can hence only represent very coarse 3D geometry or are limited to a restricted domain. In this paper, we propose Occupancy Networks, a new representation for learning-based 3D~reconstruction methods. Occupancy networks implicitly represent the 3D surface as the continuous decision boundary of a deep neural network classifier. In contrast to existing approaches, our representation encodes a description of the 3D output at infinite resolution without excessive memory footprint. We validate that our representation can efficiently encode 3D structure and can be inferred from various kinds of input. Our experiments demonstrate competitive results, both qualitatively and quantitatively, for the challenging tasks of 3D reconstruction from single images, noisy point clouds and coarse discrete voxel grids. We believe that occupancy networks will become a useful tool in a wide variety of learning-based 3D tasks.},
blog = {https://autonomousvision.github.io/occupancy-networks/},
code = {https://github.com/LMescheder/Occupancy-Networks},
img = {assets/img/publications/onet.jpg},
pdf = {http://www.cvlibs.net/publications/Mescheder2019CVPR.pdf},
supp = {http://www.cvlibs.net/publications/Mescheder2019CVPR_supplementary.pdf},
video = {http://www.youtube.com/watch?v=w1Qo3bOiPaE&t=6s&vq=hd1080&autoplay=1},
poster = {http://www.cvlibs.net/publications/Mescheder2019CVPR_poster.pdf},
html = {https://avg.is.mpg.de/publications/occupancy-networks},
award = {Oral Presentation, Best Paper Finalist},
}