My academic research focuses on computer vision and deep learning, more specifically self-supervised learning, perception for robotics, and computational photography. I believe that there exists a great deal of natural structure inherent to visual data that we can exploit and learn from without requiring all too much manual annotation, or even none whatsoever. Check out some of my projects below for an overview of what I have done so far and what keeps me busy today. (Note that this page may sometimes be out of date ;-))
Learning Object Permanence (2020-)
Dissecting Image Crops (2019-2020)
(With Prof. Carl Vondrick; in ICCV 2021)
The elementary operation of cropping underpins nearly every computer vision system, ranging from data augmentation and translation invariance to computational photography and representation learning. This paper investigates the subtle traces introduced by this operation. For example, despite refinements to camera optics, lenses will leave behind certain clues, notably chromatic aberration and vignetting. Photographers also leave behind other clues relating to image aesthetics and scene composition. We study how to detect these traces, and investigate the impact that cropping has on the image distribution. While our aim is to dissect the fundamental impact of spatial crops, there are also a number of practical implications to our work, such as detecting image manipulations and equipping neural network researchers with a better understanding of shortcut learning.
Image Outpainting using GANs (2019)
(In a course by Prof. Peter Belhumeur)
Inpainting, which attempts to recover holes within photos, is an active line of research and typically uses Generative Adversarial Networks. However, the opposite problem of predicting what pixels reside outside the borders of an image (hence the term outpainting) is something I would also deem worthy of investigation. Many architectural aspects and ideas can moreover be naturally adopted from inpainting. Check out the submission on arXiv below to see which hallucinations our model has come up with!
FPGA-based Simultaneous Localization and Mapping using High-Level Synthesis (2018-2019)
(With Prof. Erik D’Hollander and Prof. Bart Goossens)
This master’s thesis at Ghent University was an exploration of a robotic navigation algorithm called Simultaneous Localization and Mapping (SLAM), which helps autonomous agents track their location within an environment while at the same time creating a dense model of that very environment. In an effort to implement a 3D variant of this application on embedded systems (more specifically, Field-Programmable Gate Arrays or FPGAs), I devised guidelines to facilitate the translation of platform-agnostic C/C++ code to FPGA-optimized designs using the High-Level Synthesis (HLS) design technique. Furthermore, I implemented and compared several mappings of typical software blocks to efficient dataflow architectures on programmable hardware.
Camera Spoofing for Digital Forensics (2018)
(With Hannes Mareen and Tamir Mazaev)
Every camera model carries a fingerprint with it called sensor pattern noise, which can be used to trace a given photo back to the original device that captured it. In this course research project, we used Generative Adversarial Networks (more specifically, CycleGAN) to transform noise patterns embedded within image patches from one device to another, in an attempt to fool classic camera identification tools typically used in media forensics. The hypothesis of this succeeding remains plausible, as our experiments indicated that the fingerprint had been effectively removed (as in the above graph), but not fully transferred across domains.