<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en"><generator uri="https://jekyllrb.com/" version="4.3.3">Jekyll</generator><link href="https://valeoai.github.io//feed.xml" rel="self" type="application/atom+xml"/><link href="https://valeoai.github.io//" rel="alternate" type="text/html" hreflang="en"/><updated>2026-04-27T18:38:42+00:00</updated><id>https://valeoai.github.io//feed.xml</id><title type="html">valeo.ai</title><subtitle>valeo.ai is an international research team based in Paris working at the intersection of computer vision, AI, and autonomous driving. We publish at top venues, release open-source code, and shape the AI future of Valeo. </subtitle><entry><title type="html">valeo.ai at CVPR 2026</title><link href="https://valeoai.github.io//posts/cvpr-2026" rel="alternate" type="text/html" title="valeo.ai at CVPR 2026"/><published>2026-04-16T00:00:00+00:00</published><updated>2026-04-16T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-cvpr-2026</id><content type="html" xml:base="https://valeoai.github.io//posts/cvpr-2026"><![CDATA[<p>The <a href="https://cvpr.thecvf.com/">IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR)</a> is a key event for researchers and engineers working on computer vision and machine learning. At the 2026 edition, the <a href="../">valeo.ai</a> team will present nine <a href="https://valeoai.github.io/publications/">papers</a> in the main conference and one in the Findings track.</p> <p>The team will be at CVPR to present these works, exchange ideas, and share our exciting ongoing research. We look forward to seeing you in Denver!</p> <hr/> <h3 id="naf-zero-shot-feature-upsampling-via-neighborhood-attention-filtering">NAF: Zero-Shot Feature Upsampling via Neighborhood Attention Filtering</h3> <h3 id="highlight">Highlight</h3> <h5 id="authors-loick-chambon-paul-couairon-eloi-zablocki-alexandre-boulch-nicolas-thome-matthieu-cord">Authors: <a href="https://loickch.github.io/">Loick Chambon</a>, <a href="https://scholar.google.fr/citations?user=yQRnP7YAAAAJ&amp;hl=fr">Paul Couairon</a>, <a href="https://eloiz.github.io">Eloi Zablocki</a>, <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="https://thome.isir.upmc.fr">Nicolas Thome</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2511.18452">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/NAF">Code</a>] </h5> <p><img src="/assets/img/publications/naf.gif" alt="naf_teaser" height="100%" width="100%"/></p> <p>Vision Foundation Models produce spatially downsampled representations that create challenges for pixel-level tasks. We introduce Neighborhood Attention Filtering (NAF), a method that learns adaptive spatial-and-content weights through Cross-Scale Neighborhood Attention and Rotary Position Embeddings, guided solely by high-resolution input images. NAF operates as a zero-shot upsampler compatible with any VFM without retraining, achieving state-of-the-art performance across multiple downstream tasks while maintaining efficiency at 18 FPS for 2K feature maps. The method also demonstrates strong results in image restoration applications.</p> <hr/> <h3 id="driving-on-registers">Driving on Registers</h3> <h5 id="authors-ellington-kirby-alexandre-boulch-yihong-xu-yuan-yin-gilles-puy-eloi-zablocki-andrei-bursuc-spyros-gidaris-renaud-marlet-florent-bartoccioni-anh-quan-cao-nermin-samet-tuan-hung-vu-matthieu-cord">Authors: <a href="https://ellingtonkirby.github.io/">Ellington Kirby</a>, <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="https://scholar.google.fr/citations?user=vMLRRVkAAAAJ">Yihong Xu</a>, <a href="https://yuan-yin.github.io/">Yuan Yin</a>, <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="https://eloiz.github.io">Eloi Zablocki</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://gidariss.github.io/">Spyros Gidaris</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a>, <a href="https://f-barto.github.io/">Florent Bartoccioni</a>, <a href="https://anhquancao.github.io/">Anh-Quan Cao</a>, <a href="https://nerminsamet.github.io/">Nermin Samet</a>, <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2601.05083">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/DrivoR">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/driving-on-registers/">Project page</a>] </h5> <p><img src="/assets/img/publications/drivor.gif" alt="drivor_teaser" height="100%" width="100%"/></p> <p>We present DrivoR, a simple and efficient transformer-based architecture for end-to-end autonomous driving. Our approach builds on pretrained Vision Transformers (ViTs) and introduces camera-aware register tokens that compress multi-camera features into a compact scene representation, significantly reducing downstream computation without sacrificing accuracy. These tokens drive two lightweight transformer decoders that generate and then score candidate trajectories. The scoring decoder learns to mimic an oracle and predicts interpretable sub-scores representing aspects such as safety, comfort, and efficiency, enabling behavior-conditioned driving at inference. Despite its minimal design, DrivoR outperforms or matches strong contemporary baselines across NAVSIM-v1, NAVSIM-v2, and the photorealistic closed-loop HUGSIM benchmark. Our results show that a pure-transformer architecture, combined with targeted token compression, is sufficient for accurate, efficient, and adaptive end-to-end driving.</p> <hr/> <h3 id="occany-generalized-unconstrained-urban-3d-occupancy">OccAny: Generalized Unconstrained Urban 3D Occupancy</h3> <h5 id="authors-anh-quan-cao-tuan-hung-vu">Authors: <a href="https://anhquancao.github.io/">Anh-Quan Cao</a>, <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2603.23502">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/OccAny">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/OccAny/">Project page</a>] </h5> <p><img src="/assets/img/publications/teaser_occany.gif" alt="occany_teaser" height="100%" width="100%"/></p> <p>Relying on in-domain annotations and precise sensor-rig priors, existing 3D occupancy prediction methods are limited in both scalability and out-of-domain generalization. While recent visual geometry foundation models exhibit strong generalization capabilities, they were mainly designed for general purposes and lack one or more key ingredients required for urban occupancy prediction, namely metric prediction, geometry completion in cluttered scenes and adaptation to urban scenarios. We address this gap and present OccAny, the first unconstrained urban 3D occupancy model capable of operating on out-of-domain uncalibrated scenes to predict and complete metric occupancy coupled with segmentation features. OccAny is versatile and can predict occupancy from sequential, monocular, or surround-view images. Extensive experiments demonstrate that OccAny outperforms all visual geometry baselines on the 3D occupancy prediction task, while remaining competitive with in-domain self-supervised methods across three input settings on two established urban occupancy prediction datasets.</p> <hr/> <h3 id="franca-nested-matryoshka-clustering-for-scalable-visual-representation-learning">Franca: Nested Matryoshka Clustering for Scalable Visual Representation Learning</h3> <h5 id="authors-shashanka-venkataramanan-valentinos-pariza-mohammadreza-salehi-lukas-knobel-spyros-gidaris-elias-ramzi-andrei-bursuc-yuki-m-asano">Authors: <a href="https://shashankvkt.github.io/">Shashanka Venkataramanan</a>, Valentinos Pariza, <a href="https://scholar.google.com/citations?user=kpT3gcsAAAAJ&amp;hl=en">Mohammadreza Salehi</a>, Lukas Knobel, <a href="https://gidariss.github.io/">Spyros Gidaris</a>, <a href="https://elias-ramzi.github.io/">Elias Ramzi</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://yukimasano.github.io/">Yuki M. Asano</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2507.14137">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/Franca">Code</a>] </h5> <p><img src="/assets/img/publications/2026_franca/franca.png" alt="franca_teaser" height="100%" width="100%"/></p> <p>We present Franca, the first fully open-source (data, code, weights) vision foundation model that matches and in many cases surpasses the performance of state-of-the-art proprietary models, e.g., DINOv2, CLIP, SigLIPv2. Our approach is grounded in a transparent training pipeline using publicly available data: ImageNet-21K and a subset of ReLAION-2B. We introduce a parameter-efficient, multi-head clustering projector based on nested Matryoshka representations that progressively refines features into increasingly fine-grained clusters without increasing model size. Additionally, we propose a novel positional disentanglement strategy that explicitly removes positional biases from dense representations, leading to consistent gains on several downstream benchmarks. Our contributions establish a new standard for transparent, high-performance vision models and open a path toward more reproducible and generalizable foundation models.</p> <hr/> <h3 id="mad-motion-appearance-decoupling-for-efficient-driving-world-models">MAD: Motion Appearance Decoupling for efficient Driving World Models</h3> <h5 id="authors-ahmad-rahimi-valentin-gerard-eloi-zablocki-matthieu-cord-alexandre-alahi">Authors: <a href="https://scholar.google.com/citations?user=Lc1LR18AAAAJ&amp;hl=en">Ahmad Rahimi</a>, <a href="https://scholar.google.com/citations?user=zDeI0iEAAAAJ&amp;hl=en">Valentin Gerard</a>, <a href="https://eloiz.github.io">Eloi Zablocki</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>, Alexandre Alahi</h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2601.09452">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/vita-epfl/MAD-World-Model-Code/">Code</a>] &nbsp;&nbsp; [<a href="https://vita-epfl.github.io/MAD-World-Model/">Project page</a>] </h5> <p><img src="/assets/img/publications/mad.gif" alt="mad_teaser" height="100%" width="100%"/></p> <p>Recent video diffusion models generate photorealistic, temporally coherent videos, yet they fall short as reliable world models for autonomous driving, where structured motion and physically consistent interactions are essential. We propose an efficient adaptation framework that converts generalist video diffusion models into controllable driving world models with minimal supervision. The key idea is to decouple motion learning from appearance synthesis. First, the model is adapted to predict structured motion in a simplified form: videos of skeletonized agents and scene elements, focusing learning on physical and social plausibility. Then, the same backbone is reused to synthesize realistic RGB videos conditioned on these motion sequences. This two-stage process mirrors a reasoning-rendering paradigm: first infer dynamics, then render appearance. Our experiments show this decoupled approach is exceptionally efficient: adapting SVD, we match prior SOTA models with less than 6% of their compute. Scaling to LTX, our MAD-LTX model outperforms all open-source competitors, and supports a comprehensive suite of text, ego, and object controls.</p> <hr/> <h3 id="stablemtl-repurposing-latent-diffusion-models-for-multi-task-learning-from-partially-annotated-synthetic-datasets">StableMTL: Repurposing Latent Diffusion Models for Multi-Task Learning from Partially Annotated Synthetic Datasets</h3> <h5 id="authors-anh-quan-cao-ivan-lopes-raoul-de-charette">Authors: <a href="https://anhquancao.github.io/">Anh-Quan Cao</a>, Ivan Lopes, <a href="https://team.inria.fr/rits/membres/raoul-de-charette/">Raoul de Charette</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2506.08013">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/astra-vision/StableMTL">Code</a>] </h5> <p><img src="/assets/img/publications/stablemtl.png" alt="stablemtl_teaser" height="100%" width="100%"/></p> <p>Multi-task learning for dense prediction is limited by the need for extensive annotation for every task, though recent works have explored training with partial task labels. Leveraging the generalization power of diffusion models, we extend the partial learning setup to a zero-shot setting, training a multi-task model on multiple synthetic datasets, each labeled for only a subset of tasks. Our method, StableMTL, repurposes image generators for latent regression, adapting a denoising framework with task encoding, per-task conditioning and a tailored training scheme. Instead of per-task losses requiring careful balancing, a unified latent loss is adopted, enabling seamless scaling to more tasks. To encourage inter-task synergy, we introduce a multi-stream model with a task-attention mechanism that converts N-to-N task interactions into efficient 1-to-N attention, promoting effective cross-task sharing. StableMTL outperforms baselines on 7 tasks across 8 benchmarks.</p> <hr/> <h3 id="lidas-lighting-driven-dynamic-active-sensing-for-nighttime-perception">LiDAS: Lighting-driven Dynamic Active Sensing for Nighttime Perception</h3> <h5 id="authors-simon-de-moreau-andrei-bursuc-hafid-el-idrissi-fabien-moutarde">Authors: <a href="https://simondemoreau.github.io/">Simon de Moreau</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, Hafid El-Idrissi, Fabien Moutarde</h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2512.08912">Paper</a>] &nbsp;&nbsp; [<a href="https://simondemoreau.github.io/LiDAS/">Project page</a>] </h5> <p><img src="/assets/img/publications/2026_lidas/lidas.png" alt="lidas_teaser" height="100%" width="100%"/></p> <p>Camera-based perception in autonomous driving suffers a steep performance drop at night, when low light degrades image quality and existing solutions either rely on costly hardware upgrades or post-hoc image enhancement. We introduce LiDAS, a closed-loop active illumination system that dynamically predicts the optimal lighting pattern for visual perception, concentrating light on objects of interest while reducing it in empty regions. LiDAS integrates seamlessly with standard perception models and high-definition headlights, enabling zero-shot nighttime generalization of daytime-trained networks. On a realistic nighttime simulator and on real driving sequences, LiDAS yields +18.7% mAP50 and +5.0% mIoU over standard low-beam at equal power, while also enabling up to 40% energy savings at matched performance. Our approach turns commodity headlights into active perception devices, paving the way for robust nighttime autonomous perception.</p> <hr/> <h3 id="lam3c-3d-sans-3d-scans-scalable-pre-training-from-video-generated-point-clouds">LAM3C: 3D sans 3D Scans: Scalable Pre-training from Video-Generated Point Clouds</h3> <h5 id="authors-ryosuke-yamada-kohsuke-ide-yoshihiro-fukuhara-hirokatsu-kataoka-gilles-puy-andrei-bursuc-yuki-m-asano">Authors: <a href="https://ryosuke-yamada.github.io/">Ryosuke Yamada</a>, Kohsuke Ide, Yoshihiro Fukuhara, Hirokatsu Kataoka, <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://yukimasano.github.io/">Yuki M. Asano</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2512.23042">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/ryosuke-yamada/lam3c">Code</a>] &nbsp;&nbsp; [<a href="https://ryosuke-yamada.github.io/lam3c/">Project page</a>] </h5> <p><img src="/assets/img/publications/2026_lam3c/lam3c.png" alt="lam3c_teaser" height="100%" width="100%"/></p> <p>We investigate the use of unlabeled videos for learning 3D representations without ever resorting to a 3D sensor. To this end, we introduce LAM3C, a self-supervised learning framework that operates directly on point clouds reconstructed from video. To support this, we curate RoomTours, a new dataset of 49,219 scenes generated from web-collected room-walkthrough videos using a feed-forward reconstruction model. Pre-training on this video-derived data introduces challenges: noisy geometry and incomplete coverage destabilize standard SSL objectives. We address these by proposing a noise-regularized loss that stabilizes representation learning by enforcing local geometric smoothness and ensuring feature stability under noisy point clouds. Despite never seeing real 3D scans, LAM3C surpasses prior self-supervised methods on indoor semantic and instance segmentation, demonstrating that unlabeled videos are a powerful and scalable resource for 3D self-supervised learning.</p> <hr/> <h3 id="icm-attention-may-i-have-your-decision-localizing-generative-choices-in-diffusion-models">ICM: Attention, May I Have Your Decision? Localizing Generative Choices in Diffusion Models</h3> <h5 id="authors-katarzyna-zaleska-lukasz-popek-monika-wysoczanska-kamil-deja">Authors: Katarzyna Zaleska, Lukasz Popek, <a href="https://wysoczanska.github.io/">Monika Wysoczanska</a>, Kamil Deja</h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2604.06052">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/kzaleskaa/icm">Code</a>] </h5> <p><img src="/assets/img/publications/2026_icm/icm.png" alt="icm_teaser" height="100%" width="100%"/></p> <p>Text-to-image diffusion models routinely face ambiguous prompts that under-specify visual details, forcing them to make implicit decisions about unspecified attributes. We posit that this decision-making is computationally localized within the model’s architecture rather than being uniformly distributed. We introduce a probing technique that identifies the layers exhibiting the highest attribute separability, and find that self-attention layers are the locus where these implicit choices are resolved. Building on this insight, we propose ICM (Implicit Choice-Modification), a targeted steering method that modifies only the identified self-attention layers to control the model’s implicit decisions. Experiments demonstrate that ICM achieves superior debiasing performance with fewer artifacts than existing approaches, providing both a new tool for controllable generation and a better understanding of where generative choices live in diffusion models.</p> <hr/> <h3 id="pom-a-linear-time-replacement-for-attention-with-the-polynomial-mixer">PoM: A Linear-Time Replacement for Attention with the Polynomial Mixer</h3> <p class="page-description">CVPR 2026 Findings</p> <h5 id="authors-david-picard-nicolas-dufour-lucas-degeorge-arijit-ghosh-davide-allegro-tom-ravaud-yohann-perron-corentin-sautier-zeynep-sonat-baltaci-fei-meng-syrine-kalleli-marta-lopez-rauhut-thibaut-loiseau-segolene-albouy-raphael-baena-elliot-vincent-loic-landrieu">Authors: <a href="https://davidpicard.github.io/">David Picard</a>, <a href="https://nicolas-dufour.github.io/">Nicolas Dufour</a>, Lucas Degeorge, Arijit Ghosh, Davide Allegro, Tom Ravaud, Yohann Perron, <a href="https://csautier.github.io/">Corentin Sautier</a>, Zeynep Sonat Baltaci, Fei Meng, Syrine Kalleli, Marta Lopez-Rauhut, Thibaut Loiseau, Segolene Albouy, Raphael Baena, Elliot Vincent, <a href="https://loiclandrieu.com/">Loic Landrieu</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2604.06129">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/davidpicard/pom">Code</a>] </h5> <p><img src="/assets/img/publications/2026_pom/teaser.png" alt="pom_teaser" height="100%" width="100%"/></p> <p>This paper introduces the Polynomial Mixer (PoM), a novel token mixing mechanism with linear complexity that serves as a drop-in replacement for self-attention. PoM aggregates input tokens into a compact representation through a learned polynomial function, from which each token retrieves contextual information. We prove that PoM satisfies the contextual mapping property, ensuring that transformers equipped with PoM remain universal sequence-to-sequence approximators. We replace standard self-attention with PoM across five diverse domains: text generation, handwritten text recognition, image generation, 3D modeling, and Earth observation. PoM matches the performance of attention-based models while drastically reducing computational cost when working with long sequences.</p>]]></content><author><name></name></author><category term="3d-perception"/><category term="driving"/><category term="foundation"/><category term="deep-learning"/><category term="generalization"/><category term="generative-model"/><summary type="html"><![CDATA[The IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR) is a key event for researchers and engineers working on computer vision and machine learning. At the 2026 edition, the valeo.ai team will present nine papers in the main conference and one in the Findings track.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/cvpr2026.png"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/cvpr2026.png" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at NeurIPS 2025</title><link href="https://valeoai.github.io//posts/neurips-2025" rel="alternate" type="text/html" title="valeo.ai at NeurIPS 2025"/><published>2025-12-02T00:00:00+00:00</published><updated>2025-12-02T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-neurips-2025</id><content type="html" xml:base="https://valeoai.github.io//posts/neurips-2025"><![CDATA[<p>The <a href="https://neurips.cc/">Neural Information Processing Systems Conference (NeurIPS)</a> is a major inter-disciplinary event that brings together researchers and practicioners in machine learning, computer vision, natural language processing, optimization, statistics, but also neuroscience, natural sciences, social sciences, etc. This year, at the 39th edition of NeurIPS, the <a href="../">valeo.ai</a> team will present 5 papers in the main conference and 1 in the workshops. We are honored to announce that our <a href="https://valeoai.github.io/publications/ipa/">IPA</a> paper on efficient foundation model adaptation has received the outstanding paper award at the <a href="https://sites.google.com/view/ccfm-neurips2025">CCFM workshop</a>. Our team contributed to the technical program committee with multiple reviewers out of whom 1 was awarded top reviewer and 2 as top area chairs.</p> <p>The team will be at NeurIPS to present these works, exchange ideas, and share our exciting ongoing research. We look forward to seeing you in San Diego!</p> <p><img src="/assets/img/posts/2025_neurips_valeoai_papers.png" alt="valeo.ai papers at NeurIPS 2025" height="100%" width="100%"/></p> <hr/> <h3 id="jafar-jack-up-any-feature-at-any-resolution">JAFAR: Jack up Any Feature at Any Resolution</h3> <h5 id="authors-paul-couairon-loick-chambon--louis-serrano-jean-emmanuel-haugeard-matthieu-cord-nicolas-thome">Authors: <a href="https://scholar.google.fr/citations?user=yQRnP7YAAAAJ&amp;hl=fr">Paul Couairon</a>, <a href="https://loickch.github.io/">Loick Chambon</a> , <a href="https://scholar.google.com/citations?user=fKlo-lUAAAAJ&amp;hl=fr">Louis Serrano</a>, <a>Jean-Emmanuel Haugeard</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>, <a href="https://thome.isir.upmc.fr">Nicolas Thome</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2506.11136">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/PaulCouairon/JAFAR">Code</a>] </h5> <p><img src="/assets/img/publications/2025_jafar/teaser.png" alt="jafar_teaser" height="100%" width="100%"/></p> <p>Foundation Vision Encoders have become essential for a wide range of dense vision tasks. However, their low-resolution spatial feature outputs necessitate feature upsampling to produce the high-resolution modalities required for downstream tasks. In this work, we introduce JAFAR—a lightweight and flexible feature upsampler that enhances the spatial resolution of visual features from any Foundation Vision Encoder to an arbitrary target resolution. JAFAR employs an attention-based module designed to promote semantic alignment between high-resolution queries—derived from low-level image features—and semantically enriched low-resolution keys, using Spatial Feature Transform (SFT) modulation. Notably, despite the absence of high-resolution supervision, we demonstrate that learning at low upsampling ratios and resolutions generalizes remarkably well to significantly higher output scales. Extensive experiments show that JAFAR effectively recovers fine-grained spatial details and consistently outperforms existing feature upsampling methods across a diverse set of downstream tasks.</p> <hr/> <h3 id="dino-foresight-looking-into-the-future-with-dino">DINO-Foresight: Looking into the Future with DINO</h3> <h5 id="authors-efstathios-karypidis-ioannis-kakogeorgiou-spyros-gidaris-nikos-komodakis">Authors: <a href="https://archimedesai.gr/en/researchers/stathis-karypidis">Efstathios Karypidis</a>, <a href="https://scholar.google.com/citations?user=B_dKcz4AAAAJ">Ioannis Kakogeorgiou</a>, <a href="https://gidariss.github.io/">Spyros Gidaris</a>, <a href="https://www.csd.uoc.gr/~komod/">Nikos Komodakis</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2412.11673">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/Sta8is/DINO-Foresight">Code</a>] </h5> <p><img src="/assets/img/publications/2025_dinoforesight/teaser.png" alt="foresight_teaser" height="100%" width="100%"/></p> <p>Predicting future dynamics is crucial for applications like autonomous driving and robotics, where understanding the environment is key. Existing pixel-level methods are computationally expensive and often focus on irrelevant details. To address these challenges, we introduce DINO-Foresight, a novel framework that operates in the semantic feature space of pretrained Vision Foundation Models (VFMs). Our approach trains a masked feature transformer in a self-supervised manner to predict the evolution of VFM features over time. By forecasting these features, we can apply off-the-shelf, task-specific heads for various scene understanding tasks. In this framework, VFM features are treated as a latent space, to which different heads attach to perform specific tasks for future-frame analysis. Extensive experiments show the very strong performance, robustness and scalability of our framework.</p> <hr/> <h3 id="learning-to-steer-input-dependent-steering-for-multimodal-llms">Learning to Steer: Input-dependent Steering for Multimodal LLMs</h3> <h5 id="authors-jayneel-parekh-pegah-khayatan-mustafa-shukor-arnaud-dapogny-alasdair-newson-matthieu-cord">Authors: <a href="https://jayneelparekh.github.io/">Jayneel Parekh</a>, <a href="https://pegah-kh.github.io/">Pegah Khayatan</a>, <a href="https://mustafashukor.github.io/">Mustafa Shukor</a>, <a href="https://www.linkedin.com/in/arnaud-dapogny-12653493/">Arnaud Dapogny</a>, <a href="https://sites.google.com/site/alasdairnewson/">Alasdair Newson</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2508.12815">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/jayneelparekh/learn-to-steer">Code</a>] </h5> <p><img src="/assets/img/publications/2025_l2s.png" alt="l2s_teaser" height="100%" width="100%"/></p> <p>Steering has emerged as a practical approach to enable post-hoc guidance of LLMs towards enforcing a specific behavior. However, it remains largely underexplored for multimodal LLMs (MLLMs); furthermore, existing steering techniques, such as mean steering, rely on a single steering vector, applied independently of the input query. This paradigm faces limitations when the desired behavior is dependent on the example at hand. For example, a safe answer may consist in abstaining from answering when asked for an illegal activity, or may point to external resources or consultation with an expert when asked about medical advice. In this paper, we investigate a fine-grained steering that uses an input-specific linear shift. This shift is computed using contrastive input-specific prompting. However, the input-specific prompts required for this approach are not known at test time. Therefore, we propose to train a small auxiliary module to predict the input-specific steering vector. Our approach, dubbed as L2S (Learn-to-Steer), demonstrates that it reduces hallucinations and enforces safety in MLLMs, outperforming other static baselines.</p> <hr/> <h3 id="boosting-generative-image-modeling-via-joint-image-feature-synthesis">Boosting Generative Image Modeling via Joint Image-Feature Synthesis</h3> <h5 id="authors-theodoros-kouzelis-efstathios-karypidis-ioannis-kakogeorgiou-spyros-gidaris-nikos-komodakis">Authors: <a href="https://scholar.google.com/citations?user=a5vkWc8AAAAJ&amp;hl=en">Theodoros Kouzelis</a>, <a href="https://archimedesai.gr/en/researchers/stathis-karypidis">Efstathios Karypidis</a>, <a href="https://scholar.google.com/citations?user=B_dKcz4AAAAJ">Ioannis Kakogeorgiou</a>, <a href="https://gidariss.github.io/">Spyros Gidaris</a>, <a href="https://www.csd.uoc.gr/~komod/">Nikos Komodakis</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2504.16064">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/zelaki/ReDi">Code</a>] </h5> <p><img src="/assets/img/publications/2025_redi/teaser.png" alt="redi_teaser" height="100%" width="100%"/></p> <p>Latent diffusion models (LDMs) dominate high-quality image generation, yet integrating representation learning with generative modeling remains a challenge. We introduce a novel generative image modeling framework that seamlessly bridges this gap by leveraging a diffusion model to jointly model low-level image latents (from a variational autoencoder) and high-level semantic features (from a pretrained self-supervised encoder like DINO). Our latent-semantic diffusion approach learns to generate coherent image-feature pairs from pure noise, significantly enhancing both generative quality and training efficiency, all while requiring only minimal modifications to standard Diffusion Transformer architectures. By eliminating the need for complex distillation objectives, our unified design simplifies training and unlocks a powerful new inference strategy: Representation Guidance, which leverages learned semantics to steer and refine image generation. Evaluated in both conditional and unconditional settings, our method delivers substantial improvements in image quality and training convergence speed, establishing a new direction for representation-aware generative modeling.</p> <hr/> <h3 id="multi-token-prediction-needs-registers">Multi-Token Prediction Needs Registers</h3> <h5 id="authors-anastasios-gerontopoulos-spyros-gidaris-nikos-komodakis">Authors: <a href="https://scholar.google.com/citations?user=VPTaLcUAAAAJ&amp;hl=en">Anastasios Gerontopoulos</a>, <a href="https://gidariss.github.io/">Spyros Gidaris</a>, <a href="https://www.csd.uoc.gr/~komod/">Nikos Komodakis</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2505.10518">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/nasosger/MuToR">Code</a>] </h5> <p><img src="/assets/img/publications/2025_mutor/teaser.png" alt="mutor_teaser" height="100%" width="100%"/></p> <p>Multi-token prediction has emerged as a promising objective for improving language model pretraining, but its benefits have not consistently generalized to other settings such as fine-tuning. In this paper, we propose MuToR, a simple and effective approach to multi-token prediction that interleaves learnable register tokens into the input sequence, each tasked with predicting future targets. Compared to existing methods, MuToR offers several key advantages: it introduces only a negligible number of additional parameters, requires no architectural changes–ensuring compatibility with off-the-shelf pretrained language models–and remains aligned with the next-token pretraining objective, making it especially well-suited for supervised fine-tuning. Moreover, it naturally supports scalable prediction horizons. We demonstrate the effectiveness and versatility of MuToR across a range of use cases, including supervised fine-tuning, parameter-efficient fine-tuning (PEFT), and pretraining, on challenging generative tasks in both language and vision domains.</p> <hr/> <h3 id="ipa-an-information-preserving-input-projection-framework-for-efficient-foundation-model-adaptation">IPA: An Information-Preserving Input Projection Framework for Efficient Foundation Model Adaptation</h3> <p class="page-description"><a href="https://sites.google.com/view/ccfm-neurips2025">NeurIPS 2025 Workshop on Continual and Compatible Foundation Model Updates (CCFM)</a></p> <h5 id="authors-yuan-yin-shashanka-venkataramanan-tuan-hung-vu-andrei-bursuc-matthieu-cord">Authors: <a href="https://yuan-yin.github.io/">Yuan Yin</a>, <a href="https://shashankvkt.github.io/">Shashanka Venkataramanan</a>, <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2509.04398">Paper</a>]</h5> <p><img src="/assets/img/publications/2025_ipa/ipa.png" alt="ipa_teaser" height="100%" width="100%"/></p> <p>Parameter-efficient fine-tuning (PEFT) methods, such as LoRA, reduce adaptation cost by injecting low-rank updates into pretrained weights. However, LoRA’s down-projection is randomly initialized and data-agnostic, discarding potentially useful information. Prior analyses show that this projection changes little during training, while the up-projection carries most of the adaptation, making the random input compression a performance bottleneck. We propose IPA, a feature-aware projection framework that explicitly preserves information in the reduced hidden space. In the linear case, we instantiate IPA with algorithms approximating top principal components, enabling efficient projector pretraining with negligible inference overhead. Across language and vision benchmarks, IPA consistently improves over LoRA and DoRA, achieving on average 1.5 points higher accuracy on commonsense reasoning and 2.3 points on VTAB-1k, while matching full LoRA performance with roughly half the trainable parameters when the projection is frozen.</p>]]></content><author><name></name></author><category term="foundation"/><category term="limited-supervision"/><category term="deep-learning"/><category term="generalization"/><category term="explainability"/><summary type="html"><![CDATA[The Neural Information Processing Systems Conference (NeurIPS) is a major inter-disciplinary event that brings together researchers and practicioners in machine learning, computer vision, natural language processing, optimization, statistics, but also neuroscience, natural sciences, social sciences, etc. This year, at the 39th edition of NeurIPS, the valeo.ai team will present 5 papers in the main conference and 1 in the workshops. We are honored to announce that our IPA paper on efficient foundation model adaptation has received the outstanding paper award at the CCFM workshop. Our team contributed to the technical program committee with multiple reviewers out of whom 1 was awarded top reviewer and 2 as top area chairs.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2025_neurips.jpg"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2025_neurips.jpg" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at ICCV 2025</title><link href="https://valeoai.github.io//posts/iccv-2025" rel="alternate" type="text/html" title="valeo.ai at ICCV 2025"/><published>2025-10-08T00:00:00+00:00</published><updated>2025-10-08T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-iccv-2025</id><content type="html" xml:base="https://valeoai.github.io//posts/iccv-2025"><![CDATA[<p>The <a href="https://iccv.thecvf.com/">International Conference on Computer Vision (ICCV)</a> is a leading conference that brings together researchers and practitioners in computer vision and machine learning. At the 2025 edition, the <a href="../">valeo.ai</a> team will present five papers in the main conference. We are also co-organizing the <a href="https://opendrivelab.com/iccv2025/workshop/">Learning to See: Advancing Spatial Understanding for Embodied Intelligence</a> workshop, and contributing to the <a href="https://iccv2025-found-workshop.limitlab.xyz">Foundational Data for Industrial Tech Transfer</a> workshop with a keynote on <a href="https://iccv2025-found-workshop.limitlab.xyz/program"><em>Towards openness of vision foundation models</em></a>.</p> <p>The team will be at ICCV to present these works, exchange ideas, and share our exciting ongoing research. We look forward to seeing you in Honolulu!</p> <p><img src="/assets/img/posts/valeoai_iccv.png" alt="valeo.ai papers at ICCV 2025" height="100%" width="100%"/></p> <hr/> <p><img src="/assets/img/posts/vai_at_iccv25.jpg" alt="valeo.ai team at ICCV 2025" height="100%" width="100%"/></p> <hr/> <h3 id="dip-unsupervised-dense-in-context-post-training-of-visual-representations">DIP: Unsupervised Dense In-Context Post-training of Visual Representations</h3> <h5 id="authors-sophia-sirko-galouchenko-antonin-vobecky-andrei-bursuc-nicolas-thome-spyros-gidaris">Authors: <a href="https://scholar.google.com/citations?user=3ac3PQMAAAAJ&amp;hl=fr">Sophia Sirko-Galouchenko</a>, <a href="https://vobecant.github.io/">Antonin Vobecky</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://scholar.google.com/citations?user=3ac3PQMAAAAJ&amp;hl=fr">Nicolas Thome</a>, <a href="https://gidariss.github.io/&amp;hl=en">Spyros Gidaris</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2506.18463">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/sirkosophia/DIP">Code</a>] </h5> <p><img src="/assets/img/publications/2025_dip.png" alt="dip_teaser" height="100%" width="100%"/></p> <p>We introduce DIP, a novel unsupervised post-training method designed to enhance dense image representations in large-scale pretrained vision encoders for in-context scene understanding. Unlike prior approaches that rely on complex self-distillation architectures, our method trains the vision encoder using pseudo-tasks that explicitly simulate downstream in-context scenarios, inspired by meta-learning principles. To enable post-training on unlabeled data, we propose an automatic mechanism for generating in-context tasks that combines a pretrained diffusion model and the vision encoder itself. DIP is simple, unsupervised, and computationally efficient, requiring less than 9 hours on a single A100 GPU. By learning dense representations through pseudo in-context tasks, it achieves strong performance across a wide variety of downstream real-world in-context scene understanding tasks. It outperforms both the initial vision encoder and prior methods, offering a practical and effective solution for improving dense representations.</p> <hr/> <h3 id="gaussrender-learning-3d-occupancy-with-gaussian-rendering">GaussRender: Learning 3D Occupancy with Gaussian Rendering</h3> <h5 id="authors-loïck-chambon-éloi-zablocki-alexandre-boulch-mickaël-chen-matthieu-cord">Authors: <a href="https://loickch.github.io/">Loïck Chambon</a>, <a href="https://eloiz.github.io">Éloi Zablocki</a>, <a href="https://boulch.eu/">Alexandre Boulch</a>, <a href="https://scholar.google.com/citations?user=QnRpMJAAAAAJ">Mickaël Chen</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2502.05040">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/GaussRender">Code</a>] </h5> <p><img src="/assets/img/publications/2025_gaussrender/teaser.png" alt="gaussrender_teaser" height="100%" width="100%"/></p> <p>Understanding the 3D geometry and semantics of driving scenes is critical for safe autonomous driving. Recent advances in 3D occupancy prediction improve scene representation but often suffer from spatial inconsistencies, leading to floating artifacts and poor surface localization. Existing voxel-wise losses (e.g., cross-entropy) fail to enforce geometric coherence. GaussRender is a module that improves 3D occupancy learning by enforcing projective consistency. The key idea is to project both predicted and ground-truth 3D occupancy into 2D camera views for supervision, penalizing inconsistent 3D configurations and enforcing coherent 3D structure. To achieve this efficiently, GaussRender leverages differentiable rendering with Gaussian splatting. It integrates seamlessly with existing architectures, requires no inference-time modifications, and significantly improves geometric fidelity across multiple benchmarks (SurroundOcc-nuScenes, Occ3D-nuScenes, SSCBench-KITTI360) and 3D occupancy models (TPVFormer, SurroundOcc, Symphonies).</p> <hr/> <div class="row"> <div class="col-sm mt-3 mt-md-0" style="padding-right: 5px;padding-left: 5px;"> <img src="../../assets/img/publications/2025_gaussrender/demo_scene_0003.gif" class="img-fluid rounded z-depth-1"/> <div class="caption"> Scene visualization 1 </div> </div> <div class="col-sm mt-3 mt-md-0" style="padding-right: 5px;padding-left: 5px;"> <img src="../../assets/img/publications/2025_gaussrender/demo_scene_0013.gif" class="img-fluid rounded z-depth-1"/> <div class="caption"> Scene visualization 2 </div> </div> </div> <div class="caption"> GaussRender is a 3D Occupancy module that can be plugged into any 3D Occupancy model to enhance its predictions and ensure 2D-3D consistency while improving mIoU, IoU, and RayIoU. </div> <hr/> <div class="row"> <div class="col-sm mt-3 mt-md-0"> <img src="../../assets/img/publications/2025_gaussrender/pipeline.png" class="img-fluid rounded z-depth-1"/> <div class="caption"> GaussRender can be plugged into any model. The core idea is to transform voxels into Gaussians before performing a depth and semantic rendering. </div> </div> </div> <h4 align="center">Results</h4> <p>GaussRender can be plugged into any 3D model. Dedicated experiments on multiple 3D benchmarks (SurroundOcc-nuScenes, Occ3D-nuScenes, SSCBench-KITTI360) and models (TPVFormer, SurroundOcc, Symphonies) demonstrate its performance.</p> <h5>Occ3D-nuScenes</h5> <table> <thead> <tr> <th>Models</th> <th><a href="https://arxiv.org/abs/2502.05040">TPVFormer (ours)</a></th> <th><a href="https://arxiv.org/abs/2302.07817">TPVFormer</a></th> <th><a href="https://arxiv.org/abs/2502.05040">SurroundOcc (ours)</a></th> <th><a href="https://arxiv.org/abs/2303.09551">SurroundOcc</a></th> <th><a href="https://arxiv.org/abs/2304.05316">OccFormer</a></th> <th><a href="https://arxiv.org/abs/2309.09502">RenderOcc</a></th> </tr> <tr> <th>Type</th> <th>w/ GaussRender</th> <th>base</th> <th>w/ GaussRender</th> <th>base</th> <th>base</th> <th>base</th> </tr> </thead> <tbody> <tr> <td>mIoU</td> <td><strong>30.48 🥇</strong> <span style="color: green;">(+2.65)</span></td> <td>27.83</td> <td>30.38 🥈 <span style="color: green;">(+1.17)</span></td> <td>29.21</td> <td>21.93</td> <td>26.11</td> </tr> <tr> <td>RayIoU</td> <td><strong>38.3 🥇</strong> <span style="color: green;">(+1.1)</span></td> <td>37.2</td> <td>37.5 🥈 <span style="color: green;">(+2.0)</span></td> <td>35.5</td> <td>-</td> <td>19.5</td> </tr> </tbody> </table> <h5>SurroundOcc-nuScenes</h5> <table> <thead> <tr> <th>Models</th> <th><a href="https://arxiv.org/abs/2502.05040">TPVFormer (ours)</a></th> <th><a href="https://arxiv.org/abs/2302.07817">TPVFormer</a></th> <th><a href="https://arxiv.org/abs/2502.05040">SurroundOcc (ours)</a></th> <th><a href="https://arxiv.org/abs/2303.09551">SurroundOcc</a></th> <th><a href="https://arxiv.org/abs/2304.05316">OccFormer</a></th> <th><a href="https://arxiv.org/abs/2412.04384">GaussianFormerv2</a></th> </tr> <tr> <th>Type</th> <th>w/ GaussRender</th> <th>base</th> <th>w/ GaussRender</th> <th>base</th> <th>base</th> <th>base</th> </tr> </thead> <tbody> <tr> <td>IoU</td> <td>32.05 🥈 <span style="color: green;">(+1.19)</span></td> <td>30.86</td> <td><strong>32.61 🥇</strong> <span style="color: green;">(+1.12)</span></td> <td>31.49</td> <td>31.39</td> <td>30.56</td> </tr> <tr> <td>mIoU</td> <td>20.58 🥈 <span style="color: green;">(+3.48)</span></td> <td>17.10</td> <td><strong>20.82 🥇</strong> <span style="color: green;">(+0.52)</span></td> <td>20.30</td> <td>19.03</td> <td>20.02</td> </tr> </tbody> </table> <h5>SSCBench-KITTI360</h5> <table> <thead> <tr> <th>Models</th> <th><a href="https://arxiv.org/abs/2502.05040">SurroundOcc (ours)</a></th> <th><a href="https://arxiv.org/abs/2303.09551">SurroundOcc</a></th> <th><a href="https://arxiv.org/abs/2502.05040">Symphonies (ours)</a></th> <th><a href="https://arxiv.org/abs/2306.15670">Symphonies</a></th> <th><a href="https://arxiv.org/abs/2304.05316">OccFormer</a></th> <th><a href="https://arxiv.org/abs/2112.00726">MonoScene</a></th> </tr> <tr> <th>Type</th> <th>w/ GaussRender</th> <th>base</th> <th>w/ GaussRender</th> <th>base</th> <th>base</th> <th>base</th> </tr> </thead> <tbody> <tr> <td>IoU</td> <td>38.62 <span style="color: green;">(+0.11)</span></td> <td>38.51</td> <td><strong>44.08 🥇</strong> <span style="color: green;">(+0.68)</span></td> <td>43.40 🥈</td> <td>40.27</td> <td>37.87</td> </tr> <tr> <td>mIoU</td> <td>13.34 <span style="color: green;">(+0.26)</span></td> <td>13.08</td> <td><strong>18.11 🥇</strong> <span style="color: green;">(+0.29)</span></td> <td>17.82 🥈</td> <td>13.81</td> <td>12.31</td> </tr> </tbody> </table> <hr/> <h3 id="mosic-optimal-transport-motion-trajectories-for-dense-self-supervised-learning">MoSiC: Optimal-Transport Motion Trajectories for Dense Self-Supervised Learning</h3> <h5 id="authors-mohammadreza-salehi-shashanka-venkataramanan-ioana-simion-efstratios-gavves-cees-snoek-yuki-asano">Authors: <a href="https://scholar.google.com/citations?user=kpT3gcsAAAAJ&amp;hl=en">Mohammadreza Salehi</a>, <a href="https://shashankvkt.github.io/">Shashanka Venkataramanan</a>, Ioana Simion, <a href="https://www.egavves.com/">Efstratios Gavves</a>, <a href="https://www.ceessnoek.info/">Cees Snoek</a>, <a href="https://yukimasano.github.io/">Yuki Asano</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2506.08694">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/SMSD75/MoSiC/tree/main">Code</a>] </h5> <p><img src="/assets/img/publications/2025_mosic.png" alt="mosic_teaser" height="100%" width="100%"/></p> <p>Dense self-supervised learning has shown great promise for learning pixel- and patch-level representations, but extending it to videos remains challenging due to complex motion dynamics. Existing approaches struggle under object deformations, occlusions, and camera movement, leading to inconsistent feature learning over time. In this work, we introduce MoSiC, a motion-guided self-supervised framework that clusters dense point tracks to learn spatiotemporally consistent representations. Using an off-the-shelf point tracker, we extract long-range motion trajectories and optimize feature clustering with a momentum-encoder-based optimal transport mechanism. Temporal coherence is enforced by propagating cluster assignments along tracked points, ensuring feature consistency across views despite viewpoint changes. By leveraging motion as an implicit supervisory signal and initializing from strong image-pretrained models, MoSiC learns robust representations that generalize across frames. Our approach improves state-of-the-art performance by 1% to 6% across six image and video datasets and four evaluation benchmarks.</p> <hr/> <h3 id="floss-free-lunch-in-open-vocabulary-semantic-segmentation">FLOSS: Free Lunch in Open-vocabulary Semantic Segmentation</h3> <h5 id="authors-yasser-benigmim-mohammad-fahes-tuan-hung-vu-andrei-bursuc-raoul-de-charette">Authors: <a href="https://yasserben.github.io/">Yasser Benigmim</a>, <a href="https://mfahes.github.io/">Mohammad Fahes</a>, <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://team.inria.fr/rits/membres/raoul-de-charette/">Raoul de Charette</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2504.10487">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/yasserben/FLOSS">Code</a>] </h5> <p><img src="/assets/img/publications/2025_floss.jpg" alt="floss_teaser" height="100%" width="100%"/></p> <p>In Open-Vocabulary Semantic Segmentation (OVSS), class-wise text embeddings are usually averaged over multiple templates (e.g., “a photo of <class>", "a sketch of <class>") to form classifiers. We show that for each class, there exist single-template classifiers—termed class-experts—that outperform conventional averaged classifiers. To identify these class-experts without labeled data, we leverage class-wise prediction entropy and select the lowest-entropy classifiers as the most reliable. We then fuse the outputs of these class-experts using a novel plug-and-play process called FLOSS. FLOSS is complementary to existing OVSS methods, requiring no additional labels or training, yet consistently improves performance. Extensive experiments demonstrate that FLOSS enhances state-of-the-art OVSS models, generalizes across datasets with distribution shifts, and yields strong improvements in low-data scenarios with only a few unlabeled images.</class></class></p> <hr/> <h3 id="analyzing-fine-tuning-representation-shift-for-multimodal-llms-steering-alignment">Analyzing Fine-tuning Representation Shift for Multimodal LLMs Steering Alignment</h3> <h5 id="authors-pegah-khayatan-mustafa-shukor-jayneel-parekh-matthieu-cord">Authors: <a href="https://pegah-kh.github.io/">Pegah Khayatan</a>, <a href="https://mustafashukor.github.io/">Mustafa Shukor</a>, <a href="https://jayneelparekh.github.io/">Jayneel Parekh</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h5> <h5 align="center"> [<a href="https://arxiv.org/abs/2501.03012">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/mshukor/xl-vlms/">Code</a>] </h5> <p><img src="/assets/img/publications/2025_xlxvlm.png" alt="xlvlm_teaser" height="100%" width="100%"/></p> <p>Multimodal LLMs have achieved remarkable proficiency in understanding multimodal inputs, yet little attention has been given to explaining how their internal representations evolve during training. Most explainability research focuses only on final model states, ignoring dynamic representational shifts. In this work, we systematically analyze the evolution of hidden state representations during fine-tuning, revealing how models adapt to new multimodal tasks. Using a concept-based approach, we map hidden states to interpretable visual and textual concepts, enabling us to trace concept changes across modalities as training progresses. We introduce shift vectors to capture these changes, allowing recovery of fine-tuned concepts from the original model. Furthermore, we demonstrate practical applications in model steering, such as adjusting answer types, caption styles, or biasing responses without additional training. This work provides novel insights into multimodal representation adaptation and offers tools for interpreting and controlling fine-tuned multimodal LLMs.</p>]]></content><author><name></name></author><category term="multi-sensor"/><category term="3d-perception"/><category term="foundation"/><category term="limited-supervision"/><category term="zero-shot"/><category term="deep-learning"/><category term="generalization"/><category term="explainability"/><summary type="html"><![CDATA[The International Conference on Computer Vision (ICCV) is a leading conference that brings together researchers and practitioners in computer vision and machine learning. At the 2025 edition, the valeo.ai team will present five papers in the main conference. We are also co-organizing the Learning to See: Advancing Spatial Understanding for Embodied Intelligence workshop, and contributing to the Foundational Data for Industrial Tech Transfer workshop with a keynote on Towards openness of vision foundation models.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/ICCV_2025.png"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/ICCV_2025.png" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at ICLR 2025</title><link href="https://valeoai.github.io//posts/2025-04-01-valeoai-at-iclr-2025/" rel="alternate" type="text/html" title="valeo.ai at ICLR 2025"/><published>2025-04-01T00:00:00+00:00</published><updated>2025-04-01T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-iclr-2025</id><content type="html" xml:base="https://valeoai.github.io//posts/2025-04-01-valeoai-at-iclr-2025/"><![CDATA[<p>The <a href="https://iclr.cc/">International Conference on Learning Representations (ICLR)</a> is a leading conference that brings together researchers and practitioners in deep learning, representation learning, and artificial intelligence. It covers a wide range of topics, including optimization, generative models, interpretability, robustness. This year, at the thirteen edition of ICLR, the <a href="../../">valeo.ai</a> team will present 5 papers in the main conference.</p> <p>We will be happy to discuss more about these projects and ideas, and share our exciting ongoing research. Take a quick view of our papers below and come meet us at the posters or catch us for a coffee in the hallways.</p> <hr/> <h2 id="halton-scheduler-for-masked-generative-image-transformer">Halton Scheduler For Masked Generative Image Transformer</h2> <h3 id="authors-victor-besnier--mickael-chen--david-hurych--eduardo-valle--matthieu-cord">Authors: <a href="https://scholar.google.com/citations?hl=fr&amp;user=n_C2h-QAAAAJ">Victor Besnier</a>   <a href="https://www.linkedin.com/in/mickael-chen-ml/">Mickael Chen</a>   <a href="https://scholar.google.com/citations?user=XY1PVwYAAAAJ&amp;hl=fr&amp;oi=ao">David Hurych</a>   <a href="https://eduardovalle.com/">Eduardo Valle</a>   <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2503.17076">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/Halton-MaskGIT">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/2025_halton_maskgit">Project page</a>]</h4> <p><img src="/assets/img/publications/2025_halton_maskgit/schematics.png" alt="halton_overview" height="100%" width="100%"/></p> <p>Masked Generative Image Transformers (MaskGIT) have gained popularity for their fast and efficient image generation capabilities. However, the sampling strategy used to progressively <em>"unmask"</em> tokens in these models plays a crucial role in determining image quality and diversity. Our new research paper, introduces the <strong>Halton Scheduler</strong>—a novel approach that significantly enhances MaskGIT's image generation performance.</p> <h3>From Confidence to Halton: What’s New?</h3> <p>Traditional MaskGIT uses a Confidence scheduler, which selects tokens based on logit distribution but tends to cluster token selection, leading to reduced image diversity. The Halton Scheduler addresses this by leveraging <strong>low-discrepancy sequences</strong>, the Halton sequence, to distribute token selection more uniformly across the image.</p> <div style="text-align: center;"> <img src="../../assets/img/publications/2025_halton_maskgit/imagenet_quali.png" alt="Halton exemple on ImageNet" style="max-width: 100%; height: auto; border-radius: 5px;"/> <p style="font-size: 14px; color: #555;">MaskGIT using our Halton scheduler on ImageNet 256.</p> </div> <h3>Key Insights and Benefits</h3> <ul> <li><strong>Improved Image Quality and Diversity:</strong> The Halton scheduler reduces clustering of sampled tokens, enhancing image sharpness and background richness.</li> <li><strong>No Retraining Required:</strong> This scheduler can be integrated as a drop-in replacement for the existing MaskGIT sampling strategy.</li> <li><strong>Faster and More Balanced Sampling:</strong> By reducing token correlation, the Halton Scheduler allows MaskGIT to progressively add fine details while avoiding local sampling errors.</li> </ul> <div style="text-align: center;"> <img src="../../assets/img/publications/2025_halton_maskgit/txt2img_halton.jpg" alt="Halton exemple" style="max-width: 100%; height: auto; border-radius: 5px;"/> <p style="font-size: 14px; color: #555;">Figure 2: MaskGIT using our Halton scheduler for text-to-image.</p> </div> <div style="text-align: center;"> <img src="../../assets/img/publications/2025_halton_maskgit/txt2img_conf.jpg" alt="Confidence exemple" style="max-width: 100%; height: auto; border-radius: 5px;"/> <p style="font-size: 14px; color: #555;">Figure 3: MaskGIT using the Confidence scheduler for text-to-image.</p> </div> <h3>Results: ImageNet and COCO Benchmarks</h3> <p>On benchmark datasets like ImageNet (256×256) and COCO, the Halton Scheduler outperforms the baseline Confidence scheduler:</p> <ul> <li><strong>Reduced Fréchet Inception Distance (FID):</strong> Indicating better image realism.</li> <li><strong>Improved Precision and Recall:</strong> Reflecting a more diverse image generation.</li> </ul> <hr/> <h2 id="llm-wrapper-black-box-semantic-aware-adaptation-of-vision-language-models-for-referring-expression-comprehension">LLM-wrapper: Black-Box Semantic-Aware Adaptation of Vision-Language Models for Referring Expression Comprehension</h2> <h3 id="authors-amaia-cardiel--éloi-zablocki--elias-ramzi--oriane-siméoni--matthieu-cord">Authors: Amaia Cardiel    <a href="https://eloiz.github.io">Éloi Zablocki</a>    <a href="https://elias-ramzi.github.io/">Elias Ramzi</a>    <a href="https://osimeoni.github.io/">Oriane Siméoni</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2409.11919">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/LLM_wrapper">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/llm_wrapper/">Project page</a>]</h4> <p><img src="/assets/img/publications/2024_llm_wrapper/llm_wrapper_pipeline.png" alt="llm_wrapper_overview" height="100%" width="100%"/></p> <p>Vision Language Models (VLMs) have demonstrated remarkable capabilities in various open-vocabulary tasks, yet their zero-shot performance lags behind task-specific fine-tuned models, particularly in complex tasks like Referring Expression Comprehension (REC). Fine-tuning usually requires “white-box” access to the model’s architecture and weights, which is not always feasible due to proprietary or privacy concerns. In this work, we propose LLM-wrapper, a method for “black-box” adaptation of VLMs for the REC task using Large Language Models (LLMs). LLM-wrapper capitalizes on the reasoning abilities of LLMs, improved with a light fine-tuning, to select the most relevant bounding box matching the referring expression, from candidates generated by a zero-shot black-box VLM. Our approach offers several advantages: it enables the adaptation of closed-source models without needing access to their internal workings, it is versatile as it works with any VLM, it transfers to new VLMs and datasets, and it allows for the adaptation of an ensemble of VLMs. We evaluate LLM-wrapper on multiple datasets using different VLMs and LLMs, demonstrating significant performance improvements and highlighting the versatility of our method. While LLM-wrapper is not meant to directly compete with standard white-box fine-tuning, it offers a practical and effective alternative for black-box VLM adaptation.</p> <p><img src="/assets/img/publications/2024_llm_wrapper/llm_wrapper_talk2car.png" alt="llm_wrapper_results" height="100%" width="100%"/></p> <hr/> <h2 id="moca-self-supervised-representation-learning-by-predicting-masked-online-codebook-assignments">MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments</h2> <h3 id="authors-spyros-gidaris-andrei-bursuc--oriane-siméoni--antonin-vobecky--nikos-komodakis--matthieu-cord--patrick-pérez">Authors: <a href="https://gidariss.github.io/&amp;hl=en">Spyros Gidaris</a>   <a href="https://abursuc.github.io/">Andrei Bursuc</a>   <a href="https://osimeoni.github.io/">Oriane Siméoni</a>    <a href="https://vobecant.github.io/">Antonin Vobecky</a>    <a href="https://www.csd.uoc.gr/~komod/">Nikos Komodakis</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>    <a href="https://ptrckprz.github.io/">Patrick Pérez</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2312.15297">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/MOCA">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/moca/">Project page</a>]</h4> <p><img src="/assets/img/publications/2024_moca/moca-teaser.png" alt="moca_teaser" height="100%" width="100%"/></p> <p>Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks for very large fully-annotated datasets. Different classes of self-supervised learning offer representations with either good contextual reasoning properties, e.g., using masked image modeling strategies, or invariance to image perturbations, e.g., with contrastive methods. In this work, we propose a single-stage and standalone method, MOCA, which unifies both desired properties using novel mask-and-predict objectives defined with high-level features (instead of pixel-level details). Moreover, we show how to effectively employ both learning paradigms in a synergistic and computation-efficient way. Doing so, we achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols with a training that is at least 3 times faster than prior methods.</p> <hr/> <h2 id="learning-a-neural-solver-for-parametric-pdes-to-enhance-physics-informed-methods">Learning a Neural Solver for Parametric PDEs to Enhance Physics-Informed Methods</h2> <h3 id="authors-lise-le-boudec--emmanuel-de-bezenac--louis-serrano--ramon-daniel-regueiro-espino--yuan-yin--patrick-gallinari">Authors: <a href="https://2ailesb.github.io/">Lise Le Boudec</a>    <a href="https://scholar.google.fr/citations?user=KvZw5gYAAAAJ">Emmanuel de Bezenac</a>    <a href="https://scholar.google.com/citations?user=fKlo-lUAAAAJ">Louis Serrano</a>    <a href="https://rd-regueiroespino.github.io/">Ramon Daniel Regueiro-Espino</a>    <a href="https://yuan-yin.github.io">Yuan Yin</a>    <a href="https://pages.isir.upmc.fr/gallinari/">Patrick Gallinari</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2410.06820">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/2ailesB/neural-parametric-solver">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/neural-parametric-solver/">Project page</a>]</h4> <p><img src="/assets/img/publications/2025_neural_parametric_solver.png" alt="neural_solver_pde" height="100%" width="100%"/></p> <p>Physics-informed deep learning often faces optimization challenges due to the complexity of solving partial differential equations (PDEs), which involve exploring large solution spaces, require numerous iterations, and can lead to unstable training. These challenges arise particularly from the ill-conditioning of the optimization problem, caused by the differential terms in the loss function. To address these issues, we propose learning a solver, i.e., solving PDEs using a physics-informed iterative algorithm trained on data. Our method learns to condition a gradient descent algorithm that automatically adapts to each PDE instance, significantly accelerating and stabilizing the optimization process and enabling faster convergence of physics-aware models. Furthermore, while traditional physics-informed methods solve for a single PDE instance, our approach addresses parametric PDEs. Specifically, our method integrates the physical loss gradient with the PDE parameters to solve over a distribution of PDE parameters, including coefficients, initial conditions, or boundary conditions. We demonstrate the effectiveness of our method through empirical experiments on multiple datasets, comparing training and test-time optimization performance.</p> <hr/> <h2 id="toddlerdiffusion-interactive-structured-image-generation-with-cascaded-schrödinger-bridge">ToddlerDiffusion: Interactive Structured Image Generation with Cascaded Schrödinger Bridge</h2> <h3 id="authors-eslam-abdelrahman--liangbing-zhao--vincent-tao-hu--matthieu-cord--patrick-perez--mohamed-elhoseiny">Authors: Eslam Abdelrahman    Liangbing Zhao    Vincent Tao Hu    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>    <a href="https://ptrckprz.github.io/">Patrick Perez</a>    Mohamed Elhoseiny</h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2311.14542">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/toddlerdiffusion/code">Code</a>] &nbsp;&nbsp; [<a href="https://toddlerdiffusion.github.io/website/">Project page</a>]</h4> <p><img src="/assets/img/publications/2025_toddler.PNG" alt="toddlerdiffusion" height="100%" width="100%"/></p> <p>Diffusion models break down the challenging task of generating data from high-dimensional distributions into a series of easier denoising steps. Inspired by this paradigm, we propose a novel approach that extends the diffusion framework into modality space, decomposing the complex task of RGB image generation into simpler, interpretable stages. Our method, termed ToddlerDiffusion, cascades modality-specific models, each responsible for generating an intermediate representation, such as contours, palettes, and detailed textures, ultimately culminating in a high-quality RGB image. Instead of relying on the naive LDM concatenation conditioning mechanism to connect the different stages together, we employ Schrödinger Bridge to determine the optimal transport between different modalities. Although employing a cascaded pipeline introduces more stages, which could lead to a more complex architecture, each stage is meticulously formulated for efficiency and accuracy, surpassing Stable-Diffusion (LDM) performance. Modality composition not only enhances overall performance but enables emerging proprieties such as consistent editing, interaction capabilities, high-level interpretability, and faster convergence and sampling rate. Extensive experiments on diverse datasets, including LSUN-Churches, ImageNet, CelebHQ, and LAION-Art, demonstrate the efficacy of our approach, consistently outperforming state-of-the-art methods. For instance, ToddlerDiffusion achieves notable efficiency, matching LDM performance on LSUN-Churches while operating 2× faster with a 3× smaller architecture.</p>]]></content><author><name></name></author><category term="limited-supervision"/><category term="reliability"/><category term="foundation"/><category term="robustness"/><category term="generalization"/><category term="deep-learning"/><summary type="html"><![CDATA[The International Conference on Learning Representations (ICLR) is a leading conference that brings together researchers and practitioners in deep learning, representation learning, and artificial intelligence. It covers a wide range of topics, including optimization, generative models, interpretability, robustness. This year, at the thirteen edition of ICLR, the valeo.ai team will present 5 papers in the main conference.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2025_iclr.jpg"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2025_iclr.jpg" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at NeurIPS 2024</title><link href="https://valeoai.github.io//posts/2024-12-04-valeoai-at-neurips-2024/" rel="alternate" type="text/html" title="valeo.ai at NeurIPS 2024"/><published>2024-12-04T00:00:00+00:00</published><updated>2024-12-04T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-neurips-2024</id><content type="html" xml:base="https://valeoai.github.io//posts/2024-12-04-valeoai-at-neurips-2024/"><![CDATA[<p>The <a href="https://neurips.cc/">Neural Information Processing Systems Conference (NeurIPS)</a> is a major inter-disciplinary event that brings together researchers and practicioners in machine learning, computer vision, natural language processing, optimization, statistics, but also neuroscience, natural sciences, social sciences, etc. This year, at the thirty-eigth edition of NeurIPS, the <a href="../">valeo.ai</a> team will present 7 papers in the main conference.</p> <p>We will be happy to discuss more about these projects and ideas, and share our exciting ongoing research. Take a quick view of our papers below and come meet us at the posters or catch us for a coffee in the hallways.</p> <hr/> <h2 id="no-train-all-gain-self-supervised-gradients-improve-deep-frozen-representations">No Train, all Gain: Self-Supervised Gradients Improve Deep Frozen Representations</h2> <h3 id="authors-walter-simoncini---spyros-gidaris----andrei-bursuc--yuki-m-asano">Authors: <a href="https://walter.ashita.nl/">Walter Simoncini</a>    <a href="https://gidariss.github.io/&amp;hl=en">Spyros Gidaris</a>    <a href="https://abursuc.github.io/">Andrei Bursuc</a>    <a href="https://yukimasano.github.io/">Yuki M. Asano</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2407.10964">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/WalterSimoncini/fungivision">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/fungi">Project page</a>]</h4> <p><img src="/assets/img/publications/2024_fungi/overview.jpg" alt="fungi_overview" height="100%" width="100%"/></p> <p>This paper introduces FUNGI: Features from UNsupervised GradIents, a method to enhance the features of vision encoders by leveraging self-supervised gradients. Our method is simple: given any pretrained model, we first compute gradients from various self-supervised objectives for each input. These are projected to a lower dimension and then concatenated with the model’s embedding. The resulting features are evaluated on k-nearest neighbor classification over 11 datasets from vision, 5 from natural language processing, and 2 from audio. Across backbones spanning various sizes and pretraining strategies, FUNGI features provide consistent performance improvements over the embeddings. We also show that using FUNGI features can benefit linear classification and image retrieval, and that they significantly improve the retrieval-based in-context scene understanding abilities of pretrained models, for example improving upon DINO by +17% for semantic segmentation — without any training.</p> <hr/> <h2 id="manipose-manifold-constrained-multi-hypothesis-3d-human-pose-estimation">ManiPose: Manifold-Constrained Multi-Hypothesis 3D Human Pose Estimation</h2> <h3 id="authors-cédric-rommel--victor-letzelter--nermin-samet--renaud-marlet---matthieu-cord--patrick-pérez--eduardo-valle">Authors: <a href="https://cedricrommel.github.io/">Cédric Rommel</a>    <a href="https://scholar.google.com/citations?user=YhTdZh8AAAAJ&amp;hl=en&amp;oi=ao">Victor Letzelter</a>    <a href="https://nerminsamet.github.io/">Nermin Samet</a>    <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>    <a href="https://ptrckprz.github.io/">Patrick Pérez</a>    <a href="https://eduardovalle.com/">Eduardo Valle</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2312.06386">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/cedricrommel/manipose">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/manipose/">Project page</a>]</h4> <p>We propose ManiPose, a manifold-constrained multi-hypothesis model for human-pose 2D-to-3D lifting. We provide theoretical and empirical evidence that, due to the depth ambiguity inherent to monocular 3D human pose estimation, traditional regression models suffer from pose-topology consistency issues, which standard evaluation metrics (MPJPE, P-MPJPE and PCK) fail to assess. ManiPose addresses depth ambiguity by proposing multiple candidate 3D poses for each 2D input, each with its estimated plausibility. Unlike previous multi-hypothesis approaches, ManiPose forgoes generative models, greatly facilitating its training and usage. By constraining the outputs to lie on the human pose manifold, ManiPose guarantees the consistency of all hypothetical poses, in contrast to previous works. We showcase the performance of ManiPose on real-world datasets, where it outperforms state-of-the-art models in pose consistency by a large margin while being very competitive on the MPJPE metric.</p> <p><img src="/assets/img/publications/2024_manipose/ManiPose_teaser.png" alt="manipose_overview" height="90%" width="90%"/></p> <hr/> <h2 id="annealed-multiple-choice-learning-overcoming-limitations-of-winner-takes-all-with-annealing">Annealed Multiple Choice Learning: Overcoming limitations of Winner-takes-all with annealing</h2> <h3 id="authors-david-perera--victor-letzelter--théo-mariotte---adrien-cortés--mickael-chen--slim-essid--gaël-richard">Authors: <a href="https://adasp.telecom-paris.fr/members/">David Perera</a>    <a href="https://scholar.google.com/citations?user=YhTdZh8AAAAJ&amp;hl=en&amp;oi=ao">Victor Letzelter</a>    <a href="https://scholar.google.com/citations?user=q3HZFcwAAAAJ">Théo Mariotte </a>    <a href="https://www.linkedin.com/in/c1adrien/">Adrien Cortés</a>    <a href="https://www.linkedin.com/in/mickael-chen-ml/">Mickael Chen</a>    <a href="https://slimessid.github.io/research/">Slim Essid</a>    <a href="https://www.telecom-paris.fr/gael-richard">Gaël Richard</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2407.15580">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/Victorletzelter/annealed_mcl">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/annealing/">Project page</a>]</h4> <p>We introduce Annealed Multiple Choice Learning (aMCL) which combines simulated annealing with MCL. MCL is a learning framework handling ambiguous tasks by predicting a small set of plausible hypotheses. These hypotheses are trained using the Winner-takes-all (WTA) scheme, which promotes the diversity of the predictions. However, this scheme may converge toward an arbitrarily suboptimal local minimum, due to the greedy nature of WTA. We overcome this limitation using annealing, which enhances the exploration of the hypothesis space during training. We leverage insights from statistical physics and information theory to provide a detailed description of the model training trajectory. Additionally, we validate our algorithm by extensive experiments on synthetic datasets, on the standard UCI benchmark, and on speech separation.</p> <p><img src="/assets/img/publications/2024_annealing/amcl_gif.gif" alt="annealing_overview" height="100%" width="100%"/></p> <hr/> <h2 id="geps-boosting-generalization-in-parametric-pde-neural-solvers-through-adaptive-conditioning">GEPS: Boosting Generalization in Parametric PDE Neural Solvers through Adaptive Conditioning</h2> <h3 id="authors-armand-kassaï-koupaï--jorge-mifsut-benet--yuan-yin--jean-noël-vittaut--patrick-gallinari">Authors: <a href="https://itsakk.github.io/">Armand Kassaï Koupaï</a>    <a href="https://www.isir.upmc.fr/personnel/mifsutbenet/">Jorge Mifsut-Benet</a>    <a href="https://yuan-yin.github.io">Yuan Yin</a>    <a href="https://webia.lip6.fr/~vittaut/">Jean-Noël Vittaut</a>    <a href="https://pages.isir.upmc.fr/gallinari/">Patrick Gallinari</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2410.23889">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/itsakk/geps">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/geps/">Project page</a>]</h4> <p>Solving parametric partial differential equations (PDEs) presents significant challenges for data-driven methods due to the sensitivity of spatio-temporal dynamics to variations in PDE parameters. Machine learning approaches often struggle to capture this variability. To address this, data-driven approaches learn parametric PDEs by sampling a very large variety of trajectories with varying PDE parameters. We first show that incorporating conditioning mechanisms for learning parametric PDEs is essential and that among them, \textit{adaptive conditioning}, allows stronger generalization. As existing adaptive conditioning methods do not scale well with respect to the number of PDE parameters, we propose GEPS, a simple adaptation mechanism to boost GEneralization in Pde Solvers via a first-order optimization and low-rank rapid adaptation of a small set of context parameters. We demonstrate the versatility of our approach for both fully data-driven and for physics-aware neural solvers. Validation performed on a whole range of spatio-temporal forecasting problems demonstrates excellent performance for generalizing to unseen conditions including initial conditions, PDE coefficients, forcing terms and solution domain.</p> <p><img src="/assets/img/publications/2024_geps/geps.png" alt="geps_example" height="100%" width="100%"/></p> <hr/> <h2 id="a-concept-based-explainability-framework-for-large-multimodal-models">A Concept-Based Explainability Framework for Large Multimodal Models</h2> <h3 id="authors-jayneel-parekh--pegah-khayatan--mustafa-shukor--alasdair-newson--matthieu-cord">Authors: <a href="https://jayneelparekh.github.io/">Jayneel Parekh</a>    <a href="https://pegah-kh.github.io/">Pegah Khayatan</a>    <a href="https://mustafashukor.github.io/">Mustafa Shukor</a>    <a href="https://sites.google.com/site/alasdairnewson/">Alasdair Newson</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2406.08074">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/mshukor/xl-vlms">Code</a>] &nbsp;&nbsp; [<a href="https://jayneelparekh.github.io/LMM_Concept_Explainability/">Project page</a>]</h4> <p>Large multimodal models (LMMs) combine unimodal encoders and large language models (LLMs) to perform multimodal tasks. Despite recent advancements towards the interpretability of these models, understanding internal representations of LMMs remains largely a mystery. In this paper, we present a novel framework for the interpretation of LMMs. We propose a dictionary learning based approach, applied to the representation of tokens. The elements of the learned dictionary correspond to our proposed concepts. We show that these concepts are well semantically grounded in both vision and text. Thus we refer to these as “multi-modal concepts”. We qualitatively and quantitatively evaluate the results of the learnt concepts. We show that the extracted multimodal concepts are useful to interpret representations of test samples. Finally, we evaluate the disentanglement between different concepts and the quality of grounding concepts visually and textually.</p> <p><img src="/assets/img/publications/2024_neurips/concept_xai.PNG" alt="concept_overview" height="100%" width="100%"/></p> <hr/> <h2 id="diffcut-catalyzing-zero-shot-semantic-segmentation-with-diffusion-features-and-recursive-normalized-cut">DiffCut: Catalyzing Zero-Shot Semantic Segmentation with Diffusion Features and Recursive Normalized Cut</h2> <h3 id="authors-paul-couairon--mustafa-shukor--jean-emmanuel-haugeard--matthieu-cord--nicolas-thomea">Authors: <a href="https://scholar.google.fr/citations?user=yQRnP7YAAAAJ">Paul Couairon</a>    <a href="https://mustafashukor.github.io/">Mustafa Shukor</a>    <a href="https://dblp.org/pid/92/6849.html">Jean-Emmanuel Haugeard</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>    <a href="https://thome.isir.upmc.fr/">Nicolas Thome/a&gt;</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2406.02842">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/PaulCouairon/DiffCut">Code</a>] &nbsp;&nbsp; [<a href="https://diffcut-segmentation.github.io/">Project page</a>]</h4> <p>Foundation models have emerged as powerful tools across various domains including language, vision, and multimodal tasks. While prior works have addressed unsupervised image segmentation, they significantly lag behind supervised models. In this paper, we use a diffusion UNet encoder as a foundation vision encoder and introduce DiffCut, an unsupervised zero-shot segmentation method that solely harnesses the output features from the final self-attention block. Through extensive experimentation, we demonstrate that the utilization of these diffusion features in a graph based segmentation algorithm, significantly outperforms previous state-of-the-art methods on zero-shot segmentation. Specifically, we leverage a recursive Normalized Cut algorithm that softly regulates the granularity of detected objects and produces well-defined segmentation maps that precisely capture intricate image details. Our work highlights the remarkably accurate semantic knowledge embedded within diffusion UNet encoders that could then serve as foundation vision encoders for downstream tasks.</p> <p><img src="/assets/img/publications/2024_neurips/diffcut.PNG" alt="diffcut_overview" height="100%" width="100%"/></p> <hr/> <h2 id="implicit-multimodal-alignment-on-the-generalization-of-frozen-llms-to-multimodal-inputs">Implicit Multimodal Alignment: On the Generalization of Frozen LLMs to Multimodal Inputs</h2> <h3 id="authors-mustafa-shukor--matthieu-cord">Authors: <a href="https://mustafashukor.github.io/">Mustafa Shukor</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2405.16700">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/mshukor/ima-lmms">Code</a>] &nbsp;&nbsp; [<a href="https://ima-lmms.github.io/">Project page</a>]</h4> <p>Large Language Models (LLMs) have demonstrated impressive performance on multimodal tasks, without any multimodal finetuning. They are the building block for Large Multimodal Models, yet, we still lack a proper understanding of their success. In this work, we expose frozen LLMs to image, video, audio and text inputs and analyse their internal representation aiming to understand their generalization beyond textual inputs. Findings. Perceptual tokens (1) are easily distinguishable from textual ones inside LLMs, with significantly different representations, and complete translation to textual tokens does not exist. Yet, (2) both perceptual and textual tokens activate similar LLM weights. Despite being different, (3) perceptual and textual tokens are implicitly aligned inside LLMs, we call this the implicit multimodal alignment (IMA), and argue that this is linked to architectural design, helping LLMs to generalize. This provide more evidence to believe that the generalization of LLMs to multimodal inputs is mainly due to their architecture. Implications. (1) We find a positive correlation between the implicit alignment score and the task performance, suggesting that this could act as a proxy metric for model evaluation and selection. (2) A negative correlation exists regarding hallucinations, revealing that this problem is mainly due to misalignment between the internal perceptual and textual representations. (3) Perceptual tokens change slightly throughout the model, thus, we propose different approaches to skip computations (e.g. in FFN layers), and significantly reduce the inference cost. (4) Due to the slowly changing embeddings across layers, and the high overlap between textual and multimodal activated weights, we compress LLMs by keeping only 1 subnetwork that works well across a wide range of multimodal tasks.</p> <p><img src="/assets/img/publications/2024_neurips/ima_llm.PNG" alt="implicit_overview" height="100%" width="100%"/></p> <hr/>]]></content><author><name></name></author><category term="3d-perception"/><category term="multi-sensor"/><category term="limited-supervision"/><category term="reliability"/><category term="motion-forecasting"/><category term="robustness"/><category term="generalization"/><category term="driving"/><category term="deep-learning"/><summary type="html"><![CDATA[The Neural Information Processing Systems Conference (NeurIPS) is a major inter-disciplinary event that brings together researchers and practicioners in machine learning, computer vision, natural language processing, optimization, statistics, but also neuroscience, natural sciences, social sciences, etc. This year, at the thirty-eigth edition of NeurIPS, the valeo.ai team will present 7 papers in the main conference.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2023_neurips/logo_neurips.svg"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2023_neurips/logo_neurips.svg" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at ECCV 2024</title><link href="https://valeoai.github.io//posts/2024-09-25-valeoai-at-eccv-2024/" rel="alternate" type="text/html" title="valeo.ai at ECCV 2024"/><published>2024-09-25T00:00:00+00:00</published><updated>2024-09-25T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-eccv-2024</id><content type="html" xml:base="https://valeoai.github.io//posts/2024-09-25-valeoai-at-eccv-2024/"><![CDATA[<p>The <a href="https://eccv.ecva.net/">European Conference on Computer Vision (ECCV)</a> is a biennial landmark conference for the increasingly large community of researchers in computer vision and machine learning from both academia and industry. At the 2024 edition the valeo.ai team will present five papers in the main conference and four in the workshops. We are also organizing two tutorials (<a href="https://uqtutorial.github.io/">Bayesian Odyssey</a> and <a href="https://shashankvkt.github.io/eccv2024-SSLBIG-tutorial.github.io/">Time is precious: Self-Supervised Learning Beyond Images</a>), the <a href="https://uncertainty-cv.github.io/2024/">Uncertainty Quantification for Computer Vision</a> workshop, a talk at the <a href="https://sites.google.com/view/omnilabel-workshop-eccv24/program">OmniLabel workshop</a>, and the <a href="https://github.com/valeoai/bravo_challenge">BRAVO challenge</a>. Our team has also contributed to the reviewing process, with seven reviewers, three area chairs, and two outstanding reviewer awards. The team will be at ECCV to present these works and will be happy to discuss more about these projects and ideas, and share our exciting ongoing research.</p> <p><img src="/assets/img/posts/2024_eccv/valeoai_eccv.jfif" alt="valeo.ai team at ECCV 2024" height="100%" width="100%"/></p> <h2 id="train-till-you-drop-towards-stable-and-robust-source-free-unsupervised-3d-domain-adaptation">Train Till You Drop: Towards Stable and Robust Source-free Unsupervised 3D Domain Adaptation</h2> <h3 id="authors-bjoern-michele--alexandre-boulch--tuan-hung-vu--gilles-puy--renaud-marlet--nicolas-courty">Authors: <a href="https://bjoernmichele.com">Bjoern Michele</a>    <a href="https://boulch.eu/">Alexandre Boulch</a>    <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>    <a href="https://sites.google.com/site/puygilles/">Gilles Puy</a>    <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a>    <a href="https://people.irisa.fr/Nicolas.Courty/">Nicolas Courty</a></h3> <h4 align="center"> [<a href="https://arxiv.org/abs/2409.04409">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/TTYD">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/ttyd/">Project page</a>]</h4> <p><img src="/assets/img/publications/2024_ttyd/featured.png" alt="ttdy_overview" height="100%" width="100%"/></p> <div class="caption"><b>Evolution of the performance of baselines without degradation prevention strategies as they train over 20k iterations.</b> Our method (TTYDcore) uses an unsupervised criterion to stop training. The horizontal dotted line illustrates that we keep the model obtained at the stopping point (marked with a cross). </div> <p>We tackle the challenging problem of source-free unsupervised domain adaptation (SFUDA) for 3D semantic segmentation. It amounts to performing domain adaptation on an unlabeled target domain without any access to source data; the available information is a model trained to achieve good performance on the source domain. A common issue with existing SFUDA approaches is that performance degrades after some training time, which is a by-product of an under-constrained and ill-posed problem. We discuss two strategies to alleviate this issue. First, we propose a sensible way to regularize the learning problem. Second, we introduce a novel criterion based on agreement with a reference model. It is used (1) to stop the training when appropriate and (2) as validator to select hyperparameters without any knowledge on the target domain. Our contributions are easy to implement and readily amenable for all SFUDA methods, ensuring stable improvements over all baselines. We validate our findings on various 3D lidar settings, achieving state-of-the-art performance.</p> <p><img src="/assets/img/posts/2024_eccv/ttyd_results.PNG" alt="ttyd_results" height="100%" width="100%"/></p> <div class="caption"><b>Examples of results with TTYDstop</b>: ground truth (GT), initial model trained only on source data, training with our training scheme when using our stopping criterion, and “full” training for 20k iterations. Notable errors due to degradation are marked with a dashed rectangle. </div> <hr/> <h2 id="unitraj-a-unified-framework-for-scalable-vehicle-trajectory-prediction">UniTraj: A Unified Framework for Scalable Vehicle Trajectory Prediction</h2> <h4 id="authors-lan-feng--mohammadhossein-bahari--kaouther-messaoud-ben-amor--éloi-zablocki--matthieu-cord--alexandre-alahi">Authors: <a href="https://alan-lanfeng.github.io/">Lan Feng</a>    <a href="https://mohammadhossein-bahari.github.io/">Mohammadhossein Bahari</a>    <a href="https://scholar.google.com/citations?user=X0teZIAAAAAJ">Kaouther Messaoud Ben Amor</a>    <a href="https://scholar.google.fr/citations?user=dOkbUmEAAAAJ">Éloi Zablocki</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>    <a href="https://people.epfl.ch/alexandre.alahi">Alexandre Alahi</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2403.15098">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/vita-epfl/unitraj">Code</a>] &nbsp;&nbsp; [<a href="https://vita-epfl.github.io/UniTraj/">Project page</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=2IzuUtiNA_4">Video</a>]</h4> <p>Vehicle trajectory prediction has increasingly relied on data-driven solutions, but their ability to scale to different data domains and the impact of larger dataset sizes on their generalization remain under-explored. While these questions can be studied by employing multiple datasets, it is challenging due to several discrepancies, e.g., in data formats, map resolution, and semantic annotation types. To address these challenges, we introduce UniTraj, a comprehensive framework that unifies various datasets, models, and evaluation criteria, presenting new opportunities for the vehicle trajectory prediction field. In particular, using UniTraj, we conduct extensive experiments and find that model performance significantly drops when transferred to other datasets. However, enlarging data size and diversity can substantially improve performance, leading to a new state-of-the-art result for the nuScenes dataset. We provide insights into dataset characteristics to explain these findings.</p> <p><img src="/assets/img/publications/2024_unitraj/unitraj.PNG" alt="unitraj_overview" height="100%" width="100%"/></p> <div class="caption"><b>Overview of UniTraj: a unified platform for comprehensive research in trajectory prediction.</b> </div> <hr/> <h2 id="lost-and-found-overcoming-detector-failures-in-online-multi-object-tracking">Lost and Found: Overcoming Detector Failures in Online Multi-Object Tracking</h2> <h4 id="authors-lorenzo-vaquero--yihong-xu--xavier-alameda-pineda--víctor-m-brea--manuel-mucientes">Authors: <a href="https://citius.gal/team/lorenzo-vaquero-otal/">Lorenzo Vaquero</a>    <a href="https://github.com/yihongXU">Yihong Xu</a>    <a href="https://xavirema.eu/">Xavier Alameda-Pineda</a>    <a href="https://citius.gal/team/victor-manuel-brea-sanchez/">Víctor M. Brea</a>    <a href="https://persoal.citius.usc.es/manuel.mucientes/">Manuel Mucientes</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2407.10151">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/lorenzovaquero/BUSCA">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/busca/">Project page</a>]</h4> <p>Multi-object tracking (MOT) endeavors to precisely estimate the positions and identities of multiple objects over time. The prevailing approach, tracking-by-detection (TbD), first detects objects and then links detections, resulting in a simple yet effective method. However, contemporary detectors may occasionally miss some objects in certain frames, causing trackers to cease tracking prematurely. To tackle this issue, we propose BUSCA, meaning ‘to search’, a versatile framework compatible with any online TbD system, enhancing its ability to persistently track those objects missed by the detector, primarily due to occlusions. Remarkably, this is accomplished without modifying past tracking results or accessing future frames, i.e., in a fully online manner. BUSCA generates proposals based on neighboring tracks, motion, and learned tokens. Utilizing a decision Transformer that integrates multimodal visual and spatiotemporal information, it addresses the object-proposal association as a multi-choice question-answering task. BUSCA is trained independently of the underlying tracker, solely on synthetic data, without requiring fine-tuning. Through BUSCA, we showcase consistent performance enhancements across five different trackers and establish a new state-of-the-art baseline across three different benchmarks.</p> <p><img src="/assets/img/publications/2024_busca/busca.png" alt="busca_overview" height="90%" width="90%"/></p> <div class="caption"><b>Overview of BUSCA</b>: Enhancing multi-object trackers by finding undetected objects. </div> <hr/> <h2 id="clip-dinoiser-teaching-clip-a-few-dino-tricks-for-open-vocabulary-semantic-segmentation">CLIP-DINOiser: Teaching CLIP a few DINO tricks for open-vocabulary semantic segmentation</h2> <h4 id="authors-monika-wysoczańska-oriane-siméoni--michaël-ramamonjisoa-andrei-bursuc-tomasz-trzciński--patrick-pérez">Authors: <a href="https://wysoczanska.github.io/">Monika Wysoczańska</a>  <a href="https://osimeoni.github.io/">Oriane Siméoni</a>   <a href="https://michaelramamonjisoa.github.io/">Michaël Ramamonjisoa</a>  <a href="https://abursuc.github.io/">Andrei Bursuc</a>  <a href="https://scholar.google.com/citations?hl=en&amp;user=bJMRBFoAAAAJ">Tomasz Trzciński</a>   <a href="https://ptrckprz.github.io/">Patrick Pérez</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2312.12359">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/wysoczanska/clip_dinoiser/">Code</a>] &nbsp;&nbsp; [<a href="https://wysoczanska.github.io/CLIP_DINOiser/">Project page</a>]</h4> <p>The popular CLIP model displays impressive zero-shot capabilities thanks to its seamless interaction with arbitrary text prompts. However, its lack of spatial awareness makes it unsuitable for dense computer vision tasks, e.g., semantic segmentation, without an additional fine-tuning step that often uses annotations and can potentially suppress its original open-vocabulary properties. Meanwhile, self-supervised representation methods have demonstrated good localization properties without human-made annotations nor explicit supervision. In this work, we take the best of both worlds and propose an open-vocabulary semantic segmentation method, which does not require any annotations.</p> <p><img src="/assets/img/publications/2024_dinoiser/dinoiser-examples.png" alt="dinoiser_example" height="100%" width="100%"/></p> <div class="caption"><b>Examples of open-vocabulary semantic segmentation results obtained with our method CLIP-DINOiser on ‘in-the-wild’ images vs. those of MaskCLIP.</b> </div> <p>We propose to locally improve dense MaskCLIP features, which are computed with a simple modification of CLIP’s last pooling layer, by integrating localization priors extracted from self-supervised features from DINO. By doing so, we greatly improve the performance of MaskCLIP and produce smooth outputs. Moreover, we show that the used self-supervised feature properties can directly be learnt from CLIP features. Our method CLIP-DINOiser needs only a single forward pass of CLIP and two light convolutional layers at inference, no extra supervision nor extra memory and reaches state-of-the-art results on challenging and fine-grained benchmarks such as COCO, Pascal Context, Cityscapes and ADE20k.</p> <p><img src="/assets/img/posts/2024_eccv/dinoiser_overview.PNG" alt="dinoiser_overview" height="100%" width="100%"/></p> <div class="caption"><b>Overview of CLIP-DINOiser.</b> We use DINO as a teacher which ‘teaches’ CLIP how to extract localization information with similar patch correlations. At inference, an input image is forwarded through the frozen CLIP image backbone and MaskCLIP projection. The produced features are then improved with our pooling strategy which is guided by correlations predicted with a trained convolutional layer applied on CLIP.</div> <hr/> <h2 id="reliability-in-semantic-segmentation-can-we-use-synthetic-data">Reliability in Semantic Segmentation: Can We Use Synthetic Data</h2> <h4 id="authors-thibaut-loiseau---tuan-hung-vu---mickael-chen--patrick-pérez--matthieu-cord">Authors: <a href="https://imagine-lab.enpc.fr/staff-members/thibaut-loiseau/">Thibaut Loiseau</a>    <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>    <a href="https://scholar.google.fr/citations?user=QnRpMJAAAAAJ">Mickael Chen</a>    <a href="https://ptrckprz.github.io/">Patrick Pérez</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2312.09231">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/GenVal">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/genval">Project page</a>]</h4> <p>Assessing the reliability of perception models to covariate shifts and out-of-distribution (OOD) detection is crucial for safety-critical applications such as autonomous vehicles. By nature of the task, however, the relevant data is difficult to collect and annotate. In this paper, we challenge cutting-edge generative models to automatically synthesize data for assessing reliability in semantic segmentation. By fine-tuning Stable Diffusion, we perform zero-shot generation of synthetic data in OOD domains or inpainted with OOD objects. Synthetic data is employed to provide an initial assessment of pretrained segmenters, thereby offering insights into their performance when confronted with real edge cases. Through extensive experiments, we demonstrate a high correlation between the performance on synthetic data and the performance on real OOD data, showing the validity approach. Furthermore, we illustrate how synthetic data can be utilized to enhance the calibration and OOD detection capabilities of segmenters.</p> <p><img src="/assets/img/publications/2024_genval/genval-overview.PNG" alt="genval_overview" height="100%" width="100%"/></p> <div class="caption"><b>Assessing 40 pretrained segmenters under covariate shifts.</b> Segmentation models under scrutiny were trained on Cityscapes train set only (in-domain data). They are evaluated on (i) Cityscapes validation set, (ii) real OOD data, and (iii) proposed synthetic data. We observe a strong correlation between results on (ii) and (iii). </div> <hr/> <h2 id="valeo4cast-a-modular-approach-to-end-to-end-forecasting">Valeo4Cast: A Modular Approach to End-to-End Forecasting</h2> <p class="page-description"><a href="https://sites.google.com/view/road-eccv2024/home">ECCV 2024 ROAD++ Workshop</a></p> <p class="page-description"><a href="https://www.argoverse.org/E2EForecasting.html">Winning solution in Argoverse 2 Unified Detection, Tracking and Forecasting Challenge</a></p> <h4 id="authors-yihong-xu-éloi-zablocki--alexandre-boulch-gilles-puy---mickaël-chen-florent-bartoccioni--nermin-samet---oriane-siméoni---spyros-gidaris---tuan-hung-vu-andrei-bursuc--eduardo-valle-renaud-marlet--matthieu-cord">Authors: <a href="https://scholar.google.fr/citations?user=vMLRRVkAAAAJ">Yihong Xu</a>, <a href="https://scholar.google.fr/citations?user=dOkbUmEAAAAJ">Éloi Zablocki</a>, <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="https://scholar.google.com/citations?user=QnRpMJAAAAAJ">Mickaël Chen</a>, <a href="https://f-barto.github.io/">Florent Bartoccioni</a>, <a href="https://nerminsamet.github.io/">Nermin Samet</a>, <a href="https://osimeoni.github.io/">Oriane Siméoni</a>, <a href="https://gidariss.github.io/">Spyros Gidaris</a>, <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://scholar.google.com/citations?user=lxWPqWAAAAAJ">Eduardo Valle</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2406.08113">Paper</a>] &nbsp;&nbsp; [<a href="https://eval.ai/web/challenges/challenge-page/2006/leaderboard/4752">leaderboard</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/valeo4cast/">page</a>]</h4> <p>Motion forecasting is crucial in autonomous driving systems to anticipate the future trajectories of surrounding agents such as pedestrians, vehicles, and traffic signals. In end-to-end forecasting, the model must jointly detect from sensor data (cameras or LiDARs) the position and past trajectories of the different elements of the scene and predict their future location. We depart from the current trend of tackling this task via end-to-end training from perception to forecasting and we use a modular approach instead. Following a recent study, we individually build and train detection, tracking, and forecasting modules. We then only use consecutive finetuning steps to integrate the modules better and alleviate compounding errors. Our study reveals that this simple yet effective approach significantly improves performance on the end-to-end forecasting benchmark. Consequently, our solution ranks first in the Argoverse 2 end-to-end Forecasting Challenge held at CVPR 2024 Workshop on Autonomous Driving (WAD), with 63.82 mAPf. We surpass forecasting results by +17.1 points over last year’s winner and by +13.3 points over this year’s runner-up. This remarkable performance in forecasting can be explained by our modular paradigm, which integrates finetuning strategies and significantly outperforms the end-to-end-trained counterparts.</p> <p><img src="/assets/img/publications/2024_valeo4cast/valeo4cast.PNG" alt="valeo4cast_overview" height="80%" width="80%"/></p> <div class="caption"><b>Valeo4Cast overview.</b> </div> <hr/> <h2 id="pafuse-part-based-diffusion-for-3d-whole-body-pose-estimation">PAFUSE: Part-based Diffusion for 3D Whole-Body Pose Estimation</h2> <p class="page-description"><a href="https://sites.google.com/view/t-cap-2024/home">ECCV 2024 Workshop Towards a Complete Analysis of People (T-CAP)</a></p> <h4 id="authors-nermin-samet--cédric-rommel--david-picard--eduardo-valle">Authors: <a href="https://nerminsamet.github.io/">Nermin Samet</a>    <a href="https://cedricrommel.github.io/">Cédric Rommel</a>    <a href="https://davidpicard.github.io/">David Picard</a>    <a href="https://eduardovalle.com/">Eduardo Valle</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2407.10220">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/PAFUSE">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/pafuse">Project page</a>]</h4> <p>We introduce a novel approach for 3D whole-body pose estimation, addressing the challenge of scale- and deformability- variance across body parts brought by the challenge of extending the 17 major joints on the human body to fine-grained keypoints on the face and hands. In addition to addressing the challenge of exploiting motion in unevenly sampled data, we combine stable diffusion to a hierarchical part representation which predicts the relative locations of fine-grained keypoints within each part (e.g., face) with respect to the part’s local reference frame. On the H3WB dataset, our method greatly outperforms the current state of the art, which fails to exploit the temporal information. We also show considerable improvements compared to other spatiotemporal 3D human-pose estimation approaches that fail to account for the body part specificities.</p> <p><img src="/assets/img/posts/2024_eccv/pafuse.PNG" alt="pafuse_overview" height="100%" width="100%"/></p> <div class="caption"><b>Overview of PAFUSE.</b> </div> <hr/> <h2 id="llm-wrapper-black-box-semantic-aware-adaptation-of-vision-language-foundation-models">LLM-wrapper: Black-Box Semantic-Aware Adaptation of Vision-Language Foundation Models</h2> <p class="page-description"><a href="https://sites.google.com/view/eval-fomo-24/home">ECCV 2024 Workshop on Emergent Visual Abilities and Limits of Foundation Models (Eval-FoMo)</a></p> <h4 id="authors-amaia-cardiel--éloi-zablocki--oriane-siméoni--elias-ramzi--matthieu-cord">Authors: Amaia Cardiel    <a href="https://scholar.google.fr/citations?user=dOkbUmEAAAAJ">Éloi Zablocki</a>    <a href="https://osimeoni.github.io/">Oriane Siméoni</a>    <a href="https://elias-ramzi.github.io/">Elias Ramzi</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2409.11919">Paper</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/llm_wrapper">Project page</a>]</h4> <p>Vision Language Models (VLMs) have shown impressive performances on numerous tasks but their zero-shot capabilities can be limited compared to dedicated or fine-tuned models. Yet, fine-tuning VLMs comes with strong limitations as it requires a ‘white-box’ access to the model’s architecture and weights while some recent models are proprietary (e.g., Grounding DINO 1.5). It also requires expertise to design the fine-tuning objectives and optimize the hyper-parameters, which are specific to each VLM and downstream task. In this work, we propose LLM-wrapper, a novel approach to adapt VLMs in a ‘black-box’ and semantic-aware manner by leveraging large language models (LLMs) so as to reason on their outputs.</p> <p><img src="/assets/img/publications/2024_llm_wrapper/llm_wrapper.PNG" alt="llm-wrapper_overview" height="80%" width="80%"/></p> <div class="caption"><b>Overview of LLM-Wrapper.</b> </div> <p>We demonstrate the effectiveness of LLM-wrapper on Referring Expression Comprehension (REC), a challenging open-vocabulary task that requires spatial and semantic reasoning. Our approach significantly boosts the performance of off-the-shelf models, yielding results that are on par or competitive when compared with classic VLM fine-tuning (cf ‘FT VLM’ in our main results). Despite a few failure cases due to the LLM ‘blindness’ (cf Qualitative results, bottom right)), LLM-wrapper shows better semantic, spatial and relational reasoning, as illustrated in our qualitative results below.</p> <p><img src="/assets/img/posts/2024_eccv/llm_wrapper_results.PNG" alt="llm-wrapper_results" height="90%" width="90%"/></p> <div class="caption"><b>LLM-Wrapper results.</b> </div> <hr/> <h2 id="regents-real-world-safety-critical-driving-scenario-generation-made-stable">ReGentS: Real-World Safety-Critical Driving Scenario Generation Made Stable</h2> <p class="page-description"><a href="https://coda-dataset.github.io/w-coda2024/">ECCV 2024 Workshop on Multimodal Perception and Comprehension of Corner Cases in Autonomous Driving (W-CODA)</a></p> <h4 id="authors-yuan-yin--pegah-khayatan--éloi-zablocki--alexandre-boulch--matthieu-cord">Authors: <a href="https://yuan-yin.github.io/">Yuan Yin</a>    <a href="https://pegah-kh.github.io/">Pegah Khayatan</a>    <a href="https://scholar.google.fr/citations?user=dOkbUmEAAAAJ">Éloi Zablocki</a>    <a href="https://www.boulch.eu/">Alexandre Boulch</a>    <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2409.07830">Paper</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/regents">Project page</a>]</h4> <p>Machine learning based autonomous driving systems often face challenges with safety-critical scenarios that are rare in real-world data, hindering their large-scale deployment. While increasing real-world training data coverage could address this issue, it is costly and dangerous. This work explores generating safety-critical driving scenarios by modifying complex real-world regular scenarios through trajectory optimization. We propose ReGentS, which stabilizes generated trajectories and introduces heuristics to avoid obvious collisions and optimization problems. Our approach addresses unrealistic diverging trajectories and unavoidable collision scenarios that are not useful for training robust planner. We also extend the scenario generation framework to handle real-world data with up to 32 agents. Additionally, by using a differentiable simulator, our approach simplifies gradient descent-based optimization involving a simulator, paving the way for future advancements.</p> <p><img src="/assets/img/publications/2024_regents/regents_page.png" alt="regents_overview" height="100%" width="100%"/></p> <hr/> <h2 id="the-bravo-semantic-segmentation-challenge-results-in-uncv2024">The BRAVO Semantic Segmentation Challenge Results in UNCV2024</h2> <p class="page-description"><a href="https://uncertainty-cv.github.io/2024/challenge/">ECCV 2024 Workshop on Uncertainty Quantification for Computer Vision</a></p> <h4 id="authors-tuan-hung-vu---eduardo-valle--andrei-bursuc--tommie-kerssies--daan-de-geus--gijs-dubbelman--long-qian--bingke-zhu--yingying-chen--ming-tang--jinqiao-wang--tomáš-vojíř--jan-šochman--jiří-matas--michael-smith--frank-ferrie--shamik-basu--christos-sakaridis--luc-van-gool">Authors: <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>    <a href="https://eduardovalle.com/">Eduardo Valle</a>    <a href="https://abursuc.github.io/">Andrei Bursuc</a>    <a href="">Tommie Kerssies</a>    <a href="">Daan de Geus</a>    <a href="">Gijs Dubbelman</a>    <a href="">Long Qian</a>    <a href="">Bingke Zhu</a>    <a href="">Yingying Chen</a>    <a href="">Ming Tang</a>    <a href="">Jinqiao Wang</a>    <a href="">Tomáš Vojíř</a>    <a href="">Jan Šochman</a>    <a href="">Jiří Matas</a>    <a href="">Michael Smith</a>    <a href="">Frank Ferrie</a>    <a href="">Shamik Basu</a>    <a href="">Christos Sakaridis</a>    <a href="">Luc Van Gool</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2409.15107">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/bravo_challenge">Code</a>] &nbsp;&nbsp; [<a href="https://benchmarks.elsa-ai.eu/?ch=1&amp;com=introduction">Project page</a>]</h4> <p>We propose the unified BRAVO challenge to benchmark the reliability of semantic segmentation models under realistic perturbations and unknown out-of-distribution (OOD) scenarios. We define two categories of reliability: (1) semantic reliability, which reflects the model’s accuracy and calibration when exposed to various perturbations; and (2) OOD reliability, which measures the model’s ability to detect object classes that are unknown during training. The challenge attracted nearly 100 submissions from international teams representing notable research institutions. The results reveal interesting insights into the importance of large-scale pre-training and minimal architectural design in developing robust and reliable semantic segmentation models.</p> <p><img src="/assets/img/publications/2024_bravo/bravo-overview.PNG" alt="bravo_overview" height="80%" width="80%"/></p> <div class="caption"><b>All submissions.</b> Aggregated metrics (out-of-distribution and semantic) on axes, ranking metric (BRAVO Index) on level set. More freedom on the training dataset (Task 2, in orange) did not translate into better results. </div>]]></content><author><name></name></author><category term="3d-perception"/><category term="multi-sensor"/><category term="limited-supervision"/><category term="reliability"/><category term="motion-forecasting"/><category term="robustness"/><category term="generalization"/><category term="driving"/><summary type="html"><![CDATA[The European Conference on Computer Vision (ECCV) is a biennial landmark conference for the increasingly large community of researchers in computer vision and machine learning from both academia and industry. At the 2024 edition the valeo.ai team will present five papers in the main conference and four in the workshops. We are also organizing two tutorials (Bayesian Odyssey and Time is precious: Self-Supervised Learning Beyond Images), the Uncertainty Quantification for Computer Vision workshop, a talk at the OmniLabel workshop, and the BRAVO challenge. Our team has also contributed to the reviewing process, with seven reviewers, three area chairs, and two outstanding reviewer awards. The team will be at ECCV to present these works and will be happy to discuss more about these projects and ideas, and share our exciting ongoing research.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2024_eccv/eccv_banner.png"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2024_eccv/eccv_banner.png" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at CVPR 2024</title><link href="https://valeoai.github.io//posts/2024-06-13-valeoai-at-cvpr-2024/" rel="alternate" type="text/html" title="valeo.ai at CVPR 2024"/><published>2024-06-13T00:00:00+00:00</published><updated>2024-06-13T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-cvpr-2024</id><content type="html" xml:base="https://valeoai.github.io//posts/2024-06-13-valeoai-at-cvpr-2024/"><![CDATA[<p>The <a href="https://cvpr.thecvf.com/">IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR)</a> is a key event for researchers and engineers working on computer vision and machine learning. At the 2024 edition the <a href="https://ptrckprz.github.io/valeoai/">valeo.ai</a> team will present eight <a href="https://valeoai.github.io/publications/">papers</a> in the main conference, two <a href="https://valeoai.github.io/publications/">papers</a> in workshops, and one workshop <a href="https://opendrivelab.com/cvpr2024/workshop/">keynote</a>. Also, the team will present its <a href="https://valeoai.github.io/publications/valeo4cast/">winning solution</a> to the Argoverse 2 <a href="https://www.argoverse.org/E2EForecasting.html">“Unified Detection, Tracking and Forecasting”</a> challenge held at the Workshop on Autonomous Driving. The team will be at CVPR to present these works and will be happy to discuss more about these projects and ideas, and share our exciting ongoing research. We outline our team papers below.</p> <p><img src="/assets/img/posts/2024_cvpr/valeoai_cvpr.jfif" alt="valeo.ai team at CVPR 2024" height="100%" width="100%"/></p> <h2 id="three-pillars-improving-vision-foundation-model-distillation-for-lidar">Three Pillars Improving Vision Foundation Model Distillation for Lidar</h2> <h4 id="authors-gilles-puy-spyros-gidaris-alexandre-boulch-oriane-siméoni--corentin-sautier--andrei-bursuc--patrick-pérez-renaud-marlet">Authors: <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="https://gidariss.github.io/&amp;hl=en">Spyros Gidaris</a>, <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="https://osimeoni.github.io/">Oriane Siméoni</a>, <a href="https://csautier.github.io/">Corentin Sautier</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://ptrckprz.github.io/">Patrick Pérez</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2310.17504">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/ScalR">Code</a>] &nbsp;&nbsp; [<a href="https://youtu.be/yksj5WuJY4I">Video</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/scalr/">Project page</a>]</h4> <p>Self-supervised image backbones can be used to address complex 2D tasks (e.g., semantic segmentation, object discovery) very efficiently and with little or no downstream supervision. Ideally, 3D backbones for lidar should be able to inherit these properties after distillation of these powerful 2D features. The most recent methods for image-to-lidar distillation on autonomous driving data show promising results, obtained thanks to distillation methods that keep improving. Yet, we still notice a large performance gap when measuring by linear probing the quality of distilled vs fully supervised features.</p> <p>In this work, instead of focusing only on the distillation method, we study the effect of three pillars for distillation: the 3D backbone, the pretrained 2D backbone, and the pretraining 2D+3D dataset. In particular, thanks to our scalable distillation method named ScaLR, we show that scaling the 2D and 3D backbones and pretraining on diverse datasets leads to a substantial improvement of the feature quality. This allows us to significantly reduce the gap between the quality of distilled and fully-supervised 3D features, and to improve the robustness of the pretrained backbones to domain gaps and perturbations. We show that scaling the 2D and 3D backbones, and pretraining on diverse datasets leads to considerable improvements of the feature quality. The role of these pillars is actually more important than the distillation method itself, which we simplify for easier scaling.</p> <p><img src="/assets/img/posts/2024_cvpr/scalr_overview.PNG" alt="scalr_overview" height="100%" width="100%"/></p> <div class="caption"><b>ScaLR image-to-lidar distillation method with the three pillars studied in this work.</b> </div> <p>In this work, after proposing and studying a scalable distillation method, which we call ScaLR for Scalable Lidar Representation (see Figure above), we make the following contributions.</p> <p>First, we are able to significantly reduce the gap between distilled and supervised lidar representations: on nuScenes, we increase the performance by 22.8 mIoU percentage points compared to the former best distillation method.</p> <p>Second, we show it is possible to pretrain a single backbone on a mixture of datasets, performing similarly or better than separate backbones specialized on each dataset individually. The capacity of this backbone in providing good features across multiple datasets is illustrated in the figure below. For each scene in this figure, we pick a point located on a car and present the feature correlation map with respect to this point. We notice that the most correlated points also belong to cars on all datasets, illustrating the capacity of our single pretrained backbone to correctly distinguish objects on multiple datasets.</p> <p><img src="/assets/img/posts/2024_cvpr/scalr_results.PNG" alt="scalr_results" height="100%" width="100%"/></p> <div class="caption"><b>Correlation maps with a point located on a car</b> on four different scenes extracted from nuScenes, SemanticKITTI, PandaSet-64 and PandaSet-GT, respectively. The features used to compute these maps are extracted from a single pretrained backbone on all four datasets with ScaLR. Color goes from blue to red for low and high values. </div> <p>Third, we thoroughly study the properties of our distilled features. We show that they are robust to both domain gaps and perturbations. We also show that pretraining on diverse datasets improves robustness.</p> <p>Finally, we show that a possible way to get even better features is to distill the knowledge from multiple vision foundation models at the same time, which can be easily done with our scalable distillation strategy</p> <hr/> <h2 id="pointbev-a-sparse-approach-to-bev-predictions">PointBeV: A Sparse Approach to BeV Predictions</h2> <h4 id="authors-loïck-chambon-éloi-zablocki-mickaël-chen-florent-bartoccioni--patrick-pérez---matthieu-cord">Authors: <a href="https://loickch.github.io/">Loïck Chambon</a>, <a href="https://scholar.google.fr/citations?user=dOkbUmEAAAAJ">Éloi Zablocki</a>, <a href="https://scholar.google.com/citations?user=QnRpMJAAAAAJ">Mickaël Chen</a>, <a href="https://f-barto.github.io/">Florent Bartoccioni</a>, <a href="https://ptrckprz.github.io/">Patrick Pérez</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2312.00703">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/pointbev">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/pointbev/">Project page</a>]</h4> <p>Bird’s-eye View (BeV) representations have emerged as the de-facto shared space in driving applications, offering a unified space for sensor data fusion and supporting various downstream tasks. However, conventional models use grids with fixed resolution and range and face computational inefficiencies due to the uniform allocation of resources across all cells. To address this, we propose PointBeV, a novel sparse BeV segmentation model operating on sparse BeV cells instead of dense grids. This approach offers precise control over memory usage, enabling the use of long temporal contexts and accommodating memory-constrained platforms. PointBeV employs an efficient two-pass strategy for training, enabling focused computation on regions of interest. At inference time, it can be used with various memory/performance trade-offs and flexibly adjusts to new specific use cases.</p> <p><img src="/assets/img/publications/2024_pointbev/pointbev.PNG" alt="pointbev_overview" height="100%" width="100%"/></p> <div class="caption"><b>PointBeV overview.</b> As a sparse method, PointBeV is trained using local predictions, only for sampled 2D points provided as inputs. The points of interest are lifted to form 3D pillars, with each 3D point pulling visual features. To achieve this, PointBeV incorporates an efficient feature extraction process through a Sparse Feature Pulling module, illustrated in the ‘efficient feature extraction’ block. The obtained 3D BeV features are then flattened onto the 2D BeV plane and processed using a sparse U-Net with task-dependent final heads, generating local BeV predictions. For training, we only need sparse signals. At test time, points that have not been sampled are set to zero. </div> <p>PointBeV achieves state-of-the-art results on the nuScenes dataset for vehicle, pedestrian, and lane segmentation, showcasing superior performance in static and temporal settings despite being trained solely with sparse signals. We will release our code along with two new efficient modules used in the architecture: Sparse Feature Pulling, designed for the effective extraction of features from images to BeV, and Submanifold Attention, which enables efficient temporal modeling.</p> <p><img src="/assets/img/posts/2024_cvpr/pointbev_results.PNG" alt="pointbev_results" height="50%" width="50%"/></p> <div class="caption"><b>BeV vehicle IoU vs. memory footprint on nuScenes.</b> The size of a dot represents the number of BeV points being evaluated, the smaller the better. PointBeV has the capacity to explore various trade-offs between efficiency and performance by varying the number of points being considered. The remaining points are considered as zeros in the final prediction. Using PointBeV we can achieve state-of-the-art performance with only a small portion of the points and without losing performance. The memory consumption is calculated using a 40GB A100 GPU. </div> <hr/> <h2 id="dont-drop-your-samples-coherence-aware-training-benefits-conditional-diffusion">Don’t drop your samples! Coherence-aware training benefits Conditional diffusion</h2> <h3 id="highlight">Highlight</h3> <h4 id="authors--nicolas-dufour---victor-besnier-vicky-kalogeiton-david-picard">Authors: <a href="https://nicolas-dufour.github.io/"> Nicolas Dufour </a>, <a href="https://scholar.google.com/citations?hl=fr&amp;user=n_C2h-QAAAAJ">Victor Besnier</a>, <a href="https://vicky.kalogeiton.info/">Vicky Kalogeiton</a>, <a href="https://davidpicard.github.io/">David Picard</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2405.20324">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/nicolas-dufour/CAD">Code</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=4Tu-x2-Zcxs">Video</a>] &nbsp;&nbsp; [<a href="https://nicolas-dufour.github.io/cad.html">Project page</a>]</h4> <p>Conditional diffusion models are powerful generative models that can leverage various types of conditional information, such as class labels, segmentation masks, or text captions. However, in many real-world scenarios, conditional information may be noisy or unreliable due to human annotation errors or weak alignment. In this paper, we propose the Coherence-Aware Diffusion (CAD), a novel method that integrates coherence in conditional information into diffusion models, allowing them to learn from noisy annotations without discarding data. We assume that each data point has an associated coherence score that reflects the quality of the conditional information. We then condition the diffusion model on both the conditional information and the coherence score. In this way, the model learns to ignore or discount the conditioning when the coherence is low. We show that CAD is theoretically sound and empirically effective on various conditional generation tasks. Moreover, we show that leveraging coherence generates realistic and diverse samples that respect conditional information better than models trained on cleaned datasets where samples with low coherence have been discarded.</p> <p><img src="/assets/img/publications/2024_dont_drop/teaser.png" alt="dont_drop_overview" height="90%" width="90%"/></p> <div class="caption"><b>Overview of Don't Drop your Samples.</b> </div> <hr/> <h2 id="supervised-anomaly-detection-for-complex-industrial-images">Supervised Anomaly Detection for Complex Industrial Images</h2> <h4 id="authors--aimira-baitieva--david-hurych--victor-besnier--olivier-bernard">Authors: <a href=""> Aimira Baitieva </a>, <a href="https://scholar.google.com/citations?user=XY1PVwYAAAAJ&amp;hl=fr&amp;oi=ao">David Hurych</a>, <a href="https://scholar.google.com/citations?hl=fr&amp;user=n_C2h-QAAAAJ">Victor Besnier</a>, <a href="">Olivier Bernard</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2405.04953">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/abc-125/segad">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/segad/">Project page</a>]</h4> <p>Automating visual inspection in industrial production lines is essential for increasing product quality across various industries. Anomaly detection (AD) methods serve as robust tools for this purpose. However, existing public datasets primarily consist of images without anomalies, limiting the practical application of AD methods in production settings. To address this challenge, we present (1) the Valeo Anomaly Dataset (VAD), a novel real-world industrial dataset comprising 5000 images, including 2000 instances of challenging real defects across more than 20 subclasses. Acknowledging that traditional AD methods struggle with this dataset, we introduce (2) Segmentation based Anomaly Detector (SegAD). First, SegAD leverages anomaly maps as well as segmentation maps to compute local statistics. Next, SegAD uses these statistics and an optional supervised classifier score as input features for a Boosted Random Forest (BRF) classifier, yielding the final anomaly score. Our SegAD achieves state-of-the-art performance on both VAD (+2.1% AUROC) and the VisA dataset (+0.4% AUROC). The code and the models are publicly available</p> <p><img src="/assets/img/publications/2024_segad/teaser.png" alt="segad_overview" height="100%" width="100%"/></p> <div class="caption"><b>Overview of Supervised Anomaly Detection for Complex Industrial Images</b> </div> <hr/> <h2 id="a-simple-recipe-for-language-guided-domain-generalized-segmentation">A Simple Recipe for Language-guided Domain Generalized Segmentation</h2> <h4 id="authors-mohammad-fahes--tuan-hung-vu--andrei-bursuc--patrick-pérez--raoul-de-charette">Authors: <a href="https://mfahes.github.io/">Mohammad Fahes</a>, <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://ptrckprz.github.io/">Patrick Pérez</a>, <a href="https://team.inria.fr/rits/membres/raoul-de-charette/">Raoul de Charette</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2311.17922">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/astra-vision/FAMix">Code</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=vyjtvx2El9Q">Video</a>] &nbsp;&nbsp; [<a href="https://astra-vision.github.io/FAMix/">page</a>]</h4> <p>Generalization to new domains not seen during training is one of the long-standing goals and challenges in deploying neural networks in real-world applications. Existing generalization techniques necessitate substantial data augmentation, potentially sourced from external datasets, and aim at learning invariant representations by imposing various alignment constraints. Large-scale pretraining has recently shown promising generalization capabilities, along with the potential of bridging different modalities. For instance, the recent advent of vision-language models like CLIP has opened the doorway for vision models to exploit the textual modality. In this paper, we introduce a simple framework for generalizing semantic segmentation networks by employing language as the source of randomization. Our recipe comprises three key ingredients: i) the preservation of the intrinsic CLIP robustness through minimal fine-tuning, ii) language-driven local style augmentation, and iii) randomization by locally mixing the source and augmented styles during training. Extensive experiments report state-of-the-art results on various generalization benchmarks.</p> <p><img src="/assets/img/publications/2024_famix/famix-overview.png" alt="famix_overview" height="100%" width="100%"/></p> <div class="caption"><b>Overall process of FAMix.</b> FAMix consists of two steps. (Left) Local style mining consists of dividing the low-level feature activations into patches, which are used for style mining using Prompt-driven Instance Normalization (PIN). Specifically, for each patch, the dominant class is queried from the ground truth, and the mined style is added to corresponding class-specific style bank. (Right) Training the segmentation network is performed with minimal fine-tuning of the backbone. At each iteration, the low-level feature activations are viewed as grids of patches. For each patch, the dominant class is queried using the ground truth, then a style is sampled from the corresponding style bank. Style randomization is performed by normalizing each patch in the grid by its statistics, and transferring the new style which is a mixing between the original style and the sampled one. The network is trained using only a cross-entropy loss. </div> <p><img src="/assets/img/posts/2024_cvpr/famix_results.PNG" alt="famix_results" height="100%" width="100%"/></p> <div class="caption"><b>Qualitative results.</b> Columns 1-2: Image and ground truth (GT), Columns 3-4-5: Different domain generalization methods, Column 6: Our results. </div> <hr/> <h2 id="make-me-a-bnn-a-simple-strategy-for-estimating-bayesian-uncertainty-from-pre-trained-models">Make Me a BNN: A Simple Strategy for Estimating Bayesian Uncertainty from Pre-trained Models</h2> <h4 id="authors-gianni-franchi--olivier-laurent--maxence-leguéry--andrei-bursuc-andrea-pilzer--angela-yao">Authors: <a href="https://www.ensta-paris.fr/fr/gianni-franchi">Gianni Franchi</a>, <a href="https://scholar.google.com/citations?user=RW4CQ68AAAAJ">Olivier Laurent</a>, <a href="https://scholar.google.com/citations?user=RCUoocYAAAAJ&amp;hl=en">Maxence Leguéry</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://scholar.google.it/citations?user=zooORRsAAAAJ&amp;hl=en">Andrea Pilzer</a>, <a href="https://www.comp.nus.edu.sg/~ayao/">Angela Yao</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2312.15297">Paper</a>] &nbsp;&nbsp; [<a href="https://torch-uncertainty.github.io/">Code</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=aXqVBAOXc0o">Video</a>] &nbsp;&nbsp; [<a href="https://ensta-u2is-ai.github.io/ABNN-Make-me-a-BNN/">page</a>]</h4> <p>Deep Neural Networks (DNNs) are powerful tools for various computer vision tasks, yet they often struggle with reliable uncertainty quantification — a critical requirement for real-world applications. Bayesian Neural Networks (BNN) are equipped for uncertainty estimation but cannot scale to large DNNs where they are highly unstable to train. To address this challenge, we introduce the Adaptable Bayesian Neural Network (ABNN), a simple and scalable strategy to seamlessly transform DNNs into BNNs in a post-hoc manner with minimal computational and training overheads. ABNN preserves the main predictive properties of DNNs while enhancing their uncertainty quantification abilities through simple BNN adaptation layers (attached to normalization layers) and a few fine-tuning steps on pre-trained models. We conduct extensive experiments across multiple datasets for image classification and semantic segmentation tasks, and our results demonstrate that ABNN achieves state-of-the-art performance without the computational budget typically associated with ensemble methods.</p> <p><img src="/assets/img/posts/2024_cvpr/abnn_overview.PNG" alt="abnn_overview" height="100%" width="100%"/></p> <div class="caption"><b>Illustration of the training process for the ABNN.</b> The procedure begins with training a single DNN $\omega_{\text{MAP}}$, followed by architectural adjustments on the normalization layers to transform it into an ABNN. The final step involves fine-tuning the ABNN model. </div> <hr/> <h2 id="spot-self-training-with-patch-order-permutation-for-object-centric-learning-with-autoregressive-transformers">SPOT: Self-Training with Patch-Order Permutation for Object-Centric Learning with Autoregressive Transformers</h2> <h3 id="highlight-1">Highlight</h3> <h4 id="authors-ioannis-kakogeorgiou-spyros-gidaris-konstantinos-karantzalos---nikos-komodakis">Authors: <a href="https://scholar.google.com/citations?user=B_dKcz4AAAAJ">Ioannis Kakogeorgiou</a>, <a href="https://gidariss.github.io/&amp;hl=en">Spyros Gidaris</a>, <a href="http://users.ntua.gr/karank/">Konstantinos Karantzalos</a>, <a href="https://www.csd.uoc.gr/~komod/">Nikos Komodakis</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2312.00648">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/gkakogeorgiou/spot">Code</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/spot/">page</a>]</h4> <p>Unsupervised object-centric learning aims to decompose scenes into interpretable object entities, termed slots. Slot-based auto-encoders stand out as a prominent method for this task. Within them, crucial aspects include guiding the encoder to generate object-specific slots and ensuring the decoder utilizes them during reconstruction. This work introduces two novel techniques, (i) an attention-based self-training approach, which distills superior slot-based attention masks from the decoder to the encoder, enhancing object segmentation , and (ii) an innovative patch-order permutation strategy for autoregressive transformers that strengthens the role of slot vectors in reconstruction.</p> <p><img src="/assets/img/posts/2024_cvpr/spot_archi.PNG" alt="spot_archi" height="90%" width="90%"/></p> <div class="caption"> <b>Enhancing unsupervised object-centric learning via self-training.</b> Our two-stage approach starts with exclusive training in the initial stage (not depicted) using the reconstruction loss. In the following stage, shown here, a teacher-student framework is applied. The teacher model, trained in the first stage, guides the student model with an additional loss, distilling attention masks from the teacher’s decoder to the slot-attention masks in the student’s encoder. </div> <p>The effectiveness of these strategies is showcased experimentally. The combined approach significantly surpasses prior slot-based autoencoder methods in unsupervised object segmentation, especially with complex real-world images.</p> <p><img src="/assets/img/publications/2024_spot/spot_visualizations.png" alt="spot_overview" height="100%" width="100%"/></p> <div class="caption"><b>SPOT visualizations.</b> Our novel framework enhances unsupervised object-centric learning in slot-based autoencoders using self-training and sequence permutations in the transformer decoder. It improves object-specific slot generation, excelling in complex real-world images. </div> <hr/> <h2 id="nope-novel-object-pose-estimation-from-a-single-image">NOPE: Novel Object Pose Estimation from a Single Image</h2> <h4 id="authors-van-nguyen-nguyen-thibault-groueix--georgy-ponimatkin--yinlin-hu--renaud-marlet-mathieu-salzmann-vincent-lepetit">Authors: <a href="https://nv-nguyen.github.io/">Van Nguyen Nguyen</a>, <a href="https://imagine.enpc.fr/~groueixt/">Thibault Groueix</a>, <a href="https://scholar.google.co.kr/citations?hl=en&amp;user=5G-6ubcAAAAJ">Georgy Ponimatkin</a>, <a href="https://yinlinhu.github.io/">Yinlin Hu</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a>, <a href="https://people.epfl.ch/mathieu.salzmann">Mathieu Salzmann</a>, <a href="https://vincentlepetit.github.io/">Vincent Lepetit</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2303.13612">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/nv-nguyen/nope">Code</a>] &nbsp;&nbsp; [<a href="https://nv-nguyen.github.io/nope/">page</a>]</h4> <p>TL;DR: We introduce NOPE, a simple approach to estimate relative pose of unseen objects given only a single reference image. NOPE also predicts 3D pose distribution which can be used to address pose ambiguities due to symmetries.</p> <p>The practicality of 3D object pose estimation remains limited for many applications due to the need for prior knowledge of a 3D model and a training period for new objects. To address this limitation, we propose an approach that takes a single image of a new object as input and predicts the relative pose of this object in new images without prior knowledge of the object’s 3D model and without requiring training time for new objects and categories. We achieve this by training a model to directly predict discriminative embeddings for viewpoints surrounding the object. This prediction is done using a simple U-Net architecture with attention and conditioned on the desired pose, which yields extremely fast inference. We compare our approach to state-of-the-art methods and show it outperforms them both in terms of accuracy and robustness.</p> <p><img src="/assets/img/publications/2024_nope/nope.gif" alt="nope_overview" height="55%" width="55%"/></p> <div class="caption"><b>NOPE qualitative results.</b> </div> <hr/> <h2 id="valeo4cast-a-modular-approach-to-end-to-end-forecasting">Valeo4Cast: A Modular Approach to End-to-End Forecasting</h2> <p class="page-description"><a href="https://www.argoverse.org/E2EForecasting.html">Winning solution in Argoverse 2 Unified Detection, Tracking and Forecasting Challenge, at CVPR WAD 2024</a></p> <h4 id="authors-yihong-xu-éloi-zablocki--alexandre-boulch-gilles-puy---mickaël-chen-florent-bartoccioni--nermin-samet---oriane-siméoni---spyros-gidaris---tuan-hung-vu-andrei-bursuc--eduardo-valle-renaud-marlet--matthieu-cord">Authors: <a href="https://scholar.google.fr/citations?user=vMLRRVkAAAAJ">Yihong Xu</a>, <a href="https://scholar.google.fr/citations?user=dOkbUmEAAAAJ">Éloi Zablocki</a>, <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="https://scholar.google.com/citations?user=QnRpMJAAAAAJ">Mickaël Chen</a>, <a href="https://f-barto.github.io/">Florent Bartoccioni</a>, <a href="https://nerminsamet.github.io/">Nermin Samet</a>, <a href="https://osimeoni.github.io/">Oriane Siméoni</a>, <a href="https://gidariss.github.io/">Spyros Gidaris</a>, <a href="https://tuanhungvu.github.io/">Tuan-Hung Vu</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://scholar.google.com/citations?user=lxWPqWAAAAAJ">Eduardo Valle</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2406.08113">Paper</a>] &nbsp;&nbsp; [<a href="https://eval.ai/web/challenges/challenge-page/2006/leaderboard/4752">leaderboard</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/valeo4cast/">page</a>]</h4> <p>Motion forecasting is crucial in autonomous driving systems to anticipate the future trajectories of surrounding agents such as pedestrians, vehicles, and traffic signals. In end-to-end forecasting, the model must jointly detect from sensor data (cameras or LiDARs) the position and past trajectories of the different elements of the scene and predict their future location. We depart from the current trend of tackling this task via end-to-end training from perception to forecasting and we use a modular approach instead. Following a recent study, we individually build and train detection, tracking, and forecasting modules. We then only use consecutive finetuning steps to integrate the modules better and alleviate compounding errors. Our study reveals that this simple yet effective approach significantly improves performance on the end-to-end forecasting benchmark. Consequently, our solution ranks first in the Argoverse 2 end-to-end Forecasting Challenge held at CVPR 2024 Workshop on Autonomous Driving (WAD), with 63.82 mAPf. We surpass forecasting results by +17.1 points over last year’s winner and by +13.3 points over this year’s runner-up. This remarkable performance in forecasting can be explained by our modular paradigm, which integrates finetuning strategies and significantly outperforms the end-to-end-trained counterparts.</p> <p><img src="/assets/img/publications/2024_valeo4cast/valeo4cast.PNG" alt="valeo4cast_overview" height="100%" width="100%"/></p> <div class="caption"><b>Valeo4Cast overview.</b> </div> <hr/> <h2 id="occfeat-self-supervised-occupancy-feature-prediction-for-pretraining-bev-segmentation-networks">OccFeat: Self-supervised Occupancy Feature Prediction for Pretraining BEV Segmentation Networks</h2> <p class="page-description"><a href="https://cvpr2024.wad.vision/">CVPR 2024 Workshop on Autonomous Driving (WAD)</a></p> <h4 id="authors-sophia-sirko-galouchenko-alexandre-boulch--spyros-gidaris--andrei-bursuc---antonin-vobecky---renaud-marlet--patrick-pérez">Authors: Sophia Sirko-Galouchenko, <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="https://gidariss.github.io/&amp;hl=en">Spyros Gidaris</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://vobecant.github.io/">Antonin Vobecky</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a>, <a href="https://ptrckprz.github.io/">Patrick Pérez</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2404.14027">Paper</a>] &nbsp;&nbsp; [<a href="https://valeoai.github.io/publications/occfeat/">page</a>]</h4> <p>We introduce a self-supervised pretraining method, called OccFeat, for camera-only Bird’s-Eye-View (BEV) segmentation networks. With OccFeat, we pretrain a BEV network via occupancy prediction and feature distillation tasks. Occupancy prediction provides a 3D geometric understanding of the scene to the model. However, the geometry learned is class-agnostic. Hence, we add semantic information to the model in the 3D space through distillation from a self-supervised pretrained image foundation model. Models pretrained with our method exhibit improved BEV semantic segmentation performance, particularly in low-data scenarios. Moreover, empirical results affirm the efficacy of integrating feature distillation with 3D occupancy prediction in our pretraining approach.</p> <p><img src="/assets/img/publications/2024_occfeat/occfeat_teaser.png" alt="occfeat_overview" height="100%" width="100%"/></p> <div class="caption"><b>Overview of OccFeat’s self-supervised BEV pretraining approach.</b> OccFeat attaches an auxiliary pretraining head on top of the BEV network. This head “unsplats” the BEV features to a 3D feature volume and predicts with it (a) the 3D occupancy of the scene (occupancy reconstruction loss) and (b) high-level self-supervised image features characterizing the occupied voxels (occupancy-guided distillation loss). The occupancy targets are produced by “voxelizing” Lidar points, while the self-supervised image foundation model DINOv2 provides the feature targets for the occupied voxels. The pretraining head is removed after the pretraining. </div> <p>The results show the benefit of our pretraining method, especially in low-shot regimes, e.g., when using annotations only for 1% or 10% of nuScene’s training data. Additionally, our OccFeat pretraining improves the robustness, as evaluated on the nuScenes-C benchmark.</p> <p><img src="/assets/img/posts/2024_cvpr/occfeat_results.PNG" alt="occfeat_results" height="100%" width="100%"/></p> <div class="caption"> Performance comparison in low data regime 1% annotated data of nuScenes (Left). Study on robustness. Segmentation results on nuScenes-C dataset for Vehicle classes using BEVFormer network with EN-B0 image backbone on 100% annotated data. Comparison of our OccFeat against no BEV pretraining (Right). </div> <hr/> <h2 id="what-makes-multimodal-in-context-learning-work">What Makes Multimodal In-Context Learning Work?</h2> <p class="page-description"><a href="https://prompting-in-vision.github.io/index_cvpr24.html">CVPR 2024 Workshop on Prompting in Vision</a></p> <h4 id="authors-folco-bertini-baldassini--mustafa-shukor-matthieu-cord-laure-soulier--benjamin-piwowarski">Authors: <a href="https://www.folbaeni.com/">Folco Bertini Baldassini</a>, <a href="https://scholar.google.com/citations?user=lhp9mRgAAAAJ&amp;hl=en">Mustafa Shukor</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a>, <a href="https://pages.isir.upmc.fr/soulier/">Laure Soulier</a>, <a href="https://www.piwowarski.fr/">Benjamin Piwowarski</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2404.15736">Paper</a>] &nbsp;&nbsp; [<a href="https://gitlab.com/folbaeni/multimodal-icl">code</a>] &nbsp;&nbsp; [<a href="https://multimodal-icl-folbaeni-1988a753e0abbbc71bb3967331bb69edafda92e.gitlab.io/">page</a>]</h4> <p>Large Language Models have demonstrated remarkable performance across various tasks, exhibiting the capacity to swiftly acquire new skills, such as through In-Context Learning (ICL) with minimal demonstration examples. In this work, we present a comprehensive framework for investigating Multimodal ICL (M-ICL) in the context of Large Multimodal Models. We consider the best open-source multimodal models (e.g., IDEFICS, OpenFlamingo) and a wide range of multimodal tasks. Our study unveils several noteworthy findings: (1) M-ICL primarily relies on text-driven mechanisms, showing little to no influence from the image modality. (2) When used with advanced-ICL strategy (like RICES), M-ICL is not better than a simple strategy based on majority voting over context examples. Moreover, we identify several biases and limitations of M-ICL that warrant consideration prior to deployment.</p> <p><img src="/assets/img/publications/2024_multimodal_icl/multimodal-icl.PNG" alt="icl_overview" height="60%" width="60%"/></p> <div class="caption"><b>Empirical analysis of Multimodal In-Context Learning (M-ICL) behavior.</b> </div>]]></content><author><name></name></author><category term="3d-perception"/><category term="multi-sensor"/><category term="limited-supervision"/><category term="reliability"/><category term="motion-forecasting"/><category term="driving"/><summary type="html"><![CDATA[The IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR) is a key event for researchers and engineers working on computer vision and machine learning. At the 2024 edition the valeo.ai team will present eight papers in the main conference, two papers in workshops, and one workshop keynote. Also, the team will present its winning solution to the Argoverse 2 “Unified Detection, Tracking and Forecasting” challenge held at the Workshop on Autonomous Driving. The team will be at CVPR to present these works and will be happy to discuss more about these projects and ideas, and share our exciting ongoing research. We outline our team papers below.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2024_cvpr/cvpr_banner.PNG"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2024_cvpr/cvpr_banner.PNG" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at NeurIPS 2023</title><link href="https://valeoai.github.io//posts/neurips-2023" rel="alternate" type="text/html" title="valeo.ai at NeurIPS 2023"/><published>2023-12-08T00:00:00+00:00</published><updated>2023-12-08T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-neurips-2023</id><content type="html" xml:base="https://valeoai.github.io//posts/neurips-2023"><![CDATA[<p>The <a href="https://neurips.cc/">Neural Information Processing Systems Conference (NeurIPS)</a> is a major inter-disciplinary event that brings together researchers and practicioners in machine learning, computer vision, natural language processing, optimization, statistics, but also neuroscience, natural sciences, social sciences, etc. This year, at the thirty-seventh edition of NeurIPS, the <a href="../">valeo.ai</a> team will present 4 papers in the main conference and 1 in the workshops.</p> <p>Notably, we explore perception via different sensors, e.g., audio, on the path towards increasingly autonomous systems. We also study the interaction between different sensing modalities (images, language, Lidar point clouds) and advance a tri-modal self-supervised learning algorithm for 3D semantic voxel occupancy prediction from a rig of cameras mounted on a vehicle. We further show how to obtain robust deep models starting from pre-trained foundation models finetuned with reinforcement learning from human feedback. Finally, we analyze different generative models (diffusion models, GANs) and advance a unification framework considering them as instances of Particle Models.</p> <p>We will be happy to discuss more about these projects and ideas, and share our exciting ongoing research. Take a quick view of our papers below and come meet us at the posters or catch us for a coffee in the hallways.</p> <h2 id="resilient-multiple-choice-learning-a-learned-scoring-scheme-with-application-to-audio-scene-analysis">Resilient Multiple Choice Learning: A learned scoring scheme with application to audio scene analysis</h2> <h4 id="authors-victor-letzelter-mathieu-fontaine-mickaël-chen-patrick-pérez-slim-essid-gaël-richard">Authors: Victor Letzelter, Mathieu Fontaine, Mickaël Chen, Patrick Pérez, Slim Essid, Gaël Richard</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2311.01052">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/Victorletzelter/code-rMCL">Code</a>] &nbsp;&nbsp; [<a href="../publications/rmcl/">Project page</a>]</h4> <p>In this work, we tackle ambiguous machine learning tasks, where single predictions don’t suffice due to the task’s nature or inherent uncertainties. We introduce a robust multi-hypotheses framework that is capable of deterministically offering a range of plausible predictions at inference time. Our experiments on both synthetic data and real-world audio data affirm the potential and versatility of our method. Check out the paper and the code for more details.</p> <p><img src="/assets/img/posts/2023_neurips/training_dynamics.gif" alt="rmcl_overview" height="100%" width="100%"/></p> <p>This problem involves estimating a conditional distribution that is dependent on the input. The accompanying animation illustrates the early stages in the evolution of our model’s learning process, highlighting how it progressively refines its predictions (represented by shaded blue points) to the actual data distribution (indicated by green points), which varies with the input ‘t’.</p> <hr/> <h2 id="pop-3d-open-vocabulary-3d-occupancy-prediction-from-images">POP-3D: Open-Vocabulary 3D Occupancy Prediction from Images</h2> <h4 id="authors-antonin-vobecky-oriane-siméoni-david-hurych-spyros-gidaris-andrei-bursuc-patrick-pérez-josef-sivic">Authors: Antonin Vobecky, Oriane Siméoni, David Hurych, Spyros Gidaris, Andrei Bursuc, Patrick Pérez, Josef Sivic</h4> <h4 align="center"> [<a href="https://openreview.net/forum?id=eBXM62SqKY">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/vobecant/POP3D">Code</a>] &nbsp;&nbsp; [<a href="https://vobecant.github.io/POP3D">Project page</a>]</h4> <p>POP-3D is an approach to predict open-vocabulary 3D semantic voxel occupancy map from input 2D images to enable 3D grounding, segmentation, and retrieval of free-form language queries.</p> <p><img src="/assets/img/posts/2023_neurips/pop3d-overview.png" alt="pop3d_overview" height="100%" width="100%"/></p> <div class="caption">Given surround-view images on the input, our POP-3D outputs voxel occupancy with 3D-language features, which one can query using text, e.g., to obtain zero-shot semantic segmentation. </div> <p>We design a new model architecture for open-vocabulary 3D semantic occupancy prediction. The architecture consists of a 2D-3D encoder together with occupancy prediction and 3D-language heads. The output is a dense voxel map of 3D grounded language embeddings enabling a range of open-vocabulary tasks. Next, we develop a tri-modal self-supervised learning algorithm that leverages three modalities: (i) images, (ii) language, and (iii) LiDAR point clouds and enables training the proposed architecture using a strong pre-trained vision-language model without the need for any 3D manual language annotations.</p> <p><img src="/assets/img/posts/2023_neurips/pop3d-model.png" alt="pop3d_model" height="100%" width="100%"/></p> <div class="caption">Overview of POP-3D architecture and training approach.</div> <p>Finally, we demonstrate the strengths of the proposed model quantitatively on several open-vocabulary tasks: Zero-shot 3D semantic segmentation using existing datasets; 3D grounding, and retrieval of free-form language queries, using a small dataset that we propose as an extension of nuScenes.</p> <p><img src="/assets/img/posts/2023_neurips/pop3d-qualitative.png" alt="pop3d_example" height="100%" width="100%"/></p> <hr/> <h2 id="rewarded-soups-towards-pareto-optimal-alignment-by-interpolating-weights-fine-tuned-on-diverse-rewards">Rewarded soups: towards Pareto-optimal alignment by interpolating weights fine-tuned on diverse rewards</h2> <h4 id="authors-alexandre-ramé-guillaume-couairon-mustafa-shukor-corentin-dancette-jean-baptiste-gaya-laure-soulier-matthieu-cord">Authors: Alexandre Ramé, Guillaume Couairon, Mustafa Shukor, Corentin Dancette, Jean-Baptiste Gaya, Laure Soulier, Matthieu Cord</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2306.04488">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/alexrame/rewardedsoups">Code</a>] &nbsp;&nbsp; [<a href="https://huggingface.co/spaces/alexrame/rewardedsoups">Project page</a>]</h4> <p>Foundation models are first pre-trained on vast unsupervised datasets and then fine-tuned on labeled data. Reinforcement learning, notably from human feedback (RLHF), can further align the network with the intended usage. Yet the imperfections in the proxy reward may hinder the training and lead to suboptimal results; the diversity of objectives in real-world tasks and human opinions exacerbate the issue. This paper proposes embracing the heterogeneity of diverse rewards by following a multi-policy strategy. Rather than focusing on a single a priori reward, we aim for Pareto-optimal generalization across the entire space of preferences. To this end, we propose rewarded soup, first specializing multiple networks independently (one for each proxy reward) and then interpolating their weights linearly. This succeeds empirically because we show that the weights remain linearly connected when fine-tuned on diverse rewards from a shared pre-trained initialization. We demonstrate the effectiveness of our approach for text-to-text (summarization, Q&amp;A, helpful assistant, review), text-image (image captioning, text-to-image generation, visual grounding, VQA), and control (locomotion) tasks. We hope to enhance the alignment of deep models, and how they interact with the world in all its diversity.</p> <p><img src="/assets/img/posts/2023_neurips/rewarded-soups.png" alt="rs_overview" height="100%" width="100%"/></p> <div class="caption"><b>Illustration of the different steps of our proposed rewarded soup (RS).</b> After unsupervised pre-training and supervised fine-tuning, we launch $N$ independent RL fine-tunings on the proxy rewards $\{R_i\}^{N}_{i=1}$. Then we combine the trained networks by interpolation in the weight space. The final weights are adapted at test time by selecting the coefficient $\lambda$.</div> <hr/> <h2 id="unifying-gans-and-score-based-diffusion-as-generative-particle-models">Unifying GANs and Score-Based Diffusion as Generative Particle Models</h2> <h4 id="authors-jean-yves-franceschi-mike-gartrell-ludovic-dos-santos-thibaut-issenhuth-emmanuel-de-bézenac-mickaël-chen-alain-rakotomamonjy">Authors: Jean-Yves Franceschi, Mike Gartrell, Ludovic Dos Santos, Thibaut Issenhuth, Emmanuel de Bézenac, Mickaël Chen, Alain Rakotomamonjy</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2305.16150">Paper</a>] &nbsp;&nbsp; [Code (coming soon)]</h4> <p>By describing the trajectories of GAN outputs during training with particle evolution equations, we propose an unifying framework for GAN and Diffusion Models. We provide a new insights on the role of the generator network, and as proof of concept validating our theories, we propose methods to train a generator with score-based gradient instead of a discriminator, or to use a discriminator’s gradient flow to generate instead of training a generator.</p> <p><img src="/assets/img/posts/2023_neurips/unify-gan.png" alt="unigan_overview" height="70%" width="70%"/></p> <hr/> <h2 id="evaluating-the-structure-of-cognitive-tasks-with-transfer-learning">Evaluating the structure of cognitive tasks with transfer learning</h2> <p class="page-description"><a href="https://ai4sciencecommunity.github.io/neurips23.html">NeurIPS Workshop on AI for Scientific Discovery: From Theory to Practice</a></p> <h4 id="authors-bruno-aristimunha-raphael-y-de-camargo-walter-h-lopez-pinaya-sylvain-chevallier-alexandre-gramfort-cedric-rommel">Authors: Bruno Aristimunha, Raphael Y. de Camargo, Walter H. Lopez Pinaya, Sylvain Chevallier, Alexandre Gramfort, Cedric Rommel</h4> <h4 align="center"> [<a href="https://cedricrommel.github.io/assets/pdfs/NeurIPS_2023_AI_for_Science_Workshop.pdf">Paper</a>] &nbsp;&nbsp; [Code (coming soon)]</h4> <p>Electroencephalography (EEG) decoding is a challenging task due to the limited availability of labeled data. While transfer learning is a promising technique to address this challenge, it assumes that transferable data domains and tasks are known, which is not the case in this setting. This work investigates the transferability of deep learning representations between different EEG decoding tasks.</p> <p><img src="/assets/img/posts/2023_neurips/eval-cog-tasks.png" alt="cog_overview" height="90%" width="90%"/></p> <div class="caption"><b>Learned transferability maps for both datasets.</b> Each node corresponds to a distinct cognitive task. Arrow width represents the average transfer performance when using the representations learned from a source task to decode a target task.</div> <p>We conduct extensive experiments using state-of-the-art decoding models on two recently released EEG datasets, ERPCore and M3CV, containing over 140 subjects and 11 distinct cognitive tasks.</p> <p>From an EEG processing perspective, our results can be used to leverage related datasets for alleviating EEG data scarcity with transfer learning. We show that even with a linear probing transfer method, we are able to boost by up to 28% the performance of some tasks. From a neuroscientific standpoint, our transfer maps provide insights into the hierarchical relations between cognitive tasks, hence enhancing our understanding of how these tasks are connected. We discover for example evidence that certain decoding paradigms elicit specific and narrow brain activities, while others benefit from pre-training on a broad range of representations.</p>]]></content><author><name></name></author><category term="multi-sensor"/><category term="limited-supervision"/><category term="reliability"/><category term="deep-learning"/><summary type="html"><![CDATA[The Neural Information Processing Systems Conference (NeurIPS) is a major inter-disciplinary event that brings together researchers and practicioners in machine learning, computer vision, natural language processing, optimization, statistics, but also neuroscience, natural sciences, social sciences, etc. This year, at the thirty-seventh edition of NeurIPS, the valeo.ai team will present 4 papers in the main conference and 1 in the workshops.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2023_neurips/logo_neurips.svg"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2023_neurips/logo_neurips.svg" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at ICCV 2023</title><link href="https://valeoai.github.io//posts/iccv-2023" rel="alternate" type="text/html" title="valeo.ai at ICCV 2023"/><published>2023-09-26T00:00:00+00:00</published><updated>2023-09-26T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-iccv-2023</id><content type="html" xml:base="https://valeoai.github.io//posts/iccv-2023"><![CDATA[<p>The <a href="https://iccv2023.thecvf.com/">IEEE / CVF International Conference on Computer Vision (ICCV)</a> is a landmark event for the increasingly large and diverse community of researchers in computer vision and machine learning. This year, ICCV takes place in Paris, home of the <a href="../">valeo.ai</a> team. From interns to senior researchers, the valeo.ai team will participate in mass at ICCV and will be looking forward to welcoming you and talking about the exciting progress and ideas in the field.</p> <p>At ICCV 2023 we will present 5 papers in the main conference and 3 in the workshops. We are also organizing 2 tutorials with 2 challenges (<a href="https://valeoai.github.io/bravo/">BRAVO</a> and <a href="https://uncv2023.github.io/">UNCV</a>) and a tutorial (<a href="https://abursuc.github.io/many-faces-reliability/">Many Faces of Reliability</a>). Take a quick view of our papers in the conference and come meet us at the posters, at our booth or in the hallway.</p> <h2 id="using-a-waffle-iron-for-automotive-point-cloud-semantic-segmentation">Using a Waffle Iron for Automotive Point Cloud Semantic Segmentation</h2> <h4 id="authors-gilles-puy-alexandre-boulch-renaud-marlet">Authors: Gilles Puy, Alexandre Boulch, Renaud Marlet</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2301.10100">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/WaffleIron">Code</a>] &nbsp;&nbsp; [<a href="../publications/waffleiron/">Project page</a>]</h4> <p>Semantic segmentation of point clouds delivered by lidars permits autonomous vehicles to make sense of their 3D surrounding environment. Sparse convolutions have become a de-facto tool to process these large outdoor point clouds. The top performing methods on public benchmarks, such SemanticKITTI or nuScenes, all leverage sparse convolutions. Nevertheless, despite their undeniable success and efficiency, these convolutions remain available in a limited number of deep learning frameworks and hardware platforms. In this work, we propose an alternative backbone built with tools broadly available (such as 2D and 1D convolutions) but that still reaches the level of performance of the top methods on automotive datasets.</p> <p>We propose a point-based backbone, called WaffleIron, which is essentially built using standard MLPs and dense 2D convolutions, both readily available in all deep learning frameworks thanks to their wide use in the field of computer vision. The architecture of this backbone is illustrated in the figure below. It is inspired by the recent MLP-Mixer. It takes as input a point cloud with a token associated to each point. All these point tokens are then updated by a sequence of layers, each containing a token-mixing step (made of dense 2D convolutions) and a channel-mixing step (made of a MLP shared across points).</p> <p><img src="/assets/img/posts/2023_iccv/waffleiron.png" alt="waffle_overview" height="70%" width="70%"/></p> <div class="caption">The WaffleIron backbone takes as input point tokens, provided by an embedding layer (not represented), and updates these point representations L times via a point token-mixing layer (containing the WI block) followed by a channel-mixing layer. The WI block consists of a 2D projection along one of the main axes, a feed-forward network (FFN) with two dense channel-wise 2D convolutions with a ReLU activation in the hidden layer, and a simple copy of the 2D features to the 3D points. The channel-mixing layer contains a batch-norm, a MLP shared across each point, and a residual connection. The WaffleIron backbone is free of any point downsampling or upsampling layer, farthest point sampling, nearest neighbor search, or sparse convolution. </div> <p>WaffleIron has three main hyperparameters to tune: the depth L, the width F and the resolution of the 2D grid. We show that these parameters are easy to tune: the performance increases with the network width F and depth L, until an eventual saturation; we observe stable results over a wide range of values for the resolution of the 2D grid.</p> <p>In our paper, we also provide many details on how to train WaffleIron to reach the performance of top-entries on two autonomous driving benchmarks: SemanticKITTI and nuScenes.</p> <hr/> <h2 id="pøda-prompt-driven-zero-shot-domain-adaptation">PØDA: Prompt-driven Zero-shot Domain Adaptation</h2> <h4 id="authors-mohammad-fahes-tuan-hung-vu-andrei-bursuc-patrick-pérez-raoul-de-charette">Authors: Mohammad Fahes, Tuan-Hung Vu, Andrei Bursuc, Patrick Pérez, Raoul de Charette</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2212.03241">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/astra-vision/PODA">Code</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=kataxQoPuSE">Video</a>] &nbsp;&nbsp; [<a href="https://astra-vision.github.io/PODA/">Project page</a>]</h4> <p>Domain adaptation has been vastly investigated in computer vision but still requires access to target images at train time, which might be intractable in some uncommon conditions. In this paper, we propose the task of ‘Prompt-driven Zero-shot Domain Adaptation’, where we adapt a model trained on a source domain using only a general description in natural language of the target domain, i.e., a prompt. First, we leverage a pre-trained contrastive vision-language model (CLIP) to optimize affine transformations of source features, steering them towards the target text embedding while preserving their content and semantics. To achieve this, we propose Prompt-driven Instance Normalization (PIN). Second, we show that these prompt-driven augmentations can be used to perform zero-shot domain adaptation for semantic segmentation. Experiments demonstrate that our method significantly outperforms CLIP-based style transfer baselines on several datasets for the downstream task at hand, even surpassing one-shot unsupervised domain adaptation. A similar boost is observed on object detection and image classification</p> <p><img src="/assets/img/posts/2023_iccv/poda.png" alt="poda_overview" height="80%" width="80%"/></p> <div class="caption">We perform zero-shot adaptation with natural language prompts. PØDA enables the adaptation of a segmenter model (here, DeepLabv3+ trained on the source dataset Cityscapes) to unseen conditions with only a prompt. Source-only predictions are shown as smaller segmentation masks to the left or right of the test images. </div> <hr/> <h2 id="you-never-get-a-second-chance-to-make-a-good-first-impression-seeding-active-learning-for-3d-semantic-segmentation">You Never Get a Second Chance To Make a Good First Impression: Seeding Active Learning for 3D Semantic Segmentation</h2> <h4 id="authors-nermin-samet-oriane-siméoni-gilles-puy-georgy-ponimatkin-renaud-marlet-vincent-lepetit">Authors: Nermin Samet, Oriane Siméoni, Gilles Puy, Georgy Ponimatkin, Renaud Marlet, Vincent Lepetit</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2304.11762">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/nerminsamet/seedal">Code</a>]</h4> <p>We are interested in the efficient annotation of sparse 3D point clouds (as captured indoors by depth cameras or outdoors by automotive lidars) for semantic segmentation. Active Learning (AL) iteratively selects relevant data fractions to annotate within a given budget, but requires a first fraction of the dataset (a ’seed’) to be already annotated to estimate the benefit of annotating other data fractions. We show that the choice of the seed can significantly affect the performance of many AL methods and propose a method, named SeedAL, for automatically constructing a seed that will ensure good performance for AL. Assuming that images of the point clouds are available, which is common, our method relies on powerful unsupervised image features to measure the diversity of the point clouds. It selects the point clouds for the seed by optimizing the diversity under an annotation budget, which can be done by solving a linear optimization problem. Our experiments demonstrate the effectiveness of our approach compared to random seeding and existing methods on both the S3DIS and SemanticKitti datasets.</p> <p><img src="/assets/img/posts/2023_iccv/seedal.png" alt="seedal_overview" height="70%" width="70%"/></p> <div class="caption"><b>Impact of active learning seed on performance. </b>We show the variability of results obtained with 20 different random seeds (blue dashed lines), within an initial annotation budget of 3% of the dataset, when using various active learning methods for 3D semantic segmentation of S3DIS. We compare it to the result obtained with our seed selection strategy (solid red line), named SeedAL, which performs better or on par with the best (lucky) random seeds among 20, and “protects” from very bad (unlucky) random seeds.</div> <hr/> <h2 id="ep-alm-efficient-perceptual-augmentation-of-language-models">eP-ALM: Efficient Perceptual Augmentation of Language Models</h2> <h4 id="authors-mustafa-shukor-corentin-dancette-matthieu-cord">Authors: Mustafa Shukor, Corentin Dancette, Matthieu Cord</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2303.11403">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/mshukor/eP-ALM">Code</a>] &nbsp;&nbsp; [<a href="https://mshukor.github.io/eP-ALM.github.io/">Project page</a>]</h4> <p>eP-ALM aims to augment large language models (LLMs) with perception. While most existing approaches train a large number of parameters and rely on extensive multimodal pre-training, we investigate the minimal computational effort required to adapt unimodal models to multimodal tasks. We show that by freezing more than 99% of total parameters, training only one linear projection layer and prepending only one trainable token, our approach (dubbed eP-ALM) significantly outperforms other baselines on VQA and captioning for image, video and audio modalities.</p> <p><img src="/assets/img/posts/2023_iccv/ep-alm.png" alt="epalm_overview" height="70%" width="70%"/></p> <div class="caption"><b>Illustration of the adaptation mechanism in eP-ALM.</b> The perceptual input (image/video/audio) is fed to the perceptual encoder E (e.g., ViT) and the corresponding text to the LM (e.g., OPT), which then generates a text conditioned on the perceptual input. The multimodal interaction is done via the [CLS] tokens acting as Perceptual Prompt, and are extracted from the last layers of the encoder, then injected in the last layers of LM, after passing by the Linear Connection C. The previous [CLS] token is replaced by the new one coming from a deeper layer, keeping the number of tokens fixed. The first layers (grayed) of each model are kept intact without any modality interaction. We ease the adaptation with a Soft Prompt that is prepended to the input of LM. </div> <hr/> <h2 id="zero-shot-spatial-layout-conditioning-for-text-to-image-diffusion-models">Zero-shot spatial layout conditioning for text-to-image diffusion models</h2> <h4 id="authors-guillaume-couairon-marlène-careil-matthieu-cord-stéphane-lathuilière-jakob-verbeek">Authors: Guillaume Couairon, Marlène Careil, Matthieu Cord, Stéphane Lathuilière, Jakob Verbeek</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2306.13754">Paper</a>]</h4> <p>Large-scale text-to-image diffusion models have considerably improved the state of the art in generative image modeling, and provide an intuitive and powerful user interface to drive the image generation process. In this paper, we propose ZestGuide, a “zero-shot” segmentation guidance approach that can be integrated into pre-trained text-image diffusion models, and requires no additional training. It exploits the implicit segmentation maps that can be extracted from cross-attention layers, and uses them to align generation with input masks.</p> <p><img src="/assets/img/posts/2023_iccv/zest-guide.png" alt="zest_overview" height="70%" width="70%"/></p> <div class="caption">ZestGuide generates images conditioned on segmentation maps with corresponding free-form textual descriptions. </div> <hr/> <h2 id="diffhpe-robust-coherent-3d-human-pose-lifting-with-diffusion">DiffHPE: Robust, Coherent 3D Human Pose Lifting with Diffusion</h2> <p class="page-description"><a href="https://web.northeastern.edu/smilelab/amfg2023/">ICCV Workshop on Analysis and Modeling of Faces and Gestures</a></p> <h4 id="authors-cédric-rommel-eduardo-valle-mickaël-chen-souhaiel-khalfaoui-renaud-marlet-matthieu-cord-patrick-pérez">Authors: Cédric Rommel, Eduardo Valle, Mickaël Chen, Souhaiel Khalfaoui, Renaud Marlet, Matthieu Cord, Patrick Pérez</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2309.01575">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/diffhpe">Code</a>] &nbsp;&nbsp; [<a href="../publications/diffhpe">Project page</a>]</h4> <p>Diffusion models are making waves across various domains, including computer vision, natural language processing and time-series analysis. However, its application to purely predictive tasks, such as 3D human pose estimation (3D-HPE), remains largely unexplored. While a few pioneering works have shown promising performance metrics in 3D-HPE, the understanding of the benefits of diffusion models over classical supervision — as well as key design choices — is still in its infancy. In this work, we address those concerns, providing an in-depth analysis of the effects of diffusion models on 3D-HPE.</p> <p><img src="/assets/img/posts/2023_iccv/diffhpe.gif" alt="diffhpe_overview" height="100%" width="100%"/></p> <div class="caption">Poses across the learned reverse diffusion process converge to an accurate 3D reconstruction of the corresponding 2D pose in pixel space.</div> <p>More precisely, we propose DiffHPE, a novel strategy to use diffusion models in 3D-HPE, and show that combining diffusion with pre-trained supervised models allows to outperform both pure diffusion and pure supervised models trained separately. Our analysis demonstrates not only that the diffusion framework can be used to enhance accuracy, as previously understood, but also that it can improve robustness and coherence. Namely, our experiments showcase how poses estimated with diffusion models’ display better bilateral and temporal coherence, and are more robust to occlusions, even when not perfectly trained for the latter.</p> <hr/> <h2 id="challenges-of-using-real-world-sensory-inputs-for-motion-forecasting-in-autonomous-driving">Challenges of Using Real-World Sensory Inputs for Motion Forecasting in Autonomous Driving</h2> <p class="page-description"><a href="https://sites.google.com/view/road-plus-plus">ROAD++: The Second Workshop and Challenge on Event Detection for Situation Awareness in Autonomous Driving</a></p> <h4 id="authors-yihong-xu-loïck-chambon-éloi-zablocki-mickaël-chen-matthieu-cord-patrick-pérez">Authors: Yihong Xu, Loïck Chambon, Éloi Zablocki, Mickaël Chen, Matthieu Cord, Patrick Pérez</h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2306.09281">Paper</a>] &nbsp;&nbsp; [<a href="../publications/real-world-forecasting/">Project page</a>]</h4> <p>Motion forecasting is crucial in enabling autonomous vehicles to anticipate the future trajectories of surrounding agents. To do so, it requires solving mapping, detection, tracking, and then forecasting problems, in a multi-step pipeline. In this complex system, advances in conventional forecasting methods have been made using curated data, i.e., with the assumption of perfect maps, detection, and tracking. This paradigm, however, ignores any errors from upstream modules. Meanwhile, an emerging end-to-end paradigm, that tightly integrates the perception and forecasting architectures into joint training, promises to solve this issue. So far, however, the evaluation protocols between the two methods were incompatible and their comparison was not possible. In fact, and perhaps surprisingly, conventional forecasting methods are usually not trained nor tested in real-world pipelines (e.g., with upstream detection, tracking, and mapping modules). In this work, we aim to bring forecasting models closer to real-world deployment. First, we propose a unified evaluation pipeline for forecasting methods with real-world perception inputs, allowing us to compare the performance of conventional and end-to-end methods for the first time. Second, our in-depth study uncovers a substantial performance gap when transitioning from curated to perception-based data. In particular, we show that this gap (1) stems not only from differences in precision but also from the nature of imperfect inputs provided by perception modules, and that (2) is not trivially reduced by simply finetuning on perception outputs. Based on extensive experiments, we provide recommendations for critical areas that require improvement and guidance towards more robust motion forecasting in the real world. We will release an evaluation library to benchmark models under standardized and practical conditions.</p> <p><img src="/assets/img/posts/2023_iccv/e2e_forecasting.png" alt="forecast_overview" height="90%" width="90%"/></p> <div class="caption"><b>Study overview.</b> We study the challenges of deploying motion forecasting models into the real world when only predicted perception inputs are available. We compare: (1) (top) "conventional methods" (i.e., methods trained on curated input data) where (middle) we directly replace the curated inputs with real-world data, and (2) (bottom) "end-to-end methods" that are trained and used with perception modules. In the real-world setting, evaluation is challenging as the past tracks are estimated with arbitrary identities, making it difficult to establish a direct correspondence to GT identities. Therefore, we propose a matching process (purple) to assign predictions to GT and thus evaluate forecasting performances. Moreover, we study in depth the impact changing from curated data (green) to real-world (orange) mapping, or detection and tracking errors to motion forecasting. </div> <hr/> <h2 id="pop-3d-open-vocabulary-3d-occupancy-prediction-from-images">POP-3D: Open-Vocabulary 3D Occupancy Prediction from Images</h2> <p class="page-description"><a href="https://opensun3d.github.io/">ICCV 2023 Workshop on Open-Vocabulary 3D Scene Understanding (OpenSUN 3D)</a></p> <h4 id="authors-antonin-vobecky-oriane-siméoni-david-hurych-spyros-gidaris-andrei-bursuc-patrick-pérez-josef-sivic">Authors: Antonin Vobecky, Oriane Siméoni, David Hurych, Spyros Gidaris, Andrei Bursuc, Patrick Pérez, Josef Sivic</h4> <h4 align="center"> [<a href="https://data.ciirc.cvut.cz/public/projects/2023POP3D/resources/pop3d_paper.pdf">Paper</a>]</h4> <p>We propose an approach to predict a 3D semantic voxel occupancy map from input 2D images with features allowing 3D grounding, segmentation and retrieval of free-form language queries. To this end: We design a new architecture that consists of a 2D-3D encoder together with occupancy prediction and 3D-language heads; We develop a tri-modal self-supervised training that leverages three modalities – images, language and LiDAR point clouds– and enables learning the proposed architecture using a strong pre-trained vision-language model without the need for any 3D manual annotations. We quantitatively evaluate the proposed model on the task of zero-shot 3D semantic segmentation using existing datasets and show results on the tasks of 3D grounding and retrieval of free-form language queries.</p> <p><img src="/assets/img/posts/2023_iccv/pop3d.png" alt="forecast_overview" height="100%" width="100%"/></p> <div class="caption"><b>Method overview.</b>Given surround-view images, POP-3D produces a voxel grid of text-aligned features that support open-vocabulary downstream tasks such as zero-shot occupancy segmentation or text-based grounding and retrieval. </div>]]></content><author><name></name></author><category term="3d-perception"/><category term="multi-sensor"/><category term="limited-supervision"/><category term="reliability"/><category term="domain-adaptation"/><summary type="html"><![CDATA[The IEEE / CVF International Conference on Computer Vision (ICCV) is a landmark event for the increasingly large and diverse community of researchers in computer vision and machine learning. This year, ICCV takes place in Paris, home of the valeo.ai team. From interns to senior researchers, the valeo.ai team will participate in mass at ICCV and will be looking forward to welcoming you and talking about the exciting progress and ideas in the field.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2023_iccv/iccv_logo.svg"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2023_iccv/iccv_logo.svg" xmlns:media="http://search.yahoo.com/mrss/"/></entry><entry><title type="html">valeo.ai at CVPR 2023</title><link href="https://valeoai.github.io//posts/cvpr-2023" rel="alternate" type="text/html" title="valeo.ai at CVPR 2023"/><published>2023-06-14T00:00:00+00:00</published><updated>2023-06-14T00:00:00+00:00</updated><id>https://valeoai.github.io//posts/valeoai-at-cvpr-2023</id><content type="html" xml:base="https://valeoai.github.io//posts/cvpr-2023"><![CDATA[<p>The <a href="https://cvpr2023.thecvf.com/">IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR)</a> is a key event for researchers and engineers working on computer vision and machine learning. At the 2023 edition the <a href="../">valeo.ai</a> team will present six <a href="../publications/">papers</a> in the main conference, one workshop <a href="https://vision4allseason.net/">keynote</a> and organize a <a href="https://osimeoni.github.io/object-localization-for-free/">tutorial</a>. The team will be at CVPR to present these works and will be happy to discuss more about these projects and ideas, and share our exciting ongoing research. We outline four of our team papers below.</p> <h2 id="octet-object-aware-counterfactual-explanations">OCTET: Object-aware Counterfactual Explanations</h2> <h4 id="authors-mehdi-zemni-mickaël-chen-éloi-zablocki-hédi-ben-younes-patrick-pérez-matthieu-cord">Authors: Mehdi Zemni, <a href="https://scholar.google.com/citations?user=QnRpMJAAAAAJ&amp;hl=fr&amp;oi=sra">Mickaël Chen</a>, <a href="https://scholar.google.com/citations?user=dOkbUmEAAAAJ&amp;hl=fr">Éloi Zablocki</a>, <a href="https://scholar.google.com/citations?hl=fr&amp;user=IFLcfvUAAAAJ">Hédi Ben-Younes</a>, <a href="https://ptrckprz.github.io/">Patrick Pérez</a>, <a href="https://cord.isir.upmc.fr/">Matthieu Cord</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2211.12380">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/octet">Code</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=Xfq0uRcw9jQ">Video</a>] &nbsp;&nbsp; [<a href="../publications/octet/">Project page</a>]</h4> <p>Nowadays, deep vision models are being widely deployed in safety-critical applications, e.g., autonomous driving, and explainability of such models is becoming a pressing concern. Among explanation methods, counterfactual explanations aim to find minimal and interpretable changes to the input image that would also change the output of the model to be explained. Such explanations point end-users at the main factors that impact the decision of the model. However, previous methods struggle to explain decision models trained on images with many objects, e.g., urban scenes, which are more difficult to work with but also arguably more critical to explain. In this work, we propose to tackle this issue with an object-centric framework for counterfactual explanation generation. Our method, inspired by recent generative modeling works, encodes the query image into a latent space that is structured in a way to ease object-level manipulations. Doing so, it provides the end-user with control over which search directions (e.g., spatial displacement of objects, style modification, etc.) are to be explored during the counterfactual generation. We conduct a set of experiments on counterfactual explanation benchmarks for driving scenes, and we show that our method can be adapted beyond classification, e.g., to explain semantic segmentation models. To complete our analysis, we design and run a user study that measures the usefulness of counterfactual explanations in understanding a decision model.</p> <p><img src="/assets/img/posts/2023_cvpr/octet.png" alt="octet_overview" height="80%" width="80%"/></p> <div class="caption"><b>Counterfactual explanations generated by OCTET.</b> Given a classifier that predicts whether or not it is possible to go left, and a query image (top left), OCTET produces a counterfactual explanation where the most influential features that led to the decision are changed (top right). On the bottom row, we show that OCTET can also operate under different settings that result in different focused explanations. We report the prediction made by the decision model at the top left of each image. </div> <hr/> <h2 id="also-automotive-lidar-self-supervision-by-occupancy-estimation">ALSO: Automotive Lidar Self-supervision by Occupancy estimation</h2> <h4 id="authors-alexandre-boulch-corentin-sautier-björn-michele-gilles-puy-renaud-marlet">Authors: <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="https://scholar.google.com/citations?user=xYDkHEsAAAAJ">Corentin Sautier</a>, <a href="https://scholar.google.com/citations?user=xQcKnXkAAAAJ&amp;hl=en">Björn Michele</a>, <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2212.05867">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/ALSO">Code</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=GGIBKlMvphw">Video</a>] &nbsp;&nbsp; [<a href="../publications/also/">Project page</a>]</h4> <p>We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds. The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled, and to use the underlying latent vectors as input to the perception head. The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information that can be used to boost an actual perception task. This principle has a very simple formulation, which makes it both easy to implement and widely applicable to a large range of 3D sensors and deep networks performing semantic segmentation or object detection. In fact, it supports a single-stream pipeline, as opposed to most contrastive learning approaches, allowing training on limited resources. We conducted extensive experiments on various autonomous driving datasets, involving very different kinds of lidars, for both semantic segmentation and object detection. The results show the effectiveness of our method to learn useful representations without any annotation, compared to existing approaches.</p> <p><img src="/assets/img/posts/2023_cvpr/also.png" alt="also_overview" height="100%" width="100%"/></p> <div class="caption"><b>ALSO overview.</b> The backbone to pre-train produces latent vectors for each input point. At pre-training time, the latent vector are fed into an volumetric occupancy head that classifies query points as full or empty. At semantic training or test time, the same latent vectors are fed into a semantic head, e.g., for semantic segmentation or object detection. </div> <hr/> <h2 id="unsupervised-object-localization-observing-the-background-to-discover-objects">Unsupervised Object Localization: Observing the Background to Discover Objects</h2> <h4 id="authors-oriane-siméoni-chloé-sekkat-gilles-puy-antonin-vobecky-éloi-zablocki-patrick-pérez">Authors: <a href="https://osimeoni.github.io/">Oriane Siméoni</a>, <a href="https://github.com/chloeskt">Chloé Sekkat</a>, <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="https://vobecant.github.io/">Antonin Vobecky</a>, <a href="https://scholar.google.com/citations?user=dOkbUmEAAAAJ&amp;hl=fr">Éloi Zablocki</a>, <a href="https://ptrckprz.github.io/">Patrick Pérez</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2212.07834">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/FOUND">Code</a>] &nbsp;&nbsp; [<a href="https://youtu.be/jfYQfFcrJBE">Video</a>] &nbsp;&nbsp; [<a href="../publications/found">Project page</a>]</h4> <p>Recent advances in self-supervised visual representation learning have paved the way for unsupervised methods tackling tasks such as object discovery and instance segmentation. However, discovering objects in an image with no supervision is a very hard task; what are the desired objects, when to separate them into parts, how many are there, and of what classes? The answers to these questions depend on the tasks and datasets of evaluation. In this work, we take a different approach and propose to look for the background instead. This way, the salient objects emerge as a by-product without any strong assumption on what an object should be. We propose FOUND, a simple model made of a single conv1 × 1 initialized with coarse background masks extracted from self-supervised patch-based representations. After fast training and refining these seed masks, the model reaches state-of-the-art results on unsupervised saliency detection and object discovery benchmarks. Moreover, we show that our approach yields good results in the unsupervised semantic segmentation retrieval task.</p> <p><img src="/assets/img/posts/2023_cvpr/found.png" alt="found_overview" height="65%" width="65%"/></p> <div class="caption"><b>Overview of FOUND. </b>In the first stage (green upperpart), a background mask is discovered by mining a seed patch through a reweighting of the self-attention maps of a frozen DINO self-supervised features. This seed is then used to find similar patches likely belonging to the background. In the second stage (blue lower part), we train a lightweight 1 × 1 convolutional layer that produces refined masks from the self-supervised features. It is trained in a self-supervised fashion to predict both smoothed inverse coarse masks of the first step, and smoothed binarized version of its own output. Blue arrows denote where the gradients flow (in the reverse direction).</div> <hr/> <h2 id="rangevit-towards-vision-transformers-for-3d-semantic-segmentation-in-autonomous-driving">RangeViT: Towards Vision Transformers for 3D Semantic Segmentation in Autonomous Driving</h2> <h4 id="authors-angelika-ando-spyros-gidaris-andrei-bursuc-gilles-puy-alexandre-boulch-renaud-marlet">Authors: Angelika Ando, <a href="https://gidariss.github.io/&amp;hl=en">Spyros Gidaris</a>, <a href="https://abursuc.github.io/">Andrei Bursuc</a>, <a href="https://sites.google.com/site/puygilles/home">Gilles Puy</a>, <a href="https://www.boulch.eu/">Alexandre Boulch</a>, <a href="http://imagine.enpc.fr/~marletr/">Renaud Marlet</a></h4> <h4 align="center"> [<a href="https://arxiv.org/abs/2301.10222">Paper</a>] &nbsp;&nbsp; [<a href="https://github.com/valeoai/rangevit">Code</a>] &nbsp;&nbsp; [<a href="https://www.youtube.com/watch?v=urd2ZIJ70WY">Video</a>] &nbsp;&nbsp; [<a href="../publications/rangevit/">Project page</a>]</h4> <p>Semantic segmentation of LiDAR point clouds permits vehicles to perceive their surrounding 3D environment independently of the lighting condition, providing useful information to build safe and reliable vehicles. A common approach to segment large scale LiDAR point clouds is to project the points on a 2D surface and then to use regular CNNs, originally designed for images, to process the projected point clouds. These projection-based methods usually benefit from fast computations and, when combined with techniques which use other point cloud representations, achieve state-of-the-art results.</p> <p>Today, projection-based methods leverage 2D CNNs but recent advances in computer vision show that vision transformers (ViTs) have achieved state-of-the-art results in many image-based benchmarks. Despite the absence of almost any domain-specific inductive bias apart from the image tokenization process, ViTs have a strong representation learning capacity and achieve excellent results on various image perception tasks, such as image classification, object detection or semantic segmentation. Inspired by this success of ViTs for image understanding, in this work, we show that projection-based methods for 3D semantic segmentation can benefit from these latest improvements on ViTs when combined with three key ingredients, all described in our <a href="https://arxiv.org/abs/2301.10222">paper</a>.</p> <p><img src="/assets/img/posts/2023_cvpr/rangevit-teaser.png" alt="rangevit_teaser" height="75%" width="75%"/></p> <div class="caption"><b>Exploiting vision transformer (ViT) architectures and weights for LiDAR point cloud semantic segmentation.</b> We leverage the flexibility of transformer-based architectures to re-purpose them with minimal changes for processing sparse point clouds in autonomous driving tasks. The common ViT backbone across modalities allows to effectively transfer weights pre-trained on large image repositories towards improving point cloud segmentation performance with fine-tuning. </div> <p>We answer positively but only after combining them with three key ingredients: (a) ViTs are notoriously hard to train and require a lot of training data to learn powerful representations. By preserving the same backbone architecture as for RGB images, we can exploit the knowledge from long training on large image collections that are much cheaper to acquire and annotate than point clouds. We reach our best results with pre-trained ViTs on large image datasets. (b) We compensate for ViTs’ lack of inductive bias by substituting a tailored non-linear convolutional stem for the classical linear embedding layer. (c) We refine pixel-wise predictions with a convolutional decoder and a skip connection from the convolutional stem to combine low-level but fine-grained features of the convolutional stem with the high-level but coarse predictions of the ViT encoder. With these ingredients, we show that our method, called RangeViT, outperforms prior projection-based methods on nuScenes and SemanticKITTI.</p> <p><img src="/assets/img/posts/2023_cvpr/rangevit-overview.png" alt="rangevit_overview" height="100%" width="100%"/></p> <div class="caption"><b> Overview of RangeViT architecture.</b> First, the point cloud is projected in a 2D space with range projection. Then, the produced range image is processed by the convolutional stem, the ViT encoder and the decoder to obtain a 2D feature map. It is then processed by a 3D refiner layer for 3D point-wise predictions. Note that there is a single skip connection between the convolutional stem and the decoder. </div> <p>In summary, our work offers the following contributions:</p> <ul> <li>Exploiting the powerful representation learning capacity of vision transformers for LiDAR semantic segmentation.</li> <li>Unifying the network architectures for processing LiDAR point clouds and images, enabling advancements in one domain to benefit both.</li> <li>Demonstrating the utilization of pre-trained ViTs on large-scale natural image datasets for LiDAR point cloud segmentation.</li> </ul> <p>We believe that this finding is highly intriguing. The RangeViT approach can leverage off-the-shelf pre-trained ViT models, enabling direct benefits from ongoing and future advances in training ViT models with natural RGB images - a rapidly growing research field.</p>]]></content><author><name></name></author><category term="3d-perception"/><category term="multi-sensor"/><category term="limited-supervision"/><category term="reliability"/><summary type="html"><![CDATA[The IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR) is a key event for researchers and engineers working on computer vision and machine learning. At the 2023 edition the valeo.ai team will present six papers in the main conference, one workshop keynote and organize a tutorial. The team will be at CVPR to present these works and will be happy to discuss more about these projects and ideas, and share our exciting ongoing research. We outline four of our team papers below.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://valeoai.github.io//assets/img/posts/2023_cvpr/cvpr_banner.svg"/><media:content medium="image" url="https://valeoai.github.io//assets/img/posts/2023_cvpr/cvpr_banner.svg" xmlns:media="http://search.yahoo.com/mrss/"/></entry></feed>