Paper published in a journal (Scientific congresses, symposiums and conference proceedings)
Category-level Meta-learned NeRF Priors for Efficient Object Mapping
EJAZ, Saad; BAVLE, Hriday; RIBEIRO, Laura et al.
2025In IEEE International Conference on Intelligent Robots and Systems
Peer Reviewed verified by ORBi
 

Files


Full Text
2503.01582v3.pdf
Author postprint (8.74 MB)
Download

All documents in ORBilu are protected by a user license.

Send to



Details



Keywords :
robotics; computer vision; object mapping; SLAM
Abstract :
[en] In 3D object mapping, category-level priors enable efficient object reconstruction and canonical pose estimation, requiring only a single prior per semantic category (e.g., chair, book, laptop, etc.). DeepSDF has been used predominantly as a category-level shape prior, but it struggles to reconstruct sharp geometry and is computationally expensive. In contrast, NeRFs capture fine details but have yet to be effectively integrated with category-level priors in a real-time multi-object mapping framework. To bridge this gap, we introduce PRENOM, a Prior-based Efficient Neural Object Mapper that integrates category-level priors with object-level NeRFs to enhance reconstruction efficiency and enable canonical object pose estimation. PRENOM gets to know objects on a first-name basis by meta-learning on synthetic reconstruction tasks generated from open-source shape datasets. To account for object category variations, it employs a multi-objective genetic algorithm to optimize the NeRF architecture for each category, balancing reconstruction quality and training time. Additionally, prior-based probabilistic ray sampling directs sampling toward expected object regions, accelerating convergence and improving reconstruction quality under constrained resources. Experimental results highlight the ability of PRENOM to achieve high-quality reconstructions while maintaining computational feasibility. Specifically, comparisons with prior-free NeRF-based approaches on a synthetic dataset show a 21% lower Chamfer distance. Furthermore, evaluations against other approaches using shape priors on a noisy real-world dataset indicate a 13% improvement averaged across all reconstruction metrics, and comparable pose and size estimation accuracy, while being trained for 5× less time. Code available at: https://github.com/snt-arg/PRENOM
Research center :
Interdisciplinary Centre for Security, Reliability and Trust (SnT) > ARG - Automation & Robotics
Disciplines :
Computer science
Author, co-author :
EJAZ, Saad  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > Automation
BAVLE, Hriday  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust > Automation > Team Holger VOOS
RIBEIRO, Laura  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > Automation
VOOS, Holger  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > Automation
SANCHEZ LOPEZ, Jose Luis  ;  University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > Automation
External co-authors :
no
Language :
English
Title :
Category-level Meta-learned NeRF Priors for Efficient Object Mapping
Publication date :
2025
Event name :
2025 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
Event place :
Hangzhou, China
Event date :
19/10/2025 - 25/10/2025
Audience :
International
Journal title :
IEEE International Conference on Intelligent Robots and Systems
ISSN :
2153-0858
Publisher :
Institute of Electrical and Electronics Engineers, New York, United States - New York
Peer reviewed :
Peer Reviewed verified by ORBi
FnR Project :
FNR17387634 - DEUS - Deep Understanding Of The Situation For Robots, 2022 (01/09/2023-31/08/2026) - Jose-luis Sanchez-lopez
Funding text :
This research was funded, in whole or in part, by the Luxembourg National Research Fund (FNR) under the DEUS Project (Ref. C22/IS/17387634/DEUS) and the MR-Cobot Project (Ref. 18883697/MR-Cobot)
Available on ORBilu :
since 03 December 2025

Statistics


Number of views
13 (9 by Unilu)
Number of downloads
8 (0 by Unilu)

OpenCitations
 
0
OpenAlex citations
 
0

Bibliography


Similar publications



Contact ORBilu