Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/179095
Title: Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks
Authors: Yang, Xiaofeng
Liu, Fayao
Lin, Guosheng
Keywords: Computer and Information Science
Issue Date: 2024
Source: Yang, X., Liu, F. & Lin, G. (2024). Neural radiance selector: find the best 2D representations of 3D data for CLIP based 3D tasks. Knowledge-Based Systems, 299, 112002-. https://dx.doi.org/10.1016/j.knosys.2024.112002
Project: MOE-T2EP 20220-0007
RG14/22
M21K3c0130
Journal: Knowledge-Based Systems
Abstract: Representing the world in 3D space provides vivid texture and depth information. However, 3D datasets currently do not match the scale of 2D datasets. There is a growing trend in representing 3D data as multi-view 2D images and using large-scale 2D models, to solve 3D tasks. In this work, we present the Neural Radiance Selector, a method that automatically selects the optimal 2D representations of 3D data. Instead of indiscriminately sampling multi-view 2D images, we define the optimal 2D views as those capable of reconstructing the entire 3D scene with a conditional neural radiance field. We propose two distinct methods for 3D point cloud data and 3D implicit models to achieve faster inference. We demonstrate the efficacy of our methods in various 3D tasks, including zero-shot 3D point cloud classification, 3D implicit model classification, and language-guided NeRF editing.
URI: https://hdl.handle.net/10356/179095
ISSN: 0950-7051
DOI: 10.1016/j.knosys.2024.112002
Schools: School of Computer Science and Engineering 
Rights: © 2024 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.
Fulltext Permission: none
Fulltext Availability: No Fulltext
Appears in Collections:SCSE Journal Articles

SCOPUSTM   
Citations 50

2
Updated on May 7, 2025

Page view(s)

94
Updated on May 6, 2025

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.