3D scene generation for zero-shot learning using ChatGPT guided language prompts

dc.contributor.authorAhmadi, Saharen
dc.contributor.authorCheraghian, Alien
dc.contributor.authorChowdhury, Townim Faisalen
dc.contributor.authorSaberi, Mortezaen
dc.contributor.authorRahman, Shafinen
dc.date.accessioned2025-05-23T07:25:33Z
dc.date.available2025-05-23T07:25:33Z
dc.date.issued2024en
dc.description.abstractZero-shot learning in the realm of 3D point cloud data remains relatively unexplored compared to its 2D image counterpart. This domain introduces fresh challenges due to the absence of robust pre-trained feature extraction models. To tackle this, we introduce a prompt-guided method for 3D scene generation and supervision, enhancing the network's ability to comprehend the intricate relationships between seen and unseen objects. Initially, we utilize basic prompts resembling scene annotations generated from one or two point cloud objects. Recognizing the limited diversity of basic prompts, we employ ChatGPT to expand them, enriching the contextual information within the descriptions. Subsequently, leveraging these descriptions, we arrange point cloud objects’ coordinates to fabricate augmented 3D scenes. Lastly, employing contrastive learning, we train our proposed architecture end-to-end, utilizing pairs of 3D scenes and prompt-based captions. We posit that 3D scenes facilitate more efficient object relationships than individual objects, as demonstrated by the effectiveness of language models like BERT in contextual understanding. Our prompt-guided scene generation method amalgamates data augmentation and prompt-based annotation, thereby enhancing 3D ZSL performance. We present ZSL and generalized ZSL results on both synthetic (ModelNet40, ModelNet10, and ShapeNet) and real-scanned (ScanOjbectNN) 3D object datasets. Furthermore, we challenge the model by training with synthetic data and testing with real-scanned data, achieving state-of-the-art performance compared to existing 2D and 3D ZSL methods in the literature. Codes and models are available at: https://github.com/saharahmadisohraviyeh/ChatGPT_ZSL_3D.en
dc.description.sponsorshipThis work was supported by the Conference Travel and Research Grants (CTRG) 2023\u20132024 from North South University, under Grant ID: CTRG-23-SEPS-20.en
dc.description.statusPeer-revieweden
dc.identifier.issn1077-3142en
dc.identifier.scopus85208261901en
dc.identifier.urihttp://www.scopus.com/inward/record.url?scp=85208261901&partnerID=8YFLogxKen
dc.identifier.urihttps://hdl.handle.net/1885/733751735
dc.language.isoenen
dc.rightsPublisher Copyright: © 2024 Elsevier Inc.en
dc.sourceComputer Vision and Image Understandingen
dc.subjectContrastive learningen
dc.subjectDeep learningen
dc.subjectPoint cloud objecten
dc.subjectZero-shot learningen
dc.title3D scene generation for zero-shot learning using ChatGPT guided language promptsen
dc.typeJournal articleen
dspace.entity.typePublicationen
local.contributor.affiliationCheraghian, Ali; School of Computing, ANU College of Systems and Society, The Australian National Universityen
local.contributor.affiliationChowdhury, Townim Faisal; University of Adelaideen
local.contributor.affiliationSaberi, Morteza; University of Technology Sydneyen
local.contributor.affiliationRahman, Shafin; North South Universityen
local.identifier.citationvolume249en
local.identifier.doi10.1016/j.cviu.2024.104211en
local.identifier.pure45b06ca5-0ab8-4d4f-8f84-955bbb527f6aen
local.identifier.urlhttps://www.scopus.com/pages/publications/85208261901en
local.type.statusPublisheden

Downloads