Abstract / Description of output
We present Visual-Language Fields (VL-Fields), a neural implicit spatial representation that enables open-vocabulary semantic queries. Our model encodes and fuses the geometry of a scene with vision-language trained latent features by distilling information from a language-driven segmentation model. VL-Fields is trained without requiring any prior knowledge of the scene object classes, which makes it a promising representation for the field of robotics. Our model outperformed the similar CLIP-Fields model in the task of semantic segmentation by almost 10%.
Original language | English |
---|---|
Title of host publication | Workshop on Representations, Abstractions, and Priors for Robot Learning Workshop at International Conference on Robotics and Automation 2023 |
Pages | 1-6 |
Number of pages | 6 |
Publication status | Published - 29 May 2023 |
Event | 2023 IEEE International Conference on Robotics and Automation - London, United Kingdom Duration: 29 May 2023 → 2 Jun 2023 https://www.icra2023.org |
Conference
Conference | 2023 IEEE International Conference on Robotics and Automation |
---|---|
Abbreviated title | ICRA 2023 |
Country/Territory | United Kingdom |
City | London |
Period | 29/05/23 → 2/06/23 |
Internet address |