Online Shape Modeling of Resident Space Objects Through Implicit Scene Understanding
Published: 27 November 2021
Abstract: Neural networks have become state-of-the-art computer vision tools for tasks that learn implicit representations of geometrical scenes. This paper proposes a two-part network architecture that exploits a view-synthesis network to understand a context scene and a graph convolutional network to generate a shape body model of an object within the field of view of a spacecraft’s optical navigation sensors. Once the first part of the network’s architecture understands the spacecraft’s environment, it can generate images from novel observations. The second part uses a multiview set of images to construct a 3D graph-based representation of the object. The proposed network pipeline produces shape models with accuracies that compete with state-of-the-art methods currently used for missions to small bodies. The network pipeline can be trained for multi-environment missions. Moreover, the onboard implementation may be more cost-effective than the current state-of-the-art.
Keywords: models / Neural / Shape Modeling / spacecraft / implicit / scenes / architecture / representations / graph / art
Scifeed alert for new publicationsNever miss any articles matching your research from any publisher
- Get alerts for new papers matching your research
- Find out the new papers from selected authors
- Updated daily for 49'000+ journals and 6000+ publishers
- Define your Scifeed now
Click here to see the statistics on "Journal of Aerospace Information Systems" .