Links
- The Publications Warehouse does not have links to digital versions of this publication at this time
- Download citation as: RIS | Dublin Core
Abstract
The rise of neural networks has opened the door for automatic analysis of remote sensing data. A challenge to using this machinery for computational sustainability is the necessity of massive labeled data sets, which can be cost-prohibitive for many non-profit organizations. The primary motivation for this work is one such problem; the efficient management of invasive species -- invading flora and fauna that are estimated to cause damages in the billions of dollars annually. As an ongoing collaboration with the New York Natural Heritage Program, we consider the use of unsupervised deep learning techniques for dimensionality reduction of remote sensing images, which can reduce sample complexity for downstream tasks and decreases the need for large labeled data sets. We consider spatially augmenting contrastive learning by training neural networks to correctly classify two nearby patches of a landscape as such. We demonstrate that this approach improves upon previous methods and naive classification for a large-scale data set of remote sensing images derived from invasive species observations obtained over 30 years. Additionally, we simulate deployment in the field via active learning and evaluate this method on another important challenge in computational sustainability -- landcover classification -- and again find that it outperforms previous baselines.
Publication type | Article |
---|---|
Publication Subtype | Journal Article |
Title | Accelerating ecological sciences from above: Spatial contrastive learning for remote sensing |
Series title | Proceedings of the AAAI Conference on Artificial Intelligence |
Volume | 35 |
Issue | 17 |
Year Published | 2021 |
Language | English |
Publisher | Association for the Advancement of Artificial Intelligence |
Contributing office(s) | Coop Res Unit Leetown |
Description | 10 p. |
Google Analytic Metrics | Metrics page |