Neural Networks and Photogrammetry for Analysis of Marine Remote Sensing Data

Project Summary

KC and Patrick led two hands-on data workshops for ENVIRON 335: Drones in Marine Biology, Ecology, and Conservation. These labs were intended to introduce students to examples of how drones are currently being used as a remote sensing tool to monitor marine megafauna and their environments, and how machine learning can be used to efficiently analyze remote sensing datasets. The first lab specifically focused on how drones are being used to collect aerial images of whales to measure changes in body condition to help monitor populations. Students were introduced to the methods for making accurate measurements and then received an opportunity to measure whales themselves. The second lab then introduced analysis methods using computer vision and deep neural networks to detect, count, and measure objects of interest in remote sensing data. This work provided students in the environmental sciences an introduction to new techniques in machine learning and remote sensing that can be powerful multipliers of effort when analyzing large environmental datasets.

Themes and Categories
Year
2019

Graduate Students: KC Bierlich and Patrick Gray

Faculty: Dr. David Johnston

Undergraduate Course: "Drones in Marine Biology, Ecology, and Conservation" (ENVIRON 335)

Course Summary

Learning Objectives

Lab 1:

  • Understand how to accurately measure objects from aerial images depending on different camera/sensors and the altitude of the photo.
  • Understand why it is important to measure body condition and health of top predators, such as whales, in an ecosystem.
  • Hands-on activity to measure the total length of blue, humpback, and minke whales from California and Antarctica using drone imagery.
  • Compare results to classmates/class and discuss reasons for variation in measurements.

Lab 2:

  • Understand basic remote sensing technology, temporal, spectral, spatial, and radiometric characteristics of different sensors.
  • Learn some intuitive ideas about machine learning and applications in marine science
  • Build up a library of examples of how remote sensing and machine learning can be used for marine science.
  • Learn more specifically how deep learning functions and gain an understanding of the theory behind convolutional neural networks. Use this knowledge and the general remote sensing knowledge to develop a neural network model to classify animals and everyday objects.

The first lab consisted a short lecture on the theory of photogrammetry and its conservation importance and then students worked directly with the drone imagery using ImageJ for hands-on photogrammetry. The second lab session began with a short crash course in machine learning and a summary of applications in marine remote sensing. Students then had time to familiarize themselves with the datasets and the python tools we used in a guided format using Jupyter notebooks stored on Github and run on the Google Colaboratory environment. We then trained our neural network models as a group but with each student running it independently on Colab.

Artificial intelligence-based techniques are bringing incredible advances to difficult research problems, but these methods are often inaccessible to non-technical ecologists. Developing an intuitive understanding of neural networks, without the complicated setup and coding challenges, was an essential goal of this, with the hope that students will be empowered enough to bring these new techniques into their own research.

Datasets

We used three datasets during this project. Lab 1 used images of blue, humpback, and Antarctic minke whales collected by our research group in Antarctica and California. Lab 2’s initial Neural Network Development was done on the publicly available MINST and CIFAR datasets. Students used these to experiment with model design and then as a final component of the project we demonstrated how a neural network could be used to identify and measure cetaceans using the same dataset as Lab 1.

Course Materials

Drone images of whales that students measured can be found here.

Students also downloaded ImageJ, a photogrammetry software, that they used for measuring. All code from Lab 2 can be found here. An overview of both deep learning and remote sensing intended to be an hour long lecture can be found here.

Graphics

Equation for measuring whales
Figure 1) Equation and example of measuring whales using aerial photogrammetry.
Drone imagery examples
Figure 2) Example of drone imagery in raw form and after being analyzed by a convolutional neural network to identify the species and measure each animal.
CIFAR dataset images
Figure 3) CIFAR Dataset for neural network model development

 

Related Projects

A large and growing trove of patient, clinical, and organizational data is collected as a part of the “Help Desk” program at Durham’s Lincoln Community Health Center. Help Desk is a group of student volunteers who connect with patients over the phone and help them navigate to community resources (like food assistance programs, legal aid, or employment centers). Data-driven approaches to identifying service gaps, understanding the patient population, and uncovering unseen trends are important for improving patient health and advocating for the necessity of these resources. Disparities in food security, economic stability, education, neighborhood and physical environment, community and social context, and access to the healthcare system are crucial social determinants of health, which studies indicate account for nearly 70% of all health outcomes.

We led a 75-minute class session for the Marine Mammals course at the Duke University Marine Lab that introduced students to strengths and challenges of using aerial imagery to survey wildlife populations, and the growing use of machine learning to address these "big data" tasks.

Most phenomena that data scientists seek to analyze are either spatially or temporally correlated. Examples of spatial and temporal correlation include political elections, contaminant transfer, disease spread, housing market, and the weather. A question of interest is how to incorporate the spatial correlation information into modeling such phenomena.

 

In this project, we focus on the impact of environmental attributes (such as greenness, tree cover, temperature, etc.) along with other socio-demographics and home characteristics on housing prices by developing a model that takes into account the spatial autocorrelation of the response variable. To this aim, we introduce a test to diagnose spatial autocorrelation and explain how to integrate spatial autocorrelation into a regression model

 

 

In this data exploration, students are provided with data collected from remote sensing, census, and Zillow sources. Students are tasked with conducting a regression analysis of real-estate estimates against environmental amenities and other control variables which may or may not include the spatial autocorrelation information.