Using interactive data visualization to make sense of large datasets

1 minute read

Published:

As sensor technology improves, data volumes grow. We now live in a sea of data collected by our phones, smartwatches, and home assistants like Alexa. Science is not any different, new sensors are enabling the collection of large datasets that can be mined for new scientific discoveries. In plant science, sensor technology is being applied to study how plants grow under drought conditions.


NOTE: Access the workshop notebook here.

Phenomics: A case study in big data

We will be using data collected by the Field Scanalyzer at the University of Arizona Maricopa Agricultural Center. The Field Scanalyzer covers over an hectare of land - capturing data from over 20,000 plants over a growing season. The Field scanalyzer is equipped with stereo RGB and thermal cameras, a PSII chlorophyll fluorescence imager, and a pair of 3D laser scanners (pictured below).

Collectively, these sensors capture 20 terabytes (TBs) in a three-month period, which makes converting these raw data into information a difficult task. Accomplishing extraction of information requires leveraging machine learning, high performance computers, and distributed computing.

These multiple sources of data provide a fine-scale information of plant growth under drought (decreased water) conditions. Today, we will use some of these data to learn interactive visualization using Python!


Workshop materials


Survey

Please provide your feedback to improve future workshops here: https://bit.ly/2022-ds2f.


Additional materials


Acknowledgements

This program is funded by the University of Arizona Libraries: https://data.library.arizona.edu/ds2f.

With special thanks to:

  • Jeffrey Oliver
  • Megan Senseney
  • Jim Martin
  • Yvonne Mery
  • Leslie Sult
  • Cheryl Casey