Phenomenal: a software framework for model-assisted analysis of high throughput plant phenotyping data
Plant high-throughput phenotyping aims at capturing the genetic variability of plant response to environmental factors for thousands of plants, hence identifying heritable traits for genomic selection and predicting the genetic values of allelic combinations in different environment. This first implies the automation of the measurement of a large number of traits to characterize plant growth, plant development and plant functioning. It also requires a fluent and versatile interaction between data and continuously evolving plant response models, that are essential in the analysis of the marker x environment interaction and in the integration of processes for predicting crop performance [1]. In the frame of the Phenome high throughput phenotyping infrastructure, we develop Phenomenal: a software framework dedicated to the analysis of high throughput phenotyping data and models. It is based on the OpenAlea platform [2] that provides methods and softwares for the modelling of plants, together with a user-friendly interface for the design and execution of scientific workflows. OpenAlea is also part of the InfraPhenoGrid infrastructure that allows high throughput computation and recording of provenance during the execution [3]. Figure 1: The 3D plant reconstruction and segmentation pipeline. Muti-view plants images from PhenoArch are binarised and used to reconstruct plants in3D. The 3D skeleton is extracted and separated into stem (central vertical elements) and leaves. 3D voxels are segmented by propagating skeleton segmentation.Phenomenal currently consists of 2D image analysis workflows built with standard image libraries (OpenCV, Scikit.Image), algorithms for 3D reconstruction (Figure 1), segmentation and tracking of plant organs for maize (under development), and workflows for estimation of light interception by plants during their growth. The ADEL model [4], a 3D FSPM model for maize architectural development, is used to help segmenting plant images and to automate the mapping between segmented 3D objects and plant organs defined in the model. The 3D reconstructed model is combined with meteorological data to feed a light distribution model [5] and estimate light use efficiency. In the future, we plan to scale this approach to the analysis of more complex plants (wheat, rice) and more complex images acquired in the field. We also want to automatize the connection between plant response models and the data extracted from images, to get directly an estimation of the crop performance in a large range of context.
Main Authors: | , , , , , , , , |
---|---|
Format: | conference_item biblioteca |
Language: | eng |
Published: |
IAMPS
|
Online Access: | http://agritrop.cirad.fr/590723/ http://agritrop.cirad.fr/590723/1/ID590723.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|