WP5 EO data processing modules implementation
WP5 is intended to complete the development of the RS-IUS component of the proposed EO Data for Habitat Monitoring (EODHaM) system.
A preliminary design of two-stage RS-IUS component of the EODHaM is reported as
Figure 2a and Figure 2b in Part B of the proposal. Some Italian sampling sites
and the Wales (Western-Europe) test site will be used as training data for the
improvement and development of RS-IUS modules based on the following
reasons: - Hyperspectral data are also available for some areas.
- Different land cover categories characterize the set of the selected
sampling areas, ranging from forest land to eroded soil, to wetlands. (See
Annex 2)
In order to achieve its objectives, WP5 will consist of the following tasks:
Task 5.1 RS_IUS first stage module improvement
This task is devoted to: - Enhancement of the RS-IUS
first-stage SRC module efficiency will be improved by providing source
to source code transformation (from the Interactive Data Language [IDL],
licensed by ITT Industries, Inc., to the c++ programming language) and
enlarge the SRC applicability domain by guaranteeing software
maintenance and upgrading, if any.SRC automatically provides as output
spectral-based semi-concepts suitable for filling in the well known
information gap between concepts in the (3-D) scene and sub-symbolic
features in the (2-D) image. A spectral-based semi-concept is a semantic
conjecture based solely on the per-pixel (non-contextual) color
(spectral, i.e., chromatic and achromatic) properties. For example, if
the "color" (spectral signature) of a pixel is, say, green in the
visible electromagnetic spectrum, then that pixel is likely to belong to
a spectral-based semi-concept called vegetation whose information
granularity is, unfortunately, equal or coarser than that of land cover
classes (concepts) such as, say, forest or grassland. In other words, a
spectral-based semi-concept is equivalent to a (land cover) class set.
By definition, different spectral-based semi-concepts are mutually
exclusive, i.e., they must feature no spectral overlap. While land cover
classes are provided with a superior semantic meaning, but are
difficult to be detected automatically, spectral-based semi-concepts are
provided with an inferior semantic meaning, but can be detected
automatically as shown in recent literature. In operational terms SRC
requires neither user-defined parameters nor reference data samples to
run, it is accurate, near real-time (e.g., SRC requires less than 5
minutes to map a Landsat image) and robust to changes in the input
dataset acquired across time, space and sensors. SRC will provide at any
resolution, i.e HR and VHR, from different sensor a preliminary
spectral map (primal sketch). SRC maps each input pixel onto a discrete
and finite set (56 at the moment of spectral sub-categories (i.e.,
semi-concepts based on spectral properties exclusively) belonging to six
exhaustive and mutually exclusive super-categories (semantic strata)
provided with a symbolic meaning as follows: 1) either water or shadow,
2) either snow or ice, 3) clouds, 4) vegetation, 5) either bare soil or
built-up, 6) outliers.
- Up-scale of RS-IUS of SRC version for
mapping hyper-spectral imagery. Partner 11 will cooperate with Partner
12 to software design and fast prototyping of this up-scaled SRC version
- Inputs,
i.e available EO and ancillary data, derive from Task 4.2 and Task 4.1.
Outputs: SRC classifier, at the base of RS-IUS system, is operational.
For this reason it will be possible to immediately provide preliminary
output spectral maps of Italian and Welsh sampling sites at HR to feed
WP6, Task 5.2 for specific class discrimination and Task 5.4 for change
detection by comparing same strata at different dates. As an example,
burnt forest pixels belonging at date t1 to vegetation stratum will have
at data t2, with t1 < t2, , a semantic label of either bare soil or
built-up.
The vegetation stratum (and its mask image) will
be used in WP6 to immediately localize area to be investigated at finer
resolution. As a feedback WP6 will have the opportunity to indicate
areas to be monitored at finer resolution. As a result the processing of
VHR data will start in the same Task 5.1. Additional outputs are sets
of vegetation indices, soil indices to feed WP6.
Task 5.2 RS_IUS second stage modules development
Task 5.2 is devoted to the development of the deductive learning
semantic-nets module 1.3 of Figure 2.a (See also Section 1.2.2.1) by
identifying and developing the RS-IUS second-stage battery of stratified
hierarchical application-specific RS image processing modules to be
employed in series with SRC. These modules can be developed
independently, which reduces the risk of a major project breakdown.
Interpreting complex scenes from EO images often requires, or can
benefit from, a knowledge about the scene at hand. This knowledge may
provide information about the objects contained in the scene, but also
on their spatial arrangement. This type of structural information may be
crucial to differentiate between objects that have similar appearance
in the images, or to disambiguate confusing situations. Examples can be
found in many domains. The key idea in this WP is to derive procedures
that are able of translating the spatial arrangements in the knowledge
domain (or a model of a scene) as constraints for classification
enhancement and semantic labelling. Hence we obtain from a spectral-rule
based classification a semantically interpreted scene. In the context
of this project, Partner 12 use an AC-n like algorithm [Lecoutre 2009]
to maintain arc- or node-consistency between a graph derived from the
spectral-rule based classification and the model-derived scene graph.
Such an approach is closely related to graph matching methods with the
main difference that one can design in the context of CSPs a bunch of
specialized algorithms that are tractable. The problem is still however
open, and the context of this project presents an opportunity to tackle
such issue.These includes:
- Stratified context-sensitive feature extraction modules. Target contextual features in the (2-D) image domain are listed below.
- Stratified multi-scale texture features.
- Segment-based
geometric attributes, e.g. area, perimeter, compactness, straightness
of boundaries, elongation, rectangularity, number of vertices, etc.
- Stratified multi-scale morphological attributes (e.g., differential morphological profile).
- Spatial non-topological relationships between segments, e.g. distance, angle/orientation, etc.
- Spatial topological relationships between segments, e.g., adjacency, inclusion, etc.
- To date several context-sensitive image feature extraction modules have
been developed, however additional research work is required to
guarantee the full operability of this RS-IUS second stage. The
extracted features will be the input of the RS_IUS second classification
stage
- Stratified land cover class-specific fuzzy rule-based
classification modules (semantic nets). Each hierarchical processing
chain consisting of RS image processing software modules must be
analyzed, designed, implemented and validated. Their output products can
be either single-date LC maps, multi-temporal LCC maps or continuous
physical variables (e.g., leaf area index, biomass, etc.).
Outputs of Task 5.2 will include: - LC
maps at different spatial resolution to generate habitat map; to feed
WP6 for modeling at the habitat level; to feed Task 5.4 for change
detection and class transition evaluation.
- class /object
specific context-sensitive features (textural, morphological, geometric,
etc.) useful for habitat characterization and indicator extraction in
WP6.
- classification modules to feed WP 3.
Task 5.3 Integration of algorithms capable of inductive learning from labelled and unlabeled data and quality evaluation modules
The objective of this WP is twofold: Existing inductive
(data-driven) learning modules for data segmentation and classification
will be included in the proposed RS-IUS at second stage to incorporate
the "stratified" or "layered" approach. - Supervised and
unsupervised data learning neural networks (distributed systems)
developed by Partner 1 in recent years will be included in the RS-IUS
second-stage sub-system .
- Multi-sensor data fusion techniques (to be enforced at the preliminary semantic rather than signal level).
- Multiple classifier combination modules already developed by Partner 1
- Algorithms for Hyperspectral image classification developed by Partner 11.
Tools
for products quality evaluation will be identified and used to define
output product accuracies, uncertainty. Interface modules for
input/output will be developed. For each output map, the classification
matrix will provide a measure of confidence of the classification
quality in terms of overall classification accuracy in percentage (OA)
and, for each class, in terms of both user and producer accuracy, k
coefficient, but also other techniques already developed by Partner 1
based on the technique of multiple reference cluster maps [Baraldi et
al. , 2005]. To face the well-known (but often forgotten) issue of
non-injectivity of any quality index, product quality evaluation will be
based on the combination of multiple sources of evidence. The
description of product quality evaluation procedures will be included in
D5.5 of Task5.2. Outputs will be additional modules for
multi-source image classification. The modules will feed WP3 for
integration. This will reduce the risk of failure in providing land
cover maps to Users. Additional modules for land cover maps quality
evaluation will fed WP3 and WP7.
Task 5.4 Change detection modules
This task is devoted to change detection modules improvement
and development. Supervised and unsupervised change detection techniques
have been already developed by Partner 1 within previous
LEWIS-EVG1-CT-2001-00055 project, where change maps where used as
intermediate input layer of a landslide early warning system. Change
detection techniques will be improved and validated on the base of the
ground truth provided by WP4. The availability of an historical
time-serie of HR and VHR images and some no-change ground truth can
allow the detection of land cover/use changes and the trend of each land
cover/use class in time. The information derived by change probability
matrices, evaluated on the base of pixel class label frequency, will
provide specific probability class-transitions and then the trend of
each class. Inputs to this task will be the maps from Task 5.1, Task 5.2 or Task 5.3 as well as data and info from WP2 and WP4.
|