0% found this document useful (0 votes)
3 views25 pages

H5py Python

The document discusses the application of the Python programming language and the HDF5 interface (H5py) in processing satellite remote sensing data. It highlights the advantages of using Python for both numerical and non-numerical data processing, emphasizing its flexibility, ease of development, and compatibility with HDF5. The document also provides examples of Python code for data aggregation and manipulation using H5py, illustrating its practical use in remote sensing applications.

Uploaded by

Expert_Modeller
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
3 views25 pages

H5py Python

The document discusses the application of the Python programming language and the HDF5 interface (H5py) in processing satellite remote sensing data. It highlights the advantages of using Python for both numerical and non-numerical data processing, emphasizing its flexibility, ease of development, and compatibility with HDF5. The document also provides examples of Python code for data aggregation and manipulation using H5py, illustrating its practical use in remote sensing applications.

Uploaded by

Expert_Modeller
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 25

The Python Programming Language and HDF5: H5py.

Application to Satellite Remote Sensing Data Processing.

Needs of satellite remote sensing data processing and


programming languages.
Introduction to some Python elements.
Introduction to Numerical Python with Numpy.
Tying them together with H5py.
Examples.
Daniel Kahn, Science Systems and Applications Inc.
HDF/HDF-EOS Workshop XIII, Riverdale, MD
3 Nov 2009
Where credit is due:

Andrew Collette, UCLA, author of h5py

Francesc Alted, author of PyTables and the


HDF5 1.6 Pyrex definitions (i.e Python-C defs
used in H5py)

Andrew also requested acknowledgment for Darren


Dale (Cornell) and Armando Sole (ESRF) for their
suggestions, patches and encouragement.
Modern remote sensing satellite data and data processing
needs have several characteristics.
1) The data need to be subjected to numerical algorithms. The
algorithms vary in their sophistication but few, if any, require
supercomputer architectures. E.g. Computing a zonal mean.
(Fortran, C, C++)
2) The data need to be subjected to non-numerical “bookkeeping”
algorithms. Associated metadata needs to be read, matched, sliced,
diced, acted upon and committed to databases. (Perl)
3) Modern systems automatically the schedule processing on to
lots of inexpensive, commodity CPUs. This complicates using
managed (i.e. licensed) commercial runtime environments, e.g.
IDL, Matlab, Mathematica.
4) Satellite processing systems are dynamic. They need
customization several times over the course of each instrument's
life. Development time needs to be short. (Perl, IDL, Matlab,
Mathematica)
5) Data processing systems need to track data provenance.
A Remote Sensing Data Processing Challenge...

Problem:
We lack a programming language which spans the breath of
our remote sensing data processing needs, i.e. Which can
handle both numerical and non-numerical processing
(including HDF5) and that has a short development cycle.

Goal:
Find a programming language which allows fast development
and is adept at numerical processing as well as non-numerical
processing. “Adept” doesn't mean “best”, just pretty good.
...and the Python Programming Language
Why Python fits the bill better than:
Fortran
As an interpreted language with dynamic variable bindings
Python offers more flexibility during development. The
multidimensional array extension (numpy) provides efficient and
competitive array operations. It is more similar to IDL or
Matlab...
IDL or Matlab
No license required which can save money on several levels.
Python language can be more well suited to non-numerical
data processing tasks such as are encountered in production
processing of remote sensing. It is more similar to Perl...
Perl/PDL
Nicer syntax for array operations.
More complete HDF5 support (h5py vs PDL HDF5).
Some elements of Python
Python is an interpreted language. Python
statements can easily be run from an interactive
prompt. Python programs are written in ASCII files.

Python has many of the same elements as programming


languages used for numerical processing (I.e. loops,
conditionals, etc) plus...
Name spaces

Lists

Dictionaries
Python has name spaces
Name spaces allow related functions and data to be grouped together
in a way that will not conflict with unrelated functions and data which
happens to be named the same.
Name space are very useful, but are only relevant here because
Python name spaces correspond to Python modules. The
numerical and HDF5 routines we will see later are implemented
as Python modules.
For example, to open an HDF5 file in Python we first import the h5py
module and then we open the file using the File function of the h5py
module.
import h5py # First import the module.
FileID = h5py.File('hdf5_test.h5','r')#Now use the File function,
#note the module name in red
.
.
.

We could import other modules simultaneously which have an unrelated


function called File and there would be no conflict or error.
Python has Lists:
These are 1D arrays which are useful for solving many problems.
List index notation looks like 1D array notation in many languages.

>>> MyList = [1.0,70] # initialize variable to 2 element list


>>> MyList.append("Demo Data!") # Append an element, NOT like fortran
>>> MyList.append(MyList[0] + MyList[1]) # Append sum of first 2 elements
>>> print MyList
[1.0, 70, 'Demo Data!', 71.0]

Python has Dictionaries:


(Aka hash tables) These map symbols to objects in analogy to
how a 1D list maps an integer index value to an object.

>>> MyDictionary = {} # Initialize empty Dictionary


>>> MyDictionary[0] = 1
>>> MyDictionary['Gadzooks'] = 70
>>> MyDictionary['Kind of Data'] = 'Demo Data!'
>>> MyDictionary['Result'] = MyDictionary[0] + MyDictionary['Gadzooks']
>>> print MyDictionary
{0: 1, 'Kind of Data': 'Demo Data!', 'Result': 71, 'Gadzooks': 70}
Python Lists vs Python Dictionaries vs Arrays in Fortran

Index to a List must be an integer, but a Dictionary can be


indexed by an integer or string.

MyList[Integer] vs. MyDictionary[Integer or String]

The objects referenced by a List or Dictionary can be any Python


object. This is in contrast to more traditional languages, eg. The
elements Fortran array must be of a particular type such as
real*4.

MyList = [34,”String Datum”]


or
MyDictionary={'FirstKey':34,'SecondKey':”String Datum”}
The List and Dictionary data structures make it possible to write
flexible programs very quickly. However, they are not good for
numerical computation. The variety of objects (number, string, etc)
which they can reference makes computation slow.

For example, adding array elements: MyList[0] + MyList[1]


Python must check at run time if MyList[0] and MyList[1] are
objects that can be added together, and not, say, a number and a
string. In a loop this check is performed at every iteration!

What to do?

Enter Numpy
Numpy is a package for use with Python which provides
multidimensional arrays of numeric (and other) types and extends
Python syntax to provide a familiar interface to the arrays.

Numpy extends Python syntax to allow the expression of vector


operations similar to those in IDL, Matlab, or Fortran (>=90).
Numpy Example: Create, Print, and add two 2D arrays

Build from Python lists of lists of elements (the module name is numpy)...
>>> a = numpy.array([[1,2,3],[4,5,6]])
>>> print a
[[1 2 3]
[4 5 6]]
Build from dimension sizes...
>>> b = numpy.ones([2,3])
>>> print b
[[ 1. 1. 1.]
[ 1. 1. 1.]]
Print a selected element...
>>> print a[1,2]
6

Example: Add two dimensional arrays.


>>> print a+b
[[ 2. 3. 4.]
[ 5. 6. 7.]]
>>>
The HDF5 connection: H5py

H5py is an Python-HDF5 interface is a Python module


written by Andrew Collette. Its design allows the use of
familiar Python structures for working with HDF5 files.

The interface maps Python syntax for familiar Python


objects to similar objects in the HDF5 file.
Here is a familiar example HDF5 file
from the HDFView distribution:
Here is how to read the 3D int array using
h5py.

>>> import h5py


>>> fid = h5py.File('hdf5_test.h5','r')
>>> group = fid['arrays']
>>> The3DArray = group['3D int array'].value
>>> fid.close()
>>> The3DArray
array([[[ 174, 27, 0, ..., 102, 71, 100009
[ 171, 27, 0, ..., 194, 79, 100109
[ 172, 27, 0, ..., 102, 55, 100209
...,
Equivalence of HDF5 Groups and Python Dictionaries
Print value of dictionary entry:

>>> MyDictionary =
{'RandArray':numpy.random.random([2,2])}
>>> print MyDictionary['RandArray']
[[ 0.82066938 0.39219683]
[ 0.86546443 0.91276533]]

Print value of HDF5 file entry:

>>> fid = h5py.File('RandomArray.h5','r')


>>> print fid['RandArray'].value
[[ 0.1 3.14152908]
[ 2.71828008 0. ]]
Simple Real world example:

Goal: Retrieve HDF5 file from Configuration Management (CM)


and insert CM metadata into HDF5 file.

We want to place the CM path and revision number inside the


HDF5 file as an attribute.
Python script to retrieve file from CM and store Rev
number as attribute.
#! /usr/bin/env python

import sys
import os
import h5py

Rev = sys.argv[1] # Specifiy CM path on command line


SVNFilepath = sys.argv[2] # Specify revision number on
#comand line.

command = 'svn export -r ' + Rev + ' ' + SVNFilepath #Subversion


# Command
InStream = os.popen(command,'r')
ExportString = InStream.read()
ExportReturnCode = InStream.close()
Elements = SVNFilepath.split('/')

# HDF5 code

fid = h5py.File(Elements[-1]) # Elements[-1] is file name

fid.attrs['SVN Path and Revision'] = SVNFilepath + '@' + Rev

fid.close()

H5py code in red. Note the minimal effort coding HDF5 calls.
Another real world example (if NPOESS is your real world).

We have had several occasions to do data aggregation on HDF5 files for the
OMPS Limb instrument.

Our retrieval code (Fortran) processes an orbit of data as 480 distinct pieces
and places the results into 480 distinct HDF5 files. We wish to aggregate the
files such that an N dimensional array in a unaggregated file becomes a N+1
dimensional array in the aggregated HDF5 file.

The Fortran code is (mostly) the product of the retrieval scientist, while the
aggregation is a requirement of the production data processing system. It makes
sense to aggregate as a post-processing step to the Fortran code so as to
minimize changes to the original Fortran code.
Aggregation Algorithm

1) Input is list of HDF5 files


2) Analyze structure of one file to generate list of fully qualified dataset names,
dimensions, and type (In the code I use a Python Dictionary and not a list).
3) Assume all files have that structure. Read corresponding datasets (of dim N)
from each file into aggregation variable (of dim N+1).
4) After corresponding datasets have been read from all files write aggregation
variable to HDF5 file.
5) Repeat 3 until all datasets have been aggregated.

Array 4
File 1 Array File 2 Array File 3 Array File 4 Array Array 3
Array 2
Array 1
+ + +

Schematic of Array Aggregation for One Dataset


# Data Aggregator. This script takes a set of HDF5 files as input.
# The files are expected to have identical struture. All the
# corresponding arrays in the input files are combined into an array
# which has dimensions N+1, where N is the number of dimensions in the
# original, constituent arrays.

import sys
import h5py
import numpy

Files = sys.argv[1:] # Get file names from command line

# First step is to select one file and create a map of the shape and
# data type of all datasets. This is naturally done via a recursive
# function, called VisitAllObjects
FirstFid = h5py.File(Files[0],'r') # Open first HDF5 file

FileInfo = {} # Initialize FileInfo to be a dictionary. We will use it to build a mapping from


# dataset name to a tuple containing shape and type of dataset.
# EVALUATE HDF5 HIERARCHY
# Evaluating a a hierarchy is naturally a recursive process so we define a function....
def VisitAllObjects(Group,Path):
for i in Group.items():
if isinstance(i[1],h5py.Group):
VisitAllObjects(i[1],Path + '/' + i[0])
else:
DatasetName = Path + '/' + i[0]
FileInfo[DatasetName] = (Group[DatasetName].shape, Group[DatasetName].dtype)

VisitAllObjects(FirstFid,'')
FirstFid.close()
# Print dataset paths and info to screen
for (k,v) in FileInfo.items():
print k,v
# AGGREGATE DATA
# Now that we understand the file structure we can perform the aggregation.
OutputFileID = h5py.File('AggregatedData.h5','w')
NumberOfFiles = len(Files)

# Here is the meat of the code. The outer loop is over datasets, the inner
over all files.
for Dataset in FileInfo.keys():
AggregatedData =
numpy.ndarray(FileInfo[Dataset][0]+(NumberOfFiles,),dtype =
FileInfo[Dataset][1])
for FileNumber in range(NumberOfFiles):
# Open file, read data into aggregation array, and close
fid = h5py.File(Files[FileNumber],'r')
AggregatedData[...,FileNumber] = fid[Dataset].value
fid.close()
Path = Dataset.split('/')
map((lambda(x): OutputFileID.require_group(x)), Path[1:-1])
#OutputFileID[Dataset] = AggregatedData

OutputFileID.create_dataset(Dataset,data=AggregatedData,compression=5,chunks=
FileInfo[Dataset][0]+(1,))

OutputFileID.close()
Original, Unaggregated Data Field
$ h5dump -H -d /ANCILLARY_DATA/GeopotentialHeight_NCEP \
Data/OMPS_LP_SDR_20041121_55_146_-69_-119.h5
HDF5 "Data/OMPS_LP_SDR_20041121_55_146_-69_-119.h5" {
DATASET "/ANCILLARY_DATA/GeopotentialHeight_NCEP" {
DATATYPE H5T_IEEE_F32LE
DATASPACE SIMPLE { ( 5, 3, 21 ) / ( 5, 3, 21 ) }
ATTRIBUTE "Title" {
DATATYPE H5T_STRING {
STRSIZE 25;
}
Note 3 dimensions
DATASPACE SCALAR
}
ATTRIBUTE "Units" {
DATATYPE H5T_STRING {
STRSIZE 2;
}
DATASPACE SCALAR
}
ATTRIBUTE "_FillValue" {
DATATYPE H5T_IEEE_F32LE
DATASPACE SIMPLE { ( 1 ) / ( 1 ) }
}
}
Aggregated Data Field

$ h5dump -H -d /ANCILLARY_DATA/GeopotentialHeight_NCEP \
AggregatedData.h5
HDF5 "AggregatedData.h5" {
DATASET "/ANCILLARY_DATA/GeopotentialHeight_NCEP" {
DATATYPE H5T_IEEE_F32LE
DATASPACE SIMPLE { ( 5, 3, 21, 4 ) / ( 5, 3, 21, 4 ) }
}
}

Now we have 4 dimensions. The new


dimension has extent 4, corresponding to
the number of input files.

Note that none of the attributes were copied.


This is a bug, but easily fixed.
To fix the bug, I assume attributes (like “Units”) are not
aggregated, and I take attributes and values from first file.

New code consists of only 2+ additional lines, shown below


in green.
def VisitAllObjects(Group,Path):
for i in Group.items():
if isinstance(i[1],h5py.Group):
VisitAllObjects(i[1],Path + '/' + i[0])
else:
DatasetName = Path + '/' + i[0]
FileInfo[DatasetName] = (Group[DatasetName].shape,
Group[DatasetName].dtype,
Group[DatasetName].attrs.listitems())

And also....
DS=OutputFileID.create_dataset(Dataset,data=AggregatedData,compression=5,chunks=F
[DS.attrs.__setitem__(Attribute[0],Attribute[1]) for Attribute in
FileInfo[Dataset][2]]
Fixed output:
$ h5dump -H -d /ANCILLARY_DATA/GeopotentialHeight_NCEP \
AggregatedDataWithAttributes.h5
HDF5 "AggregatedDataWithAttributes.h5" {
DATASET "/ANCILLARY_DATA/GeopotentialHeight_NCEP" {
DATATYPE H5T_IEEE_F32LE
DATASPACE SIMPLE { ( 5, 3, 21, 4 ) / ( 5, 3, 21, 4 ) }
ATTRIBUTE "Title" {
DATATYPE H5T_STRING {
STRSIZE 25;
CTYPE H5T_C_S1;
}
DATASPACE SCALAR
}
Now we have attributes.
ATTRIBUTE "Units" {
DATATYPE H5T_STRING {
STRSIZE 2;
CTYPE H5T_C_S1;
}
DATASPACE SCALAR
}
ATTRIBUTE "_FillValue" {
DATATYPE H5T_IEEE_F32LE
DATASPACE SIMPLE { ( 1 ) / ( 1 ) }
}
}
}
Summary:
Python offers a high degree of flexibility for code
development, combined with the ability to do easy
text, numerical array and HDF5 coding make it a
good candidate for solving problems in satellite
remote sensing data processing.

Few, if any, other computer languages offer this


combination of benefits making it uniquely suited for
this task.
Future Work:
Tracking module version provenance is likely one of the
outstanding questions for Python use in production.

Acknowledgment:
Curt Tilmes at NASA Goddard funded this work via
contract NNG06HX18C task 614.5-01-07

You might also like