SlideShare a Scribd company logo
1 of 83
Download to read offline
Introduction to HDF5 Data
Model, Programming Model
and Library APIs
HDF and HDF-EOS Workshop VI
December 4, 2002

1
December 4, 2002
• Morning session (lecture format):
– Introduction to HDF5
– Programming Model
– Introduction to Library APIs

• Afternoon concurrent hands-on sessions
– Introduction to HDF5: files, groups, datasets, attributes
– Advanced HDF5: hyperslab selections, compound
datatypes, object and region references
– Parallel HDF5
– High-Level APIs

2
Goals This Morning
• Introduce HDF5
• Provide a basic knowledge of how data can be
organized in HDF5 & how it is used by applications.
• To provide some examples of how to read and write
HDF5 files

3
Goals This Afternoon
• Help you to start working with the HDF5 Library
– Login NCSA machines
– Run examples
– Create your own programs

4
What is HDF5?

5
What is HDF5?
• File format for storing scientific data
– To store and organize all kinds of data
– To share data , to port files from one platform to another
– To overcome a limit on number and size of the objects in
the file

• Software for accessing scientific data
–
–
–
–
–

Flexible I/O library (parallel, remote, etc.)
Efficient storage
Available on almost all platforms
C, F90, C++ , Java APIs
Tools (HDFView, utilities)

6
Example HDF5 file
“/” (root)
“/foo”
3-D array
lat | lon | temp
----|-----|----12 | 23 | 3.1
15 | 24 | 4.2
17 | 21 | 3.6

Table

palette

Raster image
Raster image

2-D array
HDF5 Software

Tools & Applications

HDF5 I/O Library

HDF File
8
Overview
HDF5 Data Model
& I/O Library

9
HDF5 Data Model

10
HDF5 file
• HDF5 file – container for storing scientific data
• Primary Objects
– Groups
– Datasets

• Additional means to organize data
– Attributes
– Sharable objects
– Storage and access properties

11
HDF5 Dataset
• HDF5 dataset – data array and metadata
• Data array
– ordered collection of identically typed data items
distinguished by their indices

• Metadata
– Dataspace – rank, dimensions, other spatial info about
dataset
– Datatype
– Attribute list – user-defined metadata
– Special storage options – how array is organized

12
Dataset Components

Metadata
Dataspace

Rank

Dimensions

3

Dim_1 = 4
Dim_2 = 5
Dim_3 = 7

Datatype
IEEE 32-bit float

Attributes
Storage info

Time = 32.4

Chunked

Pressure = 987

Compressed

Temp = 56

Data
Dataspaces
• Dataspace – spatial info
about a dataset
– Rank and dimensions
• Permanent part of dataset
definition
– Subset of points, for partial I/O
• Needed only during I/O
operations

Rank = 2
Dimensions = 4x6

• Apply to datasets in
memory or in the file

14
Sample Mappings between File Dataspaces
and Memory Dataspaces

(a) Hyperslab from a 2D array to
the corner of a smaller 2D array

(c) A sequence of points from a
2D array to a sequence of points
in a 3D array.

(b) Regular series of blocks from a
2D array to a contiguous sequence at
a certain offset in a 1D array

(d) Union of hyperslabs in file to union
of hyperslabs in memory.

15
Datatypes (array elements)
• Datatype – how to interpret a data element
– Permanent part of the dataset definition

• HDF5 atomic types
–
–
–
–
–
–

normal integer & float
user-definable integer and float (e.g. 13-bit integer)
variable length types (e.g. strings)
pointers - references to objects/dataset regions
enumeration - names mapped to integers
array

• HDF5 compound types
– Comparable to C structs
– Members can be atomic or compound types
16
HDF5 dataset: array of records
3

5

Dimensionality: 5 x 3
int8

int4

int16

Datatype:

Record

2x3x2 array of float32
Attributes
• Attribute – data of the form “name = value”,
attached to an object
• Operations are scaled-down versions of the
dataset operations
– Not extendible
– No compression
– No partial I/O

• Optional for the dataset definition
• Can be overwritten, deleted, added during the
“life” of a dataset

18
Special Storage Options
Better subsetting
access time;
extendable

chunked

Improves storage
efficiency,
transmission speed

compressed

Arrays can be
extended in any
direction

extendable
File B
Dataset “Fred”

External
file

File A
Metadata for Fred

Data for Fred

Metadata in one file,
raw data in another.
Groups
• Group – a mechanism for describing
collections of related objects
• Every file starts with a
root group
• Can have attributes
• Similar to UNIX
directories,
but cycles are
allowed

“/”

20
HDF5 objects are identified and
located by their pathnames

/ (root)
/x
/foo
/foo/temp
/foo/bar/temp

foo
temp

“/”

x

bar
temp

21
Groups & members of groups can
be shared
tom

P

R

“/”

harry

dick
P

/tom/P
/dick/R
/harry/P
22
HDF5 I/O Library

23
Structure of HDF5 Library

Applications
Object API
Library internals
Virtual file I/O
File or other “storage”
24
Structure of HDF5 Library
Object API (C, Fortran 90, Java, C++)

•Specify objects and transformation and storage properties
•Invoke data movement operations and data transformations

Library internals (C)
• Performs data transformations and other prep for I/O
• Configurable transformations (compression, etc.)

Virtual file I/O (C only)

•Perform byte-stream I/O operations (open/close, read/write, seek)
•User-implementable I/O (stdio, network, memory, etc.)
25
Virtual file I/O layer
• A public API for writing I/O drivers
• Allows HDF5 to interface to disk, the network, memory, or a
user-defined device

Virtual file I/O drivers
Stdio

File Family

MPI I/O

Memory

Network

“Storage”

File

File
Family

Memory

Network

26
Intro to
HDF5 API
Programming model for sequential access

27
Goals
• Describe the HDF5 programming model
• Give a feel for what it’ like to use the general
s
HDF5 API
• Review some of the key concepts of HDF5

28
General API Topics
•
•
•
•

General info about HDF5 programming
Creating an HDF5 file
Creating a dataset
Writing and reading a dataset

29
The General HDF5 API
• Currently has C, Fortran 90, Java and C++ bindings.
• C routines begin with prefix H5*, where * is a
single letter indicating the object on which the
operation is to be performed.
• Full functionality

Example APIs:
H5D : Dataset interface e.g.. H5Dread
H5F : File interface
e.g.. H5Fopen
H5S : dataSpace interfacee.g.. H5Sclose
30
The General Paradigm
• Properties (called creation and access property
lists) of objects are defined (optional)
• Objects are opened or created
• Objects then accessed
• Objects finally closed

31
Order of Operations
• The library imposes an order on the operations by
argument dependencies
Example: A file must be opened before a dataset
because the dataset open call requires a file handle
as an argument
• Objects can be closed in any order, and reusing a
closed object will result in an error

32
HDF5 C Programming Issues
For portability, HDF5 library has its own defined types:
hid_t:
hsize_t:

object identifiers (native integer)
size used for dimensions (unsigned long or
unsigned long long)
hssize_t: for specifying coordinates and sometimes for
dimensions (signed long or signed long long)
herr_t:
function return value

For C, include #include hdf5.h at the top of your HDF5
application.
33
h5dump
Command-line Utility for Viewing HDF5 Files

h5dump [-h] [-bb] [-header] [-a ] [-d <names>] [-g <names>]
[-l <names>] [-t <names>] <file>
-h
-header
-a <names>
-d <names>
-g <names>
-l <names>
-t <names>

Print information on this command.
Display header only; no data is displayed.
Display the specified attribute(s).
Display the specified dataset(s).
Display the specified group(s) and all the members.
Displays the value(s) of the specified soft link(s).
Display the specified named datatype(s).

<names> is one or more appropriate object names.

34
Example of h5dump Output
HDF5 "dset.h5" {
GROUP "/" {
DATASET "dset" {
DATATYPE { H5T_STD_I32BE }
DATASPACE { SIMPLE ( 4, 6 ) / ( 4, 6 ) }
DATA {
“/”
1, 2, 3, 4, 5, 6,
7, 8, 9, 10, 11, 12,
13, 14, 15, 16, 17, 18,
19, 20, 21, 22, 23, 24
‘
dset’
}
}
}
}
35
Creating an HDF5 File
Steps to Create a File
•
•
•

Specify File Creation and Access Property
Lists, if necessary
Create a file
Close the file and the property lists, if
necessary

37
Property Lists
• A property list is a collection of values that can
be passed to HDF5 functions at lower layers of
the library
• File Creation Property List
– Controls file metadata
– Size of the user-block, sizes of file data structures, etc.
– Specifying H5P_DEFAULT uses the default values

• Access Property List
– Controls different methods of performing I/O on files
– Unbuffered I/O, parallel I/O, etc.
– Specifying H5P_DEFAULT uses the default values.

38
hid_t H5Fcreate (const char *name, unsigned flags,
hid_t create_id, hid_t access_id)

name
flags
create_id
access_id

IN:
IN:
IN:
IN:

Name of the file to access
File access flags
File creation property list identifier
File access property list identifier
herr_t H5Fclose (hid_t file_id)

file_id

IN:

Identifier of the file to terminate access to
Example 1
Create a new file using
default properties
1
2

hid_t
herr_t

file_id;
status;

3

file_id = H5Fcreate ("file.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

4

status = H5Fclose (file_id);
Example 1

1
2

hid_t
herr_t

file_id;
status;

3

file_id = H5Fcreate ("file.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

4

status = H5Fclose (file_id);

Terminate access to
the File
42
h5_crtfile.c

1
2
3
4
5
6
7
8
9
10
11
12
13
14

#include <hdf5.h>
#define FILE "file.h5"
main() {
hid_t
herr_t

file_id;
status;

/* file identifier */

/* Create a new file using default properties. */
file_id = H5Fcreate (FILE, H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);
/* Terminate access to the file. */
status = H5Fclose (file_id);
}
Example 1: h5dump Output

HDF5 "file.h5" {
GROUP "/" {
}
}

‘
/’

44
Create a Dataset
Dataset Components

Metadata

Data

Dataspace
Rank

Dimensions

3

Dim_1 = 4
Dim_2 = 5
Dim_3 = 7

Datatype
IEEE 32-bit float

Attributes
Storage info

Time = 32.4

Chunked

Pressure = 987

Compressed

Temp = 56

46
Steps to Create a Dataset
•
•

•
•
•

Obtain location ID where dataset is to be
created
Define dataset characteristics (datatype,
dataspace, dataset creation property list, if
necessary)
Create the dataset
Close the datatype, dataspace, and property
list, if necessary
Close the dataset

47
Step 1
Step 1. Obtain the location identifier
where the dataset is to be created
Location Identifier: the file or group identifier in
which to create a dataset

48
Step 2

Step 2. Define the dataset characteristics
– datatype (e.g. integer)
– dataspace (2 dimensions: 100x200)
– dataset creation properties (e.g. chunked and
compressed)

49
Standard Predefined Datatypes
Examples:
H5T_IEEE_F64LE Eight-byte, little-endian, IEEE floating-point
H5T_IEEE_F32BE Four-byte, big-endian, IEEE floating point
H5T_STD_I32LE Four-byte, little-endian, signed two's
complement integer
H5T_STD_U16BE Two-byte, big-endian, unsigned integer
NOTE:

• These datatypes

(DT) are the same on all platforms
• These are DT handles generated at run-time
• Used to describe DT in the HDF5 calls
• DT are not used to describe application data buffers
50
Standard Predefined Datatypes
Examples:
H5T_IEEE_F64LE Eight-byte, little-endian, IEEE floating-point
H5T_IEEE_F32BE Four-byte, big-endian, IEEE floating point
H5T_STD_I32LE Four-byte, little-endian, signed two's
complement integer
H5T_STD_U16BE Two-byte, big-endian, unsigned integer
Architecture

Programming
Type

51
Native Predefined Datatypes
Examples of predefined native types in C:
H5T_NATIVE_INT
H5T_NATIVE_FLOAT
H5T_NATIVE_UINT
H5T_NATIVE_LONG
H5T_NATIVE_CHAR

(int)
(float )
(unsigned int)
(long )
(char )

NOTE:
• These datatypes are NOT the same on all platforms
• These are DT handles generated at run-time
52
Dataspaces

• Dataspace: size and shape of dataset and subset
– Dataset
• Rank: number of dimension
• Dimensions: sizes of all dimensions
• Permanent – part of dataset definition
– Subset
• Size, shape and position of selected elements
• Needed primarily during I/O operations
• Not permanent
• (Subsetting not covered in this tutorial)

• Applies to arrays in memory or in the file
53
Creating a Simple Dataspace
hid_t H5Screate_simple (int rank,
const hsize_t * dims,
const hsize_t *maxdims)
rank
dims
maxdims

IN: Number of dimensions of dataspace
IN: An array of the size of each dimension
IN: An array of the maximum size of each dimension
A value of H5S_UNLIMITED specifies the
unlimited dimension.
A value of NULL specifies that dims and maxdims
are the same.
Dataset Creation Property List
The dataset creation property list contains
information on how to organize data in storage.

Chunked

Chunked &
compressed
55
Property List Example

• Creating a dataset with ``deflate'' compression
create_plist_id = H5Pcreate(H5P_DATASET_CREATE);
H5Pset_chunk(create_ plist_id, ndims, chunk_dims);
H5Pset_deflate(create_ plist_id, 9);

56
Remaining Steps to Create a Dataset

•
•
•

Create the dataset
Close the datatype, dataspace, and
property list, if necessary
Close the dataset

57
hid_t H5Dcreate (hid_t loc_id, const char *name,
hid_t type_id, hid_t space_id,
hid_t create_plist_id)
loc_id

IN: Identifier of file or group to create the dataset
within
name
IN: The name of (the link to) the dataset to create
type_id
IN: Identifier of datatype to use when creating the
dataset
space_id
IN: Identifier of dataspace to use when creating
the dataset
create_plist_id IN: Identifier of the dataset creation property list (or
H5P_DEFAULT)
Example 2 – Create an empty 4x6 dataset
1
2
3

hid_t
hsize_t
herr_t

file_id, dataset_id, dataspace_id;
dims[2];
status;

Create a new file
4

file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

5
6
7

dims[0] = 4;
dims[1] = 6;
dataspace_id = H5Screate_simple (2, dims, NULL);

8

dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE,
dataspace_id, H5P_DEFAULT);

9 status = H5Dclose (dataset_id);
10 status = H5Sclose (dataspace_id);
11 status = H5Fclose (file_id);
Example 2 – Create an empty 4x6 dataset
1
2
3

hid_t
hsize_t
herr_t

file_id, dataset_id, dataspace_id;
dims[2];
status;

4

file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

Create a dataspace
5
6
7

dims[0] = 4;
dims[1] = 6;
dataspace_id = H5Screate_simple (2, dims, NULL);

8

dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE,
dataspace_id, H5P_DEFAULT);

9 status = H5Dclose (dataset_id);
10 status = H5Sclose (dataspace_id);
11 status = H5Fclose (file_id);
Example 2 – Create an empty 4x6 dataset
1
2
3

hid_t
hsize_t
herr_t

file_id, dataset_id, dataspace_id;
dims[2];
status;

4

file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

Create a dataspace
5
6
7

8

current dims
rank
dims[0] = 4;
dims[1] = 6;
dataspace_id = H5Screate_simple (2, dims, NULL);

dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE,
dataspace_id, H5P_DEFAULT);

9 status = H5Dclose (dataset_id);
10 status = H5Sclose (dataspace_id);
11 status = H5Fclose (file_id);

Set maxdims
to current
dims
Example 2 – Create an empty 4x6 dataset
1
2
3

hid_t
hsize_t
herr_t

file_id, dataset_id, dataspace_id;
dims[2];
status;

4

file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

5
6
7

dims[0] = 4;
dims[1] = 6;
dataspace_id = H5Screate_simple (2, dims, NULL);

Create a dataset
8

dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE,
dataspace_id, H5P_DEFAULT);

9 status = H5Dclose (dataset_id);
10 status = H5Sclose (dataspace_id);
11 status = H5Fclose (file_id);
Example 2 – Create an empty 4x6 dataset
1
2
3

hid_t
hsize_t
herr_t

file_id, dataset_id, dataspace_id;
dims[2];
status;

4

file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

5
6
7

Pathname
dims[0] = 4;
dims[1] = 6;
dataspace_id = H5Screate_simple (2, dims, NULL);
Create a dataset

8

Datatype

dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE,
dataspace_id, H5P_DEFAULT);

Dataspace
9 status = H5Dclose (dataset_id);
10 status = H5Sclose (dataspace_id);
11 status = H5Fclose (file_id);

Property list
(default)
Example 2 – Create an empty 4x6 dataset
1
2
3

hid_t
hsize_t
herr_t

file_id, dataset_id, dataspace_id;
dims[2];
status;

4

file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC,
H5P_DEFAULT, H5P_DEFAULT);

5
6
7

dims[0] = 4;
dims[1] = 6;
dataspace_id = H5Screate_simple (2, dims, NULL);

8

dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE,
dataspace_id, H5P_DEFAULT);

Terminate access to dataset, dataspace, & file
9 status = H5Dclose (dataset_id);
10 status = H5Sclose (dataspace_id);
11 status = H5Fclose (file_id);
Example2: h5dump Output
An empty 4x6 dataset

HDF5 "dset.h5" {
GROUP "/" {
DATASET "dset" {
DATATYPE { H5T_STD_I32BE }
DATASPACE { SIMPLE ( 4, 6 ) / ( 4, 6 ) }
DATA {
0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0
}
}
}
}

“/”

‘
dset’

65
Writing and Reading Datasets
Dataset I/O
• Dataset I/O involves
– reading or writing
– all or part of a dataset
– Compressed/uncompressed

• During I/O operations data is translated between
the source & destination (file-memory, memoryfile)
– Datatype conversion
• data types (e.g. 16-bit integer => 32-bit integer) of the
same class
– Dataspace conversion
• dataspace (e.g. 10x20 2d array => 200 1d array)
67
Partial I/O
• Selected elements (called selections) from source are
mapped (read/written) to the selected elements in
destination
• Selection
– Selections in memory can differ from selection in file
– Number of selected elements is always the same in source and
destination

• Selection can be
– Hyperslabs (contiguous blocks, regularly spaced blocks)
– Points
– Results of set operations (union, difference, etc.) on hyperslabs
or points
68
Reading Dataset into Memory from File
File
2D array of 16-bit ints

Memory
3D array of 32-bit ints

69
Reading Dataset into Memory from File
File

Memory

2D array of 16-bit ints

3D array of 32-bit ints

2-d array

Regularly
spaced series
of cubes
70
Reading Dataset into Memory from File
File

Memory

2D array of 16-bit ints

3D array of 32-bit ints

2-d array

Regularly
spaced series
of cubes

The only restriction is that
the number of selected
elements on the left be the
same as on the right.
71
Reading Dataset into Memory from File
File

Memory

2D array of 16-bit ints

3D array of 32-bit ints

Read

72
Steps for Dataset Writing/Reading

•
•
•

If necessary, open the file to obtain the file ID
Open the dataset to obtain the dataset ID
Specify
–
–
–
–
–

•
•

Memory datatype
! Library “knows” file datatype – do not need to
specify !
Memory dataspace
File dataspace
Transfer properties (optional)

Perform the desired operation on the dataset
Close dataspace, datatype and property lists
73
Data Transfer Property List

The data transfer property list is used to control
various aspects of the I/O, such as caching hints
or collective I/O information.

74
hid_t H5Dopen (hid_t loc_id, const char *name)
loc_id

IN:

name

IN:

Identifier of the file or group in which to
open a dataset
The name of the dataset to access

NOTE: File datatype and dataspace are known when
a dataset is opened
herr_t H5Dwrite (hid_t dataset_id, hid_t mem_type_id,
hid_t mem_space_id, hid_t file_space_id,
hid_t xfer_plist_id, const void * buf )
dataset_id
mem_type_id
mem_space_id
file_space_id
xfer_plist_id
buf

IN: Identifier of the dataset to write to
IN: Identifier of memory datatype of the dataset
IN: Identifier of the memory dataspace
(or H5S_ALL)
IN: Identifier of the file dataspace (or H5S_ALL)
IN: Identifier of the data transfer properties to use
(or H5P_DEFAULT)
IN: Buffer with data to be written to the file
Example 3 – Writing to an existing dataset
1
2
3

hid_t
herr_t
int

file_id, dataset_id;
status;
i, j, dset_data[4][6];

4
5
6

for (i = 0; i < 4; i++)
for (j = 0; j < 6; j++)
dset_data[i][j] = i * 6 + j + 1;

7
8

file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT);
dataset_id = H5Dopen (file_id, " dset");

9

status = H5Dwrite (dataset_id, H5T_NATIVE_INT,
H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
Example 3 – Writing to an existing dataset
1
2
3

hid_t
herr_t
int

file_id, dataset_id;
status;
i, j, dset_data[4][6];

Initialize buffer
4
5
6

for (i = 0; i < 4; i++)
for (j = 0; j < 6; j++)
dset_data[i][j] = i * 6 + j + 1;

7
8

file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT);
dataset_id = H5Dopen (file_id, " dset");

9

status = H5Dwrite (dataset_id, H5T_NATIVE_INT,
H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
Example 3 – Writing to an existing dataset
1
2
3

hid_t
herr_t
int

file_id, dataset_id;
status;
i, j, dset_data[4][6];

4
5
6

for (i = 0; i < 4; i++)
for (j = 0; j < 6; j++)
dset_data[i][j] = i * 6 + j + 1;

Open existing file and dataset
7
8

file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT);
dataset_id = H5Dopen (file_id, " dset");

9

status = H5Dwrite (dataset_id, H5T_NATIVE_INT,
H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
Example 3 – Writing to an existing dataset
1
2
3

hid_t
herr_t
int

file_id, dataset_id;
status;
i, j, dset_data[4][6];

4
5
6

for (i = 0; i < 4; i++)
for (j = 0; j < 6; j++)
dset_data[i][j] = i * 6 + j + 1;

7
8

file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT);
dataset_id = H5Dopen (file_id, " dset");

Write to dataset
9

status = H5Dwrite (dataset_id, H5T_NATIVE_INT,
H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
Example 3: h5dump Output
HDF5 "dset.h5" {
GROUP "/" {
DATASET "dset" {
DATATYPE { H5T_STD_I32BE }
DATASPACE { SIMPLE ( 4, 6 ) / ( 4, 6 ) }
DATA {
1, 2, 3, 4, 5, 6,
7, 8, 9, 10, 11, 12,
13, 14, 15, 16, 17, 18,
19, 20, 21, 22, 23, 24
}
}
}
}
81
For more information…
HDF • HDF website
– http://hdf.ncsa.uiuc.edu/

5

• HDF5 Information Center
– http://hdf.ncsa.uiuc.edu/HDF5/

• HDF Helpdesk
– hdfhelp@ncsa.uiuc.edu

• HDF users mailing list
– hdfnews@ncsa.uiuc.edu

82
Thank you

83

More Related Content

What's hot

STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...
STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...
STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...The HDF-EOS Tools and Information Center
 
Mba admission in india
Mba admission in indiaMba admission in india
Mba admission in indiaEdhole.com
 
An IDL-Based Validation Toolkit: Extensions to use the HDF-EOS Swath Format
An IDL-Based  Validation Toolkit: Extensions to  use the HDF-EOS Swath FormatAn IDL-Based  Validation Toolkit: Extensions to  use the HDF-EOS Swath Format
An IDL-Based Validation Toolkit: Extensions to use the HDF-EOS Swath FormatThe HDF-EOS Tools and Information Center
 
Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...
Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...
Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...Research Data Alliance
 
Interoperability with netCDF-4 - Experience with NPP and HDF-EOS5 products
Interoperability with netCDF-4 - Experience with NPP and HDF-EOS5 productsInteroperability with netCDF-4 - Experience with NPP and HDF-EOS5 products
Interoperability with netCDF-4 - Experience with NPP and HDF-EOS5 productsThe HDF-EOS Tools and Information Center
 

What's hot (20)

Caching and Buffering in HDF5
Caching and Buffering in HDF5Caching and Buffering in HDF5
Caching and Buffering in HDF5
 
Performance Tuning in HDF5
Performance Tuning in HDF5 Performance Tuning in HDF5
Performance Tuning in HDF5
 
STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...
STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...
STARE-PODS: A Versatile Data Store Leveraging the HDF Virtual Object Layer fo...
 
HDF-EOS 2/5 to netCDF Converter
HDF-EOS 2/5 to netCDF ConverterHDF-EOS 2/5 to netCDF Converter
HDF-EOS 2/5 to netCDF Converter
 
Data Analytics using MATLAB and HDF5
Data Analytics using MATLAB and HDF5Data Analytics using MATLAB and HDF5
Data Analytics using MATLAB and HDF5
 
H5Coro: The Cloud-Optimized Read-Only Library
H5Coro: The Cloud-Optimized Read-Only LibraryH5Coro: The Cloud-Optimized Read-Only Library
H5Coro: The Cloud-Optimized Read-Only Library
 
Advanced HDF5 Features
Advanced HDF5 FeaturesAdvanced HDF5 Features
Advanced HDF5 Features
 
Implementing HDF5 in MATLAB
Implementing HDF5 in MATLABImplementing HDF5 in MATLAB
Implementing HDF5 in MATLAB
 
Status of HDF-EOS, Related Software and Tools
Status of HDF-EOS, Related Software and ToolsStatus of HDF-EOS, Related Software and Tools
Status of HDF-EOS, Related Software and Tools
 
Product Designer Hub - Taking HPD to the Web
Product Designer Hub - Taking HPD to the WebProduct Designer Hub - Taking HPD to the Web
Product Designer Hub - Taking HPD to the Web
 
HDF4 and HDF5 Performance Preliminary Results
HDF4 and HDF5 Performance Preliminary ResultsHDF4 and HDF5 Performance Preliminary Results
HDF4 and HDF5 Performance Preliminary Results
 
HDF5 FastQuery
HDF5 FastQueryHDF5 FastQuery
HDF5 FastQuery
 
Mba admission in india
Mba admission in indiaMba admission in india
Mba admission in india
 
Introduction to NetCDF-4
Introduction to NetCDF-4Introduction to NetCDF-4
Introduction to NetCDF-4
 
An IDL-Based Validation Toolkit: Extensions to use the HDF-EOS Swath Format
An IDL-Based  Validation Toolkit: Extensions to  use the HDF-EOS Swath FormatAn IDL-Based  Validation Toolkit: Extensions to  use the HDF-EOS Swath Format
An IDL-Based Validation Toolkit: Extensions to use the HDF-EOS Swath Format
 
Efficiently serving HDF5 via OPeNDAP
Efficiently serving HDF5 via OPeNDAPEfficiently serving HDF5 via OPeNDAP
Efficiently serving HDF5 via OPeNDAP
 
Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...
Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...
Enabling Precise Identification and Citability of Dynamic Data: Recommendatio...
 
HDF and netCDF Data Support in ArcGIS
HDF and netCDF Data Support in ArcGISHDF and netCDF Data Support in ArcGIS
HDF and netCDF Data Support in ArcGIS
 
Interoperability with netCDF-4 - Experience with NPP and HDF-EOS5 products
Interoperability with netCDF-4 - Experience with NPP and HDF-EOS5 productsInteroperability with netCDF-4 - Experience with NPP and HDF-EOS5 products
Interoperability with netCDF-4 - Experience with NPP and HDF-EOS5 products
 
MATLAB, netCDF, and OPeNDAP
MATLAB, netCDF, and OPeNDAPMATLAB, netCDF, and OPeNDAP
MATLAB, netCDF, and OPeNDAP
 

Similar to Introduction to HDF5 Data Model, Programming Model and Library APIs

HDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, Datatypes
HDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, DatatypesHDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, Datatypes
HDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, DatatypesThe HDF-EOS Tools and Information Center
 
A quick start guide to using HDF5 files in GLOBE Claritas
A quick start guide to using HDF5 files in GLOBE ClaritasA quick start guide to using HDF5 files in GLOBE Claritas
A quick start guide to using HDF5 files in GLOBE ClaritasGuy Maslen
 
Hadoop and object stores can we do it better
Hadoop and object stores  can we do it betterHadoop and object stores  can we do it better
Hadoop and object stores can we do it bettergvernik
 
Hadoop and object stores: Can we do it better?
Hadoop and object stores: Can we do it better?Hadoop and object stores: Can we do it better?
Hadoop and object stores: Can we do it better?gvernik
 
Data Archiving and Sharing
Data Archiving and SharingData Archiving and Sharing
Data Archiving and SharingC. Tobin Magle
 
Big Data Architecture Workshop - Vahid Amiri
Big Data Architecture Workshop -  Vahid AmiriBig Data Architecture Workshop -  Vahid Amiri
Big Data Architecture Workshop - Vahid Amiridatastack
 
Hadoop & Complex Systems Research
Hadoop & Complex Systems ResearchHadoop & Complex Systems Research
Hadoop & Complex Systems ResearchDr. Mirko Kämpf
 
Development of the irods rados plugin @ iRODS User group meeting 2014
Development of the irods rados plugin @ iRODS User group meeting 2014Development of the irods rados plugin @ iRODS User group meeting 2014
Development of the irods rados plugin @ iRODS User group meeting 2014mgrawinkel
 
Hops - Distributed metadata for Hadoop
Hops - Distributed metadata for HadoopHops - Distributed metadata for Hadoop
Hops - Distributed metadata for HadoopJim Dowling
 

Similar to Introduction to HDF5 Data Model, Programming Model and Library APIs (20)

Introduction to HDF5 Data Model, Programming Model and Library APIs
Introduction to HDF5 Data Model, Programming Model and Library APIsIntroduction to HDF5 Data Model, Programming Model and Library APIs
Introduction to HDF5 Data Model, Programming Model and Library APIs
 
HDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, Datatypes
HDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, DatatypesHDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, Datatypes
HDF5 Advanced Topics - Object's Properties, Storage Methods, Filters, Datatypes
 
Hdf5 intro
Hdf5 introHdf5 intro
Hdf5 intro
 
HDF5 Advanced Topics
HDF5 Advanced TopicsHDF5 Advanced Topics
HDF5 Advanced Topics
 
A quick start guide to using HDF5 files in GLOBE Claritas
A quick start guide to using HDF5 files in GLOBE ClaritasA quick start guide to using HDF5 files in GLOBE Claritas
A quick start guide to using HDF5 files in GLOBE Claritas
 
Cloud-Optimized HDF5 Files
Cloud-Optimized HDF5 FilesCloud-Optimized HDF5 Files
Cloud-Optimized HDF5 Files
 
Parallel Computing with HDF Server
Parallel Computing with HDF ServerParallel Computing with HDF Server
Parallel Computing with HDF Server
 
Advanced HDF5 Features
Advanced HDF5 FeaturesAdvanced HDF5 Features
Advanced HDF5 Features
 
Tthornton code4lib
Tthornton code4libTthornton code4lib
Tthornton code4lib
 
HDF5 Advanced Topics - Datatypes and Partial I/O
HDF5 Advanced Topics - Datatypes and Partial I/OHDF5 Advanced Topics - Datatypes and Partial I/O
HDF5 Advanced Topics - Datatypes and Partial I/O
 
Hadoop and object stores can we do it better
Hadoop and object stores  can we do it betterHadoop and object stores  can we do it better
Hadoop and object stores can we do it better
 
Hadoop and object stores: Can we do it better?
Hadoop and object stores: Can we do it better?Hadoop and object stores: Can we do it better?
Hadoop and object stores: Can we do it better?
 
Data Archiving and Sharing
Data Archiving and SharingData Archiving and Sharing
Data Archiving and Sharing
 
Big Data Architecture Workshop - Vahid Amiri
Big Data Architecture Workshop -  Vahid AmiriBig Data Architecture Workshop -  Vahid Amiri
Big Data Architecture Workshop - Vahid Amiri
 
HDF for the Cloud
HDF for the CloudHDF for the Cloud
HDF for the Cloud
 
Hadoop ppt1
Hadoop ppt1Hadoop ppt1
Hadoop ppt1
 
Hadoop & Complex Systems Research
Hadoop & Complex Systems ResearchHadoop & Complex Systems Research
Hadoop & Complex Systems Research
 
Development of the irods rados plugin @ iRODS User group meeting 2014
Development of the irods rados plugin @ iRODS User group meeting 2014Development of the irods rados plugin @ iRODS User group meeting 2014
Development of the irods rados plugin @ iRODS User group meeting 2014
 
Hadoop
HadoopHadoop
Hadoop
 
Hops - Distributed metadata for Hadoop
Hops - Distributed metadata for HadoopHops - Distributed metadata for Hadoop
Hops - Distributed metadata for Hadoop
 

More from The HDF-EOS Tools and Information Center

More from The HDF-EOS Tools and Information Center (20)

Accessing HDF5 data in the cloud with HSDS
Accessing HDF5 data in the cloud with HSDSAccessing HDF5 data in the cloud with HSDS
Accessing HDF5 data in the cloud with HSDS
 
The State of HDF
The State of HDFThe State of HDF
The State of HDF
 
Highly Scalable Data Service (HSDS) Performance Features
Highly Scalable Data Service (HSDS) Performance FeaturesHighly Scalable Data Service (HSDS) Performance Features
Highly Scalable Data Service (HSDS) Performance Features
 
Creating Cloud-Optimized HDF5 Files
Creating Cloud-Optimized HDF5 FilesCreating Cloud-Optimized HDF5 Files
Creating Cloud-Optimized HDF5 Files
 
HDF5 OPeNDAP Handler Updates, and Performance Discussion
HDF5 OPeNDAP Handler Updates, and Performance DiscussionHDF5 OPeNDAP Handler Updates, and Performance Discussion
HDF5 OPeNDAP Handler Updates, and Performance Discussion
 
Hyrax: Serving Data from S3
Hyrax: Serving Data from S3Hyrax: Serving Data from S3
Hyrax: Serving Data from S3
 
Accessing Cloud Data and Services Using EDL, Pydap, MATLAB
Accessing Cloud Data and Services Using EDL, Pydap, MATLABAccessing Cloud Data and Services Using EDL, Pydap, MATLAB
Accessing Cloud Data and Services Using EDL, Pydap, MATLAB
 
HDF - Current status and Future Directions
HDF - Current status and Future DirectionsHDF - Current status and Future Directions
HDF - Current status and Future Directions
 
HDFEOS.org User Analsys, Updates, and Future
HDFEOS.org User Analsys, Updates, and FutureHDFEOS.org User Analsys, Updates, and Future
HDFEOS.org User Analsys, Updates, and Future
 
HDF - Current status and Future Directions
HDF - Current status and Future Directions HDF - Current status and Future Directions
HDF - Current status and Future Directions
 
MATLAB Modernization on HDF5 1.10
MATLAB Modernization on HDF5 1.10MATLAB Modernization on HDF5 1.10
MATLAB Modernization on HDF5 1.10
 
HDF for the Cloud - Serverless HDF
HDF for the Cloud - Serverless HDFHDF for the Cloud - Serverless HDF
HDF for the Cloud - Serverless HDF
 
HDF5 <-> Zarr
HDF5 <-> ZarrHDF5 <-> Zarr
HDF5 <-> Zarr
 
HDF for the Cloud - New HDF Server Features
HDF for the Cloud - New HDF Server FeaturesHDF for the Cloud - New HDF Server Features
HDF for the Cloud - New HDF Server Features
 
Apache Drill and Unidata THREDDS Data Server for NASA HDF-EOS on S3
Apache Drill and Unidata THREDDS Data Server for NASA HDF-EOS on S3Apache Drill and Unidata THREDDS Data Server for NASA HDF-EOS on S3
Apache Drill and Unidata THREDDS Data Server for NASA HDF-EOS on S3
 
HDF5 and Ecosystem: What Is New?
HDF5 and Ecosystem: What Is New?HDF5 and Ecosystem: What Is New?
HDF5 and Ecosystem: What Is New?
 
HDF5 Roadmap 2019-2020
HDF5 Roadmap 2019-2020HDF5 Roadmap 2019-2020
HDF5 Roadmap 2019-2020
 
Leveraging the Cloud for HDF Software Testing
Leveraging the Cloud for HDF Software TestingLeveraging the Cloud for HDF Software Testing
Leveraging the Cloud for HDF Software Testing
 
Google Colaboratory for HDF-EOS
Google Colaboratory for HDF-EOSGoogle Colaboratory for HDF-EOS
Google Colaboratory for HDF-EOS
 
HDF-EOS Data Product Developer's Guide
HDF-EOS Data Product Developer's GuideHDF-EOS Data Product Developer's Guide
HDF-EOS Data Product Developer's Guide
 

Recently uploaded

The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionDilum Bandara
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxLoriGlavin3
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersRaghuram Pandurangan
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii SoldatenkoFwdays
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 

Recently uploaded (20)

The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An Introduction
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information Developers
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 

Introduction to HDF5 Data Model, Programming Model and Library APIs

  • 1. Introduction to HDF5 Data Model, Programming Model and Library APIs HDF and HDF-EOS Workshop VI December 4, 2002 1
  • 2. December 4, 2002 • Morning session (lecture format): – Introduction to HDF5 – Programming Model – Introduction to Library APIs • Afternoon concurrent hands-on sessions – Introduction to HDF5: files, groups, datasets, attributes – Advanced HDF5: hyperslab selections, compound datatypes, object and region references – Parallel HDF5 – High-Level APIs 2
  • 3. Goals This Morning • Introduce HDF5 • Provide a basic knowledge of how data can be organized in HDF5 & how it is used by applications. • To provide some examples of how to read and write HDF5 files 3
  • 4. Goals This Afternoon • Help you to start working with the HDF5 Library – Login NCSA machines – Run examples – Create your own programs 4
  • 6. What is HDF5? • File format for storing scientific data – To store and organize all kinds of data – To share data , to port files from one platform to another – To overcome a limit on number and size of the objects in the file • Software for accessing scientific data – – – – – Flexible I/O library (parallel, remote, etc.) Efficient storage Available on almost all platforms C, F90, C++ , Java APIs Tools (HDFView, utilities) 6
  • 7. Example HDF5 file “/” (root) “/foo” 3-D array lat | lon | temp ----|-----|----12 | 23 | 3.1 15 | 24 | 4.2 17 | 21 | 3.6 Table palette Raster image Raster image 2-D array
  • 8. HDF5 Software Tools & Applications HDF5 I/O Library HDF File 8
  • 11. HDF5 file • HDF5 file – container for storing scientific data • Primary Objects – Groups – Datasets • Additional means to organize data – Attributes – Sharable objects – Storage and access properties 11
  • 12. HDF5 Dataset • HDF5 dataset – data array and metadata • Data array – ordered collection of identically typed data items distinguished by their indices • Metadata – Dataspace – rank, dimensions, other spatial info about dataset – Datatype – Attribute list – user-defined metadata – Special storage options – how array is organized 12
  • 13. Dataset Components Metadata Dataspace Rank Dimensions 3 Dim_1 = 4 Dim_2 = 5 Dim_3 = 7 Datatype IEEE 32-bit float Attributes Storage info Time = 32.4 Chunked Pressure = 987 Compressed Temp = 56 Data
  • 14. Dataspaces • Dataspace – spatial info about a dataset – Rank and dimensions • Permanent part of dataset definition – Subset of points, for partial I/O • Needed only during I/O operations Rank = 2 Dimensions = 4x6 • Apply to datasets in memory or in the file 14
  • 15. Sample Mappings between File Dataspaces and Memory Dataspaces (a) Hyperslab from a 2D array to the corner of a smaller 2D array (c) A sequence of points from a 2D array to a sequence of points in a 3D array. (b) Regular series of blocks from a 2D array to a contiguous sequence at a certain offset in a 1D array (d) Union of hyperslabs in file to union of hyperslabs in memory. 15
  • 16. Datatypes (array elements) • Datatype – how to interpret a data element – Permanent part of the dataset definition • HDF5 atomic types – – – – – – normal integer & float user-definable integer and float (e.g. 13-bit integer) variable length types (e.g. strings) pointers - references to objects/dataset regions enumeration - names mapped to integers array • HDF5 compound types – Comparable to C structs – Members can be atomic or compound types 16
  • 17. HDF5 dataset: array of records 3 5 Dimensionality: 5 x 3 int8 int4 int16 Datatype: Record 2x3x2 array of float32
  • 18. Attributes • Attribute – data of the form “name = value”, attached to an object • Operations are scaled-down versions of the dataset operations – Not extendible – No compression – No partial I/O • Optional for the dataset definition • Can be overwritten, deleted, added during the “life” of a dataset 18
  • 19. Special Storage Options Better subsetting access time; extendable chunked Improves storage efficiency, transmission speed compressed Arrays can be extended in any direction extendable File B Dataset “Fred” External file File A Metadata for Fred Data for Fred Metadata in one file, raw data in another.
  • 20. Groups • Group – a mechanism for describing collections of related objects • Every file starts with a root group • Can have attributes • Similar to UNIX directories, but cycles are allowed “/” 20
  • 21. HDF5 objects are identified and located by their pathnames / (root) /x /foo /foo/temp /foo/bar/temp foo temp “/” x bar temp 21
  • 22. Groups & members of groups can be shared tom P R “/” harry dick P /tom/P /dick/R /harry/P 22
  • 24. Structure of HDF5 Library Applications Object API Library internals Virtual file I/O File or other “storage” 24
  • 25. Structure of HDF5 Library Object API (C, Fortran 90, Java, C++) •Specify objects and transformation and storage properties •Invoke data movement operations and data transformations Library internals (C) • Performs data transformations and other prep for I/O • Configurable transformations (compression, etc.) Virtual file I/O (C only) •Perform byte-stream I/O operations (open/close, read/write, seek) •User-implementable I/O (stdio, network, memory, etc.) 25
  • 26. Virtual file I/O layer • A public API for writing I/O drivers • Allows HDF5 to interface to disk, the network, memory, or a user-defined device Virtual file I/O drivers Stdio File Family MPI I/O Memory Network “Storage” File File Family Memory Network 26
  • 27. Intro to HDF5 API Programming model for sequential access 27
  • 28. Goals • Describe the HDF5 programming model • Give a feel for what it’ like to use the general s HDF5 API • Review some of the key concepts of HDF5 28
  • 29. General API Topics • • • • General info about HDF5 programming Creating an HDF5 file Creating a dataset Writing and reading a dataset 29
  • 30. The General HDF5 API • Currently has C, Fortran 90, Java and C++ bindings. • C routines begin with prefix H5*, where * is a single letter indicating the object on which the operation is to be performed. • Full functionality Example APIs: H5D : Dataset interface e.g.. H5Dread H5F : File interface e.g.. H5Fopen H5S : dataSpace interfacee.g.. H5Sclose 30
  • 31. The General Paradigm • Properties (called creation and access property lists) of objects are defined (optional) • Objects are opened or created • Objects then accessed • Objects finally closed 31
  • 32. Order of Operations • The library imposes an order on the operations by argument dependencies Example: A file must be opened before a dataset because the dataset open call requires a file handle as an argument • Objects can be closed in any order, and reusing a closed object will result in an error 32
  • 33. HDF5 C Programming Issues For portability, HDF5 library has its own defined types: hid_t: hsize_t: object identifiers (native integer) size used for dimensions (unsigned long or unsigned long long) hssize_t: for specifying coordinates and sometimes for dimensions (signed long or signed long long) herr_t: function return value For C, include #include hdf5.h at the top of your HDF5 application. 33
  • 34. h5dump Command-line Utility for Viewing HDF5 Files h5dump [-h] [-bb] [-header] [-a ] [-d <names>] [-g <names>] [-l <names>] [-t <names>] <file> -h -header -a <names> -d <names> -g <names> -l <names> -t <names> Print information on this command. Display header only; no data is displayed. Display the specified attribute(s). Display the specified dataset(s). Display the specified group(s) and all the members. Displays the value(s) of the specified soft link(s). Display the specified named datatype(s). <names> is one or more appropriate object names. 34
  • 35. Example of h5dump Output HDF5 "dset.h5" { GROUP "/" { DATASET "dset" { DATATYPE { H5T_STD_I32BE } DATASPACE { SIMPLE ( 4, 6 ) / ( 4, 6 ) } DATA { “/” 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24 ‘ dset’ } } } } 35
  • 37. Steps to Create a File • • • Specify File Creation and Access Property Lists, if necessary Create a file Close the file and the property lists, if necessary 37
  • 38. Property Lists • A property list is a collection of values that can be passed to HDF5 functions at lower layers of the library • File Creation Property List – Controls file metadata – Size of the user-block, sizes of file data structures, etc. – Specifying H5P_DEFAULT uses the default values • Access Property List – Controls different methods of performing I/O on files – Unbuffered I/O, parallel I/O, etc. – Specifying H5P_DEFAULT uses the default values. 38
  • 39. hid_t H5Fcreate (const char *name, unsigned flags, hid_t create_id, hid_t access_id) name flags create_id access_id IN: IN: IN: IN: Name of the file to access File access flags File creation property list identifier File access property list identifier
  • 40. herr_t H5Fclose (hid_t file_id) file_id IN: Identifier of the file to terminate access to
  • 41. Example 1 Create a new file using default properties 1 2 hid_t herr_t file_id; status; 3 file_id = H5Fcreate ("file.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); 4 status = H5Fclose (file_id);
  • 42. Example 1 1 2 hid_t herr_t file_id; status; 3 file_id = H5Fcreate ("file.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); 4 status = H5Fclose (file_id); Terminate access to the File 42
  • 43. h5_crtfile.c 1 2 3 4 5 6 7 8 9 10 11 12 13 14 #include <hdf5.h> #define FILE "file.h5" main() { hid_t herr_t file_id; status; /* file identifier */ /* Create a new file using default properties. */ file_id = H5Fcreate (FILE, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); /* Terminate access to the file. */ status = H5Fclose (file_id); }
  • 44. Example 1: h5dump Output HDF5 "file.h5" { GROUP "/" { } } ‘ /’ 44
  • 46. Dataset Components Metadata Data Dataspace Rank Dimensions 3 Dim_1 = 4 Dim_2 = 5 Dim_3 = 7 Datatype IEEE 32-bit float Attributes Storage info Time = 32.4 Chunked Pressure = 987 Compressed Temp = 56 46
  • 47. Steps to Create a Dataset • • • • • Obtain location ID where dataset is to be created Define dataset characteristics (datatype, dataspace, dataset creation property list, if necessary) Create the dataset Close the datatype, dataspace, and property list, if necessary Close the dataset 47
  • 48. Step 1 Step 1. Obtain the location identifier where the dataset is to be created Location Identifier: the file or group identifier in which to create a dataset 48
  • 49. Step 2 Step 2. Define the dataset characteristics – datatype (e.g. integer) – dataspace (2 dimensions: 100x200) – dataset creation properties (e.g. chunked and compressed) 49
  • 50. Standard Predefined Datatypes Examples: H5T_IEEE_F64LE Eight-byte, little-endian, IEEE floating-point H5T_IEEE_F32BE Four-byte, big-endian, IEEE floating point H5T_STD_I32LE Four-byte, little-endian, signed two's complement integer H5T_STD_U16BE Two-byte, big-endian, unsigned integer NOTE: • These datatypes (DT) are the same on all platforms • These are DT handles generated at run-time • Used to describe DT in the HDF5 calls • DT are not used to describe application data buffers 50
  • 51. Standard Predefined Datatypes Examples: H5T_IEEE_F64LE Eight-byte, little-endian, IEEE floating-point H5T_IEEE_F32BE Four-byte, big-endian, IEEE floating point H5T_STD_I32LE Four-byte, little-endian, signed two's complement integer H5T_STD_U16BE Two-byte, big-endian, unsigned integer Architecture Programming Type 51
  • 52. Native Predefined Datatypes Examples of predefined native types in C: H5T_NATIVE_INT H5T_NATIVE_FLOAT H5T_NATIVE_UINT H5T_NATIVE_LONG H5T_NATIVE_CHAR (int) (float ) (unsigned int) (long ) (char ) NOTE: • These datatypes are NOT the same on all platforms • These are DT handles generated at run-time 52
  • 53. Dataspaces • Dataspace: size and shape of dataset and subset – Dataset • Rank: number of dimension • Dimensions: sizes of all dimensions • Permanent – part of dataset definition – Subset • Size, shape and position of selected elements • Needed primarily during I/O operations • Not permanent • (Subsetting not covered in this tutorial) • Applies to arrays in memory or in the file 53
  • 54. Creating a Simple Dataspace hid_t H5Screate_simple (int rank, const hsize_t * dims, const hsize_t *maxdims) rank dims maxdims IN: Number of dimensions of dataspace IN: An array of the size of each dimension IN: An array of the maximum size of each dimension A value of H5S_UNLIMITED specifies the unlimited dimension. A value of NULL specifies that dims and maxdims are the same.
  • 55. Dataset Creation Property List The dataset creation property list contains information on how to organize data in storage. Chunked Chunked & compressed 55
  • 56. Property List Example • Creating a dataset with ``deflate'' compression create_plist_id = H5Pcreate(H5P_DATASET_CREATE); H5Pset_chunk(create_ plist_id, ndims, chunk_dims); H5Pset_deflate(create_ plist_id, 9); 56
  • 57. Remaining Steps to Create a Dataset • • • Create the dataset Close the datatype, dataspace, and property list, if necessary Close the dataset 57
  • 58. hid_t H5Dcreate (hid_t loc_id, const char *name, hid_t type_id, hid_t space_id, hid_t create_plist_id) loc_id IN: Identifier of file or group to create the dataset within name IN: The name of (the link to) the dataset to create type_id IN: Identifier of datatype to use when creating the dataset space_id IN: Identifier of dataspace to use when creating the dataset create_plist_id IN: Identifier of the dataset creation property list (or H5P_DEFAULT)
  • 59. Example 2 – Create an empty 4x6 dataset 1 2 3 hid_t hsize_t herr_t file_id, dataset_id, dataspace_id; dims[2]; status; Create a new file 4 file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); 5 6 7 dims[0] = 4; dims[1] = 6; dataspace_id = H5Screate_simple (2, dims, NULL); 8 dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE, dataspace_id, H5P_DEFAULT); 9 status = H5Dclose (dataset_id); 10 status = H5Sclose (dataspace_id); 11 status = H5Fclose (file_id);
  • 60. Example 2 – Create an empty 4x6 dataset 1 2 3 hid_t hsize_t herr_t file_id, dataset_id, dataspace_id; dims[2]; status; 4 file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); Create a dataspace 5 6 7 dims[0] = 4; dims[1] = 6; dataspace_id = H5Screate_simple (2, dims, NULL); 8 dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE, dataspace_id, H5P_DEFAULT); 9 status = H5Dclose (dataset_id); 10 status = H5Sclose (dataspace_id); 11 status = H5Fclose (file_id);
  • 61. Example 2 – Create an empty 4x6 dataset 1 2 3 hid_t hsize_t herr_t file_id, dataset_id, dataspace_id; dims[2]; status; 4 file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); Create a dataspace 5 6 7 8 current dims rank dims[0] = 4; dims[1] = 6; dataspace_id = H5Screate_simple (2, dims, NULL); dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE, dataspace_id, H5P_DEFAULT); 9 status = H5Dclose (dataset_id); 10 status = H5Sclose (dataspace_id); 11 status = H5Fclose (file_id); Set maxdims to current dims
  • 62. Example 2 – Create an empty 4x6 dataset 1 2 3 hid_t hsize_t herr_t file_id, dataset_id, dataspace_id; dims[2]; status; 4 file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); 5 6 7 dims[0] = 4; dims[1] = 6; dataspace_id = H5Screate_simple (2, dims, NULL); Create a dataset 8 dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE, dataspace_id, H5P_DEFAULT); 9 status = H5Dclose (dataset_id); 10 status = H5Sclose (dataspace_id); 11 status = H5Fclose (file_id);
  • 63. Example 2 – Create an empty 4x6 dataset 1 2 3 hid_t hsize_t herr_t file_id, dataset_id, dataspace_id; dims[2]; status; 4 file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); 5 6 7 Pathname dims[0] = 4; dims[1] = 6; dataspace_id = H5Screate_simple (2, dims, NULL); Create a dataset 8 Datatype dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE, dataspace_id, H5P_DEFAULT); Dataspace 9 status = H5Dclose (dataset_id); 10 status = H5Sclose (dataspace_id); 11 status = H5Fclose (file_id); Property list (default)
  • 64. Example 2 – Create an empty 4x6 dataset 1 2 3 hid_t hsize_t herr_t file_id, dataset_id, dataspace_id; dims[2]; status; 4 file_id = H5Fcreate ("dset.h5", H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); 5 6 7 dims[0] = 4; dims[1] = 6; dataspace_id = H5Screate_simple (2, dims, NULL); 8 dataset_id = H5Dcreate(file_id," dset",H5T_STD_I32BE, dataspace_id, H5P_DEFAULT); Terminate access to dataset, dataspace, & file 9 status = H5Dclose (dataset_id); 10 status = H5Sclose (dataspace_id); 11 status = H5Fclose (file_id);
  • 65. Example2: h5dump Output An empty 4x6 dataset HDF5 "dset.h5" { GROUP "/" { DATASET "dset" { DATATYPE { H5T_STD_I32BE } DATASPACE { SIMPLE ( 4, 6 ) / ( 4, 6 ) } DATA { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 } } } } “/” ‘ dset’ 65
  • 67. Dataset I/O • Dataset I/O involves – reading or writing – all or part of a dataset – Compressed/uncompressed • During I/O operations data is translated between the source & destination (file-memory, memoryfile) – Datatype conversion • data types (e.g. 16-bit integer => 32-bit integer) of the same class – Dataspace conversion • dataspace (e.g. 10x20 2d array => 200 1d array) 67
  • 68. Partial I/O • Selected elements (called selections) from source are mapped (read/written) to the selected elements in destination • Selection – Selections in memory can differ from selection in file – Number of selected elements is always the same in source and destination • Selection can be – Hyperslabs (contiguous blocks, regularly spaced blocks) – Points – Results of set operations (union, difference, etc.) on hyperslabs or points 68
  • 69. Reading Dataset into Memory from File File 2D array of 16-bit ints Memory 3D array of 32-bit ints 69
  • 70. Reading Dataset into Memory from File File Memory 2D array of 16-bit ints 3D array of 32-bit ints 2-d array Regularly spaced series of cubes 70
  • 71. Reading Dataset into Memory from File File Memory 2D array of 16-bit ints 3D array of 32-bit ints 2-d array Regularly spaced series of cubes The only restriction is that the number of selected elements on the left be the same as on the right. 71
  • 72. Reading Dataset into Memory from File File Memory 2D array of 16-bit ints 3D array of 32-bit ints Read 72
  • 73. Steps for Dataset Writing/Reading • • • If necessary, open the file to obtain the file ID Open the dataset to obtain the dataset ID Specify – – – – – • • Memory datatype ! Library “knows” file datatype – do not need to specify ! Memory dataspace File dataspace Transfer properties (optional) Perform the desired operation on the dataset Close dataspace, datatype and property lists 73
  • 74. Data Transfer Property List The data transfer property list is used to control various aspects of the I/O, such as caching hints or collective I/O information. 74
  • 75. hid_t H5Dopen (hid_t loc_id, const char *name) loc_id IN: name IN: Identifier of the file or group in which to open a dataset The name of the dataset to access NOTE: File datatype and dataspace are known when a dataset is opened
  • 76. herr_t H5Dwrite (hid_t dataset_id, hid_t mem_type_id, hid_t mem_space_id, hid_t file_space_id, hid_t xfer_plist_id, const void * buf ) dataset_id mem_type_id mem_space_id file_space_id xfer_plist_id buf IN: Identifier of the dataset to write to IN: Identifier of memory datatype of the dataset IN: Identifier of the memory dataspace (or H5S_ALL) IN: Identifier of the file dataspace (or H5S_ALL) IN: Identifier of the data transfer properties to use (or H5P_DEFAULT) IN: Buffer with data to be written to the file
  • 77. Example 3 – Writing to an existing dataset 1 2 3 hid_t herr_t int file_id, dataset_id; status; i, j, dset_data[4][6]; 4 5 6 for (i = 0; i < 4; i++) for (j = 0; j < 6; j++) dset_data[i][j] = i * 6 + j + 1; 7 8 file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT); dataset_id = H5Dopen (file_id, " dset"); 9 status = H5Dwrite (dataset_id, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
  • 78. Example 3 – Writing to an existing dataset 1 2 3 hid_t herr_t int file_id, dataset_id; status; i, j, dset_data[4][6]; Initialize buffer 4 5 6 for (i = 0; i < 4; i++) for (j = 0; j < 6; j++) dset_data[i][j] = i * 6 + j + 1; 7 8 file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT); dataset_id = H5Dopen (file_id, " dset"); 9 status = H5Dwrite (dataset_id, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
  • 79. Example 3 – Writing to an existing dataset 1 2 3 hid_t herr_t int file_id, dataset_id; status; i, j, dset_data[4][6]; 4 5 6 for (i = 0; i < 4; i++) for (j = 0; j < 6; j++) dset_data[i][j] = i * 6 + j + 1; Open existing file and dataset 7 8 file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT); dataset_id = H5Dopen (file_id, " dset"); 9 status = H5Dwrite (dataset_id, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
  • 80. Example 3 – Writing to an existing dataset 1 2 3 hid_t herr_t int file_id, dataset_id; status; i, j, dset_data[4][6]; 4 5 6 for (i = 0; i < 4; i++) for (j = 0; j < 6; j++) dset_data[i][j] = i * 6 + j + 1; 7 8 file_id = H5Fopen ("dset.h5", H5F_ACC_RDWR, H5P_DEFAULT); dataset_id = H5Dopen (file_id, " dset"); Write to dataset 9 status = H5Dwrite (dataset_id, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, dset_data);
  • 81. Example 3: h5dump Output HDF5 "dset.h5" { GROUP "/" { DATASET "dset" { DATATYPE { H5T_STD_I32BE } DATASPACE { SIMPLE ( 4, 6 ) / ( 4, 6 ) } DATA { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24 } } } } 81
  • 82. For more information… HDF • HDF website – http://hdf.ncsa.uiuc.edu/ 5 • HDF5 Information Center – http://hdf.ncsa.uiuc.edu/HDF5/ • HDF Helpdesk – hdfhelp@ncsa.uiuc.edu • HDF users mailing list – hdfnews@ncsa.uiuc.edu 82