Code cleanup & reorganization
Description:
Move further in the testing framework cleanup, eliminating all the
global variables (moving them into testframe.c as static variables) from the
testing framework code and moving it into the libh5test.a.
Platforms tested:
FreeBSD 4.9 (sleipnir) w & w/o thread-safety, c++ & parallel
h5committested
536 lines
16 KiB
C++
536 lines
16 KiB
C++
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
|
||
* Copyright by the Board of Trustees of the University of Illinois. *
|
||
* All rights reserved. *
|
||
* *
|
||
* This file is part of HDF5. The full HDF5 copyright notice, including *
|
||
* terms governing use, modification, and redistribution, is contained in *
|
||
* the files COPYING and Copyright.html. COPYING can be found at the root *
|
||
* of the source code distribution tree; Copyright.html can be found at the *
|
||
* root level of an installed copy of the electronic HDF5 document set and *
|
||
* is linked from the top-level documents page. It can also be found at *
|
||
* http://hdf.ncsa.uiuc.edu/HDF5/doc/Copyright.html. If you do not have *
|
||
* access to either file, you may request a copy from hdfhelp@ncsa.uiuc.edu. *
|
||
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
|
||
|
||
/*****************************************************************************
|
||
FILE
|
||
th5s.cpp - HDF5 C++ testing the functionalities associated with the
|
||
C dataspace interface (H5S)
|
||
|
||
EXTERNAL ROUTINES/VARIABLES:
|
||
|
||
***************************************************************************/
|
||
|
||
#ifdef OLD_HEADER_FILENAME
|
||
#include <iostream.h>
|
||
#else
|
||
#include <iostream>
|
||
#endif
|
||
|
||
#include "H5Cpp.h"
|
||
#include "testhdf5.h"
|
||
|
||
#ifndef H5_NO_NAMESPACE
|
||
using namespace H5;
|
||
#endif /* !H5_NO_NAMESPACE */
|
||
|
||
#define TESTFILE "th5s.h5"
|
||
#define DATAFILE "th5s1.h5"
|
||
|
||
/* 3-D dataset with fixed dimensions */
|
||
#define SPACE1_NAME "Space1"
|
||
#define SPACE1_RANK 3
|
||
#define SPACE1_DIM1 3
|
||
#define SPACE1_DIM2 15
|
||
#define SPACE1_DIM3 13
|
||
|
||
/* 4-D dataset with one unlimited dimension */
|
||
#define SPACE2_NAME "Space2"
|
||
#define SPACE2_RANK 4
|
||
#define SPACE2_DIM1 0
|
||
#define SPACE2_DIM2 15
|
||
#define SPACE2_DIM3 13
|
||
#define SPACE2_DIM4 23
|
||
#define SPACE2_MAX1 H5S_UNLIMITED
|
||
#define SPACE2_MAX2 15
|
||
#define SPACE2_MAX3 13
|
||
#define SPACE2_MAX4 23
|
||
|
||
/* Scalar dataset with simple datatype */
|
||
#define SPACE3_NAME "Scalar1"
|
||
#define SPACE3_RANK 0
|
||
unsigned space3_data=65;
|
||
|
||
/* Scalar dataset with compound datatype */
|
||
#define SPACE4_NAME "Scalar2"
|
||
#define SPACE4_RANK 0
|
||
#define SPACE4_FIELDNAME1 "c1"
|
||
#define SPACE4_FIELDNAME2 "u"
|
||
#define SPACE4_FIELDNAME3 "f"
|
||
#define SPACE4_FIELDNAME4 "c2"
|
||
size_t space4_field1_off=0;
|
||
size_t space4_field2_off=0;
|
||
size_t space4_field3_off=0;
|
||
size_t space4_field4_off=0;
|
||
struct space4_struct {
|
||
char c1;
|
||
unsigned u;
|
||
float f;
|
||
char c2;
|
||
} space4_data={'v',987123,-3.14,'g'}; /* Test data for 4th dataspace */
|
||
|
||
/*-------------------------------------------------------------------------
|
||
*
|
||
* Function: test_h5s_basic
|
||
*
|
||
* Purpose: Test basic H5S (dataspace) code
|
||
*
|
||
* Return: none
|
||
*
|
||
* Programmer: Binh-Minh Ribler (using C version)
|
||
* Mar 2001
|
||
*
|
||
* Modifications:
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
static void
|
||
test_h5s_basic(void)
|
||
{
|
||
hsize_t dims1[] = {SPACE1_DIM1, SPACE1_DIM2, SPACE1_DIM3};
|
||
hsize_t dims2[] = {SPACE2_DIM1, SPACE2_DIM2, SPACE2_DIM3,
|
||
SPACE2_DIM4};
|
||
hsize_t dims3[H5S_MAX_RANK+1];
|
||
hsize_t tmax[4];
|
||
|
||
/* Output message about test being performed */
|
||
MESSAGE(5, ("Testing Dataspace Manipulation\n"));
|
||
|
||
try
|
||
{ // beginning of first try block
|
||
|
||
/* Create file - removed this since the following operations don't
|
||
need the file to be opened */
|
||
|
||
// Create simple dataspace sid1
|
||
DataSpace sid1 (SPACE1_RANK, dims1 );
|
||
|
||
// Get simple extent npoints of the dataspace sid1 and verify it
|
||
hssize_t n; /* Number of dataspace elements */
|
||
n = sid1.getSimpleExtentNpoints();
|
||
VERIFY(n, SPACE1_DIM1 * SPACE1_DIM2 * SPACE1_DIM3,
|
||
"H5Sget_simple_extent_npoints");
|
||
|
||
// Get the logical rank of dataspace sid1 and verify it
|
||
int rank; /* Logical rank of dataspace */
|
||
rank = sid1.getSimpleExtentNdims();
|
||
VERIFY(rank, SPACE1_RANK, "H5Sget_simple_extent_ndims");
|
||
|
||
// Retrieves dimension size of dataspace sid1 and verify it
|
||
int ndims; /* Number of dimensions */
|
||
hsize_t tdims[4]; /* Dimension array to test with */
|
||
ndims = sid1.getSimpleExtentDims( tdims );
|
||
VERIFY(HDmemcmp(tdims, dims1, SPACE1_RANK * sizeof(unsigned)), 0,
|
||
"H5Sget_simple_extent_dims");
|
||
|
||
// Create simple dataspace sid2
|
||
hsize_t max2[] = {SPACE2_MAX1, SPACE2_MAX2, SPACE2_MAX3, SPACE2_MAX4};
|
||
DataSpace sid2 (SPACE2_RANK, dims2, max2);
|
||
|
||
// Get simple extent npoints of dataspace sid2 and verify it
|
||
n = sid2.getSimpleExtentNpoints();
|
||
VERIFY(n, SPACE2_DIM1 * SPACE2_DIM2 * SPACE2_DIM3 * SPACE2_DIM4,
|
||
"H5Sget_simple_extent_npoints");
|
||
|
||
// Get the logical rank of dataspace sid2 and verify it
|
||
rank = sid2.getSimpleExtentNdims();
|
||
VERIFY(rank, SPACE2_RANK, "H5Sget_simple_extent_ndims");
|
||
|
||
// Retrieves dimension size and max size of dataspace sid2 and
|
||
// verify them
|
||
ndims = sid2.getSimpleExtentDims( tdims, tmax );
|
||
VERIFY(HDmemcmp(tdims, dims2, SPACE2_RANK * sizeof(unsigned)), 0,
|
||
"H5Sget_simple_extent_dims");
|
||
VERIFY(HDmemcmp(tmax, max2, SPACE2_RANK * sizeof(unsigned)), 0,
|
||
"H5Sget_simple_extent_dims");
|
||
} // end of first try block
|
||
catch( DataSpaceIException error )
|
||
{
|
||
CHECK(FAIL, FAIL, error.getCFuncName());
|
||
}
|
||
|
||
/*
|
||
* Check to be sure we can't create a simple data space that has too many
|
||
* dimensions.
|
||
*/
|
||
try {
|
||
DataSpace manydims_ds(H5S_MAX_RANK+1, dims3, NULL);
|
||
|
||
// Should FAIL but didn't - BMR (Note 1): a new macro that skips
|
||
// the comparison b/w the 1st & 2nd args would be more appropriate,
|
||
// but VERIFY will still do - Mar 12, 01
|
||
VERIFY(manydims_ds.getId(), FAIL, "DataSpace constructor");
|
||
}
|
||
catch( DataSpaceIException error ) {} // do nothing, FAIL expected
|
||
|
||
/*
|
||
* Try reading a file that has been prepared that has a dataset with a
|
||
* higher dimensionality than what the library can handle.
|
||
*
|
||
* If this test fails and the H5S_MAX_RANK variable has changed, follow
|
||
* the instructions in space_overflow.c for regenating the th5s.h5 file.
|
||
*/
|
||
char testfile[512]="";
|
||
char *srcdir = getenv("srcdir");
|
||
if (srcdir && ((strlen(srcdir) + strlen(TESTFILE) + 1) < sizeof(testfile))){
|
||
strcpy(testfile, srcdir);
|
||
strcat(testfile, "/");
|
||
}
|
||
strcat(testfile, TESTFILE);
|
||
try { // try block for testing higher dimensionality
|
||
|
||
// Create file
|
||
H5File fid1(testfile, H5F_ACC_RDONLY);
|
||
|
||
// Try to open the dataset that has higher dimensionality than
|
||
// what the library can handle and this operation should fail.
|
||
try {
|
||
DataSet dset1 = fid1.openDataSet( "dset" );
|
||
VERIFY( dset1.getId(), FAIL, "H5File::openDataSet");
|
||
}
|
||
catch( FileIException error ) { } // do nothing, FAIL expected
|
||
} // end of try block for testing higher dimensionality
|
||
|
||
// catch exception thrown by H5File constructor
|
||
catch( FileIException error ) {
|
||
CHECK_I(FAIL, error.getCFuncName());
|
||
cout << "***cannot open the pre-created H5S_MAX_RANK test file" <<
|
||
testfile << endl;
|
||
}
|
||
|
||
// CHECK_I(ret, "H5Fclose"); // leave this here, later, fake a failure
|
||
// in the p_close see how this will handle it. - BMR
|
||
|
||
/* Verify that incorrect dimensions don't work */
|
||
dims1[0] = 0;
|
||
try {
|
||
DataSpace wrongdim_ds (SPACE1_RANK, dims1);
|
||
VERIFY(wrongdim_ds.getId(), FAIL, "DataSpace constructor");
|
||
}
|
||
catch( DataSpaceIException error ) {} // do nothing; FAIL expected
|
||
|
||
// Create a simple dataspace
|
||
DataSpace sid3 (H5S_SIMPLE);
|
||
|
||
// Attempts to use incorrect dimensions, should fail
|
||
try { sid3.setExtentSimple( SPACE1_RANK, dims1 ); }
|
||
catch( DataSpaceIException error )
|
||
{
|
||
// ret value is already < 0 for an exception to be thrown;
|
||
// also see Note 1 above
|
||
VERIFY(FAIL, FAIL, error.getCFuncName());
|
||
}
|
||
} /* test_h5s_basic() */
|
||
|
||
/*-------------------------------------------------------------------------
|
||
*
|
||
* Function: test_h5s_scalar_write
|
||
*
|
||
* Purpose: Test scalar H5S (dataspace) writing code
|
||
*
|
||
* Return: none
|
||
*
|
||
* Programmer: Binh-Minh Ribler (using C version)
|
||
* Mar 2001
|
||
*
|
||
* Modifications:
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
static void
|
||
test_h5s_scalar_write(void)
|
||
{
|
||
|
||
/* Output message about test being performed */
|
||
MESSAGE(5, ("Testing Scalar Dataspace Writing\n"));
|
||
|
||
try
|
||
{
|
||
// Create file
|
||
H5File fid1(DATAFILE, H5F_ACC_TRUNC);
|
||
|
||
/* Create scalar dataspace */
|
||
DataSpace sid1(SPACE3_RANK, NULL);
|
||
|
||
//n = H5Sget_simple_extent_npoints(sid1);
|
||
hssize_t n; /* Number of dataspace elements */
|
||
n = sid1.getSimpleExtentNpoints();
|
||
VERIFY(n, 1, "DataSpace::getSimpleExtentNpoints");
|
||
|
||
int rank; /* Logical rank of dataspace */
|
||
rank = sid1.getSimpleExtentNdims();
|
||
VERIFY(rank, SPACE3_RANK, "DataSpace::getSimpleExtentNdims");
|
||
|
||
// Retrieves dimension size of dataspace sid1 and verify it
|
||
int ndims; /* Number of dimensions */
|
||
hsize_t tdims[4]; /* Dimension array to test with */
|
||
ndims = sid1.getSimpleExtentDims( tdims );
|
||
VERIFY(ndims, 0, "DataSpace::getSimpleExtentDims");
|
||
|
||
/* Verify extent type */
|
||
H5S_class_t ext_type; /* Extent type */
|
||
ext_type = sid1.getSimpleExtentType();
|
||
VERIFY(ext_type, H5S_SCALAR, "DataSpace::getSimpleExtentType");
|
||
|
||
/* Create a dataset */
|
||
DataSet dataset = fid1.createDataSet("Dataset1", PredType::NATIVE_UINT,sid1);
|
||
|
||
dataset.write(&space3_data, PredType::NATIVE_UINT);
|
||
} // end of try block
|
||
catch (Exception error)
|
||
{
|
||
CHECK(FAIL, FAIL, error.getCFuncName());
|
||
}
|
||
} /* test_h5s_scalar_write() */
|
||
|
||
/*-------------------------------------------------------------------------
|
||
*
|
||
* Function: test_h5s_scalar_read
|
||
*
|
||
* Purpose: Test scalar H5S (dataspace) reading code
|
||
*
|
||
* Return: none
|
||
*
|
||
* Programmer: Binh-Minh Ribler (using C version)
|
||
* Mar 2001
|
||
*
|
||
* Modifications:
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
static void
|
||
test_h5s_scalar_read(void)
|
||
{
|
||
hsize_t tdims[4]; /* Dimension array to test with */
|
||
|
||
/* Output message about test being performed */
|
||
MESSAGE(5, ("Testing Scalar Dataspace Reading\n"));
|
||
|
||
try
|
||
{
|
||
/* Create file */
|
||
H5File fid1(DATAFILE, H5F_ACC_RDWR);
|
||
|
||
/* Create a dataset */
|
||
DataSet dataset = fid1.openDataSet("Dataset1");
|
||
|
||
DataSpace sid1 = dataset.getSpace();
|
||
|
||
// Get the number of dataspace elements
|
||
hssize_t n = sid1.getSimpleExtentNpoints();
|
||
VERIFY(n, 1, "H5Sget_simple_extent_npoints");
|
||
|
||
// Get the logical rank of the dataspace
|
||
int ndims = sid1.getSimpleExtentNdims();
|
||
VERIFY(ndims, SPACE3_RANK, "H5Sget_simple_extent_ndims");
|
||
|
||
ndims = sid1.getSimpleExtentDims(tdims);
|
||
VERIFY(ndims, 0, "H5Sget_simple_extent_dims");
|
||
|
||
unsigned rdata; /* Scalar data read in */
|
||
dataset.read(&rdata, PredType::NATIVE_UINT);
|
||
VERIFY(rdata, space3_data, "H5Dread");
|
||
} // end of try block
|
||
catch (Exception error)
|
||
{
|
||
// all the exceptions caused by negative returned values by C APIs
|
||
CHECK(FAIL, FAIL, error.getCFuncName());
|
||
}
|
||
|
||
} /* test_h5s_scalar_read() */
|
||
|
||
/*-------------------------------------------------------------------------
|
||
*
|
||
* Function: test_h5s_compound_scalar_write
|
||
*
|
||
* Purpose: Test scalar H5S (dataspace) writing for compound
|
||
* datatypes
|
||
*
|
||
* Return: none
|
||
*
|
||
* Programmer: Binh-Minh Ribler (using C version)
|
||
* Mar 2001
|
||
*
|
||
* Modifications:
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
static void
|
||
test_h5s_compound_scalar_write(void)
|
||
{
|
||
|
||
/* Output message about test being performed */
|
||
MESSAGE(5, ("Testing Compound Dataspace Writing\n"));
|
||
|
||
try
|
||
{
|
||
/* Create file */
|
||
H5File fid1(DATAFILE, H5F_ACC_TRUNC);
|
||
|
||
/* Create the compound datatype. */
|
||
CompType tid1(sizeof(struct space4_struct));
|
||
space4_field1_off=HOFFSET(struct space4_struct, c1);
|
||
tid1.insertMember(SPACE4_FIELDNAME1, space4_field1_off,
|
||
PredType::NATIVE_SCHAR);
|
||
space4_field2_off=HOFFSET(struct space4_struct, u);
|
||
tid1.insertMember(SPACE4_FIELDNAME2, space4_field2_off,
|
||
PredType::NATIVE_UINT);
|
||
space4_field3_off=HOFFSET(struct space4_struct, f);
|
||
tid1.insertMember(SPACE4_FIELDNAME3, space4_field3_off,
|
||
PredType::NATIVE_FLOAT);
|
||
space4_field4_off=HOFFSET(struct space4_struct, c2);
|
||
tid1.insertMember(SPACE4_FIELDNAME4, space4_field4_off,
|
||
PredType::NATIVE_SCHAR);
|
||
|
||
/* Create scalar dataspace */
|
||
DataSpace sid1(SPACE3_RANK, NULL);
|
||
|
||
// Get the number of dataspace elements
|
||
hssize_t n = sid1.getSimpleExtentNpoints();
|
||
VERIFY(n, 1, "H5Sget_simple_extent_npoints");
|
||
|
||
// Get the logical rank of the dataspace
|
||
int ndims = sid1.getSimpleExtentNdims();
|
||
VERIFY(ndims, SPACE3_RANK, "H5Sget_simple_extent_ndims");
|
||
|
||
hsize_t tdims[4]; /* Dimension array to test with */
|
||
ndims = sid1.getSimpleExtentDims(tdims);
|
||
VERIFY(ndims, 0, "H5Sget_simple_extent_dims");
|
||
|
||
/* Create a dataset */
|
||
DataSet dataset = fid1.createDataSet("Dataset1", tid1, sid1);
|
||
|
||
dataset.write(&space4_data, tid1);
|
||
} // end of try block
|
||
catch (Exception error)
|
||
{
|
||
// all the exceptions caused by negative returned values by C APIs
|
||
CHECK(FAIL, FAIL, error.getCFuncName());
|
||
}
|
||
|
||
} /* test_h5s_compound_scalar_write() */
|
||
|
||
/*-------------------------------------------------------------------------
|
||
*
|
||
* Function: test_h5s_compound_scalar_read
|
||
*
|
||
* Purpose: Test scalar H5S (dataspace) reading for compound
|
||
* datatypes
|
||
*
|
||
* Return: none
|
||
*
|
||
* Programmer: Binh-Minh Ribler (using C version)
|
||
* Mar 2001
|
||
*
|
||
* Modifications:
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
static void
|
||
test_h5s_compound_scalar_read(void)
|
||
{
|
||
hsize_t tdims[4]; /* Dimension array to test with */
|
||
|
||
/* Output message about test being performed */
|
||
MESSAGE(5, ("Testing Compound Dataspace Reading\n"));
|
||
try
|
||
{
|
||
/* Create file */
|
||
H5File fid1(DATAFILE, H5F_ACC_RDWR);
|
||
|
||
/* Create a dataset */
|
||
DataSet dataset = fid1.openDataSet("Dataset1");
|
||
|
||
DataSpace sid1 = dataset.getSpace();
|
||
|
||
// Get the number of dataspace elements
|
||
hssize_t n = sid1.getSimpleExtentNpoints();
|
||
VERIFY(n, 1, "H5Sget_simple_extent_npoints");
|
||
|
||
// Get the logical rank of the dataspace
|
||
int ndims = sid1.getSimpleExtentNdims();
|
||
VERIFY(ndims, SPACE3_RANK, "H5Sget_simple_extent_ndims");
|
||
|
||
ndims = sid1.getSimpleExtentDims(tdims);
|
||
VERIFY(ndims, 0, "H5Sget_simple_extent_dims");
|
||
|
||
// Get the datatype of this dataset.
|
||
CompType type(dataset);
|
||
|
||
struct space4_struct rdata; /* Scalar data read in */
|
||
dataset.read(&rdata, type);
|
||
|
||
// Verify read data
|
||
if(HDmemcmp(&space4_data,&rdata,sizeof(struct space4_struct)))
|
||
{
|
||
cout << "scalar data different: space4_data.c1="
|
||
<< space4_data.c1 << ", read_data4.c1=" << rdata.c1 << endl;
|
||
cout << "scalar data different: space4_data.u="
|
||
<< space4_data.u << ", read_data4.u=" << rdata.u << endl;
|
||
cout << "scalar data different: space4_data.f="
|
||
<< space4_data.f << ", read_data4.f=" << rdata.f << endl;
|
||
TestErrPrintf("scalar data different: space4_data.c1=%c, read_data4.c1=%c\n",
|
||
space4_data.c1, rdata.c2);
|
||
} /* end if */
|
||
} // end of try block
|
||
catch (Exception error)
|
||
{
|
||
// all the exceptions caused by negative returned values by C APIs
|
||
CHECK(FAIL, FAIL, error.getCFuncName());
|
||
}
|
||
} /* test_h5s_compound_scalar_read() */
|
||
|
||
/*-------------------------------------------------------------------------
|
||
*
|
||
* Function: test_h5s
|
||
*
|
||
* Purpose: Main H5S (dataspace) testing routine
|
||
*
|
||
* Return: none
|
||
*
|
||
* Programmer: Binh-Minh Ribler (using C version)
|
||
* Mar 2001
|
||
*
|
||
* Modifications:
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
void
|
||
test_h5s(void)
|
||
{
|
||
/* Output message about test being performed */
|
||
MESSAGE(5, ("Testing Dataspaces\n"));
|
||
|
||
test_h5s_basic(); /* Test basic H5S code */
|
||
test_h5s_scalar_write(); /* Test scalar H5S writing code */
|
||
test_h5s_scalar_read(); /* Test scalar H5S reading code */
|
||
test_h5s_compound_scalar_write(); /* Test compound datatype scalar H5S writing code */
|
||
test_h5s_compound_scalar_read(); /* Test compound datatype scalar H5S reading code */
|
||
} /* test_h5s() */
|
||
|
||
|
||
/*-------------------------------------------------------------------------
|
||
* Function: cleanup_h5s
|
||
*
|
||
* Purpose: Cleanup temporary test files
|
||
*
|
||
* Return: none
|
||
*
|
||
* Programmer: Albert Cheng
|
||
* July 2, 1998
|
||
*
|
||
* Modifications:
|
||
*
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
void
|
||
cleanup_h5s(void)
|
||
{
|
||
remove(DATAFILE);
|
||
}
|
||
|