[svn-r9321] Snapshot version 1.7 release 38

This commit is contained in:
HDF Admin
2004-09-26 04:47:53 -05:00
parent 74c322019f
commit 643811be02
6 changed files with 80 additions and 18 deletions

View File

@@ -1,4 +1,4 @@
HDF5 version 1.7.37 released on Sun Sep 12 04:03:26 CDT 2004
HDF5 version 1.7.38 released on Sun Sep 26 04:46:18 CDT 2004
================================================================================
@@ -180,10 +180,15 @@ Bug Fixes since HDF5-1.6.0 release
Library
-------
- Fixed parallel bug in which some processes attempted collective
I/O while others did independent I/O. Bug appeared when some
processes used point selections, and others didn't. JRM - 2004/9/15
- Corrected error where dataset region references were written in an
incorrect way on Cray machines. PVN & QAK - 2004/09/13
- The H5Tget_native_type now determines the native type for integers
based on the precision. This is to avoid cases of wrongly converting
an int to a short in machines that have a short of 8 bytes but with
32bit precision (e.g Cray SV1). PVN - 2004/09/07
based on the precision. This is to avoid cases of wrongly converting
an int to a short in machines that have a short of 8 bytes but with
32bit precision (e.g Cray SV1). PVN - 2004/09/07
- Changed H5Dread() to not overwrite data in an application's buffer
with garbage when accessing a chunked dataset with an undefined
fill value and an unwritten chunk is uncountered. QAK - 2004/08/25
@@ -627,3 +632,60 @@ ftp://hdf.ncsa.uiuc.edu/pub/outgoing/hdf5/hdf5-1.6.2/F90_source_for_Crays
* Information about building with PGI and Intel compilers is available in
INSTALL file sections 5.7 and 5.8
* On at least one system, (SDSC DataStar), the scheduler (in this case
LoadLeveler) sends job status updates to standard error when you run
any executable that was compiled with the parallel compilers.
This causes problems when running "make check" on parallel builds, as
many of the tool tests function by saving the output from test runs,
and comparing it to an exemplar.
The best solution is to reconfigure the target system so it no longer
inserts the extra text. However, this may not be practical.
In such cases, one solution is to "setenv HDF5_Make_Ignore yes" prior to
the configure and build. This will cause "make check" to continue after
detecting errors in the tool tests. However, in the case of SDSC DataStar,
it also leaves you with some 150 "failed" tests to examine by hand.
A second solution is to write a script to run serial tests and filter
out the text added by the scheduler. A sample script used on SDSC
DataStar is given below, but you will probably have to customize it
for your installation.
Observe that the basic idea is to insert the script as the first item
on the command line which executes the the test. The script then
executes the test and filters out the offending text before passing
it on.
#!/bin/csh
set STDOUT_FILE=~/bin/serial_filter.stdout
set STDERR_FILE=~/bin/serial_filter.stderr
rm -f $STDOUT_FILE $STDERR_FILE
($* > $STDOUT_FILE) >& $STDERR_FILE
set RETURN_VALUE=$status
cat $STDOUT_FILE
tail +3 $STDERR_FILE
exit $RETURN_VALUE
You get the HDF make files and test scipts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
Note that the RUNSERIAL environment variable exists so that we can
can prefix serial runs as necessary on the target system. On DataStar,
no prefix is necessary. However on an MPICH system, the prefix might
have to be set to something like "/usr/local/mpi/bin/mpirun -np 1" to
get the serial tests to run at all.
In such cases, you will have to include the regular prefix in your
filter script.