From bca28060550bbdbaf0babb97a19cd44f18c04723 Mon Sep 17 00:00:00 2001 From: Allen Byrne <50328838+byrnHDF@users.noreply.github.com> Date: Thu, 29 Aug 2024 16:27:44 -0500 Subject: [PATCH] Update URL documentation links to support site (#4781) --- CITATION.cff | 2 +- HDF5Examples/README.md | 12 +- README.md | 2 +- config/cmake/README.md.cmake.in | 2 +- doc/parallel-compression.md | 55 +- doxygen/aliases | 21 +- doxygen/dox/About.dox | 2 +- doxygen/dox/DDLBNF110.dox | 2 - doxygen/dox/DDLBNF112.dox | 2 - doxygen/dox/DDLBNF114.dox | 2 - doxygen/dox/ExamplesAPI.dox | 141 ++-- doxygen/dox/GettingStarted.dox | 4 +- doxygen/dox/IntroHDF5.dox | 6 +- doxygen/dox/LearnBasics.dox | 8 +- doxygen/dox/LearnBasics3.dox | 16 +- doxygen/dox/LearnHDFView.dox | 4 +- doxygen/dox/Overview.dox | 6 +- doxygen/dox/Specifications.dox | 2 +- doxygen/dox/TechnicalNotes.dox | 269 +++++++- .../ThreadSafe.dox} | 613 ++++++------------ doxygen/dox/ViewTools.dox | 2 +- doxygen/dox/high_level/extension.dox | 12 +- doxygen/examples/FileFormat.html | 1 - doxygen/examples/Filters.html | 450 ------------- doxygen/examples/H5.format.1.0.html | 4 +- doxygen/examples/H5.format.1.1.html | 6 +- doxygen/examples/H5DS_Spec.pdf | Bin 0 -> 746769 bytes doxygen/examples/IOFlow.html | 1 - .../LibraryReleaseVersionNumbers.html | 2 +- doxygen/examples/intro_SWMR.html | 103 --- doxygen/examples/intro_VDS.html | 72 -- doxygen/examples/tables/propertyLists.dox | 8 +- doxygen/hdf5doxy_layout.xml | 6 +- hl/src/H5DOpublic.h | 2 +- hl/src/H5DSpublic.h | 6 +- hl/src/H5LTpublic.h | 10 +- java/src/hdf/overview.html | 2 +- java/src/jni/exceptionImp.c | 6 - java/src/jni/h5Constants.c | 6 - java/src/jni/h5Imp.c | 6 - java/src/jni/h5aImp.c | 6 - java/src/jni/h5dImp.c | 6 - java/src/jni/h5eImp.c | 3 - java/src/jni/h5fImp.c | 6 - java/src/jni/h5gImp.c | 6 - java/src/jni/h5iImp.c | 6 - java/src/jni/h5jni.h | 6 - java/src/jni/h5lImp.c | 6 - java/src/jni/h5oImp.c | 6 - java/src/jni/h5pACPLImp.c | 6 - java/src/jni/h5pDAPLImp.c | 6 - java/src/jni/h5pDCPLImp.c | 6 - java/src/jni/h5pDXPLImp.c | 6 - java/src/jni/h5pFAPLImp.c | 6 - java/src/jni/h5pFCPLImp.c | 6 - java/src/jni/h5pGAPLImp.c | 6 - java/src/jni/h5pGCPLImp.c | 6 - java/src/jni/h5pImp.c | 6 - java/src/jni/h5pLAPLImp.c | 6 - java/src/jni/h5pLCPLImp.c | 6 - java/src/jni/h5pOCPLImp.c | 6 - java/src/jni/h5pOCpyPLImp.c | 6 - java/src/jni/h5pStrCPLImp.c | 6 - java/src/jni/h5plImp.c | 6 - java/src/jni/h5rImp.c | 5 - java/src/jni/h5sImp.c | 6 - java/src/jni/h5tImp.c | 6 - java/src/jni/h5util.c | 6 - java/src/jni/h5util.h | 6 - java/src/jni/h5vlImp.c | 6 - java/src/jni/h5zImp.c | 6 - java/src/jni/nativeData.c | 5 - release_docs/INSTALL | 2 +- release_docs/INSTALL_Autotools.txt | 4 +- release_docs/INSTALL_CMake.txt | 2 +- release_docs/INSTALL_parallel | 2 +- release_docs/RELEASE.txt | 6 +- release_docs/RELEASE_PROCESS.md | 66 +- src/H5Amodule.h | 24 +- src/H5Dmodule.h | 8 +- src/H5Emodule.h | 91 ++- src/H5Fmodule.h | 29 +- src/H5Gmodule.h | 2 +- src/H5PLmodule.h | 8 +- src/H5Pmodule.h | 4 +- src/H5Ppublic.h | 2 +- src/H5Smodule.h | 203 +++++- src/H5Tmodule.h | 13 +- src/H5module.h | 8 +- tools/src/h5dump/h5dump.c | 2 +- tools/test/h5dump/expected/h5dump-help.txt | 2 +- .../pbits/tnofilename-with-packed-bits.ddl | 2 +- .../expected/pbits/tpbitsIncomplete.ddl | 2 +- .../expected/pbits/tpbitsLengthExceeded.ddl | 2 +- .../expected/pbits/tpbitsLengthPositive.ddl | 2 +- .../expected/pbits/tpbitsMaxExceeded.ddl | 2 +- .../expected/pbits/tpbitsOffsetExceeded.ddl | 2 +- .../expected/pbits/tpbitsOffsetNegative.ddl | 2 +- 98 files changed, 948 insertions(+), 1607 deletions(-) rename doxygen/{examples/ThreadSafeLibrary.html => dox/ThreadSafe.dox} (62%) delete mode 100644 doxygen/examples/Filters.html create mode 100644 doxygen/examples/H5DS_Spec.pdf delete mode 100644 doxygen/examples/intro_SWMR.html delete mode 100644 doxygen/examples/intro_VDS.html diff --git a/CITATION.cff b/CITATION.cff index 4e611a57468..f7eaf133318 100644 --- a/CITATION.cff +++ b/CITATION.cff @@ -9,4 +9,4 @@ authors: website: 'https://www.hdfgroup.org' repository-code: 'https://github.com/HDFGroup/hdf5' url: 'https://www.hdfgroup.org/HDF5/' -repository-artifact: 'https://www.hdfgroup.org/downloads/hdf5/' +repository-artifact: 'https://support.hdfgroup.org/downloads/HDF5' diff --git a/HDF5Examples/README.md b/HDF5Examples/README.md index 2f0090ba02c..e70a1a792d0 100644 --- a/HDF5Examples/README.md +++ b/HDF5Examples/README.md @@ -48,17 +48,17 @@ HDF5 SNAPSHOTS, PREVIOUS RELEASES AND SOURCE CODE -------------------------------------------- Full Documentation and Programming Resources for this HDF5 can be found at - https://portal.hdfgroup.org/documentation/index.html + https://support.hdfgroup.org/documentation/HDF5/index.html Periodically development code snapshots are provided at the following URL: - - https://gamma.hdfgroup.org/ftp/pub/outgoing/hdf5/snapshots/ + + https://github.com/HDFGroup/hdf5/releases Source packages for current and previous releases are located at: - - https://portal.hdfgroup.org/downloads/ + + https://support.hdfgroup.org/releases/hdf5/downloads/ Development code is available at our Github location: - + https://github.com/HDFGroup/hdf5.git diff --git a/README.md b/README.md index 36f853c6990..1be794abcb1 100644 --- a/README.md +++ b/README.md @@ -108,7 +108,7 @@ Periodically development code snapshots are provided at the following URL: Source packages for current and previous releases are located at: - https://portal.hdfgroup.org/Downloads + https://support.hdfgroup.org/downloads/HDF5 Development code is available at our Github location: diff --git a/config/cmake/README.md.cmake.in b/config/cmake/README.md.cmake.in index 7f6af3646a2..3f541e4e8a3 100644 --- a/config/cmake/README.md.cmake.in +++ b/config/cmake/README.md.cmake.in @@ -75,6 +75,6 @@ For more information see USING_CMake_Examples.txt in the install folder. =========================================================================== Documentation for this release can be found at the following URL: - https://portal.hdfgroup.org/documentation/index.html#hdf5 + https://support.hdfgroup.org/hdf5/@HDF5_PACKAGE_NAME@-@HDF5_PACKAGE_VERSION@/documentation/doxygen/index.html Bugs should be reported to help@hdfgroup.org. diff --git a/doc/parallel-compression.md b/doc/parallel-compression.md index 48ed4c3c37d..77126d6acf6 100644 --- a/doc/parallel-compression.md +++ b/doc/parallel-compression.md @@ -64,9 +64,9 @@ H5Dwrite(..., dxpl_id, ...); The following are two simple examples of using the parallel compression feature: -[ph5_filtered_writes.c](https://github.com/HDFGroup/hdf5/blob/develop/HDF5Examples/C/H5PAR/ph5_filtered_writes.c) +[ph5_filtered_writes.c][u1] -[ph5_filtered_writes_no_sel.c](https://github.com/HDFGroup/hdf5/blob/develop/HDF5Examples/C/H5PAR/ph5_filtered_writes_no_sel.c) +[ph5_filtered_writes_no_sel.c][u2] The former contains simple examples of using the parallel compression feature to write to compressed datasets, while the @@ -79,7 +79,7 @@ participate in the collective write call. ## Multi-dataset I/O support The parallel compression feature is supported when using the -multi-dataset I/O API routines ([H5Dwrite_multi](https://hdfgroup.github.io/hdf5/develop/group___h5_d.html#gaf6213bf3a876c1741810037ff2bb85d8)/[H5Dread_multi](https://hdfgroup.github.io/hdf5/develop/group___h5_d.html#ga8eb1c838aff79a17de385d0707709915)), but the +multi-dataset I/O API routines ([H5Dwrite_multi][u3]/[H5Dread_multi][u4]), but the following should be kept in mind: - Parallel writes to filtered datasets **must** still be collective, @@ -99,7 +99,7 @@ following should be kept in mind: ## Incremental file space allocation support -HDF5's [file space allocation time](https://hdfgroup.github.io/hdf5/develop/group___d_c_p_l.html#ga85faefca58387bba409b65c470d7d851) +HDF5's [file space allocation time][u5] is a dataset creation property that can have significant effects on application performance, especially if the application uses parallel HDF5. In a serial HDF5 application, the default file space @@ -118,7 +118,7 @@ While this strategy has worked in the past, it has some noticeable drawbacks. For one, the larger the chunked dataset being created, the more noticeable overhead there will be during dataset creation as all of the data chunks are being allocated in the HDF5 file. -Further, these data chunks will, by default, be [filled](https://hdfgroup.github.io/hdf5/develop/group___d_c_p_l.html#ga4335bb45b35386daa837b4ff1b9cd4a4) +Further, these data chunks will, by default, be [filled][u6] with HDF5's default fill data value, leading to extraordinary dataset creation overhead and resulting in pre-filling large portions of a dataset that the application might have been planning @@ -126,12 +126,12 @@ to overwrite anyway. Even worse, there will be more initial overhead from compressing that fill data before writing it out, only to have it read back in, unfiltered and modified the first time a chunk is written to. In the past, it was typically suggested that parallel -HDF5 applications should use [H5Pset_fill_time](https://hdfgroup.github.io/hdf5/develop/group___d_c_p_l.html#ga6bd822266b31f86551a9a1d79601b6a2) +HDF5 applications should use [H5Pset_fill_time][u7] with a value of `H5D_FILL_TIME_NEVER` in order to disable writing of the fill value to dataset chunks, but this isn't ideal if the application actually wishes to make use of fill values. -With [improvements made](https://www.hdfgroup.org/2022/03/parallel-compression-improvements-in-hdf5-1-13-1/) +With [improvements made][u8] to the parallel compression feature for the HDF5 1.13.1 release, "incremental" file space allocation is now the default for datasets created in parallel *only if they have filters applied to them*. @@ -154,7 +154,7 @@ optimal performance out of the parallel compression feature. ### Begin with a good chunking strategy -[Starting with a good chunking strategy](https://portal.hdfgroup.org/documentation/hdf5-docs/chunking_in_hdf5.html) +[Starting with a good chunking strategy][u9] will generally have the largest impact on overall application performance. The different chunking parameters can be difficult to fine-tune, but it is essential to start with a well-performing @@ -166,7 +166,7 @@ chosen chunk size becomes a very important factor when compression is involved, as data chunks have to be completely read and re-written to perform partial writes to the chunk. -[Improving I/O performance with HDF5 compressed datasets](https://docs.hdfgroup.org/archive/support/HDF5/doc/TechNotes/TechNote-HDF5-ImprovingIOPerformanceCompressedDatasets.pdf) +[Improving I/O performance with HDF5 compressed datasets][u10] is a useful reference for more information on getting good performance when using a chunked dataset layout. @@ -220,14 +220,14 @@ chunks to end up at addresses in the file that do not align well with the underlying file system, possibly leading to poor performance. As an example, Lustre performance is generally good when writes are aligned with the chosen stripe size. -The HDF5 application can use [H5Pset_alignment](https://hdfgroup.github.io/hdf5/develop/group___f_a_p_l.html#gab99d5af749aeb3896fd9e3ceb273677a) +The HDF5 application can use [H5Pset_alignment][u11] to have a bit more control over where objects in the HDF5 file end up. However, do note that setting the alignment of objects generally wastes space in the file and has the potential to dramatically increase its resulting size, so caution should be used when choosing the alignment parameters. -[H5Pset_alignment](https://hdfgroup.github.io/hdf5/develop/group___f_a_p_l.html#gab99d5af749aeb3896fd9e3ceb273677a) +[H5Pset_alignment][u11] has two parameters that control the alignment of objects in the HDF5 file, the "threshold" value and the alignment value. The threshold value specifies that any object greater @@ -264,19 +264,19 @@ in a file, this can create significant amounts of free space in the file over its lifetime and eventually cause performance issues. -An HDF5 application can use [H5Pset_file_space_strategy](https://hdfgroup.github.io/hdf5/develop/group___f_c_p_l.html#ga167ff65f392ca3b7f1933b1cee1b9f70) +An HDF5 application can use [H5Pset_file_space_strategy][u12] with a value of `H5F_FSPACE_STRATEGY_PAGE` to enable the paged aggregation feature, which can accumulate metadata and raw data for dataset data chunks into well-aligned, configurably sized "pages" for better performance. However, note that using the paged aggregation feature will cause any setting from -[H5Pset_alignment](https://hdfgroup.github.io/hdf5/develop/group___f_a_p_l.html#gab99d5af749aeb3896fd9e3ceb273677a) +[H5Pset_alignment][u11] to be ignored. While an application should be able to get -comparable performance effects by [setting the size of these pages](https://hdfgroup.github.io/hdf5/develop/group___f_c_p_l.html#gad012d7f3c2f1e1999eb1770aae3a4963) to be equal to the value that -would have been set for [H5Pset_alignment](https://hdfgroup.github.io/hdf5/develop/group___f_a_p_l.html#gab99d5af749aeb3896fd9e3ceb273677a), +comparable performance effects by [setting the size of these pages][u13] +to be equal to the value that would have been set for [H5Pset_alignment][u11], this may not necessarily be the case and should be studied. -Note that [H5Pset_file_space_strategy](https://hdfgroup.github.io/hdf5/develop/group___f_c_p_l.html#ga167ff65f392ca3b7f1933b1cee1b9f70) +Note that [H5Pset_file_space_strategy][u12] has a `persist` parameter. This determines whether or not the file free space manager should include extra metadata in the HDF5 file about free space sections in the file. If this @@ -300,12 +300,12 @@ hid_t file_id = H5Fcreate("file.h5", H5F_ACC_TRUNC, fcpl_id, fapl_id); While the parallel compression feature requires that the HDF5 application set and maintain collective I/O at the application -interface level (via [H5Pset_dxpl_mpio](https://hdfgroup.github.io/hdf5/develop/group___d_x_p_l.html#ga001a22b64f60b815abf5de8b4776f09e)), +interface level (via [H5Pset_dxpl_mpio][u14]), it does not require that the actual MPI I/O that occurs at the lowest layers of HDF5 be collective; independent I/O may perform better depending on the application I/O patterns and parallel file system performance, among other factors. The -application may use [H5Pset_dxpl_mpio_collective_opt](https://hdfgroup.github.io/hdf5/develop/group___d_x_p_l.html#gacb30d14d1791ec7ff9ee73aa148a51a3) +application may use [H5Pset_dxpl_mpio_collective_opt][u15] to control this setting and see which I/O method provides the best performance. @@ -318,7 +318,7 @@ H5Dwrite(..., dxpl_id, ...); ### Runtime HDF5 Library version -An HDF5 application can use the [H5Pset_libver_bounds](https://hdfgroup.github.io/hdf5/develop/group___f_a_p_l.html#gacbe1724e7f70cd17ed687417a1d2a910) +An HDF5 application can use the [H5Pset_libver_bounds][u16] routine to set the upper and lower bounds on library versions to use when creating HDF5 objects. For parallel compression specifically, setting the library version to the latest available @@ -332,3 +332,20 @@ H5Pset_libver_bounds(fapl_id, H5F_LIBVER_LATEST, H5F_LIBVER_LATEST); hid_t file_id = H5Fcreate("file.h5", H5F_ACC_TRUNC, H5P_DEFAULT, fapl_id); ... ``` + +[u1]: https://github.com/HDFGroup/hdf5/blob/develop/HDF5Examples/C/H5PAR/ph5_filtered_writes.c +[u2]: https://github.com/HDFGroup/hdf5/blob/develop/HDF5Examples/C/H5PAR/ph5_filtered_writes_no_sel.c +[u3]: https://hdfgroup.github.io/hdf5/develop/group___h5_d.html#gaf6213bf3a876c1741810037ff2bb85d8 +[u4]: https://hdfgroup.github.io/hdf5/develop/group___h5_d.html#ga8eb1c838aff79a17de385d0707709915 +[u5]: https://hdfgroup.github.io/hdf5/develop/group___d_c_p_l.html#ga85faefca58387bba409b65c470d7d851 +[u6]: https://hdfgroup.github.io/hdf5/develop/group___d_c_p_l.html#ga4335bb45b35386daa837b4ff1b9cd4a4 +[u7]: https://hdfgroup.github.io/hdf5/develop/group___d_c_p_l.html#ga6bd822266b31f86551a9a1d79601b6a2 +[u8]: https://support.hdfgroup.org/documentation/HDF5/parallel-compression-improvements-in-hdf5-1-13-1 +[u9]: https://support.hdfgroup.org/documentation/HDF5/chunking_in_hdf5.html +[u10]: https://support.hdfgroup.org/documentation/HDF5/technotes/TechNote-HDF5-ImprovingIOPerformanceCompressedDatasets.pdf +[u11]: https://hdfgroup.github.io/hdf5/develop/group___f_a_p_l.html#gab99d5af749aeb3896fd9e3ceb273677a +[u12]: https://hdfgroup.github.io/hdf5/develop/group___f_c_p_l.html#ga167ff65f392ca3b7f1933b1cee1b9f70 +[u13]: https://hdfgroup.github.io/hdf5/develop/group___f_c_p_l.html#gad012d7f3c2f1e1999eb1770aae3a4963 +[u14]: https://hdfgroup.github.io/hdf5/develop/group___d_x_p_l.html#ga001a22b64f60b815abf5de8b4776f09e +[u15]: https://hdfgroup.github.io/hdf5/develop/group___d_x_p_l.html#gacb30d14d1791ec7ff9ee73aa148a51a3 +[u16]: https://hdfgroup.github.io/hdf5/develop/group___f_a_p_l.html#gacbe1724e7f70cd17ed687417a1d2a910 diff --git a/doxygen/aliases b/doxygen/aliases index 4bb6e8c0792..24a496c2203 100644 --- a/doxygen/aliases +++ b/doxygen/aliases @@ -4,17 +4,16 @@ ALIASES += THG="The HDF Group" # Default URLs (Note that md files do not use any aliases) ################################################################################ # Default URL for HDF Group Files -ALIASES += HDFURL="docs.hdfgroup.org/hdf5" +ALIASES += HDFURL="support.hdfgroup.org" # URL for archived files -ALIASES += ARCURL="docs.hdfgroup.org/archive/support/HDF5/doc" +ALIASES += ARCURL="\HDFURL/archive/support/HDF5/doc" # URL for RFCs -ALIASES += RFCURL="docs.hdfgroup.org/hdf5/rfc" +ALIASES += RFCURL="\HDFURL/hdf5/rfc" # URL for documentation -ALIASES += DSPURL="portal.hdfgroup.org/display/HDF5" -ALIASES += DOCURL="portal.hdfgroup.org/documentation/hdf5-docs" +ALIASES += DOCURL="\HDFURL/releases/hdf5/documentation" # URL for downloads -ALIASES += DWNURL="portal.hdfgroup.org/downloads" -ALIASES += AEXURL="support.hdfgroup.org/ftp/HDF5/examples" +ALIASES += DWNURL="\HDFURL/releases/hdf5/downloads" +ALIASES += AEXURL="\HDFURL/archive/support/ftp/HDF5/examples" # doxygen subdir (develop, v1_14) ALIASES += DOXURL="hdfgroup.github.io/hdf5/develop" #branch name (develop, hdf5_1_14) @@ -259,13 +258,13 @@ ALIASES += sa_metadata_ops="\sa \li H5Pget_all_coll_metadata_ops() \li H5Pget_co ALIASES += ref_cons_semantics="Enabling a Strict Consistency Semantics Model in Parallel HDF5" ALIASES += ref_file_image_ops="HDF5 File Image Operations" -ALIASES += ref_filter_pipe="Data Flow Pipeline for H5Dread()" +ALIASES += ref_filter_pipe="Data Flow Pipeline for H5Dread()" ALIASES += ref_group_impls="Group implementations in HDF5" ALIASES += ref_h5lib_relver="HDF5 Library Release Version Numbers" -ALIASES += ref_mdc_in_hdf5="Metadata Caching in HDF5" -ALIASES += ref_mdc_logging="Metadata Cache Logging" +ALIASES += ref_mdc_in_hdf5="Metadata Caching in HDF5" +ALIASES += ref_mdc_logging="Metadata Cache Logging" ALIASES += ref_news_112="New Features in HDF5 Release 1.12" -ALIASES += ref_h5ocopy="Copying Committed Datatypes with H5Ocopy()" +ALIASES += ref_h5ocopy="Copying Committed Datatypes with H5Ocopy()" ALIASES += ref_sencode_fmt_change="RFC H5Sencode() / H5Sdecode() Format Change" ALIASES += ref_vlen_strings="\Emph{Creating variable-length string datatypes}" ALIASES += ref_vol_doc="VOL documentation" diff --git a/doxygen/dox/About.dox b/doxygen/dox/About.dox index e145516a30e..73010b0c3de 100644 --- a/doxygen/dox/About.dox +++ b/doxygen/dox/About.dox @@ -83,7 +83,7 @@ as a general reference. All custom commands for this project are located in the aliases -file in the doxygen +file in the doxygen subdirectory of the main HDF5 repo. The custom commands are grouped in sections. Find a suitable section for your command or diff --git a/doxygen/dox/DDLBNF110.dox b/doxygen/dox/DDLBNF110.dox index 6d6b67ef7fd..b392526417a 100644 --- a/doxygen/dox/DDLBNF110.dox +++ b/doxygen/dox/DDLBNF110.dox @@ -1,7 +1,5 @@ /** \page DDLBNF110 DDL in BNF through HDF5 1.10 -\todo Revise this & break it up! - \section intro110 Introduction This document contains the data description language (DDL) for an HDF5 file. The diff --git a/doxygen/dox/DDLBNF112.dox b/doxygen/dox/DDLBNF112.dox index cfe34c321f9..c6463c23d5c 100644 --- a/doxygen/dox/DDLBNF112.dox +++ b/doxygen/dox/DDLBNF112.dox @@ -1,7 +1,5 @@ /** \page DDLBNF112 DDL in BNF for HDF5 1.12 through HDF5 1.14.3 -\todo Revise this & break it up! - \section intro112 Introduction This document contains the data description language (DDL) for an HDF5 file. The diff --git a/doxygen/dox/DDLBNF114.dox b/doxygen/dox/DDLBNF114.dox index 61e9157e560..baa7a57fea6 100644 --- a/doxygen/dox/DDLBNF114.dox +++ b/doxygen/dox/DDLBNF114.dox @@ -1,7 +1,5 @@ /** \page DDLBNF114 DDL in BNF for HDF5 1.14.4 and above -\todo Revise this & break it up! - \section intro114 Introduction This document contains the data description language (DDL) for an HDF5 file. The diff --git a/doxygen/dox/ExamplesAPI.dox b/doxygen/dox/ExamplesAPI.dox index c48b00e6dbb..dbd24f4d888 100644 --- a/doxygen/dox/ExamplesAPI.dox +++ b/doxygen/dox/ExamplesAPI.dox @@ -30,7 +30,7 @@ Languages are C, Fortran, Java (JHI5), Java Object Package, Python (High Level), C FORTRAN Java -JavaObj +JavaObj MATLAB PyHigh PyLow
@@ -68,7 +68,7 @@ A brief introduction to Parallel HDF5. If you are new to HDF5 please see the @re | |||
-HDF5-1.10 New Features +New Features since HDF5-1.10 |
\li \ref VDS
diff --git a/doxygen/dox/IntroHDF5.dox b/doxygen/dox/IntroHDF5.dox
index 9ef55d3a573..6f3938ed8b2 100644
--- a/doxygen/dox/IntroHDF5.dox
+++ b/doxygen/dox/IntroHDF5.dox
@@ -262,7 +262,7 @@ FORTRAN routines are similar; they begin with “h5*” and end with “_f”.
| ||
Create a file | -C Fortran C++ Java Python + | C Fortran C++ Java Python | @@ -91,7 +91,7 @@ These examples (C, C++, Fortran, Java, Python) are provided in the HDF5 source c |
Create a group | -C Fortran C++ Java Python + | C Fortran C++ Java Python | @@ -107,7 +107,7 @@ These examples (C, C++, Fortran, Java, Python) are provided in the HDF5 source c |
Create datasets in a group | -C Fortran C++ Java Python + | C Fortran C++ Java Python | @@ -131,7 +131,7 @@ These examples (C, C++, Fortran, Java, Python) are provided in the HDF5 source c |
Create a chunked and compressed dataset | -C Fortran C++ Java Python + | C Fortran C++ Java Python | diff --git a/doxygen/dox/LearnBasics3.dox b/doxygen/dox/LearnBasics3.dox index 3e9dd8ea090..d853c83d742 100644 --- a/doxygen/dox/LearnBasics3.dox +++ b/doxygen/dox/LearnBasics3.dox @@ -183,7 +183,7 @@ to a new with a new layout. \section secLBDsetLayoutSource Sources of Information Chunking in HDF5 (See the documentation on Advanced Topics in HDF5) -\see \ref sec_plist in the HDF5 \ref UG. +see \ref sec_plist in the HDF5 \ref UG. |