Python client library for HSDS
This repository contains library, test, and examples of h5pyd - a Python package for HSDS (Highly Scalable Data Service), or other HDF REST compatible web services.
The h5pyd library provides a high-level interface to the REST specification that is generally easier to use than invoking http calls directly.
This package is based on the popular h5py package and aims to be source compatible with the h5py high level interface.
- Main website: http://www.hdfgroup.org
- Source code: https://github.com/HDFGroup/h5pyd
- Forum: https://forum.hdfgroup.org/c/hsds
- Documentation: TBD (but http://docs.h5py.org/en/latest/ should be helpful)
- HSDS: https://github.com/HDFGroup/hsds
- HDF5-JSON: https://github.com/HDFGroup/hdf5-json
- h5py: https://github.com/h5py/h5py
- REST API Documentation: https://github.com/HDFGroup/hdf-rest-api
Via pip::
pip install h5pyd
From a release tarball or Git checkout::
pip install .
Run hsconfigure
to setup the connection info (endpoint, username, and password) to HSDS.
If you don't have access to an HSDS instance, you can easily setup your own HSDS instance.
See https://github.com/HDFGroup/hsds for instructions on installing and running HSDS
on locally or in the cloud.
The h5pyd package can be used without an explicit HSDS connection. Rather, the storage system will be accessed directly.
To use in direct mode, set the HS_ENDPOINT to "local" (or "local[n]" where n is the number of desired sub-processes).
For direct mode, some additional environment variables are needed to be defined:
BUCKET_NAME
- name of the S3 Bucket, Azure Container, or Posix top level folder
To use "local" mode with S3, define these variables:
AWS_S3_GATEWAY
- AWS S3 endpoint, e.g.:https://s3.us-west-2.amazonaws.com
AWS_REGION
- Region where the Lambda function is installed, e.g.:us-west-2
AWS_SECRET_ACCESS_KEY
- Your AWS secret access AWS_SECRET_ACCESS_KEYAWS_ACCESS_KEY_ID
- Your AWS access key ID
To use "local" mode with Azure, defined these variables:
AZURE_CONNECTION_STRING
- The connection string for your Azure storage account
To use "local" with Posix storage, define these variables:
ROOT_DIR
- The top level directory used for storage (i.e. the parent directory of "buckets")
Several utility applications are included with this package:
hsconfigure
- save endpoint, username, and password in config fileshsacl
- read/update ACL (access control list) for a given folder or domainhscopy
- copy a domainhsrm
- delete a domain or folderhsdiff
- compare HDF5 file with HSDS domainhsget
- create an HDF5 file from HSDS domainhsinfo
- get server status or domain detailshsload
- upload an HDF5 file to an HSDS domainhsls
- list contents of domain or folderhsmv
- change the naame of a domainhstouch
- create new domain or folder
Use the --help
option to get usage information for each command.
By default the test suite will attempt to connect to a local instance of HSDS with the
http://localhost:5101
endpoint.
Use the following environment variables as needed to modify the default configuration for the test suite:
HS_ENDPOINT
- "http://127.0.0.1:5000" for HSDS installed locally or appropriate remote endpointHS_USERNAME
- "test_user1" or your preferred usernameHS_PASSWORD
- password for above username - "test" for test_user1 with local HSDS installTEST2_USERNAME
- "test_user2" or your preffered usernameTEST2_PASSWORD
- password for above username - "test" for test_user2 with local HSDS installH5PYD_TEST_FOLDER
- server folder that will be used for generated files. Example: "/home/test_user1/h5pyd_test/". Use hstouch command to create the folder before running the test, e.g.:$ hstouch /home/test_user1/h5pyd_test/
Run: $python testall.py
to execute the test suite.
Just remove the install directory and all contents to uninstall.
Create new issues at http://github.com/HDFGroup/h5pyd/issues for any problems you find.
For general questions/feedback, please use the HSDS forum (https://forum.hdfgroup.org/c/hsds).