Skip to content

Commit

Permalink
Enforce python 3.12 flake8 standard
Browse files Browse the repository at this point in the history
Added whitespace around arithmetic operators for flake8 on python 3.12
  • Loading branch information
mattjala committed Mar 19, 2024
1 parent 9a5a180 commit cfe68cd
Show file tree
Hide file tree
Showing 28 changed files with 65 additions and 65 deletions.
2 changes: 1 addition & 1 deletion hsds/attr_sn.py
Original file line number Diff line number Diff line change
Expand Up @@ -1085,7 +1085,7 @@ async def PUT_AttributeValue(request):
if binary_data:
npoints = getNumElements(np_shape)
if npoints * item_size != len(binary_data):
msg = f"Expected: {npoints*item_size} bytes, "
msg = f"Expected: {npoints * item_size} bytes, "
msg += f"but got {len(binary_data)}"
log.warn(msg)
raise HTTPBadRequest(reason=msg)
Expand Down
2 changes: 1 addition & 1 deletion hsds/chunk_dn.py
Original file line number Diff line number Diff line change
Expand Up @@ -253,7 +253,7 @@ async def PUT_Chunk(request):
# regular chunk update
# check that the content_length is what we expect
if itemsize != "H5T_VARIABLE":
log.debug(f"expected content_length: {num_elements*itemsize}")
log.debug(f"expected content_length: {num_elements * itemsize}")
log.debug(f"actual content_length: {request.content_length}")

actual = request.content_length
Expand Down
4 changes: 2 additions & 2 deletions hsds/chunk_sn.py
Original file line number Diff line number Diff line change
Expand Up @@ -916,7 +916,7 @@ async def PUT_Value(request):

for page_number in range(len(pages)):
page = pages[page_number]
msg = f"streaming request data for page: {page_number+1} of {len(pages)}, "
msg = f"streaming request data for page: {page_number + 1} of {len(pages)}, "
msg += f"selection: {page}"
log.info(msg)
kwargs = {"page_number": page_number, "page": page}
Expand Down Expand Up @@ -1088,7 +1088,7 @@ async def GET_Value(request):
try:
for page_number in range(len(pages)):
page = pages[page_number]
msg = f"streaming response data for page: {page_number+1} "
msg = f"streaming response data for page: {page_number + 1} "
msg += f"of {len(pages)}, selection: {page}"
log.info(msg)

Expand Down
2 changes: 1 addition & 1 deletion hsds/chunklocator.py
Original file line number Diff line number Diff line change
Expand Up @@ -231,4 +231,4 @@ def main():
log.info('done')
stop_time = time.time()
log.info(f"chunklocator stop: {stop_time:.2f}")
log.info(f"chunklocator elapsed: {(stop_time-start_time):.2f}")
log.info(f"chunklocator elapsed: {(stop_time - start_time):.2f}")
8 changes: 4 additions & 4 deletions hsds/hsds_app.py
Original file line number Diff line number Diff line change
Expand Up @@ -157,7 +157,7 @@ def __init__(
sn_url = f"http+unix://{socket_url}sn_1.sock"

for i in range(dn_count):
socket_name = f"dn_{(i+1)}.sock"
socket_name = f"dn_{(i + 1)}.sock"
dn_url = f"http+unix://{socket_url}{socket_name}"
self._dn_urls.append(dn_url)
self._socket_paths.append(f"{socket_dir}{socket_name}")
Expand All @@ -167,7 +167,7 @@ def __init__(
sn_url = f"http://{host}:{sn_port}"
dn_port = 6101 # TBD: pull this from config
for i in range(dn_count):
dn_url = f"http://{host}:{dn_port+i}"
dn_url = f"http://{host}:{dn_port + i}"
self._dn_urls.append(dn_url)

# sort the ports so that node_number can be determined based on dn_url
Expand Down Expand Up @@ -296,7 +296,7 @@ def run(self):
py_exe,
cmd_path,
"--node_type=dn",
f"--log_prefix=dn{node_number+1} ",
f"--log_prefix=dn{node_number + 1} ",
]
pargs.append(f"--dn_urls={dn_urls_arg}")
pargs.append(f"--node_number={node_number}")
Expand Down Expand Up @@ -342,7 +342,7 @@ def run(self):
self.log.error(msg)
raise IOError(msg)

self.log.info(f"Ready after: {(time.time()-start_ts):4.2f} s")
self.log.info(f"Ready after: {(time.time() - start_ts):4.2f} s")
self._ready = True

def stop(self):
Expand Down
2 changes: 1 addition & 1 deletion hsds/servicenode.py
Original file line number Diff line number Diff line change
Expand Up @@ -270,7 +270,7 @@ def create_app():
if hs_username:
log.info(f"getCmdLine hs_username: {hs_username}")
if hs_password:
log.info(f"getCmdLine hs_password: {'*'*len(hs_password)}")
log.info(f"getCmdLine hs_password: {'*' * len(hs_password)}")
if hs_username:
setPassword(app, hs_username, hs_password)

Expand Down
2 changes: 1 addition & 1 deletion hsds/util/awsLambdaClient.py
Original file line number Diff line number Diff line change
Expand Up @@ -164,7 +164,7 @@ async def __aenter__(self):
msg = f"lambda.invoke({self.lambdaFunction} "
msg += f"start={start_time:.4f} "
msg += f"finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)
self.funcStats["inflight"] -= 1
msg = f"lambda.invoke - {self.funcStats['inflight']} "
Expand Down
12 changes: 6 additions & 6 deletions hsds/util/azureBlobClient.py
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ def __init__(self, app):
msg = "No connection string specified"
log.error(msg)
raise ValueError(msg)
log.info(f"Using azure_connection_string: {'*'*len(azure_connection_string)}")
log.info(f"Using azure_connection_string: {'*' * len(azure_connection_string)}")

self._client = BlobServiceClient.from_connection_string(azure_connection_string)

Expand Down Expand Up @@ -121,7 +121,7 @@ async def get_object(self, key, bucket=None, offset=0, length=-1):
finish_time = time.time()
msg = f"azureBlobClient.get_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f} "
msg += f"elapsed={finish_time - start_time:.4f} "
msg += f"bytes={len(data)}"
log.info(msg)
except CancelledError as cle:
Expand Down Expand Up @@ -179,7 +179,7 @@ async def put_object(self, key, data, bucket=None):
log.debug(f"put_object {key} returning: {rsp}")
msg = f"azureBlobClient.put_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += "elapsed={finish_time-start_time:.4f} "
msg += "elapsed={finish_time - start_time:.4f} "
msg += f"bytes={len(data)}"
log.info(msg)

Expand Down Expand Up @@ -231,7 +231,7 @@ async def delete_object(self, key, bucket=None):
finish_time = time.time()
msg = f"azureBlobClient.delete_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)

except CancelledError as cle:
Expand Down Expand Up @@ -309,7 +309,7 @@ async def is_object(self, key, bucket=None):

msg = f"azureBlobClient.is_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)

return found
Expand Down Expand Up @@ -365,7 +365,7 @@ async def get_key_stats(self, key, bucket=None):
key_stats["LastModified"] = datetime.datetime.timestamp(lm_dt)
msg = f"azureBlobClient.get_key_stats({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)

return key_stats
Expand Down
2 changes: 1 addition & 1 deletion hsds/util/dsetUtil.py
Original file line number Diff line number Diff line change
Expand Up @@ -585,7 +585,7 @@ def getSelectionList(select, dims):
raise ValueError(f"Invalid coordinate for dim {dim}")
if coord < 0 or coord >= extent:
msg = f"out of range coordinate for dim {dim}, {coord} "
msg += f"not in range: 0-{extent-1}"
msg += f"not in range: 0-{extent - 1}"
raise ValueError(msg)
if last_coord is not None and coord <= last_coord:
raise ValueError("coordinates must be increasing")
Expand Down
8 changes: 4 additions & 4 deletions hsds/util/fileClient.py
Original file line number Diff line number Diff line change
Expand Up @@ -131,7 +131,7 @@ async def get_object(self, key, bucket=None, offset=0, length=-1):
self._validateKey(key)

if length > 0:
range = f"bytes={offset}-{offset+length-1}"
range = f"bytes={offset} - {offset + length - 1}"
log.info(f"storage range request: {range}")

filepath = self._getFilePath(bucket, key)
Expand All @@ -152,7 +152,7 @@ async def get_object(self, key, bucket=None, offset=0, length=-1):
finish_time = time.time()
msg = f"fileClient.get_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f} bytes={len(data)}"
msg += f"elapsed={finish_time - start_time:.4f} bytes={len(data)}"
log.info(msg)
except FileNotFoundError:
msg = f"fileClient: {key} not found "
Expand Down Expand Up @@ -217,7 +217,7 @@ async def put_object(self, key, data, bucket=None):
finish_time = time.time()
msg = f"fileClient.put_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f} bytes={len(data)}"
msg += f"elapsed={finish_time - start_time:.4f} bytes={len(data)}"
log.info(msg)
write_rsp = self._getFileStats(filepath, data=data)
except IOError as ioe:
Expand Down Expand Up @@ -263,7 +263,7 @@ async def delete_object(self, key, bucket=None):
finish_time = time.time()
msg = f"fileClient.delete_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)

except IOError as ioe:
Expand Down
2 changes: 1 addition & 1 deletion hsds/util/idUtil.py
Original file line number Diff line number Diff line change
Expand Up @@ -305,7 +305,7 @@ def createNodeId(prefix, node_number=None):
node_id = "" # nothing too bad happens if this doesn't get set
if node_number is not None:
# just make an id based on the node_number
hash_key = f"{node_number+1:03d}"
hash_key = f"{node_number + 1:03d}"
else:
# use the container id if we are running inside docker
hash_key = getIdHash(str(uuid.uuid1()))
Expand Down
2 changes: 1 addition & 1 deletion hsds/util/lruCache.py
Original file line number Diff line number Diff line change
Expand Up @@ -309,7 +309,7 @@ def consistencyCheck(self):
if pos == 0:
raise ValueError(f"unexpected node: {node._id}")
if node._id != id_list[pos - 1]:
msg = f"expected node: {id_list[pos-1]} but found: {node._id}"
msg = f"expected node: {id_list[pos - 1]} but found: {node._id}"
raise ValueError(msg)
pos -= 1
node = node._prev
Expand Down
14 changes: 7 additions & 7 deletions hsds/util/s3Client.py
Original file line number Diff line number Diff line change
Expand Up @@ -269,7 +269,7 @@ async def get_object(self, key, bucket=None, offset=0, length=-1):

range = ""
if length > 0:
range = f"bytes={offset}-{offset+length-1}"
range = f"bytes={offset} - {offset + length - 1}"
log.info(f"storage range request: {range}")

if not bucket:
Expand All @@ -290,12 +290,12 @@ async def get_object(self, key, bucket=None, offset=0, length=-1):
data = await resp["Body"].read()
finish_time = time.time()
if offset > 0:
range_key = f"{key}[{offset}:{offset+length}]"
range_key = f"{key}[{offset}:{offset + length}]"
else:
range_key = key
msg = f"s3Client.get_object({range_key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f} "
msg += f"elapsed={finish_time - start_time:.4f} "
msg += f"bytes={len(data)}"
log.info(msg)

Expand Down Expand Up @@ -354,7 +354,7 @@ async def put_object(self, key, data, bucket=None):
finish_time = time.time()
msg = f"s3Client.put_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f} "
msg += f"elapsed={finish_time - start_time:.4f} "
msg += f"bytes={len(data)}"
log.info(msg)
s3_rsp = {
Expand Down Expand Up @@ -406,7 +406,7 @@ async def delete_object(self, key, bucket=None):
finish_time = time.time()
msg = f"s3Client.delete_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)

except ClientError as ce:
Expand Down Expand Up @@ -466,7 +466,7 @@ async def is_object(self, key, bucket=None):
raise HTTPInternalServerError()
msg = f"s3Client.is_object({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)

return found
Expand Down Expand Up @@ -521,7 +521,7 @@ async def get_key_stats(self, key, bucket=None):
key_stats["LastModified"] = LastModified
msg = f"s3Client.get_key_stats({key} bucket={bucket}) "
msg += f"start={start_time:.4f} finish={finish_time:.4f} "
msg += f"elapsed={finish_time-start_time:.4f}"
msg += f"elapsed={finish_time - start_time:.4f}"
log.info(msg)

return key_stats
Expand Down
8 changes: 4 additions & 4 deletions tests/integ/attr_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -1988,7 +1988,7 @@ def testPutAttributeMultiple(self):
# create some groups
grp_count = 3

grp_names = [f"group{i+1}" for i in range(grp_count)]
grp_names = [f"group{i + 1}" for i in range(grp_count)]
grp_ids = []

for grp_name in grp_names:
Expand Down Expand Up @@ -2022,7 +2022,7 @@ def testPutAttributeMultiple(self):
for i in range(attr_count):
value = [i * 10 + j for j in range(extent)]
data = {"type": "H5T_STD_I32LE", "shape": extent, "value": value}
attr_name = f"attr{i+1:04d}"
attr_name = f"attr{i + 1:04d}"
attributes[attr_name] = data

# write attributes to the dataset
Expand All @@ -2042,7 +2042,7 @@ def testPutAttributeMultiple(self):
for i in range(attr_count):
attr = ret_attrs[i]
self.assertTrue("name" in attr)
self.assertEqual(attr["name"], f"attr{i+1:04d}")
self.assertEqual(attr["name"], f"attr{i + 1:04d}")
self.assertTrue("value" in attr)
attr_value = attr["value"]
self.assertEqual(len(attr_value), extent)
Expand Down Expand Up @@ -2090,7 +2090,7 @@ def testPutAttributeMultiple(self):
self.assertTrue("attributes" in rspJson)
ret_attrs = rspJson["attributes"]
# expect the 4 attributes we wrote in the first post
# plus (i+1) in the second post
# plus (i + 1) in the second post
self.assertEqual(len(ret_attrs), attr_count + i + 1)
for j in range(len(ret_attrs)):
attr = ret_attrs[j]
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/link_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -1464,7 +1464,7 @@ def testPutLinkMultiple(self):
# create some groups under grp1
grp_count = 3

grp_names = [f"grp{(i+1):04d}" for i in range(grp_count)]
grp_names = [f"grp{(i + 1):04d}" for i in range(grp_count)]
grp_ids = []

for grp_name in grp_names:
Expand Down
4 changes: 2 additions & 2 deletions tests/integ/value_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -1756,7 +1756,7 @@ def testGet(self):

# read 1x4 block from dataset
row_index = 2
params = {"select": f"[{row_index}:{row_index+1}, 0:4]"}
params = {"select": f"[{row_index}:{row_index + 1}, 0:4]"}
params["nonstrict"] = 1 # SN can read directly from S3 or DN node
rsp = self.session.get(req, params=params, headers=headers)
self.assertEqual(rsp.status_code, 200)
Expand Down Expand Up @@ -3366,7 +3366,7 @@ def testIntelligentRangeGet3D(self):
z = 35
count = 10
# read 10 element, starting at index (x,y,z)
params = {"select": f"[{x}, {y}, {z}:{z+count}]"}
params = {"select": f"[{x}, {y}, {z}:{z + count}]"}
params["nonstrict"] = 1 # enable SN to invoke lambda func

# read the selection
Expand Down
4 changes: 2 additions & 2 deletions tests/perf/append/append_1d.py
Original file line number Diff line number Diff line change
Expand Up @@ -156,5 +156,5 @@ def addRow(dset):
# print out stats
if mode == "a":
end_ts = time.time()
print(f"added {count} rows in {(end_ts-start_ts):8.4f} seconds")
print(f"{count/(end_ts-start_ts):5.4f} rows/sec")
print(f"added {count} rows in {(end_ts - start_ts):8.4f} seconds")
print(f"{count / (end_ts - start_ts):5.4f} rows/sec")
4 changes: 2 additions & 2 deletions tests/perf/append/append_1d_async.py
Original file line number Diff line number Diff line change
Expand Up @@ -312,5 +312,5 @@ async def addrow(self, session, dsetid, sensor_id, seq_num):
print("num failures:", cfg["error_count"])
count = cfg["rows_added"]
print("rows added:", count)
print(f"added {count} rows in {(end_ts-start_ts):8.4f} seconds")
print(f"{count/(end_ts-start_ts):5.4f} rows/sec")
print(f"added {count} rows in {(end_ts - start_ts):8.4f} seconds")
print(f"{count / (end_ts - start_ts):5.4f} rows/sec")
6 changes: 3 additions & 3 deletions tests/perf/arrayperf/bytes_to_array.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,18 +36,18 @@
then = time.time()
buffer_size = getByteArraySize(arr)
now = time.time()
msg = f"getByteArraySize - elapsed: {(now-then):6.4f} for {count} elements, "
msg = f"getByteArraySize - elapsed: {(now - then):6.4f} for {count} elements, "
msg += f"returned {buffer_size}"
print(msg)
then = time.time()
buffer = arrayToBytes(arr)
now = time.time()
print(f"arrayToBytes - elpased: {(now-then):6.4f} for {count} elements")
print(f"arrayToBytes - elpased: {(now - then):6.4f} for {count} elements")
if len(buffer) != buffer_size:
raise ValueError(f"unexpected buffer length: {len(buffer)}")
then = time.time()
copy = bytesToArray(buffer, dt, (count,))
now = time.time()
if copy.shape[0] != count:
raise ValueError(f"unexpected array shape: {copy.shape}")
print(f"bytesToArray - elapsed: {(now-then):6.4f}")
print(f"bytesToArray - elapsed: {(now - then):6.4f}")
6 changes: 3 additions & 3 deletions tests/perf/arrayperf/bytes_to_vlen.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,19 +52,19 @@
then = time.time()
buffer_size = getByteArraySize(arr)
now = time.time()
msg = f"getByteArraySize - elapsed: {(now-then):6.4f} for {count} elements, "
msg = f"getByteArraySize - elapsed: {(now - then):6.4f} for {count} elements, "
msg += f"returned {buffer_size}"
print(msg)
then = time.time()
buffer = arrayToBytes(arr)
now = time.time()
print(f"arrayToBytes - elpased: {(now-then):6.4f} for {count} elements")
print(f"arrayToBytes - elpased: {(now - then):6.4f} for {count} elements")

# convert back to a numpy array
then = time.time()
arr_ret = bytesToArray(buffer, dt, [count, ])
now = time.time()
print(f"bytesToArray - elpased: {(now-then):6.4f} for {count} elements")
print(f"bytesToArray - elpased: {(now - then):6.4f} for {count} elements")

# verify that same original strings got returned
for i in range(count):
Expand Down
Loading

0 comments on commit cfe68cd

Please sign in to comment.