From 7e5a3e274d6fb024fc313d663ba489d1d0b17cd0 Mon Sep 17 00:00:00 2001 From: petscheit Date: Tue, 23 Jul 2024 16:52:57 +0200 Subject: [PATCH] chore: fix unit tests --- Makefile | 2 +- src/datalakes/block_sampled_datalake.cairo | 15 ++++--- .../block_sampled_datalake.cairo | 1 + tests/cairo_programs/computational_task.cairo | 9 ++-- tests/cairo_programs/test_vectors.cairo | 41 ++++++++++--------- .../txs_in_block_datalake.cairo | 13 ++++-- tools/make/launch_cairo_files.py | 4 +- 7 files changed, 49 insertions(+), 36 deletions(-) diff --git a/Makefile b/Makefile index b67d823a..42de0162 100644 --- a/Makefile +++ b/Makefile @@ -60,7 +60,7 @@ run-pie: get-program-hash: @echo "Getting hdp.cairo program's hash..." - cairo-compile ./src/hdp.cairo --output $(BUILD_DIR)/hdp.json + cairo-compile --cairo_path="packages/eth_essentials" "src/hdp.cairo" --output $(BUILD_DIR)/hdp.json cairo-hash-program --program $(BUILD_DIR)/hdp.json @echo "Program hash retrieved." diff --git a/src/datalakes/block_sampled_datalake.cairo b/src/datalakes/block_sampled_datalake.cairo index d0dc7c54..0f023dec 100644 --- a/src/datalakes/block_sampled_datalake.cairo +++ b/src/datalakes/block_sampled_datalake.cairo @@ -64,11 +64,11 @@ func init_block_sampled{ // Account Input Layout: // extract & write field_idx - let field_idx = extract_byte_at_pos([input + 26], 5, pow2_array); + let field_idx = extract_byte_at_pos([input + 30], 5, pow2_array); assert [properties] = field_idx; let (address) = extract_address{bitwise_ptr=bitwise_ptr}( - chunk_one=[input + 24], chunk_two=[input + 25], chunk_three=[input + 26] + chunk_one=[input + 28], chunk_two=[input + 29], chunk_three=[input + 30] ); // write address to properties @@ -90,7 +90,7 @@ func init_block_sampled{ // Account Slot Input Layout: let (address) = extract_address{bitwise_ptr=bitwise_ptr}( - chunk_one=[input + 24], chunk_two=[input + 25], chunk_three=[input + 26] + chunk_one=[input + 28], chunk_two=[input + 29], chunk_three=[input + 30] ); // write address to properties @@ -99,7 +99,7 @@ func init_block_sampled{ extract_and_write_slot{ range_check_ptr=range_check_ptr, bitwise_ptr=bitwise_ptr, properties=properties - }(chunks=input + 26); + }(chunks=input + 30); return ( res=BlockSampledDataLake( @@ -336,7 +336,10 @@ func extract_constant_params{range_check_ptr, bitwise_ptr: BitwiseBuiltin*}(inpu let (increment) = word_reverse_endian_64([input + 19]); return ( - chain_id=chain_id, block_range_start=block_range_start, block_range_end=block_range_end, increment=increment + chain_id=chain_id, + block_range_start=block_range_start, + block_range_end=block_range_end, + increment=increment, ); } @@ -479,4 +482,4 @@ func fetch_header_data_points{ return fetch_header_data_points( chain_id=chain_id, datalake=datalake, index=index + 1, data_points=data_points ); -} \ No newline at end of file +} diff --git a/tests/cairo_programs/block_sampled_datalake.cairo b/tests/cairo_programs/block_sampled_datalake.cairo index 6407b835..30a1ebe2 100644 --- a/tests/cairo_programs/block_sampled_datalake.cairo +++ b/tests/cairo_programs/block_sampled_datalake.cairo @@ -47,6 +47,7 @@ func test_block_sampled_datalake_decoding{ func block_sampled_datalake_eq( a: BlockSampledDataLake, b: BlockSampledDataLake, property_type: felt ) { + assert a.chain_id = b.chain_id; assert a.block_range_start = b.block_range_start; assert a.block_range_end = b.block_range_end; assert a.increment = b.increment; diff --git a/tests/cairo_programs/computational_task.cairo b/tests/cairo_programs/computational_task.cairo index d786b2a8..f47f3bce 100644 --- a/tests/cairo_programs/computational_task.cairo +++ b/tests/cairo_programs/computational_task.cairo @@ -64,9 +64,9 @@ func test_computational_task_init{ "type": "datalake_compute", "context": { "task_bytes_len": 128, - "encoded_task": ["0x25ca8521ba63d557", "0xc9f9f40f48f31e27", "0x739b20c59ba605a5", "0x813cc91cdc15ae0e", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0"], - "datalake_bytes_len": 224, - "encoded_datalake": ["0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0xf826540000000000", "0x0", "0x0", "0x0", "0x1527540000000000", "0x0", "0x0", "0x0", "0x100000000000000", "0x0", "0x0", "0x0", "0xa000000000000000", "0x0", "0x0", "0x0", "0x200000000000000", "0x1101", "0x0", "0x0", "0x0"], + "encoded_task": ["0xAE1B44980CDF67EC", "0x45E6CCA5D27DED75", "0x438632DD6582D123", "0x3DD3EFC40A866EC9", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0", "0x0"], + "datalake_bytes_len": 256, + "encoded_datalake": ["0x0", "0x0", "0x0", "0x0", "0x0","0x0","0x0", "0xA736AA0000000000", "0x0", "0x0", "0x0", "0xf826540000000000", "0x0", "0x0", "0x0", "0x1527540000000000", "0x0", "0x0", "0x0", "0x100000000000000", "0x0", "0x0", "0x0", "0xa000000000000000", "0x0", "0x0", "0x0", "0x200000000000000", "0x1101", "0x0", "0x0", "0x0"], "datalake_type": 0, "property_type": 1 } @@ -81,6 +81,7 @@ func test_computational_task_init{ local expected_datalake: BlockSampledDataLake; assert expected_datalake = BlockSampledDataLake( + chain_id=11155111, block_range_start=5515000, block_range_end=5515029, increment=1, @@ -94,7 +95,7 @@ func test_computational_task_init{ assert expected_task = ComputationalTask( chain_id=0x1, - hash=Uint256(0xB85414EBA86F94BAC1CA653D3D3CF014, 0x212F54CE9F4342F21C5D865F1641AABC), + hash=Uint256(0x38008646DD09E46B5D7C68B43B5C5DE2, 0xEA113874535324B3CEB180080880F599), datalake_ptr=datalake_ptr, datalake_type=DatalakeType.BLOCK_SAMPLED, aggregate_fn_id=AGGREGATE_FN.AVG, diff --git a/tests/cairo_programs/test_vectors.cairo b/tests/cairo_programs/test_vectors.cairo index b12d4969..445ce578 100644 --- a/tests/cairo_programs/test_vectors.cairo +++ b/tests/cairo_programs/test_vectors.cairo @@ -21,14 +21,15 @@ namespace BlockSampledDataLakeMocker { local datalake_bytes_len: felt; %{ + ids.datalake.chain_id = 11155111 ids.datalake.block_range_start = 5382810 ids.datalake.block_range_end = 5382815 ids.datalake.increment = 1 ids.datalake.property_type = 1 ids.datalake.properties = segments.gen_arg([8]) - datalake_input = [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x9a22520000000000, 0x0, 0x0, 0x0, 0x9f22520000000000, 0x0, 0x0, 0x0, 0x100000000000000, 0x0, 0x0, 0x0, 0xa000000000000000, 0x0, 0x0, 0x0, 0x200000000000000, 0x801, 0x0, 0x0, 0x0] - ids.datalake_bytes_len = 224 + datalake_input = [0x0,0x0,0x0,0x0,0x0,0x0,0x0,0xA736AA0000000000,0x0,0x0,0x0,0x9a22520000000000,0x0,0x0,0x0,0x9f22520000000000,0x0,0x0,0x0,0x100000000000000,0x0,0x0,0x0,0xa000000000000000,0x0,0x0,0x0,0x200000000000000,0x801,0x0,0x0,0x0] + ids.datalake_bytes_len = 256 segments.write_arg(ids.datalake_input, datalake_input) %} @@ -48,14 +49,15 @@ namespace BlockSampledDataLakeMocker { local datalake_bytes_len: felt; %{ + ids.datalake.chain_id = 11155111 ids.datalake.block_range_start = 4952100 ids.datalake.block_range_end = 4952120 ids.datalake.increment = 1 ids.datalake.property_type = 2 - ids.datalake.properties = segments.gen_arg([0x1, 0x7f2c6f930306d3aa736b3a6c6a98f512f74036d4]) + ids.datalake.properties = segments.gen_arg([0x1,0x7f2c6f930306d3aa736b3a6c6a98f512f74036d4]) - datalake_input = [0x0,0x0,0x0,0x0,0x0,0x0,0x0,0x24904b0000000000,0x0,0x0,0x0,0x38904b0000000000,0x0,0x0,0x0,0x100000000000000,0x0,0x0,0x0,0xa000000000000000,0x0,0x0,0x0,0x1600000000000000,0xd30603936f2c7f02,0xf5986a6c3a6b73aa,0x1d43640f712,0x0] - ids.datalake_bytes_len = 224 + datalake_input = [0x0,0x0,0x0,0x0,0x0,0x0,0x0,0xA736AA0000000000,0x0,0x0,0x0,0x24904b0000000000,0x0,0x0,0x0,0x38904b0000000000,0x0,0x0,0x0,0x100000000000000,0x0,0x0,0x0,0xa000000000000000,0x0,0x0,0x0,0x1600000000000000,0xd30603936f2c7f02,0xf5986a6c3a6b73aa,0x1d43640f712,0x0] + ids.datalake_bytes_len = 256 segments.write_arg(ids.datalake_input, datalake_input) %} @@ -75,14 +77,15 @@ namespace BlockSampledDataLakeMocker { local datalake_bytes_len: felt; %{ + ids.datalake.chain_id = 11155111 ids.datalake.block_range_start = 5382810 ids.datalake.block_range_end = 5382815 ids.datalake.increment = 1 ids.datalake.property_type = 3 - ids.datalake.properties = segments.gen_arg([0x75CeC1db9dCeb703200EAa6595f66885C962B920, 0x0, 0x2]) + ids.datalake.properties = segments.gen_arg([0x75CeC1db9dCeb703200EAa6595f66885C962B920,0x0,0x2]) - datalake_input = [0x0,0x0,0x0,0x0,0x0,0x0,0x0,0x9a22520000000000,0x0,0x0,0x0,0x9f22520000000000,0x0,0x0,0x0,0x100000000000000,0x0,0x0,0x0,0xa000000000000000,0x0,0x0,0x0,0x3500000000000000,0xb7ce9ddbc1ce7503,0x68f69565aa0e2003,0x20b962c985,0x0,0x0,0x0,0x200000000,0x0] - ids.datalake_bytes_len = 256 + datalake_input = [0x0,0x0,0x0,0x0,0x0,0x0,0x0,0xA736AA0000000000,0x0,0x0,0x0,0x9a22520000000000,0x0,0x0,0x0,0x9f22520000000000,0x0,0x0,0x0,0x100000000000000,0x0,0x0,0x0,0xa000000000000000,0x0,0x0,0x0,0x3500000000000000,0xb7ce9ddbc1ce7503,0x68f69565aa0e2003,0x20b962c985,0x0,0x0,0x0,0x200000000,0x0] + ids.datalake_bytes_len = 288 segments.write_arg(ids.datalake_input, datalake_input) %} @@ -113,7 +116,7 @@ namespace BlockSampledTaskMocker { from tools.py.utils import bytes_to_8_bytes_chunks_little # mocks python params that are available during full flow block_sampled_tasks = [{'property_type': 1 }] - task_bytes = bytes.fromhex("22B4DA4CC94620C9DFCC5AE7429AD350AC86587E6D9925A6209587EF17967F20000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") + task_bytes = bytes.fromhex("6A3B90F31FC36A592E67293D5D9359BADCD9B6E2B5E078B349B546A5AEE0904A000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") segments.write_arg(ids.tasks_bytes_len, [len(task_bytes)]) segments.write_arg(ids.task_input, bytes_to_8_bytes_chunks_little(task_bytes)) %} @@ -162,7 +165,7 @@ namespace BlockSampledTaskMocker { %{ from tools.py.utils import bytes_to_8_bytes_chunks_little - task_bytes = bytes.fromhex("22B4DA4CC94620C9DFCC5AE7429AD350AC86587E6D9925A6209587EF17967F20000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") + task_bytes = bytes.fromhex("6A3B90F31FC36A592E67293D5D9359BADCD9B6E2B5E078B349B546A5AEE0904A000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") ids.tasks_bytes_len = len(task_bytes) segments.write_arg(ids.task_input, bytes_to_8_bytes_chunks_little(task_bytes)) %} @@ -174,7 +177,7 @@ namespace BlockSampledTaskMocker { let task = ComputationalTask( chain_id=0x1, hash=Uint256( - low=0x407E98D423A7BB2DBF09B0E42601FC9B, high=0xEF8B01F35B404615F0339EEFAE7719A2 + low=0x29EDECDB24D47C8CFA6FA2C538D8C0AD, high=0x319EF071671DCEA889F113920CBB48DD ), datalake_ptr=datalake_ptr, datalake_type=DatalakeType.BLOCK_SAMPLED, @@ -205,7 +208,7 @@ namespace BlockSampledTaskMocker { %{ from tools.py.utils import bytes_to_8_bytes_chunks_little - task_bytes = bytes.fromhex("22B4DA4CC94620C9DFCC5AE7429AD350AC86587E6D9925A6209587EF17967F20000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") + task_bytes = bytes.fromhex("6A3B90F31FC36A592E67293D5D9359BADCD9B6E2B5E078B349B546A5AEE0904A000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") ids.tasks_bytes_len = len(task_bytes) segments.write_arg(ids.task_input, bytes_to_8_bytes_chunks_little(task_bytes)) %} @@ -215,7 +218,7 @@ namespace BlockSampledTaskMocker { let task = ComputationalTask( chain_id=0x1, hash=Uint256( - low=0x3CB6684D1B4B7FDEA3FBACAEA422C944, high=0x02F8516E3F7BE7FCCFDE22FB4A98DF37 + low=0xBB8F8BE052FA69FC932F586EFF3FFF82, high=0x800F013218B39FE67DF7C0D1F7246CB8 ), datalake_ptr=datalake_ptr, datalake_type=DatalakeType.BLOCK_SAMPLED, @@ -248,7 +251,7 @@ namespace BlockSampledTaskMocker { %{ from tools.py.utils import bytes_to_8_bytes_chunks_little - task_bytes = bytes.fromhex("22B4DA4CC94620C9DFCC5AE7429AD350AC86587E6D9925A6209587EF17967F20000000000000000000000000000000000000000000000000000000000000000200000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") + task_bytes = bytes.fromhex("6A3B90F31FC36A592E67293D5D9359BADCD9B6E2B5E078B349B546A5AEE0904A000000000000000000000000000000000000000000000000000000000000000200000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") ids.tasks_bytes_len = len(task_bytes) segments.write_arg(ids.task_input, bytes_to_8_bytes_chunks_little(task_bytes)) %} @@ -258,7 +261,7 @@ namespace BlockSampledTaskMocker { let task = ComputationalTask( chain_id=0x1, hash=Uint256( - low=0x9F439795EE0CA868B463479E5A905BF0, high=0x72CEFA1188B199ECEEAB39767CD32605 + low=0xCC50E918B8F9F1DF33CC8C9C86CBF4F0, high=0x133680AC8C33499C4364FEFFCB804E94 ), datalake_ptr=datalake_ptr, datalake_type=DatalakeType.BLOCK_SAMPLED, @@ -291,7 +294,7 @@ namespace BlockSampledTaskMocker { %{ from tools.py.utils import bytes_to_8_bytes_chunks_little - task_bytes = bytes.fromhex("22B4DA4CC94620C9DFCC5AE7429AD350AC86587E6D9925A6209587EF17967F20000000000000000000000000000000000000000000000000000000000000000300000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") + task_bytes = bytes.fromhex("6A3B90F31FC36A592E67293D5D9359BADCD9B6E2B5E078B349B546A5AEE0904A000000000000000000000000000000000000000000000000000000000000000300000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000") ids.tasks_bytes_len = len(task_bytes) segments.write_arg(ids.task_input, bytes_to_8_bytes_chunks_little(task_bytes)) %} @@ -301,7 +304,7 @@ namespace BlockSampledTaskMocker { let task = ComputationalTask( chain_id=0x1, hash=Uint256( - low=0x1CD2E160D860B4D1BD1E327B6AA209BD, high=0xCABA4809710EB228D6A31DE1B852DFB7 + low=0x3ADE97877E3502F427D1853837DD1B41, high=0x67905CD8E3ACC23EFF54245537FFA500 ), datalake_ptr=datalake_ptr, datalake_type=DatalakeType.BLOCK_SAMPLED, @@ -335,7 +338,7 @@ namespace BlockSampledTaskMocker { %{ from tools.py.utils import bytes_to_8_bytes_chunks_little - task_bytes = bytes.fromhex("22B4DA4CC94620C9DFCC5AE7429AD350AC86587E6D9925A6209587EF17967F200000000000000000000000000000000000000000000000000000000000000004000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000186a0") + task_bytes = bytes.fromhex("6A3B90F31FC36A592E67293D5D9359BADCD9B6E2B5E078B349B546A5AEE0904A0000000000000000000000000000000000000000000000000000000000000004000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000186a0") ids.tasks_bytes_len = len(task_bytes) segments.write_arg(ids.task_input, bytes_to_8_bytes_chunks_little(task_bytes)) %} @@ -344,7 +347,7 @@ namespace BlockSampledTaskMocker { let task = ComputationalTask( chain_id=0x1, hash=Uint256( - low=0xAE5641FEA9032C936D7E54D7CF36E2C3, high=0xA53CFAB970F9780B3C39CFAC1DD3D425 + low=0x18E95103512DFA47ABF4237FB5FBF673, high=0xE6FF175F1DAB2E8AC4315F634B27BE8E ), datalake_ptr=datalake_ptr, datalake_type=DatalakeType.BLOCK_SAMPLED, diff --git a/tests/cairo_programs/txs_in_block_datalake.cairo b/tests/cairo_programs/txs_in_block_datalake.cairo index 08d4f8d8..d511d4be 100644 --- a/tests/cairo_programs/txs_in_block_datalake.cairo +++ b/tests/cairo_programs/txs_in_block_datalake.cairo @@ -14,13 +14,14 @@ func main{range_check_ptr, bitwise_ptr: BitwiseBuiltin*, keccak_ptr: KeccakBuilt %{ encoded_datalakes = [ - "000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000005595f50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000005c0000000000000000000000000000000000000000000000000000000000000001000000000000000000000000000000000000000000000000000000000101000000000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000020100000000000000000000000000000000000000000000000000000000000000", - "000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000005595f50000000000000000000000000000000000000000000000000000000000000005000000000000000000000000000000000000000000000000000000000000005c0000000000000000000000000000000000000000000000000000000000000003000000000000000000000000000000000000000000000000000000000000010100000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000020104000000000000000000000000000000000000000000000000000000000000", - "000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000005595f50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000005a000000000000000000000000000000000000000000000000000000000000000a000000000000000000000000000000000000000000000000000000000101010100000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000020104000000000000000000000000000000000000000000000000000000000000" + "00000000000000000000000000000000000000000000000000000000000000010000000000000000000000000000000000000000000000000000000000AA36A700000000000000000000000000000000000000000000000000000000005595f50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000005c0000000000000000000000000000000000000000000000000000000000000001000000000000000000000000000000000000000000000000000000000101000000000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000020100000000000000000000000000000000000000000000000000000000000000", + "00000000000000000000000000000000000000000000000000000000000000010000000000000000000000000000000000000000000000000000000000AA36A700000000000000000000000000000000000000000000000000000000005595f50000000000000000000000000000000000000000000000000000000000000005000000000000000000000000000000000000000000000000000000000000005c0000000000000000000000000000000000000000000000000000000000000003000000000000000000000000000000000000000000000000000000000000010100000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000020104000000000000000000000000000000000000000000000000000000000000", + "00000000000000000000000000000000000000000000000000000000000000010000000000000000000000000000000000000000000000000000000000AA36A700000000000000000000000000000000000000000000000000000000005595f50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000005a000000000000000000000000000000000000000000000000000000000000000a000000000000000000000000000000000000000000000000000000000101010100000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000020104000000000000000000000000000000000000000000000000000000000000" ] expected_datalakes = [ { + "chain_id": 11155111, "target_block": 5608949, "start_index": 0, "end_index": 92, @@ -31,6 +32,7 @@ func main{range_check_ptr, bitwise_ptr: BitwiseBuiltin*, keccak_ptr: KeccakBuilt }, { + "chain_id": 11155111, "target_block": 5608949, "start_index": 5, "end_index": 92, @@ -40,6 +42,7 @@ func main{range_check_ptr, bitwise_ptr: BitwiseBuiltin*, keccak_ptr: KeccakBuilt "type": 1, }, { + "chain_id": 11155111, "target_block": 5608949, "start_index": 0, "end_index": 90, @@ -70,6 +73,7 @@ func eval{ } let (input_chunks) = alloc(); + local chain_id: felt; local input_bytes_len: felt; local target_block: felt; local start_index: felt; @@ -88,7 +92,7 @@ func eval{ input_chunks = bytes_to_8_bytes_chunks_little(input_bytes) segments.write_arg(ids.input_chunks, input_chunks) - + ids.chain_id = expected_datalakes[ids.index]["chain_id"] ids.target_block = expected_datalakes[ids.index]["target_block"] ids.start_index = expected_datalakes[ids.index]["start_index"] ids.end_index = expected_datalakes[ids.index]["end_index"] @@ -101,6 +105,7 @@ func eval{ input=input_chunks, input_bytes_len=input_bytes_len ); + assert result.chain_id = chain_id; assert result.target_block = target_block; assert result.start_index = start_index; assert result.end_index = end_index; diff --git a/tools/make/launch_cairo_files.py b/tools/make/launch_cairo_files.py index b251f915..9a73047e 100755 --- a/tools/make/launch_cairo_files.py +++ b/tools/make/launch_cairo_files.py @@ -177,8 +177,8 @@ def test(self): if test_file == "tests/cairo_programs/test_vectors.cairo": continue - if test_file != "tests/cairo_programs/aggregate_functions.cairo": - continue + # if test_file != "tests/cairo_programs/computational_task.cairo": + # continue self.filename_dot_cairo_path = test_file self.filename_dot_cairo = os.path.basename(test_file)