mirror of
https://github.com/planetmint/planetmint.git
synced 2025-03-30 15:08:31 +00:00

* adjusted hashlib imports and renamed to bigchaindb error Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * added type hints to transactions module Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved upsert_validator txs to transactions, updated imports Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * removed unused imports Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved tx validate to lib Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved from_db to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * removed from db from transaction Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved election validation to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved election methods to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved get_validators and get_recipients to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * removed unnecessary election method Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved show_election_status to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved topology check to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved election_id_to_public_key to validator_utils Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved vote methods to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved process_block to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * removed unused code from Vote Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved has election concluded to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * adjusted has_election_concluded Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * removed unused imports, added copyright notices Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved rollback_eleciton to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved on_rollback behaviour to planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * moved some validator utils to tendermint utils, election approval now handled by planetmint Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * Use planetmint-transaction pypi package Signed-off-by: cybnon <stefan.weber93@googlemail.com> * fixed docs imports Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * fixed validate call on test case Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * resolved linting errors Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * fixed mock on test case Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * adjusted CHANGELOG Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * removed duplicate transactions test suite Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * fixed pr comments Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * reordered imports to be standardized Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * removed unused imports and reordered them Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> * fixed linter error Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> Signed-off-by: Lorenz Herzberger <lorenzherzberger@gmail.com> Signed-off-by: cybnon <stefan.weber93@googlemail.com> Co-authored-by: cybnon <stefan.weber93@googlemail.com>
485 lines
18 KiB
Python
485 lines
18 KiB
Python
# # # Copyright © 2020 Interplanetary Database Association e.V.,
|
|
# # # Planetmint and IPDB software contributors.
|
|
# # # SPDX-License-Identifier: (Apache-2.0 AND CC-BY-4.0)
|
|
# # # Code is Apache-2.0 and docs are CC-BY-4.0
|
|
#
|
|
# from copy import deepcopy
|
|
# from transactions.types.assets.create import Create
|
|
# from transactions.types.assets.transfer import Transfer
|
|
#
|
|
# # import pytest
|
|
# # import pymongo
|
|
#
|
|
# # from planetmint.backend import Connection, query
|
|
#
|
|
#
|
|
# # pytestmark = pytest.mark.bdb
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_txids_filtered(signed_create_tx, signed_transfer_tx):
|
|
# from planetmint.backend import connect, query
|
|
# from planetmint.models import Transaction
|
|
# conn = connect()
|
|
#
|
|
# # create and insert two blocks, one for the create and one for the
|
|
# # transfer transaction
|
|
# conn.db.transactions.insert_one(signed_create_tx.to_dict())
|
|
# conn.db.transactions.insert_one(signed_transfer_tx.to_dict())
|
|
#
|
|
# asset_id = Transaction.get_asset_id([signed_create_tx, signed_transfer_tx])
|
|
#
|
|
# # Test get by just asset id
|
|
# txids = set(query.get_txids_filtered(conn, asset_id))
|
|
# assert txids == {signed_create_tx.id, signed_transfer_tx.id}
|
|
#
|
|
# # Test get by asset and CREATE
|
|
# txids = set(query.get_txids_filtered(conn, asset_id, Transaction.CREATE))
|
|
# assert txids == {signed_create_tx.id}
|
|
#
|
|
# # Test get by asset and TRANSFER
|
|
# txids = set(query.get_txids_filtered(conn, asset_id, Transaction.TRANSFER))
|
|
# assert txids == {signed_transfer_tx.id}
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_write_assets():
|
|
# from planetmint.backend import connect, query
|
|
# conn = connect()
|
|
#
|
|
# assets = [
|
|
# {'id': 1, 'data': '1'},
|
|
# {'id': 2, 'data': '2'},
|
|
# {'id': 3, 'data': '3'},
|
|
# # Duplicated id. Should not be written to the database
|
|
# {'id': 1, 'data': '1'},
|
|
# ]
|
|
#
|
|
# # write the assets
|
|
# for asset in assets:
|
|
# query.store_asset(conn, deepcopy(asset))
|
|
#
|
|
# # check that 3 assets were written to the database
|
|
# cursor = conn.db.assets.find({}, projection={'_id': False})\
|
|
# .sort('id', pymongo.ASCENDING)
|
|
#
|
|
# assert cursor.collection.count_documents({}) == 3
|
|
# assert list(cursor) == assets[:-1]
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_assets():
|
|
# from planetmint.backend import connect, query
|
|
# conn = connect()
|
|
#
|
|
# assets = [
|
|
# {'id': 1, 'data': '1'},
|
|
# {'id': 2, 'data': '2'},
|
|
# {'id': 3, 'data': '3'},
|
|
# ]
|
|
#
|
|
# conn.db.assets.insert_many(deepcopy(assets), ordered=False)
|
|
#
|
|
# for asset in assets:
|
|
# assert query.get_asset(conn, asset['id'])
|
|
#
|
|
# @pytest.mark.skip
|
|
# @pytest.mark.parametrize('table', ['assets', 'metadata'])
|
|
# def test_text_search(table):
|
|
# from planetmint.backend import connect, query
|
|
# conn = connect()
|
|
#
|
|
# # Example data and tests cases taken from the mongodb documentation
|
|
# # https://docs.mongodb.com/manual/reference/operator/query/text/
|
|
# objects = [
|
|
# {'id': 1, 'subject': 'coffee', 'author': 'xyz', 'views': 50},
|
|
# {'id': 2, 'subject': 'Coffee Shopping', 'author': 'efg', 'views': 5},
|
|
# {'id': 3, 'subject': 'Baking a cake', 'author': 'abc', 'views': 90},
|
|
# {'id': 4, 'subject': 'baking', 'author': 'xyz', 'views': 100},
|
|
# {'id': 5, 'subject': 'Café Con Leche', 'author': 'abc', 'views': 200},
|
|
# {'id': 6, 'subject': 'Сырники', 'author': 'jkl', 'views': 80},
|
|
# {'id': 7, 'subject': 'coffee and cream', 'author': 'efg', 'views': 10},
|
|
# {'id': 8, 'subject': 'Cafe con Leche', 'author': 'xyz', 'views': 10}
|
|
# ]
|
|
#
|
|
# # insert the assets
|
|
# conn.db[table].insert_many(deepcopy(objects), ordered=False)
|
|
#
|
|
# # test search single word
|
|
# assert list(query.text_search(conn, 'coffee', table=table)) == [
|
|
# {'id': 1, 'subject': 'coffee', 'author': 'xyz', 'views': 50},
|
|
# {'id': 2, 'subject': 'Coffee Shopping', 'author': 'efg', 'views': 5},
|
|
# {'id': 7, 'subject': 'coffee and cream', 'author': 'efg', 'views': 10},
|
|
# ]
|
|
#
|
|
# # match any of the search terms
|
|
# assert list(query.text_search(conn, 'bake coffee cake', table=table)) == [
|
|
# {'author': 'abc', 'id': 3, 'subject': 'Baking a cake', 'views': 90},
|
|
# {'author': 'xyz', 'id': 1, 'subject': 'coffee', 'views': 50},
|
|
# {'author': 'xyz', 'id': 4, 'subject': 'baking', 'views': 100},
|
|
# {'author': 'efg', 'id': 2, 'subject': 'Coffee Shopping', 'views': 5},
|
|
# {'author': 'efg', 'id': 7, 'subject': 'coffee and cream', 'views': 10}
|
|
# ]
|
|
#
|
|
# # search for a phrase
|
|
# assert list(query.text_search(conn, '\"coffee shop\"', table=table)) == [
|
|
# {'id': 2, 'subject': 'Coffee Shopping', 'author': 'efg', 'views': 5},
|
|
# ]
|
|
#
|
|
# # exclude documents that contain a term
|
|
# assert list(query.text_search(conn, 'coffee -shop', table=table)) == [
|
|
# {'id': 1, 'subject': 'coffee', 'author': 'xyz', 'views': 50},
|
|
# {'id': 7, 'subject': 'coffee and cream', 'author': 'efg', 'views': 10},
|
|
# ]
|
|
#
|
|
# # search different language
|
|
# assert list(query.text_search(conn, 'leche', language='es', table=table)) == [
|
|
# {'id': 5, 'subject': 'Café Con Leche', 'author': 'abc', 'views': 200},
|
|
# {'id': 8, 'subject': 'Cafe con Leche', 'author': 'xyz', 'views': 10}
|
|
# ]
|
|
#
|
|
# # case and diacritic insensitive search
|
|
# assert list(query.text_search(conn, 'сы́рники CAFÉS', table=table)) == [
|
|
# {'id': 6, 'subject': 'Сырники', 'author': 'jkl', 'views': 80},
|
|
# {'id': 5, 'subject': 'Café Con Leche', 'author': 'abc', 'views': 200},
|
|
# {'id': 8, 'subject': 'Cafe con Leche', 'author': 'xyz', 'views': 10}
|
|
# ]
|
|
#
|
|
# # case sensitive search
|
|
# assert list(query.text_search(conn, 'Coffee', case_sensitive=True, table=table)) == [
|
|
# {'id': 2, 'subject': 'Coffee Shopping', 'author': 'efg', 'views': 5},
|
|
# ]
|
|
#
|
|
# # diacritic sensitive search
|
|
# assert list(query.text_search(conn, 'CAFÉ', diacritic_sensitive=True, table=table)) == [
|
|
# {'id': 5, 'subject': 'Café Con Leche', 'author': 'abc', 'views': 200},
|
|
# ]
|
|
#
|
|
# # return text score
|
|
# assert list(query.text_search(conn, 'coffee', text_score=True, table=table)) == [
|
|
# {'id': 1, 'subject': 'coffee', 'author': 'xyz', 'views': 50, 'score': 1.0},
|
|
# {'id': 2, 'subject': 'Coffee Shopping', 'author': 'efg', 'views': 5, 'score': 0.75},
|
|
# {'id': 7, 'subject': 'coffee and cream', 'author': 'efg', 'views': 10, 'score': 0.75},
|
|
# ]
|
|
#
|
|
# # limit search result
|
|
# assert list(query.text_search(conn, 'coffee', limit=2, table=table)) == [
|
|
# {'id': 1, 'subject': 'coffee', 'author': 'xyz', 'views': 50},
|
|
# {'id': 2, 'subject': 'Coffee Shopping', 'author': 'efg', 'views': 5},
|
|
# ]
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_write_metadata():
|
|
# from planetmint.backend import connect, query
|
|
# conn = connect()
|
|
#
|
|
# metadata = [
|
|
# {'id': 1, 'data': '1'},
|
|
# {'id': 2, 'data': '2'},
|
|
# {'id': 3, 'data': '3'}
|
|
# ]
|
|
#
|
|
# # write the assets
|
|
# query.store_metadatas(conn, deepcopy(metadata))
|
|
#
|
|
# # check that 3 assets were written to the database
|
|
# cursor = conn.db.metadata.find({}, projection={'_id': False})\
|
|
# .sort('id', pymongo.ASCENDING)
|
|
#
|
|
# assert cursor.collection.count_documents({}) == 3
|
|
# assert list(cursor) == metadata
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_metadata():
|
|
# from planetmint.backend import connect, query
|
|
# conn = connect()
|
|
#
|
|
# metadata = [
|
|
# {'id': 1, 'metadata': None},
|
|
# {'id': 2, 'metadata': {'key': 'value'}},
|
|
# {'id': 3, 'metadata': '3'},
|
|
# ]
|
|
#
|
|
# conn.db.metadata.insert_many(deepcopy(metadata), ordered=False)
|
|
#
|
|
# for meta in metadata:
|
|
# assert query.get_metadata(conn, [meta['id']])
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_owned_ids(signed_create_tx, user_pk):
|
|
# from planetmint.backend import connect, query
|
|
# conn = connect()
|
|
#
|
|
# # insert a transaction
|
|
# conn.db.transactions.insert_one(deepcopy(signed_create_tx.to_dict()))
|
|
#
|
|
# txns = list(query.get_owned_ids(conn, user_pk))
|
|
#
|
|
# assert txns[0] == signed_create_tx.to_dict()
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_spending_transactions(user_pk, user_sk):
|
|
# from planetmint.backend import connect, query
|
|
# conn = connect()
|
|
#
|
|
# out = [([user_pk], 1)]
|
|
# tx1 = Create.generate([user_pk], out * 3)
|
|
# tx1.sign([user_sk])
|
|
# inputs = tx1.to_inputs()
|
|
# tx2 = Transfer.generate([inputs[0]], out, tx1.id).sign([user_sk])
|
|
# tx3 = Transfer.generate([inputs[1]], out, tx1.id).sign([user_sk])
|
|
# tx4 = Transfer.generate([inputs[2]], out, tx1.id).sign([user_sk])
|
|
# txns = [deepcopy(tx.to_dict()) for tx in [tx1, tx2, tx3, tx4]]
|
|
# conn.db.transactions.insert_many(txns)
|
|
#
|
|
# links = [inputs[0].fulfills.to_dict(), inputs[2].fulfills.to_dict()]
|
|
# txns = list(query.get_spending_transactions(conn, links))
|
|
#
|
|
# # tx3 not a member because input 1 not asked for
|
|
# assert txns == [tx2.to_dict(), tx4.to_dict()]
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_spending_transactions_multiple_inputs():
|
|
# from planetmint.backend import connect, query
|
|
# from transactions.common.crypto import generate_key_pair
|
|
# conn = connect()
|
|
# (alice_sk, alice_pk) = generate_key_pair()
|
|
# (bob_sk, bob_pk) = generate_key_pair()
|
|
# (carol_sk, carol_pk) = generate_key_pair()
|
|
#
|
|
# out = [([alice_pk], 9)]
|
|
# tx1 = Create.generate([alice_pk], out).sign([alice_sk])
|
|
#
|
|
# inputs1 = tx1.to_inputs()
|
|
# tx2 = Transfer.generate([inputs1[0]],
|
|
# [([alice_pk], 6), ([bob_pk], 3)],
|
|
# tx1.id).sign([alice_sk])
|
|
#
|
|
# inputs2 = tx2.to_inputs()
|
|
# tx3 = Transfer.generate([inputs2[0]],
|
|
# [([bob_pk], 3), ([carol_pk], 3)],
|
|
# tx1.id).sign([alice_sk])
|
|
#
|
|
# inputs3 = tx3.to_inputs()
|
|
# tx4 = Transfer.generate([inputs2[1], inputs3[0]],
|
|
# [([carol_pk], 6)],
|
|
# tx1.id).sign([bob_sk])
|
|
#
|
|
# txns = [deepcopy(tx.to_dict()) for tx in [tx1, tx2, tx3, tx4]]
|
|
# conn.db.transactions.insert_many(txns)
|
|
#
|
|
# links = [
|
|
# ({'transaction_id': tx2.id, 'output_index': 0}, 1, [tx3.id]),
|
|
# ({'transaction_id': tx2.id, 'output_index': 1}, 1, [tx4.id]),
|
|
# ({'transaction_id': tx3.id, 'output_index': 0}, 1, [tx4.id]),
|
|
# ({'transaction_id': tx3.id, 'output_index': 1}, 0, None),
|
|
# ]
|
|
# for li, num, match in links:
|
|
# txns = list(query.get_spending_transactions(conn, [li]))
|
|
# assert len(txns) == num
|
|
# if len(txns):
|
|
# assert [tx['id'] for tx in txns] == match
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_store_block():
|
|
# from planetmint.backend import connect, query
|
|
# from planetmint.lib import Block
|
|
# conn = connect()
|
|
#
|
|
# block = Block(app_hash='random_utxo',
|
|
# height=3,
|
|
# transactions=[])
|
|
# query.store_block(conn, block._asdict())
|
|
# cursor = conn.db.blocks.find({}, projection={'_id': False})
|
|
# assert cursor.collection.count_documents({}) == 1
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_block():
|
|
# from planetmint.backend import connect, query
|
|
# from planetmint.lib import Block
|
|
# conn = connect()
|
|
#
|
|
# block = Block(app_hash='random_utxo',
|
|
# height=3,
|
|
# transactions=[])
|
|
#
|
|
# conn.db.blocks.insert_one(block._asdict())
|
|
#
|
|
# block = dict(query.get_block(conn, 3))
|
|
# assert block['height'] == 3
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_delete_zero_unspent_outputs(db_context, utxoset):
|
|
# from planetmint.backend import query
|
|
# unspent_outputs, utxo_collection = utxoset
|
|
# delete_res = query.delete_unspent_outputs(db_context.conn)
|
|
# assert delete_res is None
|
|
# assert utxo_collection.count_documents({}) == 3
|
|
# assert utxo_collection.count_documents(
|
|
# {'$or': [
|
|
# {'transaction_id': 'a', 'output_index': 0},
|
|
# {'transaction_id': 'b', 'output_index': 0},
|
|
# {'transaction_id': 'a', 'output_index': 1},
|
|
# ]}
|
|
# ) == 3
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_delete_one_unspent_outputs(db_context, utxoset):
|
|
# from planetmint.backend import query
|
|
# unspent_outputs, utxo_collection = utxoset
|
|
# delete_res = query.delete_unspent_outputs(db_context.conn,
|
|
# unspent_outputs[0])
|
|
# assert delete_res.raw_result['n'] == 1
|
|
# assert utxo_collection.count_documents(
|
|
# {'$or': [
|
|
# {'transaction_id': 'a', 'output_index': 1},
|
|
# {'transaction_id': 'b', 'output_index': 0},
|
|
# ]}
|
|
# ) == 2
|
|
# assert utxo_collection.count_documents(
|
|
# {'transaction_id': 'a', 'output_index': 0}) == 0
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_delete_many_unspent_outputs(db_context, utxoset):
|
|
# from planetmint.backend import query
|
|
# unspent_outputs, utxo_collection = utxoset
|
|
# delete_res = query.delete_unspent_outputs(db_context.conn,
|
|
# *unspent_outputs[::2])
|
|
# assert delete_res.raw_result['n'] == 2
|
|
# assert utxo_collection.count_documents(
|
|
# {'$or': [
|
|
# {'transaction_id': 'a', 'output_index': 0},
|
|
# {'transaction_id': 'b', 'output_index': 0},
|
|
# ]}
|
|
# ) == 0
|
|
# assert utxo_collection.count_documents(
|
|
# {'transaction_id': 'a', 'output_index': 1}) == 1
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_store_zero_unspent_output(db_context, utxo_collection):
|
|
# from planetmint.backend import query
|
|
# res = query.store_unspent_outputs(db_context.conn)
|
|
# assert res is None
|
|
# assert utxo_collection.count_documents({}) == 0
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_store_one_unspent_output(db_context,
|
|
# unspent_output_1, utxo_collection):
|
|
# from planetmint.backend import query
|
|
# res = query.store_unspent_outputs(db_context.conn, unspent_output_1)
|
|
# assert res.acknowledged
|
|
# assert len(res.inserted_ids) == 1
|
|
# assert utxo_collection.count_documents(
|
|
# {'transaction_id': unspent_output_1['transaction_id'],
|
|
# 'output_index': unspent_output_1['output_index']}
|
|
# ) == 1
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_store_many_unspent_outputs(db_context,
|
|
# unspent_outputs, utxo_collection):
|
|
# from planetmint.backend import query
|
|
# res = query.store_unspent_outputs(db_context.conn, *unspent_outputs)
|
|
# assert res.acknowledged
|
|
# assert len(res.inserted_ids) == 3
|
|
# assert utxo_collection.count_documents(
|
|
# {'transaction_id': unspent_outputs[0]['transaction_id']}
|
|
# ) == 3
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_unspent_outputs(db_context, utxoset):
|
|
# from planetmint.backend import query
|
|
# cursor = query.get_unspent_outputs(db_context.conn)
|
|
# assert cursor.collection.count_documents({}) == 3
|
|
# retrieved_utxoset = list(cursor)
|
|
# unspent_outputs, utxo_collection = utxoset
|
|
# assert retrieved_utxoset == list(
|
|
# utxo_collection.find(projection={'_id': False}))
|
|
# assert retrieved_utxoset == unspent_outputs
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_store_pre_commit_state(db_context):
|
|
# from planetmint.backend import query
|
|
#
|
|
# state = dict(height=3, transactions=[])
|
|
#
|
|
# query.store_pre_commit_state(db_context.conn, state)
|
|
# cursor = db_context.conn.db.pre_commit.find({'commit_id': 'test'},
|
|
# projection={'_id': False})
|
|
# assert cursor.collection.count_documents({}) == 1
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_get_pre_commit_state(db_context):
|
|
# from planetmint.backend import query
|
|
#
|
|
# state = dict(height=3, transactions=[])
|
|
# db_context.conn.db.pre_commit.insert_one(state)
|
|
# resp = query.get_pre_commit_state(db_context.conn)
|
|
# assert resp == state
|
|
#
|
|
# @pytest.mark.skip
|
|
# def test_validator_update():
|
|
# from planetmint.backend import connect, query
|
|
#
|
|
# conn = connect()
|
|
#
|
|
# def gen_validator_update(height):
|
|
# return {'data': 'somedata', 'height': height, 'election_id': f'election_id_at_height_{height}'}
|
|
#
|
|
# for i in range(1, 100, 10):
|
|
# value = gen_validator_update(i)
|
|
# query.store_validator_set(conn, value)
|
|
#
|
|
# v1 = query.get_validator_set(conn, 8)
|
|
# assert v1['height'] == 1
|
|
#
|
|
# v41 = query.get_validator_set(conn, 50)
|
|
# assert v41['height'] == 41
|
|
#
|
|
# v91 = query.get_validator_set(conn)
|
|
# assert v91['height'] == 91
|
|
#
|
|
# @pytest.mark.skip
|
|
# @pytest.mark.parametrize('description,stores,expected', [
|
|
# (
|
|
# 'Query empty database.',
|
|
# [],
|
|
# None,
|
|
# ),
|
|
# (
|
|
# 'Store one chain with the default value for `is_synced`.',
|
|
# [
|
|
# {'height': 0, 'chain_id': 'some-id'},
|
|
# ],
|
|
# {'height': 0, 'chain_id': 'some-id', 'is_synced': True},
|
|
# ),
|
|
# (
|
|
# 'Store one chain with a custom value for `is_synced`.',
|
|
# [
|
|
# {'height': 0, 'chain_id': 'some-id', 'is_synced': False},
|
|
# ],
|
|
# {'height': 0, 'chain_id': 'some-id', 'is_synced': False},
|
|
# ),
|
|
# (
|
|
# 'Store one chain, then update it.',
|
|
# [
|
|
# {'height': 0, 'chain_id': 'some-id', 'is_synced': True},
|
|
# {'height': 0, 'chain_id': 'new-id', 'is_synced': False},
|
|
# ],
|
|
# {'height': 0, 'chain_id': 'new-id', 'is_synced': False},
|
|
# ),
|
|
# (
|
|
# 'Store a chain, update it, store another chain.',
|
|
# [
|
|
# {'height': 0, 'chain_id': 'some-id', 'is_synced': True},
|
|
# {'height': 0, 'chain_id': 'some-id', 'is_synced': False},
|
|
# {'height': 10, 'chain_id': 'another-id', 'is_synced': True},
|
|
# ],
|
|
# {'height': 10, 'chain_id': 'another-id', 'is_synced': True},
|
|
# ),
|
|
# ])
|
|
# def test_store_abci_chain(description, stores, expected):
|
|
# conn = connect()
|
|
#
|
|
# for store in stores:
|
|
# query.store_abci_chain(conn, **store)
|
|
#
|
|
# actual = query.get_latest_abci_chain(conn)
|
|
# assert expected == actual, description
|