mirror of
https://github.com/bigchaindb/bigchaindb.git
synced 2024-10-13 13:34:05 +00:00
ABCI chain migration conclusion (#2488)
* Problem: No good way to check for val set absence. Solution: Make get_validator_set/get_validators return None/[] when there are no validators yet. * Problem: Incompatible ABCI chain upgrades. Solution: Record known chains and sync through InitChain. Triggering the migration and adjusting other ABCI endpoints will follow.
This commit is contained in:
parent
fe0a4c494b
commit
230a5b2d69
@ -296,7 +296,7 @@ def get_validator_set(conn, height=None):
|
||||
.limit(1)
|
||||
)
|
||||
|
||||
return list(cursor)[0]
|
||||
return next(cursor, None)
|
||||
|
||||
|
||||
@register_query(LocalMongoDBConnection)
|
||||
@ -322,3 +322,23 @@ def get_asset_tokens_for_public_key(conn, asset_id, public_key):
|
||||
{'$project': {'_id': False}}
|
||||
]))
|
||||
return cursor
|
||||
|
||||
|
||||
@register_query(LocalMongoDBConnection)
|
||||
def store_abci_chain(conn, height, chain_id, is_synced=True):
|
||||
return conn.run(
|
||||
conn.collection('abci_chains').replace_one(
|
||||
{'height': height},
|
||||
{'height': height, 'chain_id': chain_id,
|
||||
'is_synced': is_synced},
|
||||
upsert=True,
|
||||
)
|
||||
)
|
||||
|
||||
|
||||
@register_query(LocalMongoDBConnection)
|
||||
def get_latest_abci_chain(conn):
|
||||
return conn.run(
|
||||
conn.collection('abci_chains')
|
||||
.find_one(projection={'_id': False}, sort=[('height', DESCENDING)])
|
||||
)
|
||||
|
@ -47,6 +47,7 @@ def create_indexes(conn, dbname):
|
||||
create_utxos_secondary_index(conn, dbname)
|
||||
create_pre_commit_secondary_index(conn, dbname)
|
||||
create_validators_secondary_index(conn, dbname)
|
||||
create_abci_chains_indexes(conn, dbname)
|
||||
|
||||
|
||||
@register_schema(LocalMongoDBConnection)
|
||||
@ -133,3 +134,16 @@ def create_validators_secondary_index(conn, dbname):
|
||||
conn.conn[dbname]['validators'].create_index('height',
|
||||
name='height',
|
||||
unique=True,)
|
||||
|
||||
|
||||
def create_abci_chains_indexes(conn, dbname):
|
||||
logger.info('Create `abci_chains.height` secondary index.')
|
||||
|
||||
conn.conn[dbname]['abci_chains'].create_index('height',
|
||||
name='height',
|
||||
unique=True,)
|
||||
|
||||
logger.info('Create `abci_chains.chain_id` secondary index.')
|
||||
conn.conn[dbname]['abci_chains'].create_index('chain_id',
|
||||
name='chain_id',
|
||||
unique=True)
|
||||
|
@ -380,3 +380,23 @@ def get_asset_tokens_for_public_key(connection, asset_id,
|
||||
Iterator of transaction that list given owner in conditions.
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
|
||||
@singledispatch
|
||||
def store_abci_chain(conn, height, chain_id, is_synced=True):
|
||||
"""Create or update an ABCI chain at the given height.
|
||||
Usually invoked in the beginning of the ABCI communications (height=0)
|
||||
or when ABCI client (like Tendermint) is migrated (any height).
|
||||
|
||||
Args:
|
||||
is_synced: True if the chain is known by both ABCI client and server
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
|
||||
@singledispatch
|
||||
def get_latest_abci_chain(conn):
|
||||
"""Returns the ABCI chain stored at the biggest height, if any,
|
||||
None otherwise.
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
@ -16,7 +16,7 @@ logger = logging.getLogger(__name__)
|
||||
|
||||
# Tables/collections that every backend database must create
|
||||
TABLES = ('transactions', 'blocks', 'assets', 'metadata',
|
||||
'validators', 'pre_commit', 'utxos')
|
||||
'validators', 'pre_commit', 'utxos', 'abci_chains')
|
||||
|
||||
VALID_LANGUAGES = ('danish', 'dutch', 'english', 'finnish', 'french', 'german',
|
||||
'hungarian', 'italian', 'norwegian', 'portuguese', 'romanian',
|
||||
|
@ -6,6 +6,7 @@
|
||||
with Tendermint.
|
||||
"""
|
||||
import logging
|
||||
import sys
|
||||
|
||||
from abci.application import BaseApplication
|
||||
from abci.types_pb2 import (
|
||||
@ -47,22 +48,76 @@ class App(BaseApplication):
|
||||
self.block_transactions = []
|
||||
self.validators = None
|
||||
self.new_height = None
|
||||
self.chain = self.bigchaindb.get_latest_abci_chain()
|
||||
|
||||
def log_abci_migration_error(self, chain_id, validators):
|
||||
logger.error(f'An ABCI chain migration is in process. ' +
|
||||
'Download the new ABCI client and configure it with ' +
|
||||
'chain_id={chain_id} and validators={validators}.')
|
||||
|
||||
def abort_if_abci_chain_is_not_synced(self):
|
||||
if self.chain is None or self.chain['is_synced']:
|
||||
return
|
||||
|
||||
validators = self.bigchaindb.get_validators()
|
||||
self.log_abci_migration_error(self.chain['chain_id'], validators)
|
||||
sys.exit(1)
|
||||
|
||||
def init_chain(self, genesis):
|
||||
"""Initialize chain with block of height 0"""
|
||||
"""Initialize chain upon genesis or a migration"""
|
||||
|
||||
validator_set = [vutils.decode_validator(v) for v in genesis.validators]
|
||||
block = Block(app_hash='', height=0, transactions=[])
|
||||
app_hash = ''
|
||||
height = 0
|
||||
|
||||
known_chain = self.bigchaindb.get_latest_abci_chain()
|
||||
if known_chain is not None:
|
||||
chain_id = known_chain['chain_id']
|
||||
|
||||
if known_chain['is_synced']:
|
||||
msg = f'Got invalid InitChain ABCI request ({genesis}) - ' + \
|
||||
'the chain {chain_id} is already synced.'
|
||||
logger.error(msg)
|
||||
sys.exit(1)
|
||||
|
||||
if chain_id != genesis.chain_id:
|
||||
validators = self.bigchaindb.get_validators()
|
||||
self.log_abci_migration_error(chain_id, validators)
|
||||
sys.exit(1)
|
||||
|
||||
# set migration values for app hash and height
|
||||
block = self.bigchaindb.get_latest_block()
|
||||
app_hash = '' if block is None else block['app_hash']
|
||||
height = 0 if block is None else block['height'] + 1
|
||||
|
||||
known_validators = self.bigchaindb.get_validators()
|
||||
validator_set = [vutils.decode_validator(v)
|
||||
for v in genesis.validators]
|
||||
|
||||
if known_validators and known_validators != validator_set:
|
||||
self.log_abci_migration_error(known_chain['chain_id'],
|
||||
known_validators)
|
||||
sys.exit(1)
|
||||
|
||||
block = Block(app_hash=app_hash, height=height, transactions=[])
|
||||
self.bigchaindb.store_block(block._asdict())
|
||||
self.bigchaindb.store_validator_set(1, validator_set, None)
|
||||
self.bigchaindb.store_validator_set(height + 1, validator_set, None)
|
||||
abci_chain_height = 0 if known_chain is None else known_chain['height']
|
||||
self.bigchaindb.store_abci_chain(abci_chain_height,
|
||||
genesis.chain_id, True)
|
||||
self.chain = {'height': abci_chain_height, 'is_synced': True,
|
||||
'chain_id': genesis.chain_id}
|
||||
return ResponseInitChain()
|
||||
|
||||
def info(self, request):
|
||||
"""Return height of the latest committed block."""
|
||||
|
||||
self.abort_if_abci_chain_is_not_synced()
|
||||
|
||||
r = ResponseInfo()
|
||||
block = self.bigchaindb.get_latest_block()
|
||||
if block:
|
||||
r.last_block_height = block['height']
|
||||
chain_shift = 0 if self.chain is None else self.chain['height']
|
||||
r.last_block_height = block['height'] - chain_shift
|
||||
r.last_block_app_hash = block['app_hash'].encode('utf-8')
|
||||
else:
|
||||
r.last_block_height = 0
|
||||
@ -77,6 +132,8 @@ class App(BaseApplication):
|
||||
raw_tx: a raw string (in bytes) transaction.
|
||||
"""
|
||||
|
||||
self.abort_if_abci_chain_is_not_synced()
|
||||
|
||||
logger.benchmark('CHECK_TX_INIT')
|
||||
logger.debug('check_tx: %s', raw_transaction)
|
||||
transaction = decode_transaction(raw_transaction)
|
||||
@ -95,8 +152,11 @@ class App(BaseApplication):
|
||||
req_begin_block: block object which contains block header
|
||||
and block hash.
|
||||
"""
|
||||
self.abort_if_abci_chain_is_not_synced()
|
||||
|
||||
chain_shift = 0 if self.chain is None else self.chain['height']
|
||||
logger.benchmark('BEGIN BLOCK, height:%s, num_txs:%s',
|
||||
req_begin_block.header.height,
|
||||
req_begin_block.header.height + chain_shift,
|
||||
req_begin_block.header.num_txs)
|
||||
|
||||
self.block_txn_ids = []
|
||||
@ -109,6 +169,9 @@ class App(BaseApplication):
|
||||
Args:
|
||||
raw_tx: a raw string (in bytes) transaction.
|
||||
"""
|
||||
|
||||
self.abort_if_abci_chain_is_not_synced()
|
||||
|
||||
logger.debug('deliver_tx: %s', raw_transaction)
|
||||
transaction = self.bigchaindb.is_valid_transaction(
|
||||
decode_transaction(raw_transaction), self.block_transactions)
|
||||
@ -130,7 +193,11 @@ class App(BaseApplication):
|
||||
height (int): new height of the chain.
|
||||
"""
|
||||
|
||||
height = request_end_block.height
|
||||
self.abort_if_abci_chain_is_not_synced()
|
||||
|
||||
chain_shift = 0 if self.chain is None else self.chain['height']
|
||||
|
||||
height = request_end_block.height + chain_shift
|
||||
self.new_height = height
|
||||
block_txn_hash = calculate_hash(self.block_txn_ids)
|
||||
block = self.bigchaindb.get_latest_block()
|
||||
@ -158,6 +225,8 @@ class App(BaseApplication):
|
||||
def commit(self):
|
||||
"""Store the new height and along with block hash."""
|
||||
|
||||
self.abort_if_abci_chain_is_not_synced()
|
||||
|
||||
data = self.block_txn_hash.encode('utf-8')
|
||||
|
||||
# register a new block only when new transactions are received
|
||||
|
@ -426,8 +426,7 @@ class BigchainDB(object):
|
||||
|
||||
def get_validators(self, height=None):
|
||||
result = self.get_validator_change(height)
|
||||
validators = result['validators']
|
||||
return validators
|
||||
return [] if result is None else result['validators']
|
||||
|
||||
def get_validators_by_election_id(self, election_id):
|
||||
result = backend.query.get_validator_set_by_election_id(self.connection, election_id)
|
||||
@ -448,6 +447,37 @@ class BigchainDB(object):
|
||||
'validators': validators,
|
||||
'election_id': election_id})
|
||||
|
||||
def store_abci_chain(self, height, chain_id, is_synced=True):
|
||||
return backend.query.store_abci_chain(self.connection, height,
|
||||
chain_id, is_synced)
|
||||
|
||||
def get_latest_abci_chain(self):
|
||||
return backend.query.get_latest_abci_chain(self.connection)
|
||||
|
||||
def migrate_abci_chain(self):
|
||||
"""Generate and record a new ABCI chain ID. New blocks are not
|
||||
accepted until we receive an InitChain ABCI request with
|
||||
the matching chain ID and validator set.
|
||||
|
||||
Chain ID is generated based on the current chain and height.
|
||||
`chain-X` => `chain-X-migrated-at-height-5`.
|
||||
`chain-X-migrated-at-height-5` => `chain-X-migrated-at-height-21`.
|
||||
|
||||
If there is no known chain (we are at genesis), the function returns.
|
||||
"""
|
||||
latest_chain = self.get_latest_abci_chain()
|
||||
if latest_chain is None:
|
||||
return
|
||||
|
||||
block = self.get_latest_block()
|
||||
|
||||
suffix = '-migrated-at-height-'
|
||||
chain_id = latest_chain['chain_id']
|
||||
block_height_str = str(block['height'])
|
||||
new_chain_id = chain_id.split(suffix)[0] + suffix + block_height_str
|
||||
|
||||
self.store_abci_chain(block['height'] + 1, new_chain_id, False)
|
||||
|
||||
|
||||
Block = namedtuple('Block', ('app_hash', 'height', 'transactions'))
|
||||
|
||||
|
@ -7,6 +7,9 @@ from copy import deepcopy
|
||||
import pytest
|
||||
import pymongo
|
||||
|
||||
from bigchaindb.backend import connect, query
|
||||
|
||||
|
||||
pytestmark = [pytest.mark.tendermint, pytest.mark.bdb]
|
||||
|
||||
|
||||
@ -394,3 +397,51 @@ def test_validator_update():
|
||||
|
||||
v91 = query.get_validator_set(conn)
|
||||
assert v91['height'] == 91
|
||||
|
||||
|
||||
@pytest.mark.parametrize('description,stores,expected', [
|
||||
(
|
||||
'Query empty database.',
|
||||
[],
|
||||
None,
|
||||
),
|
||||
(
|
||||
'Store one chain with the default value for `is_synced`.',
|
||||
[
|
||||
{'height': 0, 'chain_id': 'some-id'},
|
||||
],
|
||||
{'height': 0, 'chain_id': 'some-id', 'is_synced': True},
|
||||
),
|
||||
(
|
||||
'Store one chain with a custom value for `is_synced`.',
|
||||
[
|
||||
{'height': 0, 'chain_id': 'some-id', 'is_synced': False},
|
||||
],
|
||||
{'height': 0, 'chain_id': 'some-id', 'is_synced': False},
|
||||
),
|
||||
(
|
||||
'Store one chain, then update it.',
|
||||
[
|
||||
{'height': 0, 'chain_id': 'some-id', 'is_synced': True},
|
||||
{'height': 0, 'chain_id': 'new-id', 'is_synced': False},
|
||||
],
|
||||
{'height': 0, 'chain_id': 'new-id', 'is_synced': False},
|
||||
),
|
||||
(
|
||||
'Store a chain, update it, store another chain.',
|
||||
[
|
||||
{'height': 0, 'chain_id': 'some-id', 'is_synced': True},
|
||||
{'height': 0, 'chain_id': 'some-id', 'is_synced': False},
|
||||
{'height': 10, 'chain_id': 'another-id', 'is_synced': True},
|
||||
],
|
||||
{'height': 10, 'chain_id': 'another-id', 'is_synced': True},
|
||||
),
|
||||
])
|
||||
def test_store_abci_chain(description, stores, expected):
|
||||
conn = connect()
|
||||
|
||||
for store in stores:
|
||||
query.store_abci_chain(conn, **store)
|
||||
|
||||
actual = query.get_latest_abci_chain(conn)
|
||||
assert expected == actual, description
|
||||
|
@ -24,7 +24,7 @@ def test_init_creates_db_tables_and_indexes():
|
||||
collection_names = conn.conn[dbname].collection_names()
|
||||
assert set(collection_names) == {
|
||||
'transactions', 'assets', 'metadata', 'blocks', 'utxos', 'pre_commit',
|
||||
'validators'
|
||||
'validators', 'abci_chains',
|
||||
}
|
||||
|
||||
indexes = conn.conn[dbname]['assets'].index_information().keys()
|
||||
@ -46,6 +46,9 @@ def test_init_creates_db_tables_and_indexes():
|
||||
indexes = conn.conn[dbname]['validators'].index_information().keys()
|
||||
assert set(indexes) == {'_id_', 'height'}
|
||||
|
||||
indexes = conn.conn[dbname]['abci_chains'].index_information().keys()
|
||||
assert set(indexes) == {'_id_', 'height', 'chain_id'}
|
||||
|
||||
|
||||
def test_init_database_fails_if_db_exists():
|
||||
import bigchaindb
|
||||
@ -79,7 +82,8 @@ def test_create_tables():
|
||||
collection_names = conn.conn[dbname].collection_names()
|
||||
assert set(collection_names) == {
|
||||
'transactions', 'assets', 'metadata', 'blocks', 'utxos', 'validators',
|
||||
'pre_commit'}
|
||||
'pre_commit', 'abci_chains',
|
||||
}
|
||||
|
||||
|
||||
def test_create_secondary_indexes():
|
||||
|
@ -16,7 +16,8 @@ def validator_pub_key():
|
||||
@pytest.fixture
|
||||
def init_chain_request():
|
||||
addr = codecs.decode(b'9FD479C869C7D7E7605BF99293457AA5D80C3033', 'hex')
|
||||
pk = codecs.decode(b'VAgFZtYw8bNR5TMZHFOBDWk9cAmEu3/c6JgRBmddbbI=', 'base64')
|
||||
pk = codecs.decode(b'VAgFZtYw8bNR5TMZHFOBDWk9cAmEu3/c6JgRBmddbbI=',
|
||||
'base64')
|
||||
val_a = types.Validator(address=addr, power=10,
|
||||
pub_key=types.PubKey(type='ed25519', data=pk))
|
||||
|
||||
|
@ -2,17 +2,28 @@
|
||||
# SPDX-License-Identifier: (Apache-2.0 AND CC-BY-4.0)
|
||||
# Code is Apache-2.0 and docs are CC-BY-4.0
|
||||
|
||||
import codecs
|
||||
import json
|
||||
import pytest
|
||||
import random
|
||||
|
||||
from abci.types_pb2 import (
|
||||
PubKey,
|
||||
ResponseInitChain,
|
||||
RequestInitChain,
|
||||
RequestInfo,
|
||||
RequestBeginBlock,
|
||||
RequestEndBlock
|
||||
RequestEndBlock,
|
||||
Validator,
|
||||
)
|
||||
|
||||
from bigchaindb import App
|
||||
from bigchaindb.backend.localmongodb import query
|
||||
from bigchaindb.common.crypto import generate_key_pair
|
||||
from bigchaindb.core import (CodeTypeOk,
|
||||
CodeTypeError,
|
||||
)
|
||||
from bigchaindb.lib import Block
|
||||
from bigchaindb.upsert_validator.validator_utils import new_validator_set
|
||||
from bigchaindb.tendermint_utils import public_key_to_base64
|
||||
|
||||
@ -24,6 +35,173 @@ def encode_tx_to_bytes(transaction):
|
||||
return json.dumps(transaction.to_dict()).encode('utf8')
|
||||
|
||||
|
||||
def generate_address():
|
||||
return ''.join(random.choices('1,2,3,4,5,6,7,8,9,A,B,C,D,E,F'.split(','),
|
||||
k=40)).encode()
|
||||
|
||||
|
||||
def generate_validator():
|
||||
addr = codecs.decode(generate_address(), 'hex')
|
||||
pk, _ = generate_key_pair()
|
||||
pub_key = PubKey(type='ed25519', data=pk.encode())
|
||||
val = Validator(address=addr, power=10, pub_key=pub_key)
|
||||
return val
|
||||
|
||||
|
||||
def generate_init_chain_request(chain_id, vals=None):
|
||||
vals = vals if vals is not None else [generate_validator()]
|
||||
return RequestInitChain(validators=vals, chain_id=chain_id)
|
||||
|
||||
|
||||
def test_init_chain_successfully_registers_chain(b):
|
||||
request = generate_init_chain_request('chain-XYZ')
|
||||
res = App(b).init_chain(request)
|
||||
assert res == ResponseInitChain()
|
||||
chain = query.get_latest_abci_chain(b.connection)
|
||||
assert chain == {'height': 0, 'chain_id': 'chain-XYZ', 'is_synced': True}
|
||||
assert query.get_latest_block(b.connection) == {
|
||||
'height': 0,
|
||||
'app_hash': '',
|
||||
'transactions': [],
|
||||
}
|
||||
|
||||
|
||||
def test_init_chain_ignores_invalid_init_chain_requests(b):
|
||||
validators = [generate_validator()]
|
||||
request = generate_init_chain_request('chain-XYZ', validators)
|
||||
res = App(b).init_chain(request)
|
||||
assert res == ResponseInitChain()
|
||||
|
||||
validator_set = query.get_validator_set(b.connection)
|
||||
|
||||
invalid_requests = [
|
||||
request, # the same request again
|
||||
# different validator set
|
||||
generate_init_chain_request('chain-XYZ'),
|
||||
# different chain ID
|
||||
generate_init_chain_request('chain-ABC', validators),
|
||||
]
|
||||
for r in invalid_requests:
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).init_chain(r)
|
||||
# assert nothing changed - neither validator set, nor chain ID
|
||||
new_validator_set = query.get_validator_set(b.connection)
|
||||
assert new_validator_set == validator_set
|
||||
new_chain_id = query.get_latest_abci_chain(b.connection)['chain_id']
|
||||
assert new_chain_id == 'chain-XYZ'
|
||||
assert query.get_latest_block(b.connection) == {
|
||||
'height': 0,
|
||||
'app_hash': '',
|
||||
'transactions': [],
|
||||
}
|
||||
|
||||
|
||||
def test_init_chain_recognizes_new_chain_after_migration(b):
|
||||
validators = [generate_validator()]
|
||||
request = generate_init_chain_request('chain-XYZ', validators)
|
||||
res = App(b).init_chain(request)
|
||||
assert res == ResponseInitChain()
|
||||
|
||||
validator_set = query.get_validator_set(b.connection)['validators']
|
||||
|
||||
# simulate a migration
|
||||
query.store_block(b.connection, Block(app_hash='', height=1,
|
||||
transactions=[])._asdict())
|
||||
b.migrate_abci_chain()
|
||||
|
||||
# the same or other mismatching requests are ignored
|
||||
invalid_requests = [
|
||||
request,
|
||||
generate_init_chain_request('unknown', validators),
|
||||
generate_init_chain_request('chain-XYZ'),
|
||||
generate_init_chain_request('chain-XYZ-migrated-at-height-1'),
|
||||
]
|
||||
for r in invalid_requests:
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).init_chain(r)
|
||||
assert query.get_latest_abci_chain(b.connection) == {
|
||||
'chain_id': 'chain-XYZ-migrated-at-height-1',
|
||||
'is_synced': False,
|
||||
'height': 2,
|
||||
}
|
||||
new_validator_set = query.get_validator_set(b.connection)['validators']
|
||||
assert new_validator_set == validator_set
|
||||
|
||||
# a request with the matching chain ID and matching validator set
|
||||
# completes the migration
|
||||
request = generate_init_chain_request('chain-XYZ-migrated-at-height-1',
|
||||
validators)
|
||||
res = App(b).init_chain(request)
|
||||
assert res == ResponseInitChain()
|
||||
assert query.get_latest_abci_chain(b.connection) == {
|
||||
'chain_id': 'chain-XYZ-migrated-at-height-1',
|
||||
'is_synced': True,
|
||||
'height': 2,
|
||||
}
|
||||
assert query.get_latest_block(b.connection) == {
|
||||
'height': 2,
|
||||
'app_hash': '',
|
||||
'transactions': [],
|
||||
}
|
||||
|
||||
# requests with old chain ID and other requests are ignored
|
||||
invalid_requests = [
|
||||
request,
|
||||
generate_init_chain_request('chain-XYZ', validators),
|
||||
generate_init_chain_request('chain-XYZ-migrated-at-height-1'),
|
||||
]
|
||||
for r in invalid_requests:
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).init_chain(r)
|
||||
assert query.get_latest_abci_chain(b.connection) == {
|
||||
'chain_id': 'chain-XYZ-migrated-at-height-1',
|
||||
'is_synced': True,
|
||||
'height': 2,
|
||||
}
|
||||
new_validator_set = query.get_validator_set(b.connection)['validators']
|
||||
assert new_validator_set == validator_set
|
||||
assert query.get_latest_block(b.connection) == {
|
||||
'height': 2,
|
||||
'app_hash': '',
|
||||
'transactions': [],
|
||||
}
|
||||
|
||||
|
||||
def test_info(b):
|
||||
r = RequestInfo()
|
||||
app = App(b)
|
||||
|
||||
res = app.info(r)
|
||||
assert res.last_block_height == 0
|
||||
assert res.last_block_app_hash == b''
|
||||
|
||||
b.store_block(Block(app_hash='1', height=1, transactions=[])._asdict())
|
||||
res = app.info(r)
|
||||
assert res.last_block_height == 1
|
||||
assert res.last_block_app_hash == b'1'
|
||||
|
||||
# simulate a migration and assert the height is shifted
|
||||
b.store_abci_chain(2, 'chain-XYZ')
|
||||
app = App(b)
|
||||
b.store_block(Block(app_hash='2', height=2, transactions=[])._asdict())
|
||||
res = app.info(r)
|
||||
assert res.last_block_height == 0
|
||||
assert res.last_block_app_hash == b'2'
|
||||
|
||||
b.store_block(Block(app_hash='3', height=3, transactions=[])._asdict())
|
||||
res = app.info(r)
|
||||
assert res.last_block_height == 1
|
||||
assert res.last_block_app_hash == b'3'
|
||||
|
||||
# it's always the latest migration that is taken into account
|
||||
b.store_abci_chain(4, 'chain-XYZ-new')
|
||||
app = App(b)
|
||||
b.store_block(Block(app_hash='4', height=4, transactions=[])._asdict())
|
||||
res = app.info(r)
|
||||
assert res.last_block_height == 0
|
||||
assert res.last_block_app_hash == b'4'
|
||||
|
||||
|
||||
def test_check_tx__signed_create_is_ok(b):
|
||||
from bigchaindb import App
|
||||
from bigchaindb.models import Transaction
|
||||
@ -57,7 +235,6 @@ def test_check_tx__unsigned_create_is_error(b):
|
||||
assert result.code == CodeTypeError
|
||||
|
||||
|
||||
@pytest.mark.bdb
|
||||
def test_deliver_tx__valid_create_updates_db(b, init_chain_request):
|
||||
from bigchaindb import App
|
||||
from bigchaindb.models import Transaction
|
||||
@ -225,6 +402,17 @@ def test_store_pre_commit_state_in_end_block(b, alice, init_chain_request):
|
||||
assert resp['height'] == 100
|
||||
assert resp['transactions'] == [tx.id]
|
||||
|
||||
# simulate a chain migration and assert the height is shifted
|
||||
b.store_abci_chain(100, 'new-chain')
|
||||
app = App(b)
|
||||
app.begin_block(begin_block)
|
||||
app.deliver_tx(encode_tx_to_bytes(tx))
|
||||
app.end_block(RequestEndBlock(height=1))
|
||||
resp = query.get_pre_commit_state(b.connection, PRE_COMMIT_ID)
|
||||
assert resp['commit_id'] == PRE_COMMIT_ID
|
||||
assert resp['height'] == 101
|
||||
assert resp['transactions'] == [tx.id]
|
||||
|
||||
|
||||
def test_new_validator_set(b):
|
||||
node1 = {'pub_key': {'type': 'ed25519',
|
||||
@ -247,3 +435,45 @@ def test_new_validator_set(b):
|
||||
'voting_power': u['power']})
|
||||
|
||||
assert updated_validator_set == updated_validators
|
||||
|
||||
|
||||
def test_info_aborts_if_chain_is_not_synced(b):
|
||||
b.store_abci_chain(0, 'chain-XYZ', False)
|
||||
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).info(RequestInfo())
|
||||
|
||||
|
||||
def test_check_tx_aborts_if_chain_is_not_synced(b):
|
||||
b.store_abci_chain(0, 'chain-XYZ', False)
|
||||
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).check_tx('some bytes')
|
||||
|
||||
|
||||
def test_begin_aborts_if_chain_is_not_synced(b):
|
||||
b.store_abci_chain(0, 'chain-XYZ', False)
|
||||
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).info(RequestBeginBlock())
|
||||
|
||||
|
||||
def test_deliver_tx_aborts_if_chain_is_not_synced(b):
|
||||
b.store_abci_chain(0, 'chain-XYZ', False)
|
||||
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).deliver_tx('some bytes')
|
||||
|
||||
|
||||
def test_end_block_aborts_if_chain_is_not_synced(b):
|
||||
b.store_abci_chain(0, 'chain-XYZ', False)
|
||||
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).info(RequestEndBlock())
|
||||
|
||||
|
||||
def test_commit_aborts_if_chain_is_not_synced(b):
|
||||
b.store_abci_chain(0, 'chain-XYZ', False)
|
||||
|
||||
with pytest.raises(SystemExit):
|
||||
App(b).commit()
|
||||
|
@ -15,6 +15,7 @@ import pytest
|
||||
from pymongo import MongoClient
|
||||
|
||||
from bigchaindb import backend
|
||||
from bigchaindb.lib import Block
|
||||
|
||||
|
||||
pytestmark = pytest.mark.tendermint
|
||||
@ -441,3 +442,35 @@ def test_validation_with_transaction_buffer(b):
|
||||
assert not b.is_valid_transaction(create_tx, [create_tx])
|
||||
assert not b.is_valid_transaction(transfer_tx, [create_tx, transfer_tx])
|
||||
assert not b.is_valid_transaction(double_spend, [create_tx, transfer_tx])
|
||||
|
||||
|
||||
@pytest.mark.bdb
|
||||
def test_migrate_abci_chain_yields_on_genesis(b):
|
||||
b.migrate_abci_chain()
|
||||
latest_chain = b.get_latest_abci_chain()
|
||||
assert latest_chain is None
|
||||
|
||||
|
||||
@pytest.mark.bdb
|
||||
@pytest.mark.parametrize('chain,block_height,expected', [
|
||||
(
|
||||
(1, 'chain-XYZ', True),
|
||||
4,
|
||||
{'height': 5, 'chain_id': 'chain-XYZ-migrated-at-height-4',
|
||||
'is_synced': False},
|
||||
),
|
||||
(
|
||||
(5, 'chain-XYZ-migrated-at-height-4', True),
|
||||
13,
|
||||
{'height': 14, 'chain_id': 'chain-XYZ-migrated-at-height-13',
|
||||
'is_synced': False},
|
||||
),
|
||||
])
|
||||
def test_migrate_abci_chain_generates_new_chains(b, chain, block_height,
|
||||
expected):
|
||||
b.store_abci_chain(*chain)
|
||||
b.store_block(Block(app_hash='', height=block_height,
|
||||
transactions=[])._asdict())
|
||||
b.migrate_abci_chain()
|
||||
latest_chain = b.get_latest_abci_chain()
|
||||
assert latest_chain == expected
|
||||
|
@ -12,6 +12,7 @@ except ImportError:
|
||||
|
||||
import pytest
|
||||
|
||||
|
||||
pytestmark = pytest.mark.tendermint
|
||||
|
||||
|
||||
|
@ -5,6 +5,7 @@
|
||||
from functools import singledispatch
|
||||
|
||||
from bigchaindb.backend.localmongodb.connection import LocalMongoDBConnection
|
||||
from bigchaindb.backend.schema import TABLES
|
||||
|
||||
|
||||
@singledispatch
|
||||
@ -14,13 +15,8 @@ def flush_db(connection, dbname):
|
||||
|
||||
@flush_db.register(LocalMongoDBConnection)
|
||||
def flush_localmongo_db(connection, dbname):
|
||||
connection.conn[dbname].bigchain.delete_many({})
|
||||
connection.conn[dbname].blocks.delete_many({})
|
||||
connection.conn[dbname].transactions.delete_many({})
|
||||
connection.conn[dbname].assets.delete_many({})
|
||||
connection.conn[dbname].metadata.delete_many({})
|
||||
connection.conn[dbname].utxos.delete_many({})
|
||||
connection.conn[dbname].validators.delete_many({})
|
||||
for t in TABLES:
|
||||
getattr(connection.conn[dbname], t).delete_many({})
|
||||
|
||||
|
||||
def generate_block(bigchain):
|
||||
|
Loading…
x
Reference in New Issue
Block a user