mirror of
https://github.com/bigchaindb/bigchaindb.git
synced 2024-10-13 13:34:05 +00:00

Removed support for hashlocks Unskipped tests that were skipped waiting for divisible assets Fixed remaining tests
198 lines
5.9 KiB
Python
198 lines
5.9 KiB
Python
import time
|
|
from unittest.mock import patch
|
|
|
|
import rethinkdb as r
|
|
|
|
from multipipes import Pipe
|
|
|
|
|
|
def test_filter_by_assignee(b, signed_create_tx):
|
|
from bigchaindb.pipelines.block import BlockPipeline
|
|
|
|
block_maker = BlockPipeline()
|
|
|
|
tx = signed_create_tx.to_dict()
|
|
tx.update({'assignee': b.me, 'assignment_timestamp': 111})
|
|
|
|
# filter_tx has side effects on the `tx` instance by popping 'assignee'
|
|
# and 'assignment_timestamp'
|
|
filtered_tx = block_maker.filter_tx(tx)
|
|
assert filtered_tx == tx
|
|
assert 'assignee' not in filtered_tx
|
|
assert 'assignment_timestamp' not in filtered_tx
|
|
|
|
tx = signed_create_tx.to_dict()
|
|
tx.update({'assignee': 'nobody'})
|
|
|
|
assert block_maker.filter_tx(tx) is None
|
|
|
|
|
|
def test_validate_transaction(b, create_tx):
|
|
from bigchaindb.pipelines.block import BlockPipeline
|
|
|
|
block_maker = BlockPipeline()
|
|
|
|
assert block_maker.validate_tx(create_tx.to_dict()) is None
|
|
|
|
valid_tx = create_tx.sign([b.me_private])
|
|
assert block_maker.validate_tx(valid_tx.to_dict()) == valid_tx
|
|
|
|
|
|
def test_create_block(b, user_vk):
|
|
from bigchaindb.models import Transaction
|
|
from bigchaindb.pipelines.block import BlockPipeline
|
|
|
|
block_maker = BlockPipeline()
|
|
|
|
for i in range(100):
|
|
tx = Transaction.create([b.me], [([user_vk], 1)])
|
|
tx = tx.sign([b.me_private])
|
|
block_maker.create(tx)
|
|
|
|
# force the output triggering a `timeout`
|
|
block_doc = block_maker.create(None, timeout=True)
|
|
|
|
assert len(block_doc.transactions) == 100
|
|
|
|
|
|
def test_write_block(b, user_vk):
|
|
from bigchaindb.models import Block, Transaction
|
|
from bigchaindb.pipelines.block import BlockPipeline
|
|
|
|
block_maker = BlockPipeline()
|
|
|
|
txs = []
|
|
for i in range(100):
|
|
tx = Transaction.create([b.me], [([user_vk], 1)])
|
|
tx = tx.sign([b.me_private])
|
|
txs.append(tx)
|
|
|
|
block_doc = b.create_block(txs)
|
|
block_maker.write(block_doc)
|
|
expected = b.connection.run(r.table('bigchain').get(block_doc.id))
|
|
expected = Block.from_dict(expected)
|
|
|
|
assert expected == block_doc
|
|
|
|
|
|
def test_duplicate_transaction(b, user_vk):
|
|
from bigchaindb.models import Transaction
|
|
from bigchaindb.pipelines import block
|
|
block_maker = block.BlockPipeline()
|
|
|
|
txs = []
|
|
for i in range(10):
|
|
tx = Transaction.create([b.me], [([user_vk], 1)])
|
|
tx = tx.sign([b.me_private])
|
|
txs.append(tx)
|
|
|
|
block_doc = b.create_block(txs)
|
|
block_maker.write(block_doc)
|
|
|
|
# block is in bigchain
|
|
assert b.connection.run(r.table('bigchain').get(block_doc.id)) == block_doc.to_dict()
|
|
|
|
b.write_transaction(txs[0])
|
|
|
|
# verify tx is in the backlog
|
|
assert b.connection.run(r.table('backlog').get(txs[0].id)) is not None
|
|
|
|
# try to validate a transaction that's already in the chain; should not
|
|
# work
|
|
assert block_maker.validate_tx(txs[0].to_dict()) is None
|
|
|
|
# duplicate tx should be removed from backlog
|
|
assert b.connection.run(r.table('backlog').get(txs[0].id)) is None
|
|
|
|
|
|
def test_delete_tx(b, user_vk):
|
|
from bigchaindb.models import Transaction
|
|
from bigchaindb.pipelines.block import BlockPipeline
|
|
block_maker = BlockPipeline()
|
|
for i in range(100):
|
|
tx = Transaction.create([b.me], [([user_vk], 1)])
|
|
tx = tx.sign([b.me_private])
|
|
block_maker.create(tx)
|
|
# make sure the tx appears in the backlog
|
|
b.write_transaction(tx)
|
|
|
|
# force the output triggering a `timeout`
|
|
block_doc = block_maker.create(None, timeout=True)
|
|
|
|
for tx in block_doc.to_dict()['block']['transactions']:
|
|
returned_tx = b.connection.run(r.table('backlog').get(tx['id']))
|
|
returned_tx.pop('assignee')
|
|
returned_tx.pop('assignment_timestamp')
|
|
assert returned_tx == tx
|
|
|
|
returned_block = block_maker.delete_tx(block_doc)
|
|
|
|
assert returned_block == block_doc
|
|
|
|
for tx in block_doc.to_dict()['block']['transactions']:
|
|
assert b.connection.run(r.table('backlog').get(tx['id'])) is None
|
|
|
|
|
|
def test_prefeed(b, user_vk):
|
|
import random
|
|
from bigchaindb.models import Transaction
|
|
from bigchaindb.pipelines.block import initial
|
|
|
|
for i in range(100):
|
|
tx = Transaction.create([b.me], [([user_vk], 1)],
|
|
{'msg': random.random()})
|
|
tx = tx.sign([b.me_private])
|
|
b.write_transaction(tx)
|
|
|
|
backlog = initial()
|
|
|
|
assert len(list(backlog)) == 100
|
|
|
|
|
|
@patch('bigchaindb.pipelines.block.create_pipeline')
|
|
def test_start(create_pipeline):
|
|
from bigchaindb.pipelines import block
|
|
|
|
pipeline = block.start()
|
|
assert create_pipeline.called
|
|
assert create_pipeline.return_value.setup.called
|
|
assert create_pipeline.return_value.start.called
|
|
assert pipeline == create_pipeline.return_value
|
|
|
|
|
|
def test_full_pipeline(b, user_vk):
|
|
import random
|
|
from bigchaindb.models import Block, Transaction
|
|
from bigchaindb.pipelines.block import create_pipeline, get_changefeed
|
|
|
|
outpipe = Pipe()
|
|
|
|
count_assigned_to_me = 0
|
|
for i in range(100):
|
|
tx = Transaction.create([b.me], [([user_vk], 1)],
|
|
{'msg': random.random()})
|
|
tx = tx.sign([b.me_private]).to_dict()
|
|
assignee = random.choice([b.me, 'aaa', 'bbb', 'ccc'])
|
|
tx['assignee'] = assignee
|
|
tx['assignment_timestamp'] = time.time()
|
|
if assignee == b.me:
|
|
count_assigned_to_me += 1
|
|
b.connection.run(r.table('backlog').insert(tx, durability='hard'))
|
|
|
|
assert b.connection.run(r.table('backlog').count()) == 100
|
|
|
|
pipeline = create_pipeline()
|
|
pipeline.setup(indata=get_changefeed(), outdata=outpipe)
|
|
pipeline.start()
|
|
|
|
time.sleep(2)
|
|
pipeline.terminate()
|
|
|
|
block_doc = outpipe.get()
|
|
chained_block = b.connection.run(r.table('bigchain').get(block_doc.id))
|
|
chained_block = Block.from_dict(chained_block)
|
|
|
|
assert len(block_doc.transactions) == count_assigned_to_me
|
|
assert chained_block == block_doc
|
|
assert b.connection.run(r.table('backlog').count()) == 100 - count_assigned_to_me
|