mirror of
https://github.com/kaspanet/kaspad.git
synced 2025-09-14 05:20:11 +00:00

* Add StagingArea struct * Implemented staging areas in blockStore * Move blockStagingShard to separate folder * Apply staging shard to acceptanceDataStore * Update blockHeaderStore with StagingArea * Add StagingArea to BlockRelationStore * Add StagingArea to blockStatusStore * Add StagingArea to consensusStateStore * Add StagingArea to daaBlocksStore * Add StagingArea to finalityStore * Add StagingArea to ghostdagDataStore * Add StagingArea to headersSelectedChainStore and headersSelectedTipStore * Add StagingArea to multisetStore * Add StagingArea to pruningStore * Add StagingArea to reachabilityDataStore * Add StagingArea to utxoDiffStore * Fix forgotten compilation error * Update reachability manager and some more things with StagingArea * Add StagingArea to dagTopologyManager, and some more * Add StagingArea to GHOSTDAGManager, and some more * Add StagingArea to difficultyManager, and some more * Add StagingArea to dagTraversalManager, and some more * Add StagingArea to headerTipsManager, and some more * Add StagingArea to constnsusStateManager, pastMedianTimeManager * Add StagingArea to transactionValidator * Add StagingArea to finalityManager * Add StagingArea to mergeDepthManager * Add StagingArea to pruningManager * Add StagingArea to rest of ValidateAndInsertBlock * Add StagingArea to blockValidator * Add StagingArea to coinbaseManager * Add StagingArea to syncManager * Add StagingArea to blockBuilder * Update consensus with StagingArea * Add StagingArea to ghostdag2 * Fix remaining compilation errors * Update names of stagingShards * Fix forgotten stagingArea passing * Mark stagingShard.isCommited = true once commited * Move isStaged to stagingShard, so that it's available without going through store * Make blockHeaderStore count be avilable from stagingShard * Fix remaining forgotten stagingArea passing * commitAllChanges should call dbTx.Commit in the end * Fix all tests tests in blockValidator * Fix all tests in consensusStateManager and some more * Fix all tests in pruningManager * Add many missing stagingAreas in tests * Fix many tests * Fix most of all other tests * Fix ghostdag_test.go * Add comment to StagingArea * Make list of StagingShards an array * Add comment to StagingShardID * Make sure all staging shards are pointer-receiver * Undo bucket rename in block_store * Typo: isCommited -> isCommitted * Add comment explaining why stagingArea.shards is an array
164 lines
4.0 KiB
Go
164 lines
4.0 KiB
Go
package main
|
|
|
|
import (
|
|
"compress/gzip"
|
|
"fmt"
|
|
"math/rand"
|
|
"os"
|
|
"time"
|
|
|
|
"github.com/kaspanet/kaspad/domain/consensus/model"
|
|
|
|
"github.com/kaspanet/kaspad/domain/consensus"
|
|
"github.com/kaspanet/kaspad/domain/consensus/model/externalapi"
|
|
"github.com/kaspanet/kaspad/domain/consensus/model/testapi"
|
|
"github.com/kaspanet/kaspad/domain/consensus/utils/consensushashing"
|
|
"github.com/kaspanet/kaspad/domain/dagconfig"
|
|
)
|
|
|
|
func testReorg(cfg *configFlags) {
|
|
params := dagconfig.DevnetParams
|
|
params.SkipProofOfWork = true
|
|
|
|
factory := consensus.NewFactory()
|
|
tc, teardown, err := factory.NewTestConsensus(¶ms, false, "ReorgHonest")
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
defer teardown(false)
|
|
|
|
f, err := os.Open(cfg.DAGFile)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
defer f.Close()
|
|
|
|
gzipReader, err := gzip.NewReader(f)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
defer gzipReader.Close()
|
|
|
|
_, err = tc.MineJSON(gzipReader, testapi.MineJSONBlockTypeUTXOValidBlock)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
tcAttacker, teardownAttacker, err := factory.NewTestConsensus(¶ms, false, "ReorgAttacker")
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
defer teardownAttacker(false)
|
|
|
|
virtualSelectedParent, err := tc.GetVirtualSelectedParent()
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
stagingArea := model.NewStagingArea()
|
|
virtualSelectedParentGHOSTDAGData, err := tc.GHOSTDAGDataStore().Get(tc.DatabaseContext(), stagingArea, virtualSelectedParent)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
log.Infof("Selected tip blue score %d", virtualSelectedParentGHOSTDAGData.BlueScore())
|
|
|
|
sideChain := make([]*externalapi.DomainBlock, 0)
|
|
|
|
for i := uint64(0); ; i++ {
|
|
tips, err := tcAttacker.Tips()
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
block, _, err := tcAttacker.BuildBlockWithParents(tips, nil, nil)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
// We change the nonce of the first block so its hash won't be similar to any of the
|
|
// honest DAG blocks. As a result the rest of the side chain should have unique hashes
|
|
// as well.
|
|
if i == 0 {
|
|
mutableHeader := block.Header.ToMutable()
|
|
mutableHeader.SetNonce(uint64(rand.NewSource(84147).Int63()))
|
|
block.Header = mutableHeader.ToImmutable()
|
|
}
|
|
|
|
_, err = tcAttacker.ValidateAndInsertBlock(block)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
sideChain = append(sideChain, block)
|
|
|
|
if i%100 == 0 {
|
|
log.Infof("Attacker side chain mined %d blocks", i)
|
|
}
|
|
|
|
blockHash := consensushashing.BlockHash(block)
|
|
ghostdagData, err := tcAttacker.GHOSTDAGDataStore().Get(tcAttacker.DatabaseContext(), stagingArea, blockHash)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
if virtualSelectedParentGHOSTDAGData.BlueWork().Cmp(ghostdagData.BlueWork()) == -1 {
|
|
break
|
|
}
|
|
}
|
|
|
|
sideChainTipHash := consensushashing.BlockHash(sideChain[len(sideChain)-1])
|
|
sideChainTipGHOSTDAGData, err := tcAttacker.GHOSTDAGDataStore().Get(tcAttacker.DatabaseContext(), stagingArea, sideChainTipHash)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
log.Infof("Side chain tip (%s) blue score %d", sideChainTipHash, sideChainTipGHOSTDAGData.BlueScore())
|
|
|
|
doneChan := make(chan struct{})
|
|
spawn("add-sidechain-to-honest", func() {
|
|
for i, block := range sideChain {
|
|
if i%100 == 0 {
|
|
log.Infof("Validated %d blocks from the attacker chain", i)
|
|
}
|
|
_, err := tc.ValidateAndInsertBlock(block)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
}
|
|
|
|
doneChan <- struct{}{}
|
|
})
|
|
|
|
const timeout = 10 * time.Minute
|
|
select {
|
|
case <-doneChan:
|
|
case <-time.After(timeout):
|
|
fail("Adding the side chain took more than %s", timeout)
|
|
}
|
|
|
|
sideChainTipGHOSTDAGData, err = tc.GHOSTDAGDataStore().Get(tc.DatabaseContext(), stagingArea, sideChainTipHash)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
log.Infof("Side chain tip (%s) blue score %d", sideChainTipHash, sideChainTipGHOSTDAGData.BlueScore())
|
|
|
|
newVirtualSelectedParent, err := tc.GetVirtualSelectedParent()
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
if !newVirtualSelectedParent.Equal(sideChainTipHash) {
|
|
fail("No reorg happened")
|
|
}
|
|
}
|
|
|
|
func fail(format string, args ...interface{}) {
|
|
msg := fmt.Sprintf(format, args...)
|
|
fmt.Fprintln(os.Stderr, msg)
|
|
log.Criticalf(msg)
|
|
backendLog.Close()
|
|
os.Exit(1)
|
|
}
|