mirror of
https://github.com/kaspanet/kaspad.git
synced 2025-09-13 13:00:10 +00:00

* Add StagingArea struct * Implemented staging areas in blockStore * Move blockStagingShard to separate folder * Apply staging shard to acceptanceDataStore * Update blockHeaderStore with StagingArea * Add StagingArea to BlockRelationStore * Add StagingArea to blockStatusStore * Add StagingArea to consensusStateStore * Add StagingArea to daaBlocksStore * Add StagingArea to finalityStore * Add StagingArea to ghostdagDataStore * Add StagingArea to headersSelectedChainStore and headersSelectedTipStore * Add StagingArea to multisetStore * Add StagingArea to pruningStore * Add StagingArea to reachabilityDataStore * Add StagingArea to utxoDiffStore * Fix forgotten compilation error * Update reachability manager and some more things with StagingArea * Add StagingArea to dagTopologyManager, and some more * Add StagingArea to GHOSTDAGManager, and some more * Add StagingArea to difficultyManager, and some more * Add StagingArea to dagTraversalManager, and some more * Add StagingArea to headerTipsManager, and some more * Add StagingArea to constnsusStateManager, pastMedianTimeManager * Add StagingArea to transactionValidator * Add StagingArea to finalityManager * Add StagingArea to mergeDepthManager * Add StagingArea to pruningManager * Add StagingArea to rest of ValidateAndInsertBlock * Add StagingArea to blockValidator * Add StagingArea to coinbaseManager * Add StagingArea to syncManager * Add StagingArea to blockBuilder * Update consensus with StagingArea * Add StagingArea to ghostdag2 * Fix remaining compilation errors * Update names of stagingShards * Fix forgotten stagingArea passing * Mark stagingShard.isCommited = true once commited * Move isStaged to stagingShard, so that it's available without going through store * Make blockHeaderStore count be avilable from stagingShard * Fix remaining forgotten stagingArea passing * commitAllChanges should call dbTx.Commit in the end * Fix all tests tests in blockValidator * Fix all tests in consensusStateManager and some more * Fix all tests in pruningManager * Add many missing stagingAreas in tests * Fix many tests * Fix most of all other tests * Fix ghostdag_test.go * Add comment to StagingArea * Make list of StagingShards an array * Add comment to StagingShardID * Make sure all staging shards are pointer-receiver * Undo bucket rename in block_store * Typo: isCommited -> isCommitted * Add comment explaining why stagingArea.shards is an array
412 lines
11 KiB
Go
412 lines
11 KiB
Go
package ghostdag2
|
|
|
|
import (
|
|
"sort"
|
|
|
|
"github.com/kaspanet/kaspad/util/difficulty"
|
|
|
|
"math/big"
|
|
|
|
"github.com/kaspanet/kaspad/domain/consensus/model"
|
|
"github.com/kaspanet/kaspad/domain/consensus/model/externalapi"
|
|
)
|
|
|
|
type ghostdagHelper struct {
|
|
k model.KType
|
|
dataStore model.GHOSTDAGDataStore
|
|
dbAccess model.DBReader
|
|
dagTopologyManager model.DAGTopologyManager
|
|
headerStore model.BlockHeaderStore
|
|
}
|
|
|
|
// New creates a new instance of this alternative ghostdag impl
|
|
func New(
|
|
databaseContext model.DBReader,
|
|
dagTopologyManager model.DAGTopologyManager,
|
|
ghostdagDataStore model.GHOSTDAGDataStore,
|
|
headerStore model.BlockHeaderStore,
|
|
k model.KType) model.GHOSTDAGManager {
|
|
|
|
return &ghostdagHelper{
|
|
dbAccess: databaseContext,
|
|
dagTopologyManager: dagTopologyManager,
|
|
dataStore: ghostdagDataStore,
|
|
headerStore: headerStore,
|
|
k: k,
|
|
}
|
|
}
|
|
|
|
/* --------------------------------------------- */
|
|
|
|
func (gh *ghostdagHelper) GHOSTDAG(stagingArea *model.StagingArea, blockCandidate *externalapi.DomainHash) error {
|
|
myWork := new(big.Int)
|
|
maxWork := new(big.Int)
|
|
var myScore uint64
|
|
var spScore uint64
|
|
/* find the selectedParent */
|
|
blockParents, err := gh.dagTopologyManager.Parents(stagingArea, blockCandidate)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
var selectedParent = blockParents[0]
|
|
for _, parent := range blockParents {
|
|
blockData, err := gh.dataStore.Get(gh.dbAccess, stagingArea, parent)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
blockWork := blockData.BlueWork()
|
|
blockScore := blockData.BlueScore()
|
|
if blockWork.Cmp(maxWork) == 1 {
|
|
selectedParent = parent
|
|
maxWork = blockWork
|
|
spScore = blockScore
|
|
}
|
|
if blockWork.Cmp(maxWork) == 0 && ismoreHash(parent, selectedParent) {
|
|
selectedParent = parent
|
|
maxWork = blockWork
|
|
spScore = blockScore
|
|
}
|
|
}
|
|
myWork.Set(maxWork)
|
|
myScore = spScore
|
|
|
|
/* Goal: iterate blockCandidate's mergeSet and divide it to : blue, blues, reds. */
|
|
var mergeSetBlues = make([]*externalapi.DomainHash, 0)
|
|
var mergeSetReds = make([]*externalapi.DomainHash, 0)
|
|
var blueSet = make([]*externalapi.DomainHash, 0)
|
|
|
|
mergeSetBlues = append(mergeSetBlues, selectedParent)
|
|
|
|
mergeSetArr, err := gh.findMergeSet(stagingArea, blockParents, selectedParent)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
err = gh.sortByBlueWork(stagingArea, mergeSetArr)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = gh.findBlueSet(stagingArea, &blueSet, selectedParent)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
for _, mergeSetBlock := range mergeSetArr {
|
|
if mergeSetBlock.Equal(selectedParent) {
|
|
if !contains(selectedParent, mergeSetBlues) {
|
|
mergeSetBlues = append(mergeSetBlues, selectedParent)
|
|
blueSet = append(blueSet, selectedParent)
|
|
}
|
|
continue
|
|
}
|
|
err := gh.divideBlueRed(stagingArea, selectedParent, mergeSetBlock, &mergeSetBlues, &mergeSetReds, &blueSet)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
myScore += uint64(len(mergeSetBlues))
|
|
|
|
// We add up all the *work*(not blueWork) that all our blues and selected parent did
|
|
for _, blue := range mergeSetBlues {
|
|
header, err := gh.headerStore.BlockHeader(gh.dbAccess, stagingArea, blue)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
myWork.Add(myWork, difficulty.CalcWork(header.Bits()))
|
|
}
|
|
|
|
e := model.NewBlockGHOSTDAGData(myScore, myWork, selectedParent, mergeSetBlues, mergeSetReds, nil)
|
|
gh.dataStore.Stage(stagingArea, blockCandidate, e)
|
|
return nil
|
|
}
|
|
|
|
/* --------isMoreHash(w, selectedParent)----------------*/
|
|
func ismoreHash(parent *externalapi.DomainHash, selectedParent *externalapi.DomainHash) bool {
|
|
parentByteArray := parent.ByteArray()
|
|
selectedParentByteArray := selectedParent.ByteArray()
|
|
//Check if parentHash is more then selectedParentHash
|
|
for i := 0; i < len(parentByteArray); i++ {
|
|
switch {
|
|
case parentByteArray[i] < selectedParentByteArray[i]:
|
|
return false
|
|
case parentByteArray[i] > selectedParentByteArray[i]:
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
/* 1. blue = selectedParent.blue + blues
|
|
2. not connected to at most K blocks (from the blue group)
|
|
3. for each block at blue , check if not destroy
|
|
*/
|
|
|
|
/* ---------------divideBluesReds--------------------- */
|
|
func (gh *ghostdagHelper) divideBlueRed(stagingArea *model.StagingArea,
|
|
selectedParent *externalapi.DomainHash, desiredBlock *externalapi.DomainHash,
|
|
blues *[]*externalapi.DomainHash, reds *[]*externalapi.DomainHash, blueSet *[]*externalapi.DomainHash) error {
|
|
|
|
var k = int(gh.k)
|
|
counter := 0
|
|
|
|
var suspectsBlues = make([]*externalapi.DomainHash, 0)
|
|
isMergeBlue := true
|
|
//check that not-connected to at most k.
|
|
for _, block := range *blueSet {
|
|
isAnticone, err := gh.isAnticone(stagingArea, block, desiredBlock)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if isAnticone {
|
|
counter++
|
|
suspectsBlues = append(suspectsBlues, block)
|
|
}
|
|
if counter > k {
|
|
isMergeBlue = false
|
|
break
|
|
}
|
|
}
|
|
if !isMergeBlue {
|
|
if !contains(desiredBlock, *reds) {
|
|
*reds = append(*reds, desiredBlock)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// check that the k-cluster of each blue is still valid.
|
|
for _, blue := range suspectsBlues {
|
|
isDestroyed, err := gh.checkIfDestroy(stagingArea, blue, blueSet)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if isDestroyed {
|
|
isMergeBlue = false
|
|
break
|
|
}
|
|
}
|
|
if !isMergeBlue {
|
|
if !contains(desiredBlock, *reds) {
|
|
*reds = append(*reds, desiredBlock)
|
|
}
|
|
return nil
|
|
}
|
|
if !contains(desiredBlock, *blues) {
|
|
*blues = append(*blues, desiredBlock)
|
|
}
|
|
if !contains(desiredBlock, *blueSet) {
|
|
*blueSet = append(*blueSet, desiredBlock)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
/* ---------------isAnticone-------------------------- */
|
|
func (gh *ghostdagHelper) isAnticone(stagingArea *model.StagingArea, blockA, blockB *externalapi.DomainHash) (bool, error) {
|
|
isAAncestorOfAB, err := gh.dagTopologyManager.IsAncestorOf(stagingArea, blockA, blockB)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
isBAncestorOfA, err := gh.dagTopologyManager.IsAncestorOf(stagingArea, blockB, blockA)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
return !isAAncestorOfAB && !isBAncestorOfA, nil
|
|
|
|
}
|
|
|
|
/* ----------------validateKCluster------------------- */
|
|
func (gh *ghostdagHelper) validateKCluster(stagingArea *model.StagingArea, chain *externalapi.DomainHash,
|
|
checkedBlock *externalapi.DomainHash, counter *int, blueSet *[]*externalapi.DomainHash) (bool, error) {
|
|
|
|
var k = int(gh.k)
|
|
isAnticone, err := gh.isAnticone(stagingArea, chain, checkedBlock)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if isAnticone {
|
|
if *counter > k {
|
|
return false, nil
|
|
}
|
|
ifDestroy, err := gh.checkIfDestroy(stagingArea, chain, blueSet)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if ifDestroy {
|
|
return false, nil
|
|
}
|
|
*counter++
|
|
return true, nil
|
|
}
|
|
isAncestorOf, err := gh.dagTopologyManager.IsAncestorOf(stagingArea, checkedBlock, chain)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if isAncestorOf {
|
|
dataStore, err := gh.BlockData(stagingArea, chain)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if mergeSetReds := dataStore.MergeSetReds(); contains(checkedBlock, mergeSetReds) {
|
|
return false, nil
|
|
}
|
|
} else {
|
|
return true, nil
|
|
}
|
|
|
|
return false, nil
|
|
}
|
|
|
|
/*----------------contains-------------------------- */
|
|
func contains(item *externalapi.DomainHash, items []*externalapi.DomainHash) bool {
|
|
for _, r := range items {
|
|
if r.Equal(item) {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
/* ----------------checkIfDestroy------------------- */
|
|
/* find number of not-connected in his blue*/
|
|
func (gh *ghostdagHelper) checkIfDestroy(stagingArea *model.StagingArea, blockBlue *externalapi.DomainHash,
|
|
blueSet *[]*externalapi.DomainHash) (bool, error) {
|
|
|
|
// Goal: check that the K-cluster of each block in the blueSet is not destroyed when adding the block to the mergeSet.
|
|
var k = int(gh.k)
|
|
counter := 0
|
|
for _, blue := range *blueSet {
|
|
isAnticone, err := gh.isAnticone(stagingArea, blue, blockBlue)
|
|
if err != nil {
|
|
return true, err
|
|
}
|
|
if isAnticone {
|
|
counter++
|
|
}
|
|
if counter > k {
|
|
return true, nil
|
|
}
|
|
}
|
|
return false, nil
|
|
}
|
|
|
|
/* ----------------findMergeSet------------------- */
|
|
func (gh *ghostdagHelper) findMergeSet(stagingArea *model.StagingArea, parents []*externalapi.DomainHash,
|
|
selectedParent *externalapi.DomainHash) ([]*externalapi.DomainHash, error) {
|
|
|
|
allMergeSet := make([]*externalapi.DomainHash, 0)
|
|
blockQueue := make([]*externalapi.DomainHash, 0)
|
|
for _, parent := range parents {
|
|
if !contains(parent, blockQueue) {
|
|
blockQueue = append(blockQueue, parent)
|
|
}
|
|
|
|
}
|
|
for len(blockQueue) > 0 {
|
|
block := blockQueue[0]
|
|
blockQueue = blockQueue[1:]
|
|
if selectedParent.Equal(block) {
|
|
if !contains(block, allMergeSet) {
|
|
allMergeSet = append(allMergeSet, block)
|
|
}
|
|
continue
|
|
}
|
|
isancestorOf, err := gh.dagTopologyManager.IsAncestorOf(stagingArea, block, selectedParent)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if isancestorOf {
|
|
continue
|
|
}
|
|
if !contains(block, allMergeSet) {
|
|
allMergeSet = append(allMergeSet, block)
|
|
}
|
|
err = gh.insertParent(stagingArea, block, &blockQueue)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
}
|
|
return allMergeSet, nil
|
|
}
|
|
|
|
/* ----------------insertParent------------------- */
|
|
/* Insert all parents to the queue*/
|
|
func (gh *ghostdagHelper) insertParent(stagingArea *model.StagingArea, child *externalapi.DomainHash,
|
|
queue *[]*externalapi.DomainHash) error {
|
|
|
|
parents, err := gh.dagTopologyManager.Parents(stagingArea, child)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
for _, parent := range parents {
|
|
if contains(parent, *queue) {
|
|
continue
|
|
}
|
|
*queue = append(*queue, parent)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
/* ----------------findBlueSet------------------- */
|
|
func (gh *ghostdagHelper) findBlueSet(stagingArea *model.StagingArea, blueSet *[]*externalapi.DomainHash, selectedParent *externalapi.DomainHash) error {
|
|
for selectedParent != nil {
|
|
if !contains(selectedParent, *blueSet) {
|
|
*blueSet = append(*blueSet, selectedParent)
|
|
}
|
|
blockData, err := gh.dataStore.Get(gh.dbAccess, stagingArea, selectedParent)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
mergeSetBlue := blockData.MergeSetBlues()
|
|
for _, blue := range mergeSetBlue {
|
|
if contains(blue, *blueSet) {
|
|
continue
|
|
}
|
|
*blueSet = append(*blueSet, blue)
|
|
}
|
|
selectedParent = blockData.SelectedParent()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
/* ----------------sortByBlueScore------------------- */
|
|
func (gh *ghostdagHelper) sortByBlueWork(stagingArea *model.StagingArea, arr []*externalapi.DomainHash) error {
|
|
|
|
var err error = nil
|
|
sort.Slice(arr, func(i, j int) bool {
|
|
|
|
blockLeft, error := gh.dataStore.Get(gh.dbAccess, stagingArea, arr[i])
|
|
if error != nil {
|
|
err = error
|
|
return false
|
|
}
|
|
|
|
blockRight, error := gh.dataStore.Get(gh.dbAccess, stagingArea, arr[j])
|
|
if error != nil {
|
|
err = error
|
|
return false
|
|
}
|
|
|
|
if blockLeft.BlueWork().Cmp(blockRight.BlueWork()) == -1 {
|
|
return true
|
|
}
|
|
if blockLeft.BlueWork().Cmp(blockRight.BlueWork()) == 0 {
|
|
return ismoreHash(arr[j], arr[i])
|
|
}
|
|
return false
|
|
})
|
|
return err
|
|
}
|
|
|
|
/* --------------------------------------------- */
|
|
|
|
func (gh *ghostdagHelper) BlockData(stagingArea *model.StagingArea, blockHash *externalapi.DomainHash) (*model.BlockGHOSTDAGData, error) {
|
|
return gh.dataStore.Get(gh.dbAccess, stagingArea, blockHash)
|
|
}
|
|
func (gh *ghostdagHelper) ChooseSelectedParent(stagingArea *model.StagingArea, blockHashes ...*externalapi.DomainHash) (*externalapi.DomainHash, error) {
|
|
panic("implement me")
|
|
}
|
|
|
|
func (gh *ghostdagHelper) Less(blockHashA *externalapi.DomainHash, ghostdagDataA *model.BlockGHOSTDAGData, blockHashB *externalapi.DomainHash, ghostdagDataB *model.BlockGHOSTDAGData) bool {
|
|
panic("implement me")
|
|
}
|