Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
28 commits
Select commit Hold shift + click to select a range
e5083b2
core/vm: fix RequiredGas for P256Verify
allformless Dec 31, 2025
54fef3b
cmd/geth: not support OverrideBPO1 and OverrideBPO2 flags in BSC
allformless Jan 5, 2026
214f495
cmd/geth: hidden flag PruneAncientDataFlag
allformless Jan 5, 2026
a5c8d1d
cmd/utils: deprecated MetricsEnabledExpensiveFlag
allformless Jan 5, 2026
bd120a5
core/state: simplify statedb copy
allformless Sep 16, 2025
012d7bb
all: upgrade go version to 1.25 for workflows and dockers
allformless Jan 5, 2026
6a55e56
all: handle golang lint
allformless Jan 5, 2026
f16dce4
bid_simulator.go: disable blob version1 in bid simulator
zlacfzy Jan 5, 2026
9a59b84
triedb: align signature of func Size()
allformless Jan 6, 2026
8a31794
triedb/pathdb: read old journal file
allformless Jan 6, 2026
22e998f
cmd/utils: deprecated JournalFileFlag
allformless Jan 6, 2026
a1f6e34
triedb/pathdb: remove JournalFile field in config
allformless Jan 6, 2026
4a1f70a
eth: fix BroadcastTransactions
allformless Jan 6, 2026
ef741d2
core/txpool/blobpool: only accept BlobSidecarVersion0 in BSC
allformless Jan 6, 2026
b65edca
core/txpool/blobpool: not convert blobsidecar to V1 in BSC
allformless Jan 6, 2026
b5d8de5
fix: fix eip-7910 test case
flywukong Jan 6, 2026
6d79b77
fix: fix api
flywukong Jan 6, 2026
5d1a60a
legacypool_test.go: fix TestDualHeapEviction
zlacfzy Jan 6, 2026
4c01a16
eth/ethconfig: delete field JournalFileEnabled
allformless Jan 7, 2026
5e13c69
core/vm: revert comments
allformless Jan 7, 2026
d1aba50
params: fix Rules for ChainConfig
allformless Jan 7, 2026
0e5f7aa
params: fix Rules for ChainConfig
allformless Jan 7, 2026
2fbbb4f
fix: fix transaction rollback
flywukong Jan 7, 2026
21eee8c
internal/ethapi: fix eth_config
allformless Jan 7, 2026
970aaf9
core: fix core ut
zlacfzy Jan 7, 2026
6f976ff
core: fix TestWaitDeployedCornerCases
zlacfzy Jan 7, 2026
2d5ecd9
fix suit_test: only set HistoryStorageAddress nonce on BSC chains dur…
zlacfzy Jan 8, 2026
1c5f126
eth/syncer: disable debug_sync for now
allformless Jan 8, 2026
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .github/workflows/build-test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ jobs:
unit-test:
strategy:
matrix:
go-version: [1.24.x]
go-version: [1.25.x]
os: [ubuntu-latest]
runs-on: ${{ matrix.os }}
steps:
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/evm-tests.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ jobs:
evm-test:
strategy:
matrix:
go-version: [1.24.x]
go-version: [1.25.x]
os: [ubuntu-latest]
runs-on: ${{ matrix.os }}
steps:
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/lint.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ jobs:
golang-lint:
strategy:
matrix:
go-version: [1.24.x]
go-version: [1.25.x]
os: [ubuntu-latest]
runs-on: ${{ matrix.os }}
steps:
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/nancy.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ jobs:
if: github.event.pull_request.head.repo.full_name == github.repository
strategy:
matrix:
go-version: [1.24.x]
go-version: [1.25.x]
os: [ubuntu-latest]
runs-on: ${{ matrix.os }}
steps:
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/pre-release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ jobs:
name: Build Release
strategy:
matrix:
go-version: [1.24.x]
go-version: [1.25.x]
os: [ubuntu-latest, macos-latest, windows-latest]
runs-on: ${{ matrix.os }}
steps:
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ jobs:
name: Build Release
strategy:
matrix:
go-version: [1.24.x]
go-version: [1.25.x]
os: [ubuntu-latest, macos-latest, windows-latest]
runs-on: ${{ matrix.os }}
steps:
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/unit-test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ jobs:
unit-test:
strategy:
matrix:
go-version: [1.24.x]
go-version: [1.25.x]
os: [ubuntu-latest]
runs-on: ${{ matrix.os }}
steps:
Expand Down
2 changes: 1 addition & 1 deletion Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ ARG VERSION=""
ARG BUILDNUM=""

# Build Geth in a stock Go builder container
FROM golang:1.24-alpine AS builder
FROM golang:1.25-alpine AS builder

RUN apk add --no-cache make cmake gcc musl-dev linux-headers git bash build-base libc-dev
# Get dependencies - will also be cached if we won't change go.mod/go.sum
Expand Down
2 changes: 1 addition & 1 deletion Dockerfile.alltools
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ ARG VERSION=""
ARG BUILDNUM=""

# Build Geth in a stock Go builder container
FROM golang:1.24-alpine AS builder
FROM golang:1.25-alpine AS builder

RUN apk add --no-cache gcc musl-dev linux-headers git
# Get dependencies - will also be cached if we won't change go.mod/go.sum
Expand Down
2 changes: 1 addition & 1 deletion accounts/abi/bind/v2/util_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -107,7 +107,7 @@ func TestWaitDeployedCornerCases(t *testing.T) {
},
)
head, _ = backend.Client().HeaderByNumber(t.Context(), nil) // Should be child's, good enough
gasPrice = new(big.Int).Add(head.BaseFee, big.NewInt(1))
gasPrice = new(big.Int).Add(head.BaseFee, big.NewInt(params.GWei))
signer = types.LatestSigner(params.AllDevChainProtocolChanges)
code = common.FromHex("6060604052600a8060106000396000f360606040526008565b00")
ctx, cancel = context.WithCancel(t.Context())
Expand Down
2 changes: 1 addition & 1 deletion cmd/evm/internal/t8ntool/transition.go
Original file line number Diff line number Diff line change
Expand Up @@ -212,7 +212,7 @@ func applyShanghaiChecks(env *stEnv, chainConfig *params.ChainConfig) error {
if !chainConfig.IsShanghai(big.NewInt(int64(env.Number)), env.Timestamp) {
return nil
}
if chainConfig.Parlia == nil && env.Withdrawals == nil {
if chainConfig.IsNotInBSC() && env.Withdrawals == nil {
return NewError(ErrorConfig, errors.New("Shanghai config but missing 'withdrawals' in env section"))
}
return nil
Expand Down
1 change: 0 additions & 1 deletion cmd/geth/config.go
Original file line number Diff line number Diff line change
Expand Up @@ -406,7 +406,6 @@ func applyMetricConfig(ctx *cli.Context, cfg *gethConfig) {
}
if ctx.IsSet(utils.MetricsEnabledExpensiveFlag.Name) {
log.Warn("Expensive metrics will remain in BSC and may be removed in the future", "flag", utils.MetricsEnabledExpensiveFlag.Name)
cfg.Metrics.EnabledExpensive = ctx.Bool(utils.MetricsEnabledExpensiveFlag.Name)
}
if ctx.IsSet(utils.MetricsHTTPFlag.Name) {
cfg.Metrics.HTTP = ctx.String(utils.MetricsHTTPFlag.Name)
Expand Down
4 changes: 2 additions & 2 deletions cmd/geth/main.go
Original file line number Diff line number Diff line change
Expand Up @@ -78,8 +78,8 @@ var (
utils.OverrideFermi,
utils.OverrideOsaka,
utils.OverrideMendel,
utils.OverrideBPO1,
utils.OverrideBPO2,
// utils.OverrideBPO1,
// utils.OverrideBPO2,
utils.OverrideVerkle,
utils.OverrideGenesisFlag,
utils.OverrideFullImmutabilityThreshold,
Expand Down
13 changes: 0 additions & 13 deletions cmd/utils/flags.go
Original file line number Diff line number Diff line change
Expand Up @@ -397,12 +397,6 @@ var (
Value: false,
Category: flags.StateCategory,
}
JournalFileFlag = &cli.BoolFlag{
Name: "journalfile",
Usage: "Enable using journal file to store the TrieJournal instead of KVDB in pbss (default = true)",
Value: true,
Category: flags.StateCategory,
}
StateSizeTrackingFlag = &cli.BoolFlag{
Name: "state.size-tracking",
Usage: "Enable state size tracking, retrieve state size with debug_stateSize.",
Expand Down Expand Up @@ -2211,8 +2205,6 @@ func SetEthConfig(ctx *cli.Context, stack *node.Node, cfg *ethconfig.Config) {
cfg.PathSyncFlush = true
}

cfg.JournalFileEnabled = ctx.Bool(JournalFileFlag.Name)

if ctx.String(GCModeFlag.Name) == "archive" {
if cfg.TransactionHistory != 0 {
cfg.TransactionHistory = 0
Expand Down Expand Up @@ -2696,10 +2688,6 @@ func SetupMetrics(cfg *metrics.Config, options ...SetupMetricsOption) {
}
log.Info("Enabling metrics collection")
metrics.Enable()
if cfg.EnabledExpensive {
log.Info("Enabling expensive metrics collection")
metrics.EnableExpensive()
}

// InfluxDB exporter.
var (
Expand Down Expand Up @@ -3021,7 +3009,6 @@ func MakeTrieDatabase(ctx *cli.Context, stack *node.Node, disk ethdb.Database, p
}
pathConfig.JournalDirectory = stack.ResolvePath("triedb")
config.PathDB = &pathConfig
// TODO(Nathan): delete JournalFilePath ?
config.PathDB.JournalFilePath = fmt.Sprintf("%s/%s", stack.ResolvePath("chaindata"), eth.JournalFileName)
return triedb.NewDatabase(disk, config)
}
Expand Down
14 changes: 12 additions & 2 deletions cmd/utils/flags_legacy.go
Original file line number Diff line number Diff line change
Expand Up @@ -42,8 +42,10 @@ var DeprecatedFlags = []cli.Flag{
LogBacktraceAtFlag,
LogDebugFlag,
MinerNewPayloadTimeoutFlag,
PruneAncientDataFlag,
MetricsEnabledExpensiveFlag,
EnablePersonal,
PruneAncientDataFlag,
JournalFileFlag,
}

var (
Expand Down Expand Up @@ -126,8 +128,16 @@ var (
// Deprecated Dec 2024
PruneAncientDataFlag = &cli.BoolFlag{
Name: "pruneancient",
Hidden: true,
Usage: "Prune ancient data, is an optional config and disabled by default. Only keep the latest 9w blocks' data,the older blocks' data will be permanently pruned. Notice:the geth/chaindata/ancient dir will be removed, if restart without the flag, the ancient data will start with the previous point that the oldest unpruned block number. Recommends to the user who don't care about the ancient data.",
Category: flags.BlockHistoryCategory,
Category: flags.DeprecatedCategory,
}
JournalFileFlag = &cli.BoolFlag{
Name: "journalfile",
Hidden: true,
Usage: "Enable using journal file to store the TrieJournal instead of KVDB in pbss (default = true, deprecated)",
Value: true,
Category: flags.DeprecatedCategory,
}
)

Expand Down
3 changes: 1 addition & 2 deletions consensus/beacon/consensus.go
Original file line number Diff line number Diff line change
Expand Up @@ -61,8 +61,7 @@ var (
// is only used for necessary consensus checks. The legacy consensus engine can be any
// engine implements the consensus interface (except the beacon itself).
type Beacon struct {
ethone consensus.Engine // Original consensus engine used in eth1, e.g. ethash or clique
ttdblock *uint64 // Merge block-number for testchain generation without TTDs
ethone consensus.Engine // Original consensus engine used in eth1, e.g. ethash or clique
}

// New creates a consensus engine with the given embedded eth1 engine.
Expand Down
4 changes: 2 additions & 2 deletions consensus/misc/eip1559/eip1559.go
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ import (
// - gas limit check
// - basefee check
func VerifyEIP1559Header(config *params.ChainConfig, parent, header *types.Header) error {
if config.Parlia == nil {
if config.IsNotInBSC() {
// Verify that the gas limit remains within allowed bounds
parentGasLimit := parent.GasLimit
if !config.IsLondon(parent.Number) {
Expand All @@ -57,7 +57,7 @@ func VerifyEIP1559Header(config *params.ChainConfig, parent, header *types.Heade

// CalcBaseFee calculates the basefee of the header.
func CalcBaseFee(config *params.ChainConfig, parent *types.Header) *big.Int {
if config.Parlia != nil {
if config.IsInBSC() {
return new(big.Int).SetUint64(params.InitialBaseFeeForBSC)
}

Expand Down
48 changes: 21 additions & 27 deletions core/blockchain.go
Original file line number Diff line number Diff line change
Expand Up @@ -181,7 +181,6 @@ type BlockChainConfig struct {
NoTries bool // Insecure settings. Do not have any tries in databases if enabled.
PathSyncFlush bool // Whether sync flush the trienodebuffer of pathdb to disk.
JournalFilePath string // The path to store journal file which is used in pathdb
JournalFile bool // Whether to use single file to store journal data in pathdb
EnableIncr bool // Flag whether the freezer db stores incremental block and state history
IncrHistoryPath string // The path to store incremental block and chain files
IncrHistory uint64 // Amount of block and state history stored in incremental freezer db
Expand Down Expand Up @@ -284,7 +283,6 @@ func (cfg *BlockChainConfig) triedbConfig(isVerkle bool) *triedb.Config {
if cfg.StateScheme == rawdb.PathScheme {
config.PathDB = &pathdb.Config{
JournalFilePath: cfg.JournalFilePath,
JournalFile: cfg.JournalFile,
EnableIncr: cfg.EnableIncr,
IncrHistoryPath: cfg.IncrHistoryPath,
IncrHistory: cfg.IncrHistory,
Expand Down Expand Up @@ -1573,7 +1571,7 @@ func (bc *BlockChain) Stop() {
for !bc.triegc.Empty() {
triedb.Dereference(bc.triegc.PopItem())
}
if _, size, _, _ := triedb.Size(); size != 0 {
if _, size, _ := triedb.Size(); size != 0 {
log.Error("Dangling trie nodes after full cleanup")
}
}
Expand Down Expand Up @@ -1660,7 +1658,7 @@ func (bc *BlockChain) InsertReceiptChain(blockChain types.Blocks, receiptChain [
}
// check DA after cancun
lastBlk := blockChain[len(blockChain)-1]
if bc.chainConfig.Parlia != nil && bc.chainConfig.IsCancun(lastBlk.Number(), lastBlk.Time()) {
if bc.chainConfig.IsInBSC() && bc.chainConfig.IsCancun(lastBlk.Number(), lastBlk.Time()) {
if _, err := CheckDataAvailableInBatch(bc, blockChain); err != nil {
log.Debug("CheckDataAvailableInBatch", "err", err)
return 0, err
Expand Down Expand Up @@ -1968,8 +1966,8 @@ func (bc *BlockChain) writeBlockWithState(block *types.Block, receipts []*types.
}
// If we exceeded our memory allowance, flush matured singleton nodes to disk
var (
_, nodes, _, imgs = bc.triedb.Size()
limit = common.StorageSize(bc.cfg.TrieDirtyLimit) * 1024 * 1024
_, nodes, imgs = bc.triedb.Size()
limit = common.StorageSize(bc.cfg.TrieDirtyLimit) * 1024 * 1024
)
if nodes > limit || imgs > 4*1024*1024 {
bc.triedb.Cap(limit - ethdb.IdealBatchSize)
Expand Down Expand Up @@ -2206,7 +2204,7 @@ func (bc *BlockChain) insertChain(chain types.Blocks, setHead bool, makeWitness
}()

// check block data available first
if bc.chainConfig.Parlia != nil {
if bc.chainConfig.IsInBSC() {
if index, err := CheckDataAvailableInBatch(bc, chain); err != nil {
return nil, index, err
}
Expand Down Expand Up @@ -2389,8 +2387,8 @@ func (bc *BlockChain) insertChain(chain types.Blocks, setHead bool, makeWitness
if bc.snaps != nil {
snapDiffItems, snapBufItems, _ = bc.snaps.Size()
}
trieDiffNodes, trieBufNodes, trieImmutableBufNodes, _ := bc.triedb.Size()
stats.report(chain, it.index, snapDiffItems, snapBufItems, trieDiffNodes, trieBufNodes, trieImmutableBufNodes, res.status == CanonStatTy)
trieDiffNodes, trieBufNodes, _ := bc.triedb.Size()
stats.report(chain, it.index, snapDiffItems, snapBufItems, trieDiffNodes, trieBufNodes, res.status == CanonStatTy)

// Print confirmation that a future fork is scheduled, but not yet active.
bc.logForkReadiness(block)
Expand Down Expand Up @@ -2637,19 +2635,17 @@ func (bc *BlockChain) ProcessBlock(parentRoot common.Hash, block *types.Block, s
proctime := time.Since(startTime) // processing + validation + cross validation

// Update the metrics touched during block processing and validation
if metrics.EnabledExpensive() {
accountReadTimer.Update(statedb.AccountReads) // Account reads are complete(in processing)
storageReadTimer.Update(statedb.StorageReads) // Storage reads are complete(in processing)
if statedb.AccountLoaded != 0 {
accountReadSingleTimer.Update(statedb.AccountReads / time.Duration(statedb.AccountLoaded))
}
if statedb.StorageLoaded != 0 {
storageReadSingleTimer.Update(statedb.StorageReads / time.Duration(statedb.StorageLoaded))
}
accountUpdateTimer.Update(statedb.AccountUpdates) // Account updates are complete(in validation)
storageUpdateTimer.Update(statedb.StorageUpdates) // Storage updates are complete(in validation)
accountHashTimer.Update(statedb.AccountHashes) // Account hashes are complete(in validation)
accountReadTimer.Update(statedb.AccountReads) // Account reads are complete(in processing)
storageReadTimer.Update(statedb.StorageReads) // Storage reads are complete(in processing)
if statedb.AccountLoaded != 0 {
accountReadSingleTimer.Update(statedb.AccountReads / time.Duration(statedb.AccountLoaded))
}
if statedb.StorageLoaded != 0 {
storageReadSingleTimer.Update(statedb.StorageReads / time.Duration(statedb.StorageLoaded))
}
accountUpdateTimer.Update(statedb.AccountUpdates) // Account updates are complete(in validation)
storageUpdateTimer.Update(statedb.StorageUpdates) // Storage updates are complete(in validation)
accountHashTimer.Update(statedb.AccountHashes) // Account hashes are complete(in validation)
triehash := statedb.AccountHashes // The time spent on tries hashing
trieUpdate := statedb.AccountUpdates + statedb.StorageUpdates // The time spent on tries update
blockExecutionTimer.Update(ptime - (statedb.AccountReads + statedb.StorageReads)) // The time spent on EVM processing
Expand All @@ -2676,12 +2672,10 @@ func (bc *BlockChain) ProcessBlock(parentRoot common.Hash, block *types.Block, s
}

// Update the metrics touched during block commit
if metrics.EnabledExpensive() {
accountCommitTimer.Update(statedb.AccountCommits) // Account commits are complete, we can mark them
storageCommitTimer.Update(statedb.StorageCommits) // Storage commits are complete, we can mark them
snapshotCommitTimer.Update(statedb.SnapshotCommits) // Snapshot commits are complete, we can mark them
triedbCommitTimer.Update(statedb.TrieDBCommits) // Trie database commits are complete, we can mark them
}
accountCommitTimer.Update(statedb.AccountCommits) // Account commits are complete, we can mark them
storageCommitTimer.Update(statedb.StorageCommits) // Storage commits are complete, we can mark them
snapshotCommitTimer.Update(statedb.SnapshotCommits) // Snapshot commits are complete, we can mark them
triedbCommitTimer.Update(statedb.TrieDBCommits) // Trie database commits are complete, we can mark them
blockWriteTimer.Update(time.Since(wstart) - max(statedb.AccountCommits, statedb.StorageCommits) /* concurrent */ - statedb.SnapshotCommits - statedb.TrieDBCommits)
elapsed := time.Since(startTime) + 1 // prevent zero division
blockInsertTimer.Update(elapsed)
Expand Down
3 changes: 1 addition & 2 deletions core/blockchain_insert.go
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ const statsReportLimit = 8 * time.Second

// report prints statistics if some number of blocks have been processed
// or more than a few seconds have passed since the last message.
func (st *insertStats) report(chain []*types.Block, index int, snapDiffItems, snapBufItems, trieDiffNodes, trieBufNodes, trieImmutableBufNodes common.StorageSize, setHead bool) {
func (st *insertStats) report(chain []*types.Block, index int, snapDiffItems, snapBufItems, trieDiffNodes, trieBufNodes common.StorageSize, setHead bool) {
// Fetch the timings for the batch
var (
now = mclock.Now()
Expand Down Expand Up @@ -78,7 +78,6 @@ func (st *insertStats) report(chain []*types.Block, index int, snapDiffItems, sn
context = append(context, []interface{}{"triediffs", trieDiffNodes}...)
}
context = append(context, []interface{}{"triedirty", trieBufNodes}...)
context = append(context, []interface{}{"trieimutabledirty", trieImmutableBufNodes}...)

if st.ignored > 0 {
context = append(context, []interface{}{"ignored", st.ignored}...)
Expand Down
2 changes: 1 addition & 1 deletion core/blockchain_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -1678,7 +1678,7 @@ func TestTrieForkGC(t *testing.T) {
chain.TrieDB().Dereference(blocks[len(blocks)-1-i].Root())
chain.TrieDB().Dereference(forks[len(blocks)-1-i].Root())
}
if _, nodes, _, _ := chain.TrieDB().Size(); nodes > 0 { // all memory is returned in the nodes return for hashdb
if _, nodes, _ := chain.TrieDB().Size(); nodes > 0 { // all memory is returned in the nodes return for hashdb
t.Fatalf("stale tries still alive after garbase collection")
}
}
Expand Down
8 changes: 4 additions & 4 deletions core/chain_makers.go
Original file line number Diff line number Diff line change
Expand Up @@ -262,7 +262,7 @@ func (b *BlockGen) AddUncle(h *types.Header) {
h.GasLimit = parent.GasLimit
if b.cm.config.IsLondon(h.Number) {
h.BaseFee = eip1559.CalcBaseFee(b.cm.config, parent)
if b.cm.config.Parlia == nil && !b.cm.config.IsLondon(parent.Number) {
if b.cm.config.IsNotInBSC() && !b.cm.config.IsLondon(parent.Number) {
parentGasLimit := parent.GasLimit * b.cm.config.ElasticityMultiplier()
h.GasLimit = CalcGasLimit(parentGasLimit, parentGasLimit)
}
Expand Down Expand Up @@ -336,7 +336,7 @@ func (b *BlockGen) collectRequests(readonly bool) (requests [][]byte) {
statedb = statedb.Copy()
}

if b.cm.config.IsPrague(b.header.Number, b.header.Time) && b.cm.config.Parlia == nil {
if b.cm.config.IsPrague(b.header.Number, b.header.Time) && b.cm.config.IsNotInBSC() {
requests = [][]byte{}
// EIP-6110 deposits
var blockLogs []*types.Log
Expand Down Expand Up @@ -641,7 +641,7 @@ func (cm *chainMaker) makeHeader(parent *types.Block, state *state.StateDB, engi

if cm.config.IsLondon(header.Number) {
header.BaseFee = eip1559.CalcBaseFee(cm.config, parentHeader)
if cm.config.Parlia == nil && !cm.config.IsLondon(parent.Number()) {
if cm.config.IsNotInBSC() && !cm.config.IsLondon(parent.Number()) {
parentGasLimit := parent.GasLimit() * cm.config.ElasticityMultiplier()
header.GasLimit = CalcGasLimit(parentGasLimit, parentGasLimit)
}
Expand All @@ -650,7 +650,7 @@ func (cm *chainMaker) makeHeader(parent *types.Block, state *state.StateDB, engi
excessBlobGas := eip4844.CalcExcessBlobGas(cm.config, parentHeader, time)
header.ExcessBlobGas = &excessBlobGas
header.BlobGasUsed = new(uint64)
if cm.config.Parlia == nil {
if cm.config.IsNotInBSC() {
header.ParentBeaconRoot = new(common.Hash)
} else {
header.WithdrawalsHash = &types.EmptyWithdrawalsHash
Expand Down
Loading
Loading