Skip to content

Commit

Permalink
rename encode to compress
Browse files Browse the repository at this point in the history
  • Loading branch information
colinlyguo committed Aug 19, 2024
1 parent 4a93f2f commit 50895d7
Show file tree
Hide file tree
Showing 14 changed files with 48 additions and 46 deletions.
4 changes: 2 additions & 2 deletions common/forks/forks.go
Original file line number Diff line number Diff line change
Expand Up @@ -23,15 +23,15 @@ func GetHardforkName(config *params.ChainConfig, blockHeight, blockTimestamp uin

// GetCodecVersion returns the encoding codec version for the given block height and timestamp.
// It determines the appropriate codec version based on the active hardfork.
func GetCodecVersion(config *params.ChainConfig, blockHeight, blockTimestamp uint64, enableConditionalEncoding bool) encoding.CodecVersion {
func GetCodecVersion(config *params.ChainConfig, blockHeight, blockTimestamp uint64, enableConditionalCompress bool) encoding.CodecVersion {
if !config.IsBernoulli(new(big.Int).SetUint64(blockHeight)) {
return encoding.CodecV0
} else if !config.IsCurie(new(big.Int).SetUint64(blockHeight)) {
return encoding.CodecV1
} else if !config.IsDarwin(blockTimestamp) {
return encoding.CodecV2
} else {
if !enableConditionalEncoding {
if !enableConditionalCompress {
return encoding.CodecV3
} else {
return encoding.CodecV4
Expand Down
2 changes: 1 addition & 1 deletion common/version/version.go
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ import (
"runtime/debug"
)

var tag = "v4.4.45"
var tag = "v4.4.46"

var commit = func() string {
if info, ok := debug.ReadBuildInfo(); ok {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,21 +3,21 @@

ALTER TABLE chunk
ADD COLUMN codec_version SMALLINT NOT NULL DEFAULT -1,
ADD COLUMN enable_encode BOOLEAN NOT NULL DEFAULT false;
ADD COLUMN enable_compress BOOLEAN NOT NULL DEFAULT false;

ALTER TABLE batch
ADD COLUMN enable_encode BOOLEAN NOT NULL DEFAULT false;
ADD COLUMN enable_compress BOOLEAN NOT NULL DEFAULT false;

-- +goose StatementEnd

-- +goose Down
-- +goose StatementBegin

ALTER TABLE IF EXISTS chunk
DROP COLUMN IF EXISTS enable_encode,
DROP COLUMN IF EXISTS enable_compress,
DROP COLUMN IF EXISTS codec_version;

ALTER TABLE IF EXISTS batch
DROP COLUMN IF EXISTS enable_encode;
DROP COLUMN IF EXISTS enable_compress;

-- +goose StatementEnd
2 changes: 2 additions & 0 deletions go.work.sum
Original file line number Diff line number Diff line change
Expand Up @@ -740,6 +740,8 @@ github.com/russross/blackfriday v1.6.0 h1:KqfZb0pUVN2lYqZUYRddxF4OR8ZMURnJIG5Y3V
github.com/russross/blackfriday v1.6.0/go.mod h1:ti0ldHuxg49ri4ksnFxlkCfN+hvslNlmVHqNRXXJNAY=
github.com/russross/blackfriday/v2 v2.0.1/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM=
github.com/scroll-tech/da-codec v0.0.0-20240605080813-32bfc9fccde7/go.mod h1:1wWYii0OPwd5kw+xrz0PFgS420xNadrNF1x/ELJT+TM=
github.com/scroll-tech/da-codec v0.0.0-20240819100936-c6af3bbe7068 h1:oVGwhg4cCq35B04eG/S4OBXDwXiFH7+LezuH2ZTRBPs=
github.com/scroll-tech/da-codec v0.0.0-20240819100936-c6af3bbe7068/go.mod h1:D6XEESeNVJkQJlv3eK+FyR+ufPkgVQbJzERylQi53Bs=
github.com/scroll-tech/go-ethereum v1.10.14-0.20240607130425-e2becce6a1a4/go.mod h1:byf/mZ8jLYUCnUePTicjJWn+RvKdxDn7buS6glTnMwQ=
github.com/segmentio/kafka-go v0.1.0/go.mod h1:X6itGqS9L4jDletMsxZ7Dz+JFWxM6JHfPOCvTvk+EJo=
github.com/segmentio/kafka-go v0.2.0/go.mod h1:X6itGqS9L4jDletMsxZ7Dz+JFWxM6JHfPOCvTvk+EJo=
Expand Down
2 changes: 1 addition & 1 deletion rollup/conf/config.json
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@
"l1_commit_gas_limit_multiplier": 1.2
},
"chunk_proposer_config": {
"enable_conditional_encoding": false,
"enable_conditional_compress": false,
"propose_interval_milliseconds": 100,
"max_block_num_per_chunk": 100,
"max_tx_num_per_chunk": 100,
Expand Down
2 changes: 1 addition & 1 deletion rollup/go.mod
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ require (
github.com/go-resty/resty/v2 v2.7.0
github.com/holiman/uint256 v1.2.4
github.com/prometheus/client_golang v1.16.0
github.com/scroll-tech/da-codec v0.0.0-20240818122843-6901956c9910
github.com/scroll-tech/da-codec v0.0.0-20240819100936-c6af3bbe7068
github.com/scroll-tech/go-ethereum v1.10.14-0.20240626125436-418bc6f728b6
github.com/smartystreets/goconvey v1.8.0
github.com/stretchr/testify v1.9.0
Expand Down
4 changes: 2 additions & 2 deletions rollup/go.sum
Original file line number Diff line number Diff line change
Expand Up @@ -236,8 +236,8 @@ github.com/rs/cors v1.7.0 h1:+88SsELBHx5r+hZ8TCkggzSstaWNbDvThkVK8H6f9ik=
github.com/rs/cors v1.7.0/go.mod h1:gFx+x8UowdsKA9AchylcLynDq+nNFfI8FkUZdN/jGCU=
github.com/russross/blackfriday/v2 v2.1.0 h1:JIOH55/0cWyOuilr9/qlrm0BSXldqnqwMsf35Ld67mk=
github.com/russross/blackfriday/v2 v2.1.0/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM=
github.com/scroll-tech/da-codec v0.0.0-20240818122843-6901956c9910 h1:c2siFeESs8MgppUAoatnkbnQMoFTq2DPT0TEdZFPbok=
github.com/scroll-tech/da-codec v0.0.0-20240818122843-6901956c9910/go.mod h1:D6XEESeNVJkQJlv3eK+FyR+ufPkgVQbJzERylQi53Bs=
github.com/scroll-tech/da-codec v0.0.0-20240819100936-c6af3bbe7068 h1:oVGwhg4cCq35B04eG/S4OBXDwXiFH7+LezuH2ZTRBPs=
github.com/scroll-tech/da-codec v0.0.0-20240819100936-c6af3bbe7068/go.mod h1:D6XEESeNVJkQJlv3eK+FyR+ufPkgVQbJzERylQi53Bs=
github.com/scroll-tech/go-ethereum v1.10.14-0.20240626125436-418bc6f728b6 h1:Q8YyvrcPIcXQwE4ucm4bqmPh6TP6IB1GUTXripf2WyQ=
github.com/scroll-tech/go-ethereum v1.10.14-0.20240626125436-418bc6f728b6/go.mod h1:byf/mZ8jLYUCnUePTicjJWn+RvKdxDn7buS6glTnMwQ=
github.com/scroll-tech/zktrie v0.8.4 h1:UagmnZ4Z3ITCk+aUq9NQZJNAwnWl4gSxsLb2Nl7IgRE=
Expand Down
2 changes: 1 addition & 1 deletion rollup/internal/config/l2.go
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ type ChunkProposerConfig struct {
MaxRowConsumptionPerChunk uint64 `json:"max_row_consumption_per_chunk"`
GasCostIncreaseMultiplier float64 `json:"gas_cost_increase_multiplier"`
MaxUncompressedBatchBytesSize uint64 `json:"max_uncompressed_batch_bytes_size"`
EnableConditionalEncoding bool `json:"enable_conditional_encoding"`
EnableConditionalCompress bool `json:"enable_conditional_compress"`
}

// BatchProposerConfig loads batch_proposer configuration items.
Expand Down
2 changes: 1 addition & 1 deletion rollup/internal/controller/relayer/l2_relayer.go
Original file line number Diff line number Diff line change
Expand Up @@ -1068,7 +1068,7 @@ func (r *Layer2Relayer) constructCommitBatchPayloadCodecV4(dbBatch *orm.Batch, d
Chunks: chunks,
}

daBatch, createErr := codecv4.NewDABatch(batch, dbBatch.EnableEncode)
daBatch, createErr := codecv4.NewDABatch(batch, dbBatch.EnableCompress)
if createErr != nil {
return nil, nil, fmt.Errorf("failed to create DA batch: %w", createErr)
}
Expand Down
16 changes: 8 additions & 8 deletions rollup/internal/controller/watcher/batch_proposer.go
Original file line number Diff line number Diff line change
Expand Up @@ -151,7 +151,7 @@ func (p *BatchProposer) TryProposeBatch() {

func (p *BatchProposer) updateDBBatchInfo(batch *encoding.Batch, codecVersion encoding.CodecVersion, metrics *utils.BatchMetrics) error {
compatibilityBreachOccurred := false
enableEncode := true
enableCompress := true

for {
compatible, err := utils.CheckBatchCompressedDataCompatibility(batch, codecVersion)
Expand All @@ -169,7 +169,7 @@ func (p *BatchProposer) updateDBBatchInfo(batch *encoding.Batch, codecVersion en
if len(batch.Chunks) == 1 {
log.Warn("Disable encode: cannot truncate batch with only 1 chunk for compatibility", "start block number", batch.Chunks[0].Blocks[0].Header.Number.Uint64(),
"end block number", batch.Chunks[0].Blocks[len(batch.Chunks[0].Blocks)-1].Header.Number.Uint64())
enableEncode = false
enableCompress = false
break
}

Expand All @@ -183,7 +183,7 @@ func (p *BatchProposer) updateDBBatchInfo(batch *encoding.Batch, codecVersion en

// recalculate batch metrics after truncation
var calcErr error
metrics, calcErr = utils.CalculateBatchMetrics(batch, codecVersion, enableEncode)
metrics, calcErr = utils.CalculateBatchMetrics(batch, codecVersion, enableCompress)
if calcErr != nil {
return fmt.Errorf("failed to calculate batch metrics, batch index: %v, error: %w", batch.Index, calcErr)
}
Expand All @@ -194,9 +194,9 @@ func (p *BatchProposer) updateDBBatchInfo(batch *encoding.Batch, codecVersion en

p.proposeBatchUpdateInfoTotal.Inc()
err := p.db.Transaction(func(dbTX *gorm.DB) error {
dbBatch, dbErr := p.batchOrm.InsertBatch(p.ctx, batch, codecVersion, enableEncode, *metrics, dbTX)
dbBatch, dbErr := p.batchOrm.InsertBatch(p.ctx, batch, codecVersion, enableCompress, *metrics, dbTX)
if dbErr != nil {
log.Warn("BatchProposer.updateBatchInfoInDB insert batch failure", "index", batch.Index, "parent hash", batch.ParentBatchHash.Hex(), "codec version", codecVersion, "enable encode", enableEncode, "error", dbErr)
log.Warn("BatchProposer.updateBatchInfoInDB insert batch failure", "index", batch.Index, "parent hash", batch.ParentBatchHash.Hex(), "codec version", codecVersion, "enable compress", enableCompress, "error", dbErr)
return dbErr
}
if dbErr = p.chunkOrm.UpdateBatchHashInRange(p.ctx, dbBatch.StartChunkIndex, dbBatch.EndChunkIndex, dbBatch.Hash, dbTX); dbErr != nil {
Expand Down Expand Up @@ -266,7 +266,7 @@ func (p *BatchProposer) proposeBatch() error {

for i, chunk := range daChunks {
batch.Chunks = append(batch.Chunks, chunk)
metrics, calcErr := utils.CalculateBatchMetrics(&batch, codecVersion, true /* enable encode for codecv4 */)
metrics, calcErr := utils.CalculateBatchMetrics(&batch, codecVersion, true /* enable compress for codecv4 */)
if calcErr != nil {
return fmt.Errorf("failed to calculate batch metrics: %w", calcErr)
}
Expand Down Expand Up @@ -295,7 +295,7 @@ func (p *BatchProposer) proposeBatch() error {

batch.Chunks = batch.Chunks[:len(batch.Chunks)-1]

metrics, err := utils.CalculateBatchMetrics(&batch, codecVersion, true /* enable encode for codecv4 */)
metrics, err := utils.CalculateBatchMetrics(&batch, codecVersion, true /* enable compress for codecv4 */)
if err != nil {
return fmt.Errorf("failed to calculate batch metrics: %w", err)
}
Expand All @@ -305,7 +305,7 @@ func (p *BatchProposer) proposeBatch() error {
}
}

metrics, calcErr := utils.CalculateBatchMetrics(&batch, codecVersion, true /* enable encode for codecv4 */)
metrics, calcErr := utils.CalculateBatchMetrics(&batch, codecVersion, true /* enable compress for codecv4 */)
if calcErr != nil {
return fmt.Errorf("failed to calculate batch metrics: %w", calcErr)
}
Expand Down
24 changes: 12 additions & 12 deletions rollup/internal/controller/watcher/chunk_proposer.go
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ type ChunkProposer struct {
gasCostIncreaseMultiplier float64
maxUncompressedBatchBytesSize uint64

enableConditionalEncoding bool
enableConditionalCompress bool

chainCfg *params.ChainConfig

Expand Down Expand Up @@ -86,7 +86,7 @@ func NewChunkProposer(ctx context.Context, cfg *config.ChunkProposerConfig, chai
chunkTimeoutSec: cfg.ChunkTimeoutSec,
gasCostIncreaseMultiplier: cfg.GasCostIncreaseMultiplier,
maxUncompressedBatchBytesSize: cfg.MaxUncompressedBatchBytesSize,
enableConditionalEncoding: cfg.EnableConditionalEncoding,
enableConditionalCompress: cfg.EnableConditionalCompress,
chainCfg: chainCfg,

chunkProposerCircleTotal: promauto.With(reg).NewCounter(prometheus.CounterOpts{
Expand Down Expand Up @@ -174,7 +174,7 @@ func (p *ChunkProposer) updateDBChunkInfo(chunk *encoding.Chunk, codecVersion en
}

compatibilityBreachOccurred := false
enableEncode := true
enableCompress := true

for {
compatible, err := utils.CheckChunkCompressedDataCompatibility(chunk, codecVersion)
Expand All @@ -191,7 +191,7 @@ func (p *ChunkProposer) updateDBChunkInfo(chunk *encoding.Chunk, codecVersion en

if len(chunk.Blocks) == 1 {
log.Warn("Disable encode: cannot truncate chunk with only 1 block for compatibility", "block number", chunk.Blocks[0].Header.Number)
enableEncode = false
enableCompress = false
}

chunk.Blocks = chunk.Blocks[:len(chunk.Blocks)-1]
Expand All @@ -204,7 +204,7 @@ func (p *ChunkProposer) updateDBChunkInfo(chunk *encoding.Chunk, codecVersion en

// recalculate chunk metrics after truncation
var calcErr error
metrics, calcErr = utils.CalculateChunkMetrics(chunk, codecVersion, enableEncode)
metrics, calcErr = utils.CalculateChunkMetrics(chunk, codecVersion, enableCompress)
if calcErr != nil {
return fmt.Errorf("failed to calculate chunk metrics, start block number: %v, error: %w", chunk.Blocks[0].Header.Number, calcErr)
}
Expand All @@ -215,9 +215,9 @@ func (p *ChunkProposer) updateDBChunkInfo(chunk *encoding.Chunk, codecVersion en

p.proposeChunkUpdateInfoTotal.Inc()
err := p.db.Transaction(func(dbTX *gorm.DB) error {
dbChunk, err := p.chunkOrm.InsertChunk(p.ctx, chunk, codecVersion, enableEncode, *metrics, dbTX)
dbChunk, err := p.chunkOrm.InsertChunk(p.ctx, chunk, codecVersion, enableCompress, *metrics, dbTX)
if err != nil {
log.Warn("ChunkProposer.InsertChunk failed", "codec version", codecVersion, "enable encode", enableEncode, "err", err)
log.Warn("ChunkProposer.InsertChunk failed", "codec version", codecVersion, "enable compress", enableCompress, "err", err)
return err
}
if err := p.l2BlockOrm.UpdateChunkHashInRange(p.ctx, dbChunk.StartBlockNumber, dbChunk.EndBlockNumber, dbChunk.Hash, dbTX); err != nil {
Expand Down Expand Up @@ -265,12 +265,12 @@ func (p *ChunkProposer) proposeChunk() error {
}
}

codecVersion := forks.GetCodecVersion(p.chainCfg, blocks[0].Header.Number.Uint64(), blocks[0].Header.Time, p.enableConditionalEncoding)
codecVersion := forks.GetCodecVersion(p.chainCfg, blocks[0].Header.Number.Uint64(), blocks[0].Header.Time, p.enableConditionalCompress)

// Including Curie block in a sole chunk.
if p.chainCfg.CurieBlock != nil && blocks[0].Header.Number.Cmp(p.chainCfg.CurieBlock) == 0 {
chunk := encoding.Chunk{Blocks: blocks[:1]}
metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable encode for codecv4 */)
metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable compress for codecv4 */)
if calcErr != nil {
return fmt.Errorf("failed to calculate chunk metrics: %w", calcErr)
}
Expand All @@ -282,7 +282,7 @@ func (p *ChunkProposer) proposeChunk() error {
for i, block := range blocks {
chunk.Blocks = append(chunk.Blocks, block)

metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable encode for codecv4 */)
metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable compress for codecv4 */)
if calcErr != nil {
return fmt.Errorf("failed to calculate chunk metrics: %w", calcErr)
}
Expand Down Expand Up @@ -319,7 +319,7 @@ func (p *ChunkProposer) proposeChunk() error {

chunk.Blocks = chunk.Blocks[:len(chunk.Blocks)-1]

metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable encode for codecv4 */)
metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable compress for codecv4 */)
if calcErr != nil {
return fmt.Errorf("failed to calculate chunk metrics: %w", calcErr)
}
Expand All @@ -329,7 +329,7 @@ func (p *ChunkProposer) proposeChunk() error {
}
}

metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable encode for codecv4 */)
metrics, calcErr := utils.CalculateChunkMetrics(&chunk, codecVersion, true /* enable compress for codecv4 */)
if calcErr != nil {
return fmt.Errorf("failed to calculate chunk metrics: %w", calcErr)
}
Expand Down
8 changes: 4 additions & 4 deletions rollup/internal/orm/batch.go
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ type Batch struct {
ParentBatchHash string `json:"parent_batch_hash" gorm:"column:parent_batch_hash"`
BatchHeader []byte `json:"batch_header" gorm:"column:batch_header"`
CodecVersion int16 `json:"codec_version" gorm:"column:codec_version"`
EnableEncode bool `json:"enable_encode" gorm:"column:enable_encode"`
EnableCompress bool `json:"enable_compress" gorm:"column:enable_compress"`

// proof
ChunkProofsStatus int16 `json:"chunk_proofs_status" gorm:"column:chunk_proofs_status;default:1"`
Expand Down Expand Up @@ -249,7 +249,7 @@ func (o *Batch) GetBatchByIndex(ctx context.Context, index uint64) (*Batch, erro
}

// InsertBatch inserts a new batch into the database.
func (o *Batch) InsertBatch(ctx context.Context, batch *encoding.Batch, codecVersion encoding.CodecVersion, enableEncode bool, metrics rutils.BatchMetrics, dbTX ...*gorm.DB) (*Batch, error) {
func (o *Batch) InsertBatch(ctx context.Context, batch *encoding.Batch, codecVersion encoding.CodecVersion, enableCompress bool, metrics rutils.BatchMetrics, dbTX ...*gorm.DB) (*Batch, error) {
if batch == nil {
return nil, errors.New("invalid args: batch is nil")
}
Expand All @@ -270,7 +270,7 @@ func (o *Batch) InsertBatch(ctx context.Context, batch *encoding.Batch, codecVer
startChunkIndex = parentBatch.EndChunkIndex + 1
}

batchMeta, err := rutils.GetBatchMetadata(batch, codecVersion, enableEncode)
batchMeta, err := rutils.GetBatchMetadata(batch, codecVersion, enableCompress)
if err != nil {
log.Error("failed to get batch metadata", "index", batch.Index, "total l1 message popped before", batch.TotalL1MessagePoppedBefore,
"parent hash", batch.ParentBatchHash, "number of chunks", numChunks, "err", err)
Expand All @@ -290,7 +290,7 @@ func (o *Batch) InsertBatch(ctx context.Context, batch *encoding.Batch, codecVer
ParentBatchHash: batch.ParentBatchHash.Hex(),
BatchHeader: batchMeta.BatchBytes,
CodecVersion: int16(codecVersion),
EnableEncode: enableEncode,
EnableCompress: enableCompress,
ChunkProofsStatus: int16(types.ChunkProofsStatusPending),
ProvingStatus: int16(types.ProvingTaskUnassigned),
RollupStatus: int16(types.RollupPending),
Expand Down
6 changes: 3 additions & 3 deletions rollup/internal/orm/chunk.go
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ type Chunk struct {
ParentChunkStateRoot string `json:"parent_chunk_state_root" gorm:"column:parent_chunk_state_root"`
WithdrawRoot string `json:"withdraw_root" gorm:"column:withdraw_root"`
CodecVersion int16 `json:"codec_version" gorm:"column:codec_version;default:-1"`
EnableEncode bool `json:"enable_encode" gorm:"column:enable_encode"` // use for debug
EnableCompress bool `json:"enable_compress" gorm:"column:enable_compress"` // use for debug

// proof
ProvingStatus int16 `json:"proving_status" gorm:"column:proving_status;default:1"`
Expand Down Expand Up @@ -177,7 +177,7 @@ func (o *Chunk) GetChunksByBatchHash(ctx context.Context, batchHash string) ([]*
}

// InsertChunk inserts a new chunk into the database.
func (o *Chunk) InsertChunk(ctx context.Context, chunk *encoding.Chunk, codecVersion encoding.CodecVersion, enableEncode bool, metrics utils.ChunkMetrics, dbTX ...*gorm.DB) (*Chunk, error) {
func (o *Chunk) InsertChunk(ctx context.Context, chunk *encoding.Chunk, codecVersion encoding.CodecVersion, enableCompress bool, metrics utils.ChunkMetrics, dbTX ...*gorm.DB) (*Chunk, error) {
if chunk == nil || len(chunk.Blocks) == 0 {
return nil, errors.New("invalid args")
}
Expand Down Expand Up @@ -228,7 +228,7 @@ func (o *Chunk) InsertChunk(ctx context.Context, chunk *encoding.Chunk, codecVer
ParentChunkStateRoot: parentChunkStateRoot,
WithdrawRoot: chunk.Blocks[numBlocks-1].WithdrawRoot.Hex(),
CodecVersion: int16(codecVersion),
EnableEncode: enableEncode,
EnableCompress: enableCompress,
ProvingStatus: int16(types.ProvingTaskUnassigned),
CrcMax: metrics.CrcMax,
BlobSize: metrics.L1CommitBlobSize,
Expand Down
Loading

0 comments on commit 50895d7

Please sign in to comment.