diff --git a/pages/builders/chain-operators/configuration/batcher.mdx b/pages/builders/chain-operators/configuration/batcher.mdx
index 44d6c816c..5d23e5fef 100644
--- a/pages/builders/chain-operators/configuration/batcher.mdx
+++ b/pages/builders/chain-operators/configuration/batcher.mdx
@@ -10,7 +10,50 @@ import { Callout, Tabs } from 'nextra/components'
This page lists all configuration options for the op-batcher. The op-batcher posts
L2 sequencer data to the L1, to make it available for verifiers. The following
-options are from the `--help` in [v1.7.6](https://github.com/ethereum-optimism/optimism/releases/tag/v1.7.6).
+options are from the `--help` in [v1.10.0](https://github.com/ethereum-optimism/optimism/releases/tag/op-batcher%2Fv1.10.0).
+
+## Recommendations
+
+### Set your `OP_BATCHER_MAX_CHANNEL_DURATION`
+
+
+ The default value inside `op-batcher`, if not specified, is still `0`, which means channel duration tracking is disabled.
+ For very low throughput chains, this would mean to fill channels until close to the sequencing window and post the channel to `L1 SUB_SAFETY_MARGIN` L1 blocks before the sequencing window expires.
+
+
+To minimize costs, we recommend setting your `OP_BATCHER_MAX_CHANNEL_DURATION` to target 5 hours, with a value of `1500` L1 blocks. When non-zero, this parameter is the max time (in L1 blocks, which are 12 seconds each) between which batches will be submitted to the L1. If you have this set to 5 for example, then your batcher will send a batch to the L1 every 5\*12=60 seconds. When using blobs, because 130kb blobs need to be purchased in full, if your chain doesn't generate at least \~130kb of data in those 60 seconds, then you'll be posting only partially full blobs and wasting storage.
+
+* We do not recommend setting any values higher than targeting 5 hours, as batches have to be submitted within the sequencing window which defaults to 12 hours for OP chains, otherwise your chain may experience a 12 hour long chain reorg. 5 hours is the longest length of time we recommend that still sits snugly within that 12 hour window to avoid affecting stability.
+* If your chain fills up full blobs of data before the `OP_BATCHER_MAX_CHANNEL_DURATION` elapses, a batch will be submitted anyways - (e.g. even if the OP Mainnet batcher sets an `OP_BATCHER_MAX_CHANNEL_DURATION` of 5 hours, it will still be submitting batches every few minutes)
+
+
+ While setting an`OP_BATCHER_MAX_CHANNEL_DURATION` of `1500` results in the cheapest fees, it also means that your [safe head](https://github.com/ethereum-optimism/specs/blob/main/specs/glossary.md#safe-l2-head) can stall for up to 5 hours.
+
+ * This will negatively impact apps on your chain that rely on the safe head for operation. While many apps can likely operate simply by following the unsafe head, often Centralized Exchanges or third party bridges wait until transactions are marked safe before processing deposits and withdrawal.
+ * Thus a larger gap between posting batches can result in significant delays in the operation of certain types of high-security applications.
+
+
+### Configure your batcher to use multiple blobs
+
+The `op-batcher` has the capabilities to send multiple blobs per single blob transaction. This is accomplished by the use of multi-frame channels, see the [specs](https://specs.optimism.io/protocol/derivation.html#frame-format) for more technical details on channels and frames.
+
+A minimal batcher configuration (with env vars) to enable 6-blob batcher transactions is:
+
+```
+ - OP_BATCHER_BATCH_TYPE=1 # span batches, optional
+ - OP_BATCHER_DATA_AVAILABILITY_TYPE=blobs
+ - OP_BATCHER_TARGET_NUM_FRAMES=6 # 6 blobs per tx
+ - OP_BATCHER_TXMGR_MIN_BASEFEE=2.0 # 2 gwei, might need to tweak, depending on gas market
+ - OP_BATCHER_TXMGR_MIN_TIP_CAP=2.0 # 2 gwei, might need to tweak, depending on gas market
+ - OP_BATCHER_RESUBMISSION_TIMEOUT=240s # wait 4 min before bumping fees
+```
+
+This enables blob transactions and sets the target number of frames to 6, which translates to 6 blobs per transaction.
+The minimum tip cap and base fee are also lifted to 2 gwei because it is uncertain how easy it will be to get 6-blob transactions included and slightly higher priority fees should help.
+The resubmission timeout is increased to a few minutes to give more time for inclusion before bumping the fees because current transaction pool implementations require a doubling of fees for blob transaction replacements.
+
+Multi-blob transactions are particularly useful for medium to high-throughput chains, where enough transaction volume exists to fill up 6 blobs in a reasonable amount of time.
+You can use [this calculator](https://docs.google.com/spreadsheets/d/12VIiXHaVECG2RUunDSVJpn67IQp9NHFJqUsma2PndpE/edit) for your chain to determine what number of blobs are right for you, and what gas scalar configuration to use. Please also refer to guide on [Using Blobs](/builders/chain-operators/management/blobs) for chain operators.
## Global options
@@ -25,6 +68,81 @@ default value is `2m0s`.
`OP_BATCHER_ACTIVE_SEQUENCER_CHECK_DURATION=2m0s`
+### altda.da-server
+
+HTTP address of a DA Server.
+
+
+ `--altda.da-server=`
+ `--altda.da-server=http://da.example.com:1234`
+ `$OP_BATCHER_ALTDA_DA_SERVER=http://da.example.com:1234`
+
+
+### altda.da-service
+
+Use DA service type where commitments are generated by Alt-DA server. The default
+value is `false`.
+
+
+ `--altda.da-service=`
+ `--altda.da-service=true`
+ `$OP_BATCHER_ALTDA_DA_SERVER=true`
+
+
+### altda.enabled
+
+Enable Alt-DA mode
+Alt-DA Mode is a Beta feature of the MIT licensed OP Stack.
+While it has received initial review from core contributors, it is still
+undergoing testing, and may have bugs or other issues.
+The default value is `false`.
+
+
+ `--altda.enabled=`
+ `--altda.enabled=false`
+ `$OP_BATCHER_ALTDA_ENABLED=false`
+
+
+### altda.get-timeout
+
+Timeout for get requests. **0 means no timeout**.
+
+
+ `--altda.get-timeout=`
+ `--altda.get-timeout=5s`
+ `OP_BATCHER_ALTDA_GET_TIMEOUT=5s`
+
+
+### altda.max-concurrent-da-requests
+
+Maximum number of concurrent requests to the DA server.
+
+
+ `--altda.max-concurrent-da-requests=`
+ `--altda.max-concurrent-da-requests=4`
+ `OP_BATCHER_ALTDA_MAX_CONCURRENT_DA_REQUESTS=4`
+
+
+### altda.put-timeout
+
+Timeout for put requests. **0 means no timeout**.
+
+
+ `--altda.put-timeout=`
+ `--altda.put-timeout=10s`
+ `OP_BATCHER_ALTDA_PUT_TIMEOUT=10s`
+
+
+### altda.verify-on-read
+
+Verify input data matches the commitments from the DA storage service.
+
+
+ `--altda.verify-on-read=`
+ `--altda.verify-on-read=true`
+ `OP_BATCHER_ALTDA_VERIFY_ON_READ=true`
+
+
### approx-compr-ratio
The approximate compression ratio (`<=1.0`). Only relevant for ratio
@@ -168,6 +286,26 @@ The lowest log level that will be output. The default value is `INFO`.
`OP_BATCHER_LOG_LEVEL=INFO`
+### log.pid
+
+Show PID in the log.
+
+
+ `--log.pid=`
+ `--log.pid=true`
+ `OP_BATCHER_LOG_PID=true`
+
+
+### max-blocks-per-span-batch
+
+Maximum number of blocks to add to a span batch. **Default is 0 (no maximum)**.
+
+
+ `--max-blocks-per-span-batch=`
+ `--max-blocks-per-span-batch=100`
+ `OP_BATCHER_MAX_BLOCKS_PER_SPAN_BATCH=100`
+
+
### max-channel-duration
The maximum duration of L1-blocks to keep a channel open. 0 to disable. The
@@ -470,6 +608,19 @@ Signer endpoint the client will connect to.
`OP_BATCHER_SIGNER_ENDPOINT=`
+### signer.header
+
+Headers to pass to the remote signer. Format `key=value`.
+Value can contain any character allowed in an HTTP header.
+When using env vars, split multiple headers with commas.
+When using flags, provide one key-value pair per flag.
+
+
+ `--signer.header=`
+ `--signer.header="Authorization=Bearer 123abc"`
+ `OP_BATCHER_SIGNER_HEADER=Authorization=Bearer 123abc`
+
+
### signer.tls.ca
tls ca cert path. The default value is `tls/ca.crt`.
@@ -534,6 +685,63 @@ blob tx, if using Blob DA. The default value is `1`.
`OP_BATCHER_TARGET_NUM_FRAMES=1`
+### throttle-always-block-size
+
+The total DA limit to start imposing on block building **at all times**.
+
+
+ `--throttle-always-block-size=`
+ `--throttle-always-block-size=250000`
+ `OP_BATCHER_THROTTLE_ALWAYS_BLOCK_SIZE=250000`
+
+
+### throttle-block-size
+
+The total DA limit to start imposing on block building **when we are over the throttle threshold**.
+
+
+ `--throttle-block-size=`
+ `--throttle-block-size=50000`
+ `OP_BATCHER_THROTTLE_BLOCK_SIZE=50000`
+
+
+---
+
+### throttle-interval
+
+Interval between potential DA throttling actions. **Zero disables throttling**.
+
+
+ `--throttle-interval=`
+ `--throttle-interval=5s`
+ `OP_BATCHER_THROTTLE_INTERVAL=5s`
+
+
+---
+
+### throttle-threshold
+
+Threshold on `pending-blocks-bytes-current` beyond which the batcher instructs the
+block builder to start throttling transactions with larger DA demands.
+
+
+ `--throttle-threshold=`
+ `--throttle-threshold=1500000`
+ `OP_BATCHER_THROTTLE_THRESHOLD=1500000`
+
+
+---
+
+### throttle-tx-size
+
+The DA size of transactions at which throttling begins **when we are over the throttle threshold**.
+
+
+ `--throttle-tx-size=`
+ `--throttle-tx-size=10000`
+ `OP_BATCHER_THROTTLE_TX_SIZE=10000`
+
+
### txmgr.fee-limit-threshold
The minimum threshold (in GWei) at which fee bumping starts to be capped.
@@ -631,45 +839,31 @@ Print the version. The default value is false.
`--version=false`
-## Recommendations
-
-### Set your `OP_BATCHER_MAX_CHANNEL_DURATION`
-
-
- The default value inside `op-batcher`, if not specified, is still `0`, which means channel duration tracking is disabled.
- For very low throughput chains, this would mean to fill channels until close to the sequencing window and post the channel to `L1 SUB_SAFETY_MARGIN` L1 blocks before the sequencing window expires.
-
-
-To minimize costs, we recommend setting your `OP_BATCHER_MAX_CHANNEL_DURATION` to target 5 hours, with a value of `1500` L1 blocks. When non-zero, this parameter is the max time (in L1 blocks, which are 12 seconds each) between which batches will be submitted to the L1. If you have this set to 5 for example, then your batcher will send a batch to the L1 every 5\*12=60 seconds. When using blobs, because 130kb blobs need to be purchased in full, if your chain doesn't generate at least \~130kb of data in those 60 seconds, then you'll be posting only partially full blobs and wasting storage.
-
-* We do not recommend setting any values higher than targeting 5 hours, as batches have to be submitted within the sequencing window which defaults to 12 hours for OP chains, otherwise your chain may experience a 12 hour long chain reorg. 5 hours is the longest length of time we recommend that still sits snugly within that 12 hour window to avoid affecting stability.
-* If your chain fills up full blobs of data before the `OP_BATCHER_MAX_CHANNEL_DURATION` elapses, a batch will be submitted anyways - (e.g. even if the OP Mainnet batcher sets an `OP_BATCHER_MAX_CHANNEL_DURATION` of 5 hours, it will still be submitting batches every few minutes)
-
-
- While setting an`OP_BATCHER_MAX_CHANNEL_DURATION` of `1500` results in the cheapest fees, it also means that your [safe head](https://github.com/ethereum-optimism/specs/blob/main/specs/glossary.md#safe-l2-head) can stall for up to 5 hours.
+## Batcher Policy
- * This will negatively impact apps on your chain that rely on the safe head for operation. While many apps can likely operate simply by following the unsafe head, often Centralized Exchanges or third party bridges wait until transactions are marked safe before processing deposits and withdrawal.
- * Thus a larger gap between posting batches can result in significant delays in the operation of certain types of high-security applications.
-
+The batcher policy defines high-level constraints and responsibilities regarding how L2 data is posted to L1. Below are the standard guidelines for configuring the batcher within the OP Stack.
-### Configure your batcher to use multiple blobs
+| Parameter | Description | Administrator | Requirement | Notes |
+|-----------|------------|---------------|-------------|--------|
+| Data Availability Type | Specifies whether the batcher uses **blobs** or **calldata** to post transaction data to L1. | Batch submitter address | Ethereum (Blobs or Calldata) | - Alternative data availability (Alt-DA) is not yet supported in the standard configuration.
- The sequencer can switch at will between blob transactions and calldata, with no restrictions, because both are fully secured by L1. |
+| Batch Submission Frequency | Determines how frequently the batcher submits aggregated transaction data to L1 (via the batcher transaction). | Batch submitter address | Must target **1,800 L1 blocks** (6 hours on Ethereum, assuming 12s L1 block time) or lower | - Batches must be posted before the sequencing window closes (commonly 12 hours by default).
- Leave a buffer for L1 network congestion and data size to ensure that each batch is fully committed in a timely manner. |
+| Output Frequency | Defines how frequently L2 output roots are submitted to L1 (via the output oracle). | L1 Proxy Admin | **43,200 L2 blocks** (24 hours at 2s block times) or lower | - Once fault proofs are implemented, this value may become deprecated.
- It cannot be set to 0 (there must be some cadence for outputs). |
-The `op-batcher` has the capabilities to send multiple blobs per single blob transaction. This is accomplished by the use of multi-frame channels, see the [specs](https://specs.optimism.io/protocol/derivation.html#frame-format) for more technical details on channels and frames.
+### Additional Guidance
-A minimal batcher configuration (with env vars) to enable 6-blob batcher transactions is:
+* **Data Availability Types**:
+ * **Calldata** is generally simpler but can be more expensive on mainnet Ethereum, depending on gas prices.
+ * **Blobs** are typically lower cost when your chain has enough transaction volume to fill large chunks of data.
+ * The `op-batcher` can toggle between these approaches by setting the `--data-availability-type=` flag or with the `OP_BATCHER_DATA_AVAILABILITY_TYPE` env variable.
-```
- - OP_BATCHER_BATCH_TYPE=1 # span batches, optional
- - OP_BATCHER_DATA_AVAILABILITY_TYPE=blobs
- - OP_BATCHER_TARGET_NUM_FRAMES=6 # 6 blobs per tx
- - OP_BATCHER_TXMGR_MIN_BASEFEE=2.0 # 2 gwei, might need to tweak, depending on gas market
- - OP_BATCHER_TXMGR_MIN_TIP_CAP=2.0 # 2 gwei, might need to tweak, depending on gas market
- - OP_BATCHER_RESUBMISSION_TIMEOUT=240s # wait 4 min before bumping fees
-```
+* **Batch Submission Frequency** (`OP_BATCHER_MAX_CHANNEL_DURATION` and related flags):
+ * Standard OP Chains frequently target a maximum channel duration between 1–6 hours.
+ * Your chain should never exceed your L2's sequencing window (commonly 12 hours).
+ * If targeting a longer submission window (e.g., 5 or 6 hours), be aware that the [safe head](https://github.com/ethereum-optimism/specs/blob/main/specs/glossary.md#safe-l2-head) can stall up to that duration.
-This enables blob transactions and sets the target number of frames to 6, which translates to 6 blobs per transaction.
-The minimum tip cap and base fee are also lifted to 2 gwei because it is uncertain how easy it will be to get 6-blob transactions included and slightly higher priority fees should help.
-The resubmission timeout is increased to a few minutes to give more time for inclusion before bumping the fees because current transaction pool implementations require a doubling of fees for blob transaction replacements.
+* **Output Frequency**:
+ * Used to post output roots to L1 for verification.
+ * The recommended maximum is 24 hours (43,200 blocks at 2s each), though many chains choose smaller intervals.
+ * Will eventually be replaced or significantly changed by the introduction of fault proofs.
-Multi-blob transactions are particularly useful for medium to high-throughput chains, where enough transaction volume exists to fill up 6 blobs in a reasonable amount of time.
-You can use [this calculator](https://docs.google.com/spreadsheets/d/12VIiXHaVECG2RUunDSVJpn67IQp9NHFJqUsma2PndpE/edit) for your chain to determine what number of blobs are right for you, and what gas scalar configuration to use. Please also refer to guide on [Using Blobs](/builders/chain-operators/management/blobs) for chain operators.
+Include these high-level "policy" requirements when you set up or modify your `op-batcher` configuration. See the [Batcher Configuration](#global-options) reference, which explains each CLI flag and environment variable in depth.