Repair Nodes¶
The Canton platform is generally built to self-heal and automatically recover from issues. As such, if there is a situation where some degradation can be expected, there exists a mechanism that yields graceful degradation and automated recovery from said issues.
Common examples are database outages (retry until success) or network outages (failover and reconnect until success).
Canton reports such issues as warnings to alert the operator about the degradation of its dependencies, but generally, does not require any manual intervention to recover from a degradation.
However, not all situations can be foreseen and corruption of systems can always happen in unanticipated ways. Therefore, Canton can always be manually repaired. This means that whatever the corruption is, there is always a series of operational steps that can be made in order to recover the correct state of a node. If several nodes in the distributed system are affected, it may be necessary to coordinate the recovery among the affected nodes.
Conceptually, this means that Canton recovery is structured along the four layers:
- Automated self-recovery and self-healing.
- Recovery from crash or restart by re-creating a consistent state from the persisted store.
- Standard disaster recovery from a database backup in case of database outage and replay from sync domain.
- Corruption disaster recovery using repair and other console commands to re-establish a consistent state within the distributed system.
If you run into corruption issues, you need to first understand what caused the issue. Ideally, you can contact our support team to help you diagnose the issue and provide you with a custom recipe on how to recover from your issue (and prevent recurrence).
The toolbox the support engineers have at hand are:
- Exporting/importing secret keys
- Manually initializing nodes
- Exporting/importing DARs
- Exporting/importing topology transactions
- Manually adding or removing contracts from the active contract set
- Moving contracts from one sync domain to another
- Manually ignoring faulty transactions (and then using add/remove contract to repair the ACS).
All these methods are very powerful but dangerous. You should not attempt to repair your nodes on your own as you risk severe data corruption.
Keep in mind that the corruption of the system state may not have been discovered immediately; thus, the corruption may have leaked out through the APIs to the applications using the corrupted node. Bringing the node back into a correct state with respect to the other nodes in the distributed system can therefore make the application state inconsistent with the node’s state. Accordingly, the application should either re-initialize itself from the repaired state or itself offer tools to fix inconsistencies.
Preparation¶
As contracts (1) “belong to” parties and (2) are instances of Daml templates defined in Daml Archives (DARs), importing contracts to Canton also requires creating corresponding parties and uploading DARs.
- Contracts are often interdependent requiring care to honor dependencies such that the set of imported contracts is internally consistent. This requires particular attention if you choose to modify contracts prior to their import.
- Additionally, the use of divulgence in the original ledger could have likely introduced non-obvious dependencies that may impede exercising contract choices after import. As a result, such divulged contracts need to be re-divulged as part of the import (by exercising existing choices or, if there are no-side-effect-free choices that re-divulge the necessary contracts, by extending your Daml models with new choices).
- Party IDs have a strict format on Canton ledgers ending with a required “fingerprint” suffix, so at a minimum, you will need to “remap” party IDs.
- Canton contract keys do not have to be unique, so if your Daml models rely on uniqueness, consider extending the models using these strategies or limit your Canton participants to connect to a single Canton sync domain with unique contract key semantics.
- Canton does not support implicit party creation, so be sure to create all needed parties explicitly.
- In addition you could choose to spread contracts, parties, and DARs across multiple Canton Participants.
With the above requirements in mind, you are ready to plan and execute the following three-step process:
- Download parties and contracts from the existing Daml Participant Node and locate the DAR files that the contracts are based on.
- Modify the parties and contracts (at the minimum assigning Canton-conformant party IDs).
- Provision Canton participants along with at least one Canton sync domain. Then upload DARs, create parties, and finally the contracts to the Canton participants. Finally connect the participants to the sync domain(s).
Importing an actual Ledger¶
To follow along with this guide, ensure you have installed and unpacked the Canton release bundle. Install and configure a postgres database for 2 sync domains and 4 participant nodes following the steps described in Persistence using Postgres. Make sure that your Postgres has enough “max_connections” configured to serve effectively 6 Canton nodes as required by this example. Finally, run the following commands from the “canton-X.Y.Z” directory to set up the initial topology.
export CANTON=`pwd`
export CONF="$CANTON/config"
export IMPORT="$CANTON/examples/07-repair"
bin/canton \
-c $IMPORT/participant1.conf,$IMPORT/participant2.conf,$IMPORT/participant3.conf,$IMPORT/participant4.conf \
-c $IMPORT/domain-export-ledger.conf,$IMPORT/domain-import-ledger.conf \
-c $CONF/storage/postgres.conf,$IMPORT/enable-preview-commands.conf \
--bootstrap $IMPORT/import-ledger-init.canton
This sets up an “exportLedger” with a set of parties consisting of painters, house owners, and banks along with a handful of paint offer contracts and IOUs.
Define the following helper functions useful to extract parties and contracts via the ledger api:
def queryActiveContractsFromDamlLedger(
hostname: String,
port: Port,
templateFilter: Seq[TemplateId],
tls: Option[TlsClientConfig],
token: Option[String] = None,
)(implicit consoleEnvironment: ConsoleEnvironment): Seq[CreatedEvent] = {
// Helper to query the ledger api using the specified command.
def queryLedgerApi[Svc <: AbstractStub[Svc], Result](
command: GrpcAdminCommand[_, _, Result]
): Either[String, Result] =
consoleEnvironment.grpcLedgerCommandRunner
.runCommand("sourceLedger", command, ClientConfig(hostname, port, tls), token)
.toEither
(for {
// Identify all the parties on the ledger and narrow down the list to local parties.
allParties <- queryLedgerApi(
LedgerApiCommands.PartyManagementService.ListKnownParties(identityProviderId = "")
)
localParties = allParties.collect {
case PartyDetails(party, _, isLocal, _, _) if isLocal => LfPartyId.assertFromString(party)
}
// Query the ActiveContractsService for the actual contracts
acs <- queryLedgerApi(
LedgerApiCommands.AcsService
.GetActiveContracts(
localParties.toSet,
templateFilter = templateFilter,
limit = PositiveInt.MaxValue,
timeout = NonNegativeDuration.maxTimeout,
includeCreatedEventBlob = true,
)(consoleEnvironment.environment.scheduler)
)
} yield acs.map(_.event)).valueOr(err =>
throw new IllegalStateException(s"Failed to query parties, ledger id, or acs: $err")
)
}
def removeCantonSpecifics(acs: Seq[CreatedEvent]): Seq[CreatedEvent] = {
def stripPartyIdSuffix(suffixedPartyId: String): String =
suffixedPartyId.split(SafeSimpleString.delimiter).head
acs.map { event =>
ValueRemapper.convertEvent(identity, stripPartyIdSuffix)(event)
}
}
def lookUpPartyId(participant: ParticipantReference, party: String): PartyId = {
participant.parties
.list(filterParty = party + SafeSimpleString.delimiter)
.map(_.party)
.headOption
.getOrElse(throw new IllegalArgumentException("Unknown party"))
}
As the first step, export the active contract set (ACS). To illustrate how to import data from non-Canton ledgers, strip the Canton-specifics by making the party ids generic (stripping the Canton-specific suffix).
val acs =
queryActiveContractsFromDamlLedger(
exportLedger.config.ledgerApi.address,
exportLedger.config.ledgerApi.port,
TemplateId
.templateIdsFromJava(
iou.Iou.TEMPLATE_ID_WITH_PACKAGE_ID,
paint.OfferToPaintHouseByPainter.TEMPLATE_ID_WITH_PACKAGE_ID,
),
exportLedger.config.ledgerApi.tls.map(_.clientConfig),
)
val acsExported = removeCantonSpecifics(acs).toList
Step number two involves preparing the Canton participants and sync domain by uploading DARs and creating parties. Here we choose to place the house owners, painters, and banks on different participants.
Also modify the events to be based on the newly created party ids.
// Decide on which canton participants to host which parties along with their contracts.
// We place house owners, painters, and banks on separate participants.
val participants = Seq(participant1, participant2, participant3)
val partyAssignments =
Seq(participant1 -> houseOwners, participant2 -> painters, participant3 -> banks)
// Connect to domain prior to uploading dars and parties.
participants.foreach { participant =>
participant.domains.connect_local(importLedgerDomain)
participant.dars.upload(darPath)
}
// Create canton party ids and remember mapping of plain to canton party ids.
val toCantonParty: Map[String, String] =
partyAssignments.flatMap { case (participant, parties) =>
val partyMappingOnParticipant = parties.map { party =>
participant.ledger_api.parties.allocate(party, party)
party -> lookUpPartyId(participant, party).toLf
}
partyMappingOnParticipant
}.toMap
// Create traffic on all participants so that the repair commands will pick an identity snapshot that is aware of
// all party allocations
participants.foreach { participant =>
participant.health.ping(participant, workflowId = importLedgerDomain.name)
}
// Switch the ACS to be based on canton party ids.
val acsToImportToCanton =
acsExported.map(ValueRemapper.convertEvent(identity, toCantonParty(_)))
As the third step, perform the actual import to each participant filtering the contracts based on the location of contract stakeholders and witnesses.
// Disconnect from domain temporarily to allow import to be performed.
participants.foreach(_.domains.disconnect(importLedgerDomain.name))
// Pick a ledger create time according to the domain's clock.
val ledgerCreateTime = {
consoleEnvironment.environment.domains
.getRunning(importLedgerDomain.name)
.getOrElse(throw new IllegalArgumentException("No running domain node"))
.clock
.now
.toInstant
}
// Specify protocol version used in the installation e.g. com.digitalasset.canton.version.ProtocolVersion.latest
val protocolVersion = testedProtocolVersion
val contractsWithRecomputedContractIds = {
utils
.recompute_contract_ids(
participant1,
acsToImportToCanton.view
.map(WrappedCreatedEvent)
.map { event =>
utils.contract_data_to_instance(event.toContractData, ledgerCreateTime)
}
.toSeq,
protocolVersion,
)
._1
}
val createdEventsAndContractsToImport =
acsToImportToCanton zip contractsWithRecomputedContractIds
// Filter active contracts based on participant parties and upload.
partyAssignments.foreach { case (participant, rawParties) =>
val parties = rawParties.map(toCantonParty(_))
val participantAcs = createdEventsAndContractsToImport.collect {
case (event, contract)
if event.signatories.intersect(parties).nonEmpty
|| event.observers.intersect(parties).nonEmpty
|| event.witnessParties.intersect(parties).nonEmpty =>
repairContract(contract)
}
participant.repair.add(importLedgerDomain.name, participantAcs, ignoreAlreadyAdded = false)
}
Finally, reconnect to the sync domain.
participants.foreach(_.domains.reconnect(importLedgerDomain.name))
To demonstrate that the imported ledger works, let’s have each of the house owners accept one of the painters’ offers to paint their house.
def yesYouMayPaintMyHouse(
houseOwner: PartyId,
painter: PartyId,
participant: ParticipantReference,
): Unit = {
val iouContract = participant.ledger_api.javaapi.acs.await(iou.Iou.COMPANION)(houseOwner)
val bank = iouContract.data.payer
val paintProposal = participant.ledger_api.javaapi.acs
.await(paint.OfferToPaintHouseByPainter.COMPANION)(
houseOwner,
pp => pp.data.painter == painter.toProtoPrimitive && pp.data.bank == bank,
)
val cmd = paintProposal.id
.exerciseAcceptByOwner(iouContract.id)
.commands
.asScala
.toSeq
val _ = clue(
s"$houseOwner accepts paint proposal by $painter financing through $bank"
)(participant.ledger_api.javaapi.commands.submit(Seq(houseOwner), cmd))
}
// Have each house owner accept one of the paint offers to illustrate use of the imported ledger.
houseOwners.zip(painters).foreach { case (houseOwner, painter) =>
yesYouMayPaintMyHouse(
lookUpPartyId(participant1, houseOwner),
lookUpPartyId(participant1, painter),
participant1,
)
}
// Illustrate that acceptance of have resulted in
{
val paintHouseContracts = painters.map { painter =>
participant2.ledger_api.javaapi.acs
.await(paint.PaintHouse.COMPANION)(lookUpPartyId(participant2, painter))
}
assert(paintHouseContracts.size == 4)
}
This guide has demonstrated how to import data from non-Canton Daml Participant Nodes or from a Canton Participant of a lower major version as part of a Canton upgrade.
Repairing Participants¶
Canton enables interoperability of distributed participants and sync domains. Particularly in distributed settings without trust assumptions, faults in one part of the system should ideally produce minimal irrecoverable damage to other parts. For example if a sync domain is irreparably lost, the participants previously connected to that sync domain need to recover and be empowered to continue their workflows on a new sync domain.
This guide will illustrate how to replace a lost sync domain with a new sync domain providing business continuity to affected participants.
Recovering from a Lost Synchronization Domain¶
Note
Please note that the given section describes a preview feature, because using multiple sync domains is a preview feature.
Suppose that a set of participants have been conducting workflows via a sync domain that runs into trouble. In fact consider that the sync domain has gotten into such a disastrous state that the sync domain is beyond repair, for example:
- The sync domain has experienced data loss and is unable to be restored from backups or the backups are missing crucial recent history.
- The sync domain data is found to be corrupt causing participants to lose trust in the sync domain as a mediator.
Next the participant operators each examine their local state, and upon coordinating conclude that their participants’ active contracts are “mostly the same”. This sync domain-recovery repair demo illustrates how the participants can
- coordinate to agree on a set of contracts to use moving forward, serving as a new consistent state,
- copying over the agreed-upon set of contracts to a brand-new sync domain,
- “fail over” to the new sync domain,
- and finally continue running workflows on the new sync domain having recovered from the permanent loss of the old sync domain.
Repairing an actual Topology¶
To follow along with this guide, ensure you have installed and unpacked the Canton release bundle and run the following commands from the “canton-X.Y.Z” directory to set up the initial topology.
export CANTON=`pwd`
export CONF="$CANTON/config"
export REPAIR="$CANTON/examples/07-repair"
bin/canton \
-c $REPAIR/participant1.conf,$REPAIR/participant2.conf,$REPAIR/domain-repair-lost.conf,$REPAIR/domain-repair-new.conf \
-c $CONF/storage/postgres.conf,$REPAIR/enable-preview-commands.conf \
--bootstrap $REPAIR/domain-repair-init.canton
To simplify the demonstration, this not only sets up the starting topology of
- two participants, “participant1” and “participant2”, along with
- one sync domain “lostDomain” that is about to become permanently unavailable leaving “participant1” and “participant2” unable to continue executing workflows,
but also already includes the ingredients needed to recover:
- The setup includes “newDomain” that we will rely on as a replacement sync domain, and
- we already enable the “enable-preview-commands” configuration needed to make available the “repair.change_domain” command.
In practice you would only add the new sync domain once you need to recover from sync domain loss and also only then enable the repair commands.
We simulate “lostDomain” permanently disappearing by stopping the sync domain and never bringing it up again to emphasize the point that the participants no longer have access to any state from the sync domain. We also disconnect “participant1” and “participant2” from “lostDomain” to reflect that the participants have “given up” on the sync domain and recognize the need for a replacement for business continuity. The fact that we disconnect the participants “at the same time” is somewhat artificial as in practice the participants might have lost connectivity to the sync domain at different times (more on reconciling contracts below).
lostDomain.stop()
Seq(participant1, participant2).foreach { p =>
p.domains.disconnect(lostDomain.name)
// Also let the participant know not to attempt to reconnect to lostDomain
p.domains.modify(lostDomain.name, _.copy(manualConnect = true))
}
Even though the sync domain is “the node that has broken”, recovering entails repairing the participants using the “newDomain” already set up. As of now, participant repairs have to be performed in an offline fashion requiring participants being repaired to be disconnected from the the new sync domain. However we temporarily connect to the sync domain, to let the topology state initialize, and disconnect only once the parties can be used on the new sync domain.
Seq(participant1, participant2).foreach(_.domains.connect_local(newDomain))
// Wait for topology state to appear before disconnecting again.
clue("newDomain initialization timed out") {
eventually()(
(
participant1.domains.active(newDomain.name),
participant2.domains.active(newDomain.name),
) shouldBe (true, true)
)
}
// Run a few transactions on the new domain so that the topology state chosen by the repair commands
// really is the active one that we've seen
participant1.health.ping(participant2, workflowId = newDomain.name)
Seq(participant1, participant2).foreach(_.domains.disconnect(newDomain.name))
With the participants connected neither to “lostDomain” nor “newDomain”, each participant can
- locally look up the active contracts assigned to the lost sync domain using the “testing.pcs_search” command made available via the “features.enable-testing-commands” configuration,
- and invoke “repair.change_domain” (enabled via the “features.enable-preview-commands” configuration) in order to “move” the contracts to the new sync domain.
// Extract participant contracts from "lostDomain".
val contracts1 =
participant1.testing.pcs_search(lostDomain.name, filterTemplate = "^Iou", activeSet = true)
val contracts2 =
participant2.testing.pcs_search(lostDomain.name, filterTemplate = "^Iou", activeSet = true)
// Ensure that shared contracts match.
val Seq(sharedContracts1, sharedContracts2) = Seq(contracts1, contracts2).map(
_.filter { case (_isActive, contract) =>
contract.metadata.stakeholders.contains(Alice.toLf) &&
contract.metadata.stakeholders.contains(Bob.toLf)
}.toSet
)
clue("checking if contracts match") {
sharedContracts1 shouldBe sharedContracts2
}
// Finally change the contracts from "lostDomain" to "newDomain"
participant1.repair.change_domain(
contracts1.map(_._2.contractId),
lostDomain.name,
newDomain.name,
)
participant2.repair.change_domain(
contracts2.map(_._2.contractId),
lostDomain.name,
newDomain.name,
skipInactive = false,
)
Note
The code snippet above includes a check that the contracts shared among the participants match (as determined by each participant, “sharedContracts1” by “participant1” and “sharedContracts2” by “participant2). Should the contracts not match (as could happen if the participants had lost connectivity to the sync domain at different times), this check fails at soliciting the participant operators to reach an agreement on the set of contracts. The agreed-upon set of active contracts may for example be
- the intersection of the active contracts among the participants
- or perhaps the union (for which the operators can use the “repair.add” command to create the contracts missing from one participant).
Also note that both the repair commands and the “testing.pcs_search” command are currently “preview” features, and therefore their names may change.
Once each participant has associated the contracts with “newDomain”, let’s have them reconnect, and we should be able to confirm that the new sync domain is able to execute workflows from where the lost sync domain disappeared.
Seq(participant1, participant2).foreach(_.domains.reconnect(newDomain.name))
// Look up a couple of contracts moved from lostDomain
val Seq(iouAlice, iouBob) = Seq(participant1 -> Alice, participant2 -> Bob).map {
case (participant, party) =>
participant.ledger_api.javaapi.acs
.await[iou.Iou.Contract, iou.Iou.ContractId, iou.Iou](iou.Iou.COMPANION)(
party,
_.data.owner == party.toProtoPrimitive,
)
}
// Ensure that we can create new contracts
Seq(participant1 -> ((Alice, Bob)), participant2 -> ((Bob, Alice))).foreach {
case (participant, (payer, owner)) =>
participant.ledger_api.javaapi.commands.submit_flat(
Seq(payer),
new iou.Iou(
payer.toProtoPrimitive,
owner.toProtoPrimitive,
new iou.Amount(200.toBigDecimal, "USD"),
java.util.List.of(),
).create.commands.asScala.toSeq,
)
}
// Even better: Confirm that we can exercise choices on the moved contracts
Seq(participant2 -> ((Bob, iouBob)), participant1 -> ((Alice, iouAlice))).foreach {
case (participant, (owner, iou)) =>
participant.ledger_api.javaapi.commands
.submit_flat(Seq(owner), iou.id.exerciseCall().commands.asScala.toSeq)
}
In practice, we would now be in a position to remove the “lostDomain” from both participants and to disable the repair commands again to prevent accidental use of these “dangerously powerful” tools.
This guide has demonstrated how participants can recover from losing a sync domain that has been permanently lost or somehow become irreparably corrupted.
Repair Macros¶
Some operations are combined as macros, which are a series of consecutive repair commands, coded as a single command. While we discourage you from using these commands on your own, we document them here for the sake of completeness. These macros are available only in the enterprise edition.
Clone Identity¶
Many nodes can be rehydrated from a sync domain, as long as the sync domain is not pruned. In such situations, you might want to reset your node while keeping the identity and the secret keys of the node. This can be done using the repair macros.
You need local console access to the node. If you are running your production node in a container, you need to create a new configuration file that allows you to access the database of the node from an interactive console. Make sure that the normal node process is stopped and that nothing else is accessing the same database (e.g. ensure that replication is turned on). Also, make sure that the nodes are configured to not perform auto-initialization, as this would create a new identity. You ensure that by setting the corresponding auto-init configuration option to false:
canton.participants.myparticipant.init.auto-init = false
Then start Canton interactively using:
./bin/canton -c myconfig --manual-start
Starting with --manual-start
will prevent the participant from attempting to reconnect to the sync domains.
Then, you can download the identity state of the node to a directory on the machine you are running the
process:
repair.identity.download(participant, tempDirParticipant)
repair.dars.download(participant, tempDirParticipant)
participant.stop()
This will store the secret keys, the topology state and the identity on the disk in the given directory.
You can run the identity.download
command on all nodes. However, mediator and sequencer
nodes will only store their keys in files, as the sequencer’s identity is attached to the sync domain identity
and the mediator’s identity is set only later during initialization.
The dars.download
command is a convenience command to download all dars that have been
added to the participant via the console command participant.dars.upload
. Dars that
were uploaded through the Ledger API need to be manually re-uploaded to the new participant.
Once the data is stored, stop the node and then truncate the database (please back it up before). Then restart the node and upload the identity data again:
participant.start()
repair.identity.upload(participant, tempDirParticipant)
repair.dars.upload(participant, tempDirParticipant)
Please note that dar uploading is only necessary for participants.
Now, depending on the node type, you need to re-integrate the node into the sync domain. For the sync domain nodes, you need to grab the static sync domain parameters and the sync domain ID from the sync domain manager. If you have remote access to the sync domain manager, you can run
val domainId = domainManager1.id
val domainParameters = domainManager1.service.get_static_domain_parameters
You also want to grab the mediator identities for each mediator using:
val mediatorId = mediator.id
For the sequencer, rehydration works only if the sync domain uses a blockchain; the database-only sequencers cannot rehydrate. So rehydration for blockchain-based sequencers will be:
repair.identity.upload(newSequencer, tempDirSequencer)
newSequencer.initialization.initialize_from_beginning(domainId, domainParameters)
newSequencer.health.wait_for_initialized()
For the sync domain manager, it looks like:
repair.identity.upload(domainManager2, tempDirDomainManager)
domainManager2.setup.init(
SequencerConnections.single(newSequencer.sequencerConnection)
)
domainManager2.health.wait_for_initialized()
For the mediator, it would be:
repair.identity.upload(mediator, tempDirMediator)
mediator.mediator.initialize(
domainId,
mediatorId,
domainParameters,
SequencerConnections.single(newSequencer.sequencerConnection),
topologySnapshot = None,
)
mediator.health.wait_for_initialized()
For a participant, you would reconnect it to the sync domain using a normal connect:
participant.domains.connect_local(sequencer)
Note that this will replay all transactions from the sync domain. However, command deduplication will only be fully functional once the participant catches up with the sync domain. Therefore, you need to ensure that applications relying on command deduplication do not submit commands during recovery.
Importing Existing Contracts¶
You may have existing contracts, parties, and DARs in other Daml Participant Nodes (such as the Daml sandbox) that you want to import into your Canton-based participant node. To address this need, you can extract contracts and associated parties via the ledger api, modify contracts, parties, and daml archived as needed, and upload the data to Canton using the Canton Console.
You can also import existing contracts from Canton as that is useful as part of Canton upgrades across major versions with incompatible internal storage.