Skip to content

Commit 7ae5459

Browse files
committed
docs: fix broken links found in audit
- update Arbol main link to arbol.io (company rebranded) - replace broken gateway.arbolmarket.com with archive.org snapshot - replace broken waterlily.ai with archive.org snapshot
1 parent b64433e commit 7ae5459

2 files changed

Lines changed: 3 additions & 3 deletions

File tree

docs/case-studies/arbol.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ _— Ben Andre, CTO, Arbol_
1717
<img src="./images/logo-arbol.svg" alt="Arbol logo" width="220">
1818
:::
1919

20-
[Arbol](https://www.arbolmarket.com/) is a software platform that connects agricultural entities like farmers and other weather-dependent parties with investors and other capital providers to insure and protect against weather-related risks. Arbol's platform sells contracts for parametric weather protection agreements in a marketplace that's an innovative, data-driven approach to risk management, cutting out the usual legacy insurance claims process of making loss assessments on the ground. Instead, Arbol relies on tamper-proof data indexes to determine payouts, and doesn't require a defined loss to be indemnified. Arbol's platform combines parametric weather protection with blockchain-based smart contracts to provide cost-efficient, automated, and user-defined weather-related risk hedging. As with traditional crop insurance and similar legacy products, end users purchase assurance that they'll be financially protected in the case of adverse weather — but with Arbol, these end users are paid automatically if adverse conditions occur, as defined by the contract and measured by local meteorological observations tracked by Arbol's data sources.
20+
[Arbol](https://www.arbol.io/) is a software platform that connects agricultural entities like farmers and other weather-dependent parties with investors and other capital providers to insure and protect against weather-related risks. Arbol's platform sells contracts for parametric weather protection agreements in a marketplace that's an innovative, data-driven approach to risk management, cutting out the usual legacy insurance claims process of making loss assessments on the ground. Instead, Arbol relies on tamper-proof data indexes to determine payouts, and doesn't require a defined loss to be indemnified. Arbol's platform combines parametric weather protection with blockchain-based smart contracts to provide cost-efficient, automated, and user-defined weather-related risk hedging. As with traditional crop insurance and similar legacy products, end users purchase assurance that they'll be financially protected in the case of adverse weather — but with Arbol, these end users are paid automatically if adverse conditions occur, as defined by the contract and measured by local meteorological observations tracked by Arbol's data sources.
2121

2222
To build the data indexes that Arbol uses to handle its contracts, the team aggregates and standardizes billions of data files comprising decades of weather information from a wide range of reputable sources — all of which is stored on IPFS. IPFS is critical to Arbol's service model due to the inherent verifiability provided by its [content-addressed architecture](../concepts/content-addressing.md), as well as a decentralized data delivery model that facilitates Arbol's day-to-day aggregation, synchronization, and distribution of massive amounts of data.
2323

@@ -88,7 +88,7 @@ Arbol's end users enjoy the "it just works" benefits of parametric protection, b
8888

8989
8. **Pinning and syncing:** When storage nodes in the Arbol network detect that a new hash has been added to the heads file, they run the standard, recursive [`ipfs pin -r`](../reference/kubo/cli.md#ipfs-pin) command on it. Arbol's primary active nodes don't need to be large in number: The network includes a single [gateway node](../concepts/ipfs-gateway.md) that bootstraps with all the parsing/hashing nodes, and a few large storage nodes that serve as the primary data storage backup. However, data is also regularly synced with "cold nodes" — archival storage nodes that are mostly kept offline — as well as on individual IPFS nodes on Arbol's developers' and agronomists' personal computers.
9090

91-
9. **Garbage collection:** Some older Arbol datasets require [garbage collection](../concepts/glossary.md#garbage-collection) whenever new data is added, due to a legacy method of overwriting old hashes with new hashes. However, all of Arbol's newer datasets use an architecture where old hashes are preserved and new posts reference the previous post. This methodology creates a linked list of hashes, with each hash containing a reference to the previous hash. As the length of the list becomes computationally burdensome, the system consolidates intermediate nodes and adds a new route to the head, creating a [DAG (directed acyclic graph)](../concepts/merkle-dag.md) structure. Heads are always stored in a master [heads.json reference file](https://gateway.arbolmarket.com/climate/hashes/heads.json) located on Arbol's command server.
91+
9. **Garbage collection:** Some older Arbol datasets require [garbage collection](../concepts/glossary.md#garbage-collection) whenever new data is added, due to a legacy method of overwriting old hashes with new hashes. However, all of Arbol's newer datasets use an architecture where old hashes are preserved and new posts reference the previous post. This methodology creates a linked list of hashes, with each hash containing a reference to the previous hash. As the length of the list becomes computationally burdensome, the system consolidates intermediate nodes and adds a new route to the head, creating a [DAG (directed acyclic graph)](../concepts/merkle-dag.md) structure. Heads are always stored in a master [heads.json reference file](https://web.archive.org/web/20230318223234/https://gateway.arbolmarket.com/climate/hashes/heads.json) located on Arbol's command server.
9292

9393
### The tooling
9494

docs/concepts/cod.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ Bacalhau can:
2525
- Run against data [mounted anywhere](https://docs.bacalhau.org/#how-it-works) on your machine.
2626
- Integrate with services running on nodes to run jobs, such as [DuckDB](https://docs.bacalhau.org/examples/data-engineering/DuckDB/).
2727
- Operate at scale over parallel jobs and batch process petabytes of data.
28-
- Auto-generate art using a [Stable Diffusion AI model](https://www.waterlily.ai/) trained on the chosen artist’s original works.
28+
- Auto-generate art using a [Stable Diffusion AI model](https://web.archive.org/web/20250313163631/https://www.waterlily.ai/) trained on the chosen artist’s original works.
2929

3030
### More Bacalhau resources
3131

0 commit comments

Comments
 (0)