hoverboard ๐Ÿ›น

(e-ipfs mk2). Run miniswap in a Cloudflare worker handling bitswap requests via websockets. Like E-IPFS without having to manage EKS infra.

Motivation

  1. Lower bandwidth egress costs from CF.
  2. bitswap-peer is crash looping and its getting worse. The elastic in the E-IPFS has lost its snap.
  3. There is too much infra in e-ipfs for the team to support.

Hosting it on cloudflare instead of aws + eks would mean

  • ๐ŸŽ‰ Memory management easier when worker per peer/connection
  • ๐Ÿ˜Š No long lived, multi tenant processes to babysit
  • ๐Ÿ’ฐ Cheaper egress from Cloudflare

@alanshaw notes that the majority of the code is already written

The flow

๐Ÿ“ก A libp2p client (kubo, helia, ir0h) connects to
wss://hoverboard.w3s.link and sends a Want-Have message over bitswap v1.2.0 with a root CID.

๐Ÿ›น Hoverboard looks up a CAR CID for the root CID in DUDEWHERE then locates the CAR index in satnav ands stores them in KV using the clients peerID as the key.

๐Ÿ›น Hoverboard could send a Have if the block is large, but in this case it's just a small directory node so it sends the block.

๐Ÿ“ก the client sends a Want with the CIDs for all the links in the Directory.

๐Ÿ›น Hoverboard looks up the satnav index from the KV for this peer ID, and fetches the blocks from R2 and sends block messages with the batch.

repeat to fade

Interesting parts

Sub-request limit

We get to make 1000 subrequests per request to a worker. We need to check if each inbound message on the websocket from the libp2p counts as a reset on that counterโ€ฆ it does for durable objects, so if we need to we could move the work there.

If not, we can just stop at 1000, drop the connection, and let the client contact us again if they still want the blocks. Worse may be better here, as we are just another peer in a decenralised network. Clients must try reconnecting if they think we are a useful provider. This is already happening every few seconds in E-IPFS, but in an uncontrolled way, and gives us some throttling for "free".

Finding non-root blocks

If a worker has to drop a connection mid bitswap dance, and a new connection comes in from an existing peer asking for a non-root block because we already sent them some of the dag then we need a mechanism for find the CAR and index for any block we store.

Longer term, this could be done reliably by exposing satnav-as-a-service, but in the short term we can get this working by just caching the CAR CIDs and indexes we find from the first request from a new connection. Assuming that most sessions start with a request for a root CID, we already have enough info in R2 to find all the blocks for a given dag (โ€ฆthat was uploaded to us! We're not offering "find random blocks from IPFS" here)

Ideally we'd have both. Most requests could be satisfied by caching the indexes, but for true random access we'd have to either fallback to satnav-as-a-service or consider migrating (or replicating) the per block indexes to R2 instead of dynamo.

It is worth exploring a world where we map every block to the list of CARs it is stored as an extention of DUDEWHEREโ€ฆ keys in R2 pointing to empty files (or directly to the relevant indexes if some duplication of storage is tollerable) see upload-api in CF

There is also the potential to unify w3link here so that our gateway can find non-root blocks directly from R2 instead of the current triple spend of e-ipfs -> ipfs.io -> w3s.link

Maximum websocket message size

CF workers websocket impl limits inbound message size to 1Mib but this is ok, as we'd assume that inbound bitswap messages would be lists of CIDs only and unlikely to get near that limit.

Worker duration

As far as I can tell a worker on the unbound plan gets 30s of CPU spin time per request. In the durable object docs it is clear that that limit is per inbound websocket message but it's not clear if that also applies to workers without a durablie object.

Actual "duration" is uncapped, though workers are "more likely to be evicted after 30s".. which we could use some clarification on.

There is no hard limit for duration. However, after 30 seconds, there is a higher chance of eviction.
โ€“ https://developers.cloudflare.com/workers/platform/limits#bundled-usage-model

Capping infra costs

We'd be charged for each worker that is running. 128Mb of mem is allocated per worker, and we'd have to mulitply that by session duration to transfer the complete dag + bandwidth egress from CF.

By keeping track of work done per peerId we can choose to not respond to Want requests after a certain cap per hour/day etc.

Not responding or deliberately closing the connection and letting them reconnect as needed would let us cap costs. Actually slowing down our response rate would likely cost us more over time, as it would dive up session duration.

We can send Not Have if we want to be explicit, but that might send a confusing signal if we really do have the block. However the spec is clear that it's appropriate to send Not Have if you dont want to send a thing you do have.

if C sends S a Have request for data S does not have (or has but is not willing to give to C) and C has requested for DontHave responses then S should respond with DontHave
โ€“ https://github.com/ipfs/specs/blob/main/BITSWAP.md#bitswap-120-interaction-pattern

How long will it take to build

We are motivated to build it! E-IPFS is creaking ominously and none of us want to manage (or has experience of) the terraform and eks stack that is there.

We can prototype it out in 1 week, demo it, and have a better idea of the fesibilty and how long it would take to make production ready.

Publishing CID provider records

We currently hardcode "E-IPFS" via a dns multiaddr as the provider of every block uploaded to us. We need to switch that to be a dnsaddr address so that we can easily change or add multiple providers for each block from a dns txt record we control.

This would allow us to experiment with adding hoverboard as an additional source source without having to change every existing block index that has already been publised (โ€ฆgoing forwards. Existing ones will need updating as alas we didn't land this recommendation first time round)

This is currently handled by E-IPFS indexer in aws. There is an opportunity to re-write this to happen in cloudflare and store the complete index of blocks in the same place as the w3s.link gateway to save money on finding non-root blocks as discussed above.


Notes

WANT cid, cid
- increment asks from peerId
  - if 1000 drop conn ? or let CF do it.
- check denylist KV
- do we have a DUDEWHERE map CID to CARCID
  - fetch index (100Kb)
      - stream block to client (1Mb)
  - ask satnav for index info
      - have block in r2?
          - stream block from r2 to client (1Mb)
          - or stream block from s3 to client
      - no: increment nopes
          - nopes > threshold?
              - drop connection.
const peerMap = new Map([["peerIdString", {ask: 0, hit: 0, miss: 0}]]) // drop if n misses in a row.

satnav-as-as-service

how does this play with content claims?

what's in there?