Skip to content

ipni/index-provider

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Index Provider 📢

Go Reference Coverage Status

A golang implementation of index provider

This repo provides a reference index provider implementation that can be used to advertise content to indexer nodes and serve retrieval requests over graphsync both as a standalone service or embedded into an existing Golang application via a reusable library.

A list of features include:

  • provider CLI that can:
    • Run as a standalone provider daemon instance.
    • Generate and publish indexing advertisements directly from CAR files.
    • Serve retrieval requests for the advertised content over HTTP or HTTP over libp2p.
  • A Golang SDK to embed indexing integration into existing applications, which includes:

Use of the ipni-cli provides additional utility that is useful to check the functioning of an index-provider instance:

  • list advertisements published by a provider instance
  • verify ingestion of multihashes by an indexer node from CAR files, detached CARv2 indices or from an index provider's advertisement chain.

Current status 🚧

This implementation is under active development.

Background

The protocol implemented by this repository is the index provider portion of a larger indexing protocol documented here . The indexer node implementation can be found at storetheindex and go-libipni.

For more details on the ingestion protocol itself see IPNI Spec - Ingestion .

Install

Prerequisite:

To use the provider as a Go library, execute:

go get github.com/ipni/index-provider

To install the latest provider CLI, run:

go install github.com/ipni/index-provider/cmd/provider@latest

Usage

Running an standalone provider daemon

To run a provider service first initialize it by executing:

provider init

Initialization generates a default configuration for the provider instance along with a randomly generated identity keypair. The configuration is stored at user home under .index-provider/config in JSON format. The root configuration path can be overridden by setting the PROVIDER_PATH environment variable

Once initialized, start the service daemon by executing:

provider daemon

The running daemon allows advertisement for new content to the indexer nodes and retrieval of content over GraphSync. Additionally, it starts an admin HTTP server that enables administrative operations using the provider CLI tool. By default, the admin server is bound to http://localhost:3102.

You can then advertise content by importing/removing CAR files via the provider CLI, for example:

provider import car -i <path-to-car-file>

Both CARv1 and CARv2 formats are supported. Index is regenerated on the fly if one is not present.

Exposing delegated routing server from provider (Experimental)

Provider can export a Delegated Routing server. Delegated Routing allows IPFS nodes to advertise their contents to indexers alongside DHT. Delegated Routing server is off by default. To enable it, add the following configuration block to the provider config file.

{
  ...
  DelegatedRouting {
    ListenMultiaddr: "/ip4/0.0.0.0/tcp/50617 (example)"
  }
  ...
}

Configuring Kubo to advertise content onto IPNI (Experimental)

Disclaimer: PUT /routing/v1 is currently not officially supported in Kubo. Please use it at your own risk. See IPIP-378 for the latest updates.

Kubo supports HTTP delegated routing as of v0.18.0. The following section contains configuration examples and a few tips to enable Kubo to advertise its CIDs to IPNI systems like cid.contact using index-provider. Delegated Routing is still in the Experimental stage and configuration might change from version to version. This section serves as an inspiration for configuring your node to use IPNI, but for comprehensive information, refer to the Kubo documentation. Here are some important points to consider:

  • PUT /routing/v1 is currently not officially supported in Kubo. HTTP Delegated Routing supports only reads at the moment, not writes. Please use it at your own risk;
  • The index-provider delegated routing server should be running continuously as a "sidecar" to the Kubo node. While index-provider can be restarted safely, if it goes down, no new CIDs will flow from Kubo to IPNI.
  • The latest version of Kubo with HTTP delegated routing support should be used since index-provider no longer supports Reframe.
  • Kubo advertises its data in snapshots, which means that all CIDs managed by Kubo are reprovided to the configured routers every 12/24 hours (configurable). This mechanism is similar to how the Distributed Hash Table (DHT) works. During the reproviding process, there may be significant communication between the involved processes. In between reprovides, Kubo also sends new individual CIDs to the configured routers.
  • Kubo requires index-provider only for publishing its CIDs to IPNI. Kubo can perform IPNI lookups natively without the need for a sidecar (refer to Kubo docs on auto routers).
  • index-provider must be publicly reachable. IPNI will try to establish connection into it to fetch Advertisement chains. If that can't be done CIDs will not appear in IPNI. Ensure that your firewall is configured to allow incoming connections on the ProviderServer port specified in the index-provider configuration. Ensure that the index-provider is configured to advertise routable addresses in its announcements (where indexers get advertisements) and in its advertisements (where retrieval clients get content).

To configure index-provider to expose the delegated routing server, use the following configuration:

"DelegatedRouting": {
  "ListenMultiaddr": "/ip4/0.0.0.0/tcp/50617",
  "ProviderID": "PEER ID OF YOUR IPFS NODE",
  "Addrs": [] // List of multiaddresses that you'd like to be advertised to IPNI. Announce addrs are going to be advertised if not specified.
}

Configure Kubo to publish into both DHT and IPNI:

"Routing": {
    "Methods": {
      "find-peers": {
        "RouterName": "WanDHT"
      },
      "find-providers": {
        "RouterName": "ParallelHelper"
      },
      "get-ipns": {
        "RouterName": "WanDHT"
      },
      "provide": {
        "RouterName": "ParallelHelper"
      },
      "put-ipns": {
        "RouterName": "WanDHT"
      }
    },
    "Routers": {
      "IndexProvider": {
        "Parameters": {
          "Endpoint": "http://127.0.0.1:50617",
          "MaxProvideBatchSize": 10000,
          "MaxProvideConcurrency": 1
        },
        "Type": "http"
      },
      "ParallelHelper": {
        "Parameters": {
          "Routers": [
            {
              "IgnoreErrors": true,
              "RouterName": "IndexProvider",
              "Timeout": "30m"
            },
            {
              "IgnoreErrors": true,
              "RouterName": "WanDHT",
              "Timeout": "30m"
            }
          ]
        },
        "Type": "parallel"
      },
      "WanDHT": {
        "Parameters": {
          "AcceleratedDHTClient": false,
          "Mode": "auto",
          "PublicIPNetwork": true
        },
        "Type": "dht"
      }
    },
    "Type": "custom"
  },

With the above configuration, Kubo will advertise its CIDs to both DHT and IPNI and will use both DHT and IPNI for find-providers lookups. Additionally, enable the following flag in the Kubo config to enable batch re-provides (especially for larger nodes):

"Experimental": { 
 "AcceleratedDHTClient": true,
},

After adding a new file to your Kubo node, you should see index-provider logs starting to appear immediately. If that doesn't happen, it's likely that Kubo has been configured incorrectly.

index-provider publishes announcements about new advertisements on a libp2p pub/sub topic. This topic is listened by IPNI systems like cid.contact. Once a new announcement is seen, IPNI would reach out to index-provider to download advertisement chains and index the content. It's important to keep in mind:

  • There might be a delay before IPNI picks up an announcement from the libp2p pub/sub depending on the network, number of hops and etc;
  • There might be a delay before IPNI reaches out to index-provider depending on the overall business of the system;
  • If no comminication has been received from IPNI within a reasonable amount of time then most likely index-provider is not reachable from the Internet. You can verify whether it's reachable by using index-provider CLI. For example provider ls ad --provider-addr-info=/ip4/76.21.23.45/tcp/24001/p2p/12D3KooWPNbEgjdBNeaCGpsgCrPRETe4uBZf1ShFXSdN18ys (replace with the correct multiaddress and peer id of your index-provider). Remember to run this command not from the same computer where index-provider is.

Here are a few additional configuration options to consider:

  • ChunkSize: index-provider publishes advertisements with a certain number of CIDs in each chunk. An advertisement needs to accumulate enough CIDs before it gets published. You can reduce the ChunkSize parameter to publish data more quickly. The default value is 1000.
  • AdFlushFrequency: index-provider can publish advertisements before they are full based on the AdFlushFrequency parameter. In other words, an advertisement will be published either when it has reached the ChunkSize or after the specified AdFlushFrequency. Setting this value to a lower value helps in publishing data more quickly. The default is 10 minutes.

Embedding index provider integration

The root go module offers a set of reusable libraries that can be used to embed index provider support into existing application. The core provider.Interface is implemented by engine.Engine.

The provider Engine exposes a set of APIs that allows a user to programmatically announce the availability or removal of content to the indexer nodes referred to as “advertisement”. Advertisements are represented as an IPLD DAG, chained together via a link to the previous advertisement. An advertisement effectively captures the "diff" of the content that is either added or is no longer provided.

Each advertisement contains:

  • Provider ID: the libp2p peer ID of the content provider.
  • Addresses: a list of addresses from which the content can be retrieved.
  • Metadata: a blob of bytes capturing how to retrieve the data.
  • Entries: a link pointing to a list of chunked multihashes.
  • Context ID: a key for the content being advertised.
  • IsRm: flag that tells whether this advertisement is for removal of the previously published content.
  • ExtendedProviders: an optional field that is explained in the next section.

The Entries link points to the IPLD node that contains a list of mulitihashes being advertised. The list is represented as a chain of "Entry Chunk"s where each chunk contains a list of multihashes and a link to the next chunk. This is to accommodate pagination for a large number of multihashes.

The engine can be configured to dynamically look up the list of multihashes that correspond to the context ID of an advertisement. To do this, the engine requires a MultihashLister to be registered. The MultihashLister is then used to look up the list of multihashes associated to a content advertisement.

NotifyPut and NotifyRemove are convinience wrappers on top of Publish that aim to help to construct advertisements. They take care of such things as generating entry chunks, linking to the last published advertisement, signing and others. NotifyPut can be also used to update metadata for a previously published advertisement (for example in the case when a protocol has changed). That can be done by invoking NotifyPut with the same context ID but different metadata field. ErrAlreadyAdvertised will be returned if both context ID and metadata have stayed the same.

For an example on how to start up a provider engine, register a lister and advertise content, see:

Configuration for Sublishing Advertisements

See the Publisher Configuratgion document

Publishing advertisements with extended providers

Extended providers field allows for specification of provider families, in cases where a provider operates multiple PeerIDs, perhaps with different transport protocols between them, but over the same database of content.

ExtendedProviders can either be applied at the chain-level (for all previous and future CIDs published by a provider) or at a context-level (for CIDs belonging to the specified context ID). That behaviour is set by ContextID field. Multiple different behaviours can be triggered by a combination of ContextID, Metadata, ExtendedProviders and Override fields. For more information see the specification

Advertisements with ExtendedProviders can be composed manually or by using a convenience ExtendedProvidersAdBuilder and will have to be signed by the main provider as well as by all ExtendedProviders' identities. Private keys for these identities have to be provided in the xproviders.Info (objects)[https://github.com/ipni/index-provider/blob/main/engine/xproviders/xproviders.go] and
ExtendedProvidersAdBuilder will take care of the rest.

Identity of the main provider will be added to the extended providers list automatically and should not be passed in explicitly.

Some examples can be found below (assumes the readers familiriaty with the (specification)[https://github.com/ipni/storetheindex/blob/main/doc/ingest.md#extendedprovider]).

Publishing an advertisement with context-level ExtendedProviders, that will be returned only for CIDs from the specified context ID:

  adv, err := ep.NewExtendedProviderAdBuilder(providerID, priv, addrs).
    WithContextID(contextID). 
    WithMetadata(metadata). 
    WithOverride(override). 
    WithExtendedProviders(extendedProviders). 
    WithLastAdID(lastAdId). 
    BuildAndSign()

  if err != nil {
    //...
  }

  engine.Publish(ctx, *adv)
)

Constructing an advertisement with chain-level ExtendedProviders, that will be returned for every past and future CID published by the main provider:

  adv, err := ep.NewExtendedProviderAdBuilder(providerID, priv, addrs). 
    WithMetadata(metadata). 
    WithExtendedProviders(extendedProviders).
    WithLastAdID(lastAdId). 
    BuildAndSign()
)

ExtendedProviders can also be used to add a new metadata to all CIDs published by the main provider. Such advertisement need to be constructed without specifying ExtendedProviders at all. The identity of the main provider will be added to the ExtendedProviders list by the builder automatically, so the resulting advertisement will contain only the main provider in the ExtendedProviders list (yes, this is also possible:). That can be used for example to advertise new endpoint with a new protocol alongside the previously advertised one:

  adv, err := ep.NewExtendedProviderAdBuilder(providerID, priv, addrs). 
    WithMetadata(metadata). 
    WithLastAdID(lastAdId). 
    BuildAndSign()
)

On ingestion, previously published ExtendedProviders get overwritten (not merged!) by the newer ones. So in order to update ExtendedProviders, just publish a new chain/context-level advertisement with the required changes.

Examples of constructing advertisements with ExtendedProviders can be found (here)[https://github.com/ipni/index-provider/blob/main/engine/xproviders/xproviders_test.go].

provider CLI

The provider CLI can be used to interact with a running daemon via the admin server to perform a range of administrative operations. For example, the provider CLI can be used to import a CAR file and advertise its content to the indexer nodes by executing:

provider import car -i <path-to-car-file>

For usage description, execute provider --help

Storage Consumption

The index provider engine uses a given datastore to persist two general category of data:

  1. Internal advertisement mappings, and
  2. Chunked entries chain cache

If the datastore passed to the engine is reused, it is recommended to wrap it in a namespace prior to instantiating the engine.

Internal advertisement mappings

The internal advertisement mappings are purely used by the engine to efficiently handle publication requests. It generally includes:

  • mapping to the latest advertisement
  • mappings between advertisement CIDs, their context ID and their corresponding metadata.

The storage consumed by such mappings is negligible and grows linearly as a factor of the number of advertisements published.

Chunked entries chain cache

This category stores chunked entries generated by publishing an advertisement with a never seen before context ID. The chunks are stored in an LRU cache, the maximum size of which is configured by the following configuration parameters in Ingest config:

  • LinkChunkSize - The maximum number of multihashes in a chunk (defaults to 16,384)
  • LinkCacheSize - The maximum number of entries links to chace (defaults to 1024)

The exact storage usage depends on the size of multihashes. For example, using the default config to advertise 128-bit long multihashes will result in chunk sizes of 0.25MiB with maximum cache growth of 256 MiB.

To delete the cache set PurgeLinkCache to true and restart the engine.

Note that the LRU cache may grow beyond its max size if the generated chain of chunks is longer than the configured LinkChunkSize. This is to avoid partial caching of chunks within a single advertisement. The cache expansion is logged in INFO level at provider/engine logging subsystem.

Related Resources

License

SPDX-License-Identifier: Apache-2.0 OR MIT