1eade4ae58
This introduces a pin/update operation which allows to Pin a new item to cluster indicating that said pin is an update to an already-existing pin. When this is the case, all the configuration for the existing pin is copied to the new one (including allocations). The IPFS connector will then trigger pin/update directly in IPFS, allowing an efficient pinning based on DAG-differences. Since the allocations where the same for both pins, the pin/update can proceed. PinUpdate does not unpin the previous pin (it is not possible to do this atomically in cluster like it happens in IPFS). The user can manually do it after the pin/update is done. Internally, after a lot of deliberations on what the optimal way for this is, I opted for adding a `PinUpdate` option to the `PinOptions` type (carries the CID to update from). In order to carry this option from the REST API to the IPFS Connector, it is serialized in the Protobuf (and stored in the datastore). There is no other way to do this in a simple fashion since the Pin object is piece of information that is sent around. Additionally, making it a PinOption plays well with the Pin/PinPath APIs which need little changes. Effectively, you are pinning a new thing. You are just indicating that it should be configured from an existing one. Fixes #732
619 lines
16 KiB
Go
619 lines
16 KiB
Go
package ipfscluster
|
|
|
|
import (
|
|
"context"
|
|
|
|
"github.com/ipfs/ipfs-cluster/api"
|
|
"github.com/ipfs/ipfs-cluster/version"
|
|
|
|
cid "github.com/ipfs/go-cid"
|
|
peer "github.com/libp2p/go-libp2p-core/peer"
|
|
rpc "github.com/libp2p/go-libp2p-gorpc"
|
|
|
|
ocgorpc "github.com/lanzafame/go-libp2p-ocgorpc"
|
|
"go.opencensus.io/trace"
|
|
)
|
|
|
|
// RPC endpoint types w.r.t. trust level
|
|
const (
|
|
// RPCClosed endpoints can only be called by the local cluster peer
|
|
// on itself.
|
|
RPCClosed RPCEndpointType = iota
|
|
// RPCTrusted endpoints can be called by "trusted" peers.
|
|
// It depends which peers are considered trusted. For example,
|
|
// in "raft" mode, Cluster will all peers as "trusted". In "crdt" mode,
|
|
// trusted peers are those specified in the configuration.
|
|
RPCTrusted
|
|
// RPCOpen endpoints can be called by any peer in the Cluster swarm.
|
|
RPCOpen
|
|
)
|
|
|
|
// RPCEndpointType controls how access is granted to an RPC endpoint
|
|
type RPCEndpointType int
|
|
|
|
// A trick to find where something is used (i.e. Cluster.Pin):
|
|
// grep -R -B 3 '"Pin"' | grep -C 1 '"Cluster"'.
|
|
// This does not cover globalPinInfo*(...) broadcasts nor redirects to leader
|
|
// in Raft.
|
|
|
|
// newRPCServer returns a new RPC Server for Cluster.
|
|
func newRPCServer(c *Cluster) (*rpc.Server, error) {
|
|
var s *rpc.Server
|
|
|
|
authF := func(pid peer.ID, svc, method string) bool {
|
|
endpointType, ok := c.config.RPCPolicy[svc+"."+method]
|
|
if !ok {
|
|
return false
|
|
}
|
|
|
|
switch endpointType {
|
|
case RPCTrusted:
|
|
return c.consensus.IsTrustedPeer(c.ctx, pid)
|
|
case RPCOpen:
|
|
return true
|
|
default:
|
|
return false
|
|
}
|
|
}
|
|
|
|
if c.config.Tracing {
|
|
s = rpc.NewServer(
|
|
c.host,
|
|
version.RPCProtocol,
|
|
rpc.WithServerStatsHandler(&ocgorpc.ServerHandler{}),
|
|
rpc.WithAuthorizeFunc(authF),
|
|
)
|
|
} else {
|
|
s = rpc.NewServer(c.host, version.RPCProtocol, rpc.WithAuthorizeFunc(authF))
|
|
}
|
|
|
|
cl := &ClusterRPCAPI{c}
|
|
err := s.RegisterName(RPCServiceID(cl), cl)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
pt := &PinTrackerRPCAPI{c.tracker}
|
|
err = s.RegisterName(RPCServiceID(pt), pt)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
ic := &IPFSConnectorRPCAPI{c.ipfs}
|
|
err = s.RegisterName(RPCServiceID(ic), ic)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
cons := &ConsensusRPCAPI{c.consensus}
|
|
err = s.RegisterName(RPCServiceID(cons), cons)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
pm := &PeerMonitorRPCAPI{c.monitor}
|
|
err = s.RegisterName(RPCServiceID(pm), pm)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return s, nil
|
|
}
|
|
|
|
// RPCServiceID returns the Service ID for the given RPCAPI object.
|
|
func RPCServiceID(rpcSvc interface{}) string {
|
|
switch rpcSvc.(type) {
|
|
case *ClusterRPCAPI:
|
|
return "Cluster"
|
|
case *PinTrackerRPCAPI:
|
|
return "PinTracker"
|
|
case *IPFSConnectorRPCAPI:
|
|
return "IPFSConnector"
|
|
case *ConsensusRPCAPI:
|
|
return "Consensus"
|
|
case *PeerMonitorRPCAPI:
|
|
return "PeerMonitor"
|
|
default:
|
|
return ""
|
|
}
|
|
}
|
|
|
|
// ClusterRPCAPI is a go-libp2p-gorpc service which provides the internal peer
|
|
// API for the main cluster component.
|
|
type ClusterRPCAPI struct {
|
|
c *Cluster
|
|
}
|
|
|
|
// PinTrackerRPCAPI is a go-libp2p-gorpc service which provides the internal
|
|
// peer API for the PinTracker component.
|
|
type PinTrackerRPCAPI struct {
|
|
tracker PinTracker
|
|
}
|
|
|
|
// IPFSConnectorRPCAPI is a go-libp2p-gorpc service which provides the
|
|
// internal peer API for the IPFSConnector component.
|
|
type IPFSConnectorRPCAPI struct {
|
|
ipfs IPFSConnector
|
|
}
|
|
|
|
// ConsensusRPCAPI is a go-libp2p-gorpc service which provides the
|
|
// internal peer API for the Consensus component.
|
|
type ConsensusRPCAPI struct {
|
|
cons Consensus
|
|
}
|
|
|
|
// PeerMonitorRPCAPI is a go-libp2p-gorpc service which provides the
|
|
// internal peer API for the PeerMonitor component.
|
|
type PeerMonitorRPCAPI struct {
|
|
mon PeerMonitor
|
|
}
|
|
|
|
/*
|
|
Cluster component methods
|
|
*/
|
|
|
|
// ID runs Cluster.ID()
|
|
func (rpcapi *ClusterRPCAPI) ID(ctx context.Context, in struct{}, out *api.ID) error {
|
|
id := rpcapi.c.ID(ctx)
|
|
*out = *id
|
|
return nil
|
|
}
|
|
|
|
// Pin runs Cluster.pin().
|
|
func (rpcapi *ClusterRPCAPI) Pin(ctx context.Context, in *api.Pin, out *api.Pin) error {
|
|
// we do not call the Pin method directly since that method does not
|
|
// allow to pin other than regular DataType pins. The adder will
|
|
// however send Meta, Shard and ClusterDAG pins.
|
|
pin, _, err := rpcapi.c.pin(ctx, in, []peer.ID{})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pin
|
|
return nil
|
|
}
|
|
|
|
// Unpin runs Cluster.Unpin().
|
|
func (rpcapi *ClusterRPCAPI) Unpin(ctx context.Context, in *api.Pin, out *api.Pin) error {
|
|
pin, err := rpcapi.c.Unpin(ctx, in.Cid)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pin
|
|
return nil
|
|
}
|
|
|
|
// PinPath resolves path into a cid and runs Cluster.Pin().
|
|
func (rpcapi *ClusterRPCAPI) PinPath(ctx context.Context, in *api.PinPath, out *api.Pin) error {
|
|
pin, err := rpcapi.c.PinPath(ctx, in.Path, in.PinOptions)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pin
|
|
return nil
|
|
}
|
|
|
|
// UnpinPath resolves path into a cid and runs Cluster.Unpin().
|
|
func (rpcapi *ClusterRPCAPI) UnpinPath(ctx context.Context, in *api.PinPath, out *api.Pin) error {
|
|
pin, err := rpcapi.c.UnpinPath(ctx, in.Path)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pin
|
|
return nil
|
|
}
|
|
|
|
// Pins runs Cluster.Pins().
|
|
func (rpcapi *ClusterRPCAPI) Pins(ctx context.Context, in struct{}, out *[]*api.Pin) error {
|
|
cidList, err := rpcapi.c.Pins(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = cidList
|
|
return nil
|
|
}
|
|
|
|
// PinGet runs Cluster.PinGet().
|
|
func (rpcapi *ClusterRPCAPI) PinGet(ctx context.Context, in cid.Cid, out *api.Pin) error {
|
|
pin, err := rpcapi.c.PinGet(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pin
|
|
return nil
|
|
}
|
|
|
|
// Version runs Cluster.Version().
|
|
func (rpcapi *ClusterRPCAPI) Version(ctx context.Context, in struct{}, out *api.Version) error {
|
|
*out = api.Version{
|
|
Version: rpcapi.c.Version(),
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Peers runs Cluster.Peers().
|
|
func (rpcapi *ClusterRPCAPI) Peers(ctx context.Context, in struct{}, out *[]*api.ID) error {
|
|
*out = rpcapi.c.Peers(ctx)
|
|
return nil
|
|
}
|
|
|
|
// PeerAdd runs Cluster.PeerAdd().
|
|
func (rpcapi *ClusterRPCAPI) PeerAdd(ctx context.Context, in peer.ID, out *api.ID) error {
|
|
id, err := rpcapi.c.PeerAdd(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *id
|
|
return nil
|
|
}
|
|
|
|
// ConnectGraph runs Cluster.GetConnectGraph().
|
|
func (rpcapi *ClusterRPCAPI) ConnectGraph(ctx context.Context, in struct{}, out *api.ConnectGraph) error {
|
|
graph, err := rpcapi.c.ConnectGraph()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = graph
|
|
return nil
|
|
}
|
|
|
|
// PeerRemove runs Cluster.PeerRm().
|
|
func (rpcapi *ClusterRPCAPI) PeerRemove(ctx context.Context, in peer.ID, out *struct{}) error {
|
|
return rpcapi.c.PeerRemove(ctx, in)
|
|
}
|
|
|
|
// Join runs Cluster.Join().
|
|
func (rpcapi *ClusterRPCAPI) Join(ctx context.Context, in api.Multiaddr, out *struct{}) error {
|
|
return rpcapi.c.Join(ctx, in.Value())
|
|
}
|
|
|
|
// StatusAll runs Cluster.StatusAll().
|
|
func (rpcapi *ClusterRPCAPI) StatusAll(ctx context.Context, in struct{}, out *[]*api.GlobalPinInfo) error {
|
|
pinfos, err := rpcapi.c.StatusAll(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = pinfos
|
|
return nil
|
|
}
|
|
|
|
// StatusAllLocal runs Cluster.StatusAllLocal().
|
|
func (rpcapi *ClusterRPCAPI) StatusAllLocal(ctx context.Context, in struct{}, out *[]*api.PinInfo) error {
|
|
pinfos := rpcapi.c.StatusAllLocal(ctx)
|
|
*out = pinfos
|
|
return nil
|
|
}
|
|
|
|
// Status runs Cluster.Status().
|
|
func (rpcapi *ClusterRPCAPI) Status(ctx context.Context, in cid.Cid, out *api.GlobalPinInfo) error {
|
|
pinfo, err := rpcapi.c.Status(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pinfo
|
|
return nil
|
|
}
|
|
|
|
// StatusLocal runs Cluster.StatusLocal().
|
|
func (rpcapi *ClusterRPCAPI) StatusLocal(ctx context.Context, in cid.Cid, out *api.PinInfo) error {
|
|
pinfo := rpcapi.c.StatusLocal(ctx, in)
|
|
*out = *pinfo
|
|
return nil
|
|
}
|
|
|
|
// SyncAll runs Cluster.SyncAll().
|
|
func (rpcapi *ClusterRPCAPI) SyncAll(ctx context.Context, in struct{}, out *[]*api.GlobalPinInfo) error {
|
|
pinfos, err := rpcapi.c.SyncAll(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = pinfos
|
|
return nil
|
|
}
|
|
|
|
// SyncAllLocal runs Cluster.SyncAllLocal().
|
|
func (rpcapi *ClusterRPCAPI) SyncAllLocal(ctx context.Context, in struct{}, out *[]*api.PinInfo) error {
|
|
pinfos, err := rpcapi.c.SyncAllLocal(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = pinfos
|
|
return nil
|
|
}
|
|
|
|
// Sync runs Cluster.Sync().
|
|
func (rpcapi *ClusterRPCAPI) Sync(ctx context.Context, in cid.Cid, out *api.GlobalPinInfo) error {
|
|
pinfo, err := rpcapi.c.Sync(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pinfo
|
|
return nil
|
|
}
|
|
|
|
// SyncLocal runs Cluster.SyncLocal().
|
|
func (rpcapi *ClusterRPCAPI) SyncLocal(ctx context.Context, in cid.Cid, out *api.PinInfo) error {
|
|
pinfo, err := rpcapi.c.SyncLocal(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pinfo
|
|
return nil
|
|
}
|
|
|
|
// RecoverAllLocal runs Cluster.RecoverAllLocal().
|
|
func (rpcapi *ClusterRPCAPI) RecoverAllLocal(ctx context.Context, in struct{}, out *[]*api.PinInfo) error {
|
|
pinfos, err := rpcapi.c.RecoverAllLocal(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = pinfos
|
|
return nil
|
|
}
|
|
|
|
// Recover runs Cluster.Recover().
|
|
func (rpcapi *ClusterRPCAPI) Recover(ctx context.Context, in cid.Cid, out *api.GlobalPinInfo) error {
|
|
pinfo, err := rpcapi.c.Recover(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pinfo
|
|
return nil
|
|
}
|
|
|
|
// RecoverLocal runs Cluster.RecoverLocal().
|
|
func (rpcapi *ClusterRPCAPI) RecoverLocal(ctx context.Context, in cid.Cid, out *api.PinInfo) error {
|
|
pinfo, err := rpcapi.c.RecoverLocal(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *pinfo
|
|
return nil
|
|
}
|
|
|
|
// BlockAllocate returns allocations for blocks. This is used in the adders.
|
|
// It's different from pin allocations when ReplicationFactor < 0.
|
|
func (rpcapi *ClusterRPCAPI) BlockAllocate(ctx context.Context, in *api.Pin, out *[]peer.ID) error {
|
|
if rpcapi.c.config.FollowerMode {
|
|
return errFollowerMode
|
|
}
|
|
|
|
err := rpcapi.c.setupPin(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Return the current peer list.
|
|
if in.ReplicationFactorMin < 0 {
|
|
// Returned metrics are Valid and belong to current
|
|
// Cluster peers.
|
|
metrics := rpcapi.c.monitor.LatestMetrics(ctx, pingMetricName)
|
|
peers := make([]peer.ID, len(metrics), len(metrics))
|
|
for i, m := range metrics {
|
|
peers[i] = m.Peer
|
|
}
|
|
|
|
*out = peers
|
|
return nil
|
|
}
|
|
|
|
allocs, err := rpcapi.c.allocate(
|
|
ctx,
|
|
in.Cid,
|
|
in.ReplicationFactorMin,
|
|
in.ReplicationFactorMax,
|
|
[]peer.ID{}, // blacklist
|
|
[]peer.ID{}, // prio list
|
|
)
|
|
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
*out = allocs
|
|
return nil
|
|
}
|
|
|
|
// SendInformerMetric runs Cluster.sendInformerMetric().
|
|
func (rpcapi *ClusterRPCAPI) SendInformerMetric(ctx context.Context, in struct{}, out *api.Metric) error {
|
|
m, err := rpcapi.c.sendInformerMetric(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *m
|
|
return nil
|
|
}
|
|
|
|
/*
|
|
Tracker component methods
|
|
*/
|
|
|
|
// Track runs PinTracker.Track().
|
|
func (rpcapi *PinTrackerRPCAPI) Track(ctx context.Context, in *api.Pin, out *struct{}) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/tracker/Track")
|
|
defer span.End()
|
|
return rpcapi.tracker.Track(ctx, in)
|
|
}
|
|
|
|
// Untrack runs PinTracker.Untrack().
|
|
func (rpcapi *PinTrackerRPCAPI) Untrack(ctx context.Context, in *api.Pin, out *struct{}) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/tracker/Untrack")
|
|
defer span.End()
|
|
return rpcapi.tracker.Untrack(ctx, in.Cid)
|
|
}
|
|
|
|
// StatusAll runs PinTracker.StatusAll().
|
|
func (rpcapi *PinTrackerRPCAPI) StatusAll(ctx context.Context, in struct{}, out *[]*api.PinInfo) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/tracker/StatusAll")
|
|
defer span.End()
|
|
*out = rpcapi.tracker.StatusAll(ctx)
|
|
return nil
|
|
}
|
|
|
|
// Status runs PinTracker.Status().
|
|
func (rpcapi *PinTrackerRPCAPI) Status(ctx context.Context, in cid.Cid, out *api.PinInfo) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/tracker/Status")
|
|
defer span.End()
|
|
pinfo := rpcapi.tracker.Status(ctx, in)
|
|
*out = *pinfo
|
|
return nil
|
|
}
|
|
|
|
// RecoverAll runs PinTracker.RecoverAll().f
|
|
func (rpcapi *PinTrackerRPCAPI) RecoverAll(ctx context.Context, in struct{}, out *[]*api.PinInfo) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/tracker/RecoverAll")
|
|
defer span.End()
|
|
pinfos, err := rpcapi.tracker.RecoverAll(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = pinfos
|
|
return nil
|
|
}
|
|
|
|
// Recover runs PinTracker.Recover().
|
|
func (rpcapi *PinTrackerRPCAPI) Recover(ctx context.Context, in cid.Cid, out *api.PinInfo) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/tracker/Recover")
|
|
defer span.End()
|
|
pinfo, err := rpcapi.tracker.Recover(ctx, in)
|
|
*out = *pinfo
|
|
return err
|
|
}
|
|
|
|
/*
|
|
IPFS Connector component methods
|
|
*/
|
|
|
|
// Pin runs IPFSConnector.Pin().
|
|
func (rpcapi *IPFSConnectorRPCAPI) Pin(ctx context.Context, in *api.Pin, out *struct{}) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/ipfsconn/IPFSPin")
|
|
defer span.End()
|
|
return rpcapi.ipfs.Pin(ctx, in)
|
|
}
|
|
|
|
// Unpin runs IPFSConnector.Unpin().
|
|
func (rpcapi *IPFSConnectorRPCAPI) Unpin(ctx context.Context, in *api.Pin, out *struct{}) error {
|
|
return rpcapi.ipfs.Unpin(ctx, in.Cid)
|
|
}
|
|
|
|
// PinLsCid runs IPFSConnector.PinLsCid().
|
|
func (rpcapi *IPFSConnectorRPCAPI) PinLsCid(ctx context.Context, in cid.Cid, out *api.IPFSPinStatus) error {
|
|
b, err := rpcapi.ipfs.PinLsCid(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = b
|
|
return nil
|
|
}
|
|
|
|
// PinLs runs IPFSConnector.PinLs().
|
|
func (rpcapi *IPFSConnectorRPCAPI) PinLs(ctx context.Context, in string, out *map[string]api.IPFSPinStatus) error {
|
|
m, err := rpcapi.ipfs.PinLs(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = m
|
|
return nil
|
|
}
|
|
|
|
// ConfigKey runs IPFSConnector.ConfigKey().
|
|
func (rpcapi *IPFSConnectorRPCAPI) ConfigKey(ctx context.Context, in string, out *interface{}) error {
|
|
res, err := rpcapi.ipfs.ConfigKey(in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = res
|
|
return nil
|
|
}
|
|
|
|
// RepoStat runs IPFSConnector.RepoStat().
|
|
func (rpcapi *IPFSConnectorRPCAPI) RepoStat(ctx context.Context, in struct{}, out *api.IPFSRepoStat) error {
|
|
res, err := rpcapi.ipfs.RepoStat(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = *res
|
|
return err
|
|
}
|
|
|
|
// SwarmPeers runs IPFSConnector.SwarmPeers().
|
|
func (rpcapi *IPFSConnectorRPCAPI) SwarmPeers(ctx context.Context, in struct{}, out *[]peer.ID) error {
|
|
res, err := rpcapi.ipfs.SwarmPeers(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = res
|
|
return nil
|
|
}
|
|
|
|
// BlockPut runs IPFSConnector.BlockPut().
|
|
func (rpcapi *IPFSConnectorRPCAPI) BlockPut(ctx context.Context, in *api.NodeWithMeta, out *struct{}) error {
|
|
return rpcapi.ipfs.BlockPut(ctx, in)
|
|
}
|
|
|
|
// BlockGet runs IPFSConnector.BlockGet().
|
|
func (rpcapi *IPFSConnectorRPCAPI) BlockGet(ctx context.Context, in cid.Cid, out *[]byte) error {
|
|
res, err := rpcapi.ipfs.BlockGet(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = res
|
|
return nil
|
|
}
|
|
|
|
// Resolve runs IPFSConnector.Resolve().
|
|
func (rpcapi *IPFSConnectorRPCAPI) Resolve(ctx context.Context, in string, out *cid.Cid) error {
|
|
c, err := rpcapi.ipfs.Resolve(ctx, in)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = c
|
|
return nil
|
|
}
|
|
|
|
/*
|
|
Consensus component methods
|
|
*/
|
|
|
|
// LogPin runs Consensus.LogPin().
|
|
func (rpcapi *ConsensusRPCAPI) LogPin(ctx context.Context, in *api.Pin, out *struct{}) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/consensus/LogPin")
|
|
defer span.End()
|
|
return rpcapi.cons.LogPin(ctx, in)
|
|
}
|
|
|
|
// LogUnpin runs Consensus.LogUnpin().
|
|
func (rpcapi *ConsensusRPCAPI) LogUnpin(ctx context.Context, in *api.Pin, out *struct{}) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/consensus/LogUnpin")
|
|
defer span.End()
|
|
return rpcapi.cons.LogUnpin(ctx, in)
|
|
}
|
|
|
|
// AddPeer runs Consensus.AddPeer().
|
|
func (rpcapi *ConsensusRPCAPI) AddPeer(ctx context.Context, in peer.ID, out *struct{}) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/consensus/AddPeer")
|
|
defer span.End()
|
|
return rpcapi.cons.AddPeer(ctx, in)
|
|
}
|
|
|
|
// RmPeer runs Consensus.RmPeer().
|
|
func (rpcapi *ConsensusRPCAPI) RmPeer(ctx context.Context, in peer.ID, out *struct{}) error {
|
|
ctx, span := trace.StartSpan(ctx, "rpc/consensus/RmPeer")
|
|
defer span.End()
|
|
return rpcapi.cons.RmPeer(ctx, in)
|
|
}
|
|
|
|
// Peers runs Consensus.Peers().
|
|
func (rpcapi *ConsensusRPCAPI) Peers(ctx context.Context, in struct{}, out *[]peer.ID) error {
|
|
peers, err := rpcapi.cons.Peers(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*out = peers
|
|
return nil
|
|
}
|
|
|
|
/*
|
|
PeerMonitor
|
|
*/
|
|
|
|
// LatestMetrics runs PeerMonitor.LatestMetrics().
|
|
func (rpcapi *PeerMonitorRPCAPI) LatestMetrics(ctx context.Context, in string, out *[]*api.Metric) error {
|
|
*out = rpcapi.mon.LatestMetrics(ctx, in)
|
|
return nil
|
|
}
|