Capitalized DHT

License: MIT
Signed-off-by: Richard Littauer <richard.littauer@gmail.com>
This commit is contained in:
Richard Littauer 2016-02-18 17:39:26 -05:00
parent f8eefba0a3
commit ac42bb7932
8 changed files with 14 additions and 14 deletions

View File

@ -158,7 +158,7 @@ USAGE:
id Show info about ipfs peers
bootstrap Add or remove bootstrap peers
swarm Manage connections to the p2p network
dht Query the dht for values or peers
dht Query the DHT for values or peers
ping Measure the latency of a connection
diag Print diagnostics

View File

@ -319,7 +319,7 @@ var getValueDhtCmd = &cmds.Command{
Helptext: cmds.HelpText{
Tagline: "Run a 'GetValue' query through the DHT.",
ShortDescription: `
GetValue will return the value stored in the dht at the given key.
GetValue will return the value stored in the DHT at the given key.
`,
},
@ -422,7 +422,7 @@ var putValueDhtCmd = &cmds.Command{
Helptext: cmds.HelpText{
Tagline: "Run a 'PutValue' query through the DHT.",
ShortDescription: `
PutValue will store the given key value pair in the dht.
PutValue will store the given key value pair in the DHT.
`,
},

View File

@ -59,7 +59,7 @@ NETWORK COMMANDS
id Show info about ipfs peers
bootstrap Add or remove bootstrap peers
swarm Manage connections to the p2p network
dht Query the dht for values or peers
dht Query the DHT for values or peers
ping Measure the latency of a connection
diag Print diagnostics

2
dev.md
View File

@ -16,7 +16,7 @@ There are multiple subpackages:
- `path` - path resolution over merkledag data structure
- `peer` - identity + addresses of local and remote peers
- `routing` - the routing system
- `routing/dht` - the dht default routing system implementation
- `routing/dht` - the DHT default routing system implementation
- `swarm` - connection multiplexing, many peers and many transports
- `util` - various utilities

View File

@ -70,7 +70,7 @@ func (dht *IpfsDHT) handleNewMessage(s inet.Stream) {
// measure the RTT for latency measurements.
func (dht *IpfsDHT) sendRequest(ctx context.Context, p peer.ID, pmes *pb.Message) (*pb.Message, error) {
log.Debugf("%s dht starting stream", dht.self)
log.Debugf("%s DHT starting stream", dht.self)
s, err := dht.host.NewStream(ctx, ProtocolDHT, p)
if err != nil {
return nil, err
@ -108,7 +108,7 @@ func (dht *IpfsDHT) sendRequest(ctx context.Context, p peer.ID, pmes *pb.Message
// sendMessage sends out a message
func (dht *IpfsDHT) sendMessage(ctx context.Context, p peer.ID, pmes *pb.Message) error {
log.Debugf("%s dht starting stream", dht.self)
log.Debugf("%s DHT starting stream", dht.self)
s, err := dht.host.NewStream(ctx, ProtocolDHT, p)
if err != nil {
return err

View File

@ -93,7 +93,7 @@ func connect(t *testing.T, ctx context.Context, a, b *IpfsDHT) {
func bootstrap(t *testing.T, ctx context.Context, dhts []*IpfsDHT) {
ctx, cancel := context.WithCancel(ctx)
log.Debugf("bootstrapping dhts...")
log.Debugf("Bootstrapping DHTs...")
// tried async. sequential fares much better. compare:
// 100 async https://gist.github.com/jbenet/56d12f0578d5f34810b2
@ -391,7 +391,7 @@ func TestPeriodicBootstrap(t *testing.T) {
connect(t, ctx, dhts[i], dhts[(i+1)%len(dhts)])
}
t.Logf("dhts are now connected to 1-2 others.", nDHTs)
t.Logf("DHTs are now connected to 1-2 others.", nDHTs)
for _, dht := range dhts {
rtlen := dht.routingTable.Size()
if rtlen > 2 {

View File

@ -106,7 +106,7 @@ func (dht *IpfsDHT) checkLocalDatastore(k key.Key) (*pb.Record, error) {
rec := new(pb.Record)
err = proto.Unmarshal(byts, rec)
if err != nil {
log.Debug("Failed to unmarshal dht record from datastore")
log.Debug("Failed to unmarshal DHT record from datastore.")
return nil, err
}

View File

@ -42,7 +42,7 @@ func (dht *IpfsDHT) GetPublicKey(ctx context.Context, p peer.ID) (ci.PubKey, err
}
// last ditch effort: let's try the dht.
log.Debugf("pk for %s not in peerstore, and peer failed. trying dht.", p)
log.Debugf("pk for %s not in peerstore, and peer failed. Trying DHT.", p)
pkkey := routing.KeyForPublicKey(p)
val, err := dht.GetValue(ctxT, pkkey)
@ -77,14 +77,14 @@ func (dht *IpfsDHT) getPublicKeyFromNode(ctx context.Context, p peer.ID) (ci.Pub
// node doesn't have key :(
record := pmes.GetRecord()
if record == nil {
return nil, fmt.Errorf("node not responding with its public key: %s", p)
return nil, fmt.Errorf("Node not responding with its public key: %s", p)
}
// Success! We were given the value. we don't need to check
// validity because a) we can't. b) we know the hash of the
// key we're looking for.
val := record.GetValue()
log.Debug("dht got a value from other peer.")
log.Debug("DHT got a value from other peer.")
pk, err = ci.UnmarshalPublicKey(val)
if err != nil {
@ -100,7 +100,7 @@ func (dht *IpfsDHT) getPublicKeyFromNode(ctx context.Context, p peer.ID) (ci.Pub
}
// ok! it's valid. we got it!
log.Debugf("dht got public key from node itself.")
log.Debugf("DHT got public key from node itself.")
return pk, nil
}