2017-05-24 16:51:49 +02:00
|
|
|
import base64
|
|
|
|
from binascii import hexlify
|
2017-03-11 11:12:08 +01:00
|
|
|
import hashlib
|
2017-05-24 16:51:49 +02:00
|
|
|
import math
|
2017-03-11 11:12:08 +01:00
|
|
|
import time
|
|
|
|
import socket
|
2017-05-24 16:51:49 +02:00
|
|
|
import struct
|
|
|
|
import random
|
|
|
|
import traceback
|
|
|
|
|
|
|
|
from addresses import calculateInventoryHash
|
|
|
|
from debug import logger
|
|
|
|
from inventory import Inventory
|
|
|
|
import knownnodes
|
2017-03-11 11:12:08 +01:00
|
|
|
from network.advanceddispatcher import AdvancedDispatcher
|
2017-05-24 16:51:49 +02:00
|
|
|
from network.bmobject import BMObject, BMObjectInsufficientPOWError, BMObjectInvalidDataError, BMObjectExpiredError, BMObjectUnwantedStreamError, BMObjectInvalidError, BMObjectAlreadyHaveError
|
|
|
|
import network.connectionpool
|
|
|
|
from network.downloadqueue import DownloadQueue
|
2017-04-16 18:27:15 +02:00
|
|
|
from network.node import Node
|
2017-03-11 11:12:08 +01:00
|
|
|
import network.asyncore_pollchoose as asyncore
|
2017-05-27 19:09:21 +02:00
|
|
|
from network.objectracker import ObjectTracker
|
2017-03-11 11:12:08 +01:00
|
|
|
from network.proxy import Proxy, ProxyError, GeneralProxyError
|
2017-05-24 16:51:49 +02:00
|
|
|
from network.uploadqueue import UploadQueue, UploadElem, AddrUploadQueue, ObjUploadQueue
|
2017-03-11 11:12:08 +01:00
|
|
|
|
|
|
|
import addresses
|
2017-04-04 10:46:01 +02:00
|
|
|
from bmconfigparser import BMConfigParser
|
2017-05-29 00:24:07 +02:00
|
|
|
from queues import objectProcessorQueue, portCheckerQueue, UISignalQueue, invQueue
|
2017-04-16 18:27:15 +02:00
|
|
|
import shared
|
2017-05-24 16:51:49 +02:00
|
|
|
import state
|
2017-03-11 11:12:08 +01:00
|
|
|
import protocol
|
|
|
|
|
|
|
|
class BMProtoError(ProxyError): pass
|
|
|
|
|
|
|
|
|
2017-05-24 16:51:49 +02:00
|
|
|
class BMProtoInsufficientDataError(BMProtoError): pass
|
|
|
|
|
|
|
|
|
|
|
|
class BMProtoExcessiveDataError(BMProtoError): pass
|
|
|
|
|
|
|
|
|
2017-05-27 19:09:21 +02:00
|
|
|
class BMProto(AdvancedDispatcher, ObjectTracker):
|
2017-03-20 18:32:26 +01:00
|
|
|
# ~1.6 MB which is the maximum possible size of an inv message.
|
|
|
|
maxMessageSize = 1600100
|
2017-05-24 16:51:49 +02:00
|
|
|
# 2**18 = 256kB is the maximum size of an object payload
|
|
|
|
maxObjectPayloadSize = 2**18
|
2017-04-04 10:46:01 +02:00
|
|
|
# protocol specification says max 1000 addresses in one addr command
|
|
|
|
maxAddrCount = 1000
|
|
|
|
# protocol specification says max 50000 objects in one inv command
|
|
|
|
maxObjectCount = 50000
|
2017-05-27 19:09:21 +02:00
|
|
|
# address is online if online less than this many seconds ago
|
|
|
|
addressAlive = 10800
|
|
|
|
# maximum time offset
|
|
|
|
maxTimeOffset = 3600
|
|
|
|
|
2017-05-29 00:24:07 +02:00
|
|
|
def __init__(self, address=None, sock=None):
|
|
|
|
AdvancedDispatcher.__init__(self, sock)
|
|
|
|
self.isOutbound = False
|
|
|
|
# packet/connection from a local IP
|
|
|
|
self.local = False
|
2017-03-20 18:32:26 +01:00
|
|
|
|
|
|
|
def bm_proto_reset(self):
|
|
|
|
self.magic = None
|
|
|
|
self.command = None
|
2017-05-24 16:51:49 +02:00
|
|
|
self.payloadLength = 0
|
2017-03-20 18:32:26 +01:00
|
|
|
self.checksum = None
|
2017-03-11 11:12:08 +01:00
|
|
|
self.payload = None
|
2017-03-20 18:32:26 +01:00
|
|
|
self.invalid = False
|
2017-04-16 18:27:15 +02:00
|
|
|
self.payloadOffset = 0
|
2017-05-27 22:30:30 +02:00
|
|
|
self.expectBytes = protocol.Header.size
|
2017-05-24 16:51:49 +02:00
|
|
|
self.object = None
|
2017-03-11 11:12:08 +01:00
|
|
|
|
2017-03-20 18:32:26 +01:00
|
|
|
def state_bm_header(self):
|
|
|
|
if len(self.read_buf) < protocol.Header.size:
|
2017-05-24 16:51:49 +02:00
|
|
|
#print "Length below header size"
|
2017-03-11 11:12:08 +01:00
|
|
|
return False
|
2017-03-20 18:32:26 +01:00
|
|
|
self.magic, self.command, self.payloadLength, self.checksum = protocol.Header.unpack(self.read_buf[:protocol.Header.size])
|
|
|
|
self.command = self.command.rstrip('\x00')
|
|
|
|
if self.magic != 0xE9BEB4D9:
|
|
|
|
# skip 1 byte in order to sync
|
|
|
|
self.bm_proto_reset()
|
|
|
|
self.set_state("bm_header", 1)
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("Bad magic")
|
2017-05-25 23:04:33 +02:00
|
|
|
self.close()
|
|
|
|
return False
|
2017-05-27 19:09:21 +02:00
|
|
|
if self.payloadLength > BMProto.maxMessageSize:
|
2017-03-20 18:32:26 +01:00
|
|
|
self.invalid = True
|
2017-05-27 22:30:30 +02:00
|
|
|
self.set_state("bm_command", protocol.Header.size, expectBytes=self.payloadLength)
|
2017-03-20 18:32:26 +01:00
|
|
|
return True
|
|
|
|
|
|
|
|
def state_bm_command(self):
|
|
|
|
if len(self.read_buf) < self.payloadLength:
|
2017-05-24 16:51:49 +02:00
|
|
|
#print "Length below announced object length"
|
2017-03-11 11:12:08 +01:00
|
|
|
return False
|
2017-05-29 00:24:07 +02:00
|
|
|
#logger.debug("%s:%i: command %s (%ib)", self.destination.host, self.destination.port, self.command, self.payloadLength)
|
2017-03-20 18:32:26 +01:00
|
|
|
self.payload = self.read_buf[:self.payloadLength]
|
|
|
|
if self.checksum != hashlib.sha512(self.payload).digest()[0:4]:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("Bad checksum, ignoring")
|
2017-03-20 18:32:26 +01:00
|
|
|
self.invalid = True
|
2017-04-04 10:46:01 +02:00
|
|
|
retval = True
|
2017-05-29 00:47:41 +02:00
|
|
|
if not self.fullyEstablished and self.command not in ("error", "version", "verack"):
|
2017-05-24 16:51:49 +02:00
|
|
|
logger.error("Received command %s before connection was fully established, ignoring", self.command)
|
|
|
|
self.invalid = True
|
2017-03-20 18:32:26 +01:00
|
|
|
if not self.invalid:
|
|
|
|
try:
|
2017-04-04 10:46:01 +02:00
|
|
|
retval = getattr(self, "bm_command_" + str(self.command).lower())()
|
2017-03-20 18:32:26 +01:00
|
|
|
except AttributeError:
|
|
|
|
# unimplemented command
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("unimplemented command %s", self.command)
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMProtoInsufficientDataError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("packet length too short, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMProtoExcessiveDataError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("too much data, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMObjectInsufficientPOWError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("insufficient PoW, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMObjectInvalidDataError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("object invalid data, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMObjectExpiredError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("object expired, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMObjectUnwantedStreamError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("object not in wanted stream, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMObjectInvalidError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("object invalid, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except BMObjectAlreadyHaveError:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("already got object, skipping")
|
2017-05-24 16:51:49 +02:00
|
|
|
except struct.error:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("decoding error, skipping")
|
2017-03-11 11:12:08 +01:00
|
|
|
else:
|
2017-05-25 23:04:33 +02:00
|
|
|
#print "Skipping command %s due to invalid data" % (self.command)
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("Closing due to invalid command %s", self.command)
|
2017-05-25 23:04:33 +02:00
|
|
|
self.close()
|
|
|
|
return False
|
2017-04-04 10:46:01 +02:00
|
|
|
if retval:
|
|
|
|
self.set_state("bm_header", self.payloadLength)
|
|
|
|
self.bm_proto_reset()
|
|
|
|
# else assume the command requires a different state to follow
|
2017-03-20 18:32:26 +01:00
|
|
|
return True
|
2017-03-11 11:12:08 +01:00
|
|
|
|
2017-04-16 18:27:15 +02:00
|
|
|
def decode_payload_string(self, length):
|
|
|
|
value = self.payload[self.payloadOffset:self.payloadOffset+length]
|
|
|
|
self.payloadOffset += length
|
|
|
|
return value
|
|
|
|
|
|
|
|
def decode_payload_varint(self):
|
|
|
|
value, offset = addresses.decodeVarint(self.payload[self.payloadOffset:])
|
|
|
|
self.payloadOffset += offset
|
|
|
|
return value
|
|
|
|
|
|
|
|
def decode_payload_node(self):
|
2017-05-24 16:51:49 +02:00
|
|
|
services, host, port = self.decode_payload_content("Q16sH")
|
|
|
|
if host[0:12] == '\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\xFF\xFF':
|
|
|
|
host = socket.inet_ntop(socket.AF_INET, host[12:])
|
|
|
|
elif host[0:6] == '\xfd\x87\xd8\x7e\xeb\x43':
|
|
|
|
# Onion, based on BMD/bitcoind
|
|
|
|
host = base64.b32encode(host[6:]).lower() + ".onion"
|
|
|
|
else:
|
|
|
|
host = socket.inet_ntop(socket.AF_INET6, host)
|
|
|
|
if host == "":
|
|
|
|
# This can happen on Windows systems which are not 64-bit compatible
|
|
|
|
# so let us drop the IPv6 address.
|
|
|
|
host = socket.inet_ntop(socket.AF_INET, host[12:])
|
|
|
|
|
|
|
|
return Node(services, host, port)
|
2017-04-16 18:27:15 +02:00
|
|
|
|
|
|
|
def decode_payload_content(self, pattern = "v"):
|
2017-05-24 16:51:49 +02:00
|
|
|
# l = varint indicating the length of the next array
|
|
|
|
# L = varint indicating the length of the next item
|
2017-04-16 18:27:15 +02:00
|
|
|
# v = varint (or array)
|
|
|
|
# H = uint16
|
|
|
|
# I = uint32
|
|
|
|
# Q = uint64
|
|
|
|
# i = net_addr (without time and stream number)
|
|
|
|
# s = string
|
|
|
|
# 0-9 = length of the next item
|
|
|
|
# , = end of array
|
|
|
|
|
|
|
|
retval = []
|
2017-05-24 16:51:49 +02:00
|
|
|
size = None
|
2017-04-16 18:27:15 +02:00
|
|
|
insideDigit = False
|
2017-05-24 16:51:49 +02:00
|
|
|
i = 0
|
2017-04-16 18:27:15 +02:00
|
|
|
|
2017-05-24 16:51:49 +02:00
|
|
|
while i < len(pattern):
|
|
|
|
if pattern[i] in "0123456789" and (i == 0 or pattern[i-1] not in "lL"):
|
|
|
|
if size is None:
|
|
|
|
size = 0
|
2017-04-16 18:27:15 +02:00
|
|
|
size = size * 10 + int(pattern[i])
|
2017-05-24 16:51:49 +02:00
|
|
|
i += 1
|
2017-04-16 18:27:15 +02:00
|
|
|
continue
|
2017-05-24 16:51:49 +02:00
|
|
|
elif pattern[i] == "l" and size is None:
|
2017-04-16 18:27:15 +02:00
|
|
|
size = self.decode_payload_varint()
|
2017-05-24 16:51:49 +02:00
|
|
|
i += 1
|
2017-04-16 18:27:15 +02:00
|
|
|
continue
|
2017-05-24 16:51:49 +02:00
|
|
|
elif pattern[i] == "L" and size is None:
|
|
|
|
size = self.decode_payload_varint()
|
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
if size is not None:
|
2017-04-16 18:27:15 +02:00
|
|
|
if pattern[i] == "s":
|
|
|
|
retval.append(self.payload[self.payloadOffset:self.payloadOffset + size])
|
|
|
|
self.payloadOffset += size
|
2017-05-24 16:51:49 +02:00
|
|
|
i += 1
|
2017-04-16 18:27:15 +02:00
|
|
|
else:
|
2017-05-24 16:51:49 +02:00
|
|
|
if "," in pattern[i:]:
|
|
|
|
subpattern = pattern[i:pattern.index(",")]
|
|
|
|
else:
|
|
|
|
subpattern = pattern[i:]
|
|
|
|
|
2017-04-16 18:27:15 +02:00
|
|
|
for j in range(size):
|
2017-05-24 16:51:49 +02:00
|
|
|
if pattern[i-1:i] == "L":
|
|
|
|
retval.extend(self.decode_payload_content(subpattern))
|
2017-04-16 18:27:15 +02:00
|
|
|
else:
|
2017-05-24 16:51:49 +02:00
|
|
|
retval.append(self.decode_payload_content(subpattern))
|
|
|
|
i += len(subpattern)
|
|
|
|
size = None
|
2017-04-16 18:27:15 +02:00
|
|
|
else:
|
|
|
|
if pattern[i] == "v":
|
|
|
|
retval.append(self.decode_payload_varint())
|
|
|
|
if pattern[i] == "i":
|
|
|
|
retval.append(self.decode_payload_node())
|
|
|
|
if pattern[i] == "H":
|
2017-05-24 16:51:49 +02:00
|
|
|
retval.append(struct.unpack(">H", self.payload[self.payloadOffset:self.payloadOffset+2])[0])
|
2017-04-16 18:27:15 +02:00
|
|
|
self.payloadOffset += 2
|
|
|
|
if pattern[i] == "I":
|
2017-05-24 16:51:49 +02:00
|
|
|
retval.append(struct.unpack(">I", self.payload[self.payloadOffset:self.payloadOffset+4])[0])
|
2017-04-16 18:27:15 +02:00
|
|
|
self.payloadOffset += 4
|
|
|
|
if pattern[i] == "Q":
|
2017-05-24 16:51:49 +02:00
|
|
|
retval.append(struct.unpack(">Q", self.payload[self.payloadOffset:self.payloadOffset+8])[0])
|
2017-04-16 18:27:15 +02:00
|
|
|
self.payloadOffset += 8
|
2017-05-24 16:51:49 +02:00
|
|
|
i += 1
|
|
|
|
if self.payloadOffset > self.payloadLength:
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.debug("Insufficient data %i/%i", self.payloadOffset, self.payloadLength)
|
2017-05-24 16:51:49 +02:00
|
|
|
raise BMProtoInsufficientDataError()
|
2017-04-16 18:27:15 +02:00
|
|
|
return retval
|
|
|
|
|
2017-04-04 10:46:01 +02:00
|
|
|
def bm_command_error(self):
|
2017-04-16 18:27:15 +02:00
|
|
|
fatalStatus, banTime, inventoryVector, errorText = self.decode_payload_content("vvlsls")
|
2017-05-29 00:24:07 +02:00
|
|
|
logger.error("%s:%i error: %i, %s", self.destination.host, self.destination.port, fatalStatus, errorText)
|
2017-05-24 16:51:49 +02:00
|
|
|
return True
|
2017-04-16 18:27:15 +02:00
|
|
|
|
2017-04-04 10:46:01 +02:00
|
|
|
def bm_command_getdata(self):
|
2017-05-24 16:51:49 +02:00
|
|
|
items = self.decode_payload_content("L32s")
|
|
|
|
# if time.time() < self.skipUntil:
|
|
|
|
# print "skipping getdata"
|
|
|
|
# return True
|
2017-04-16 18:27:15 +02:00
|
|
|
for i in items:
|
2017-05-24 21:15:36 +02:00
|
|
|
#print "received getdata request for item %s" % (hexlify(i))
|
2017-05-24 16:51:49 +02:00
|
|
|
#logger.debug('received getdata request for item:' + hexlify(i))
|
|
|
|
#if i in ObjUploadQueue.streamElems(1):
|
|
|
|
if False:
|
2017-04-16 18:27:15 +02:00
|
|
|
self.antiIntersectionDelay()
|
|
|
|
else:
|
2017-05-25 23:04:33 +02:00
|
|
|
self.receiveQueue.put(("object", i))
|
2017-05-24 16:51:49 +02:00
|
|
|
return True
|
|
|
|
|
|
|
|
def bm_command_inv(self):
|
|
|
|
items = self.decode_payload_content("L32s")
|
|
|
|
|
2017-05-27 19:09:21 +02:00
|
|
|
if len(items) >= BMProto.maxObjectCount:
|
2017-05-24 16:51:49 +02:00
|
|
|
logger.error("Too many items in inv message!")
|
|
|
|
raise BMProtoExcessiveDataError()
|
|
|
|
else:
|
2017-05-24 21:15:36 +02:00
|
|
|
pass
|
2017-05-24 16:51:49 +02:00
|
|
|
|
|
|
|
for i in items:
|
2017-05-27 19:09:21 +02:00
|
|
|
self.receiveQueue.put(("inv", i))
|
|
|
|
self.handleReceivedInventory(i)
|
2017-05-24 16:51:49 +02:00
|
|
|
|
|
|
|
payload = addresses.encodeVarint(len(self.objectsNewToMe)) + ''.join(self.objectsNewToMe.keys())
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.CreatePacket('getdata', payload))
|
2017-05-24 16:51:49 +02:00
|
|
|
return True
|
2017-04-16 18:27:15 +02:00
|
|
|
|
2017-04-04 10:46:01 +02:00
|
|
|
def bm_command_object(self):
|
2017-05-24 16:51:49 +02:00
|
|
|
objectOffset = self.payloadOffset
|
|
|
|
nonce, expiresTime, objectType, version, streamNumber = self.decode_payload_content("QQIvv")
|
|
|
|
self.object = BMObject(nonce, expiresTime, objectType, version, streamNumber, self.payload)
|
|
|
|
|
2017-05-27 19:09:21 +02:00
|
|
|
if len(self.payload) - self.payloadOffset > BMProto.maxObjectPayloadSize:
|
2017-05-24 16:51:49 +02:00
|
|
|
logger.info('The payload length of this object is too large (%s bytes). Ignoring it.' % len(self.payload) - self.payloadOffset)
|
|
|
|
raise BMProtoExcessiveDataError()
|
|
|
|
|
|
|
|
self.object.checkProofOfWorkSufficient()
|
|
|
|
self.object.checkEOLSanity()
|
|
|
|
self.object.checkStream()
|
2017-05-30 23:53:43 +02:00
|
|
|
self.object.checkAlreadyHave()
|
|
|
|
|
|
|
|
if self.object.objectType == protocol.OBJECT_GETPUBKEY:
|
|
|
|
self.object.checkGetpubkey()
|
|
|
|
elif self.object.objectType == protocol.OBJECT_PUBKEY:
|
|
|
|
self.object.checkPubkey(self.payload[self.payloadOffset:self.payloadOffset+32])
|
|
|
|
elif self.object.objectType == protocol.OBJECT_MSG:
|
|
|
|
self.object.checkMessage()
|
|
|
|
elif self.object.objectType == protocol.OBJECT_BROADCAST:
|
|
|
|
self.object.checkBroadcast(self.payload[self.payloadOffset:self.payloadOffset+32])
|
|
|
|
# other objects don't require other types of tests
|
|
|
|
Inventory()[self.object.inventoryHash] = (
|
|
|
|
self.object.objectType, self.object.streamNumber, self.payload[objectOffset:], self.object.expiresTime, self.object.tag)
|
|
|
|
objectProcessorQueue.put((self.object.objectType,self.object.data))
|
|
|
|
#DownloadQueue().task_done(self.object.inventoryHash)
|
|
|
|
invQueue.put((self.object.streamNumber, self.object.inventoryHash, self))
|
|
|
|
#ObjUploadQueue().put(UploadElem(self.object.streamNumber, self.object.inventoryHash))
|
|
|
|
#broadcastToSendDataQueues((streamNumber, 'advertiseobject', inventoryHash))
|
2017-05-24 16:51:49 +02:00
|
|
|
return True
|
2017-04-16 18:27:15 +02:00
|
|
|
|
2017-05-27 19:09:21 +02:00
|
|
|
def _decode_addr(self):
|
|
|
|
return self.decode_payload_content("lQIQ16sH")
|
|
|
|
|
2017-04-16 18:27:15 +02:00
|
|
|
def bm_command_addr(self):
|
2017-05-27 19:09:21 +02:00
|
|
|
addresses = self._decode_addr()
|
2017-05-25 23:04:33 +02:00
|
|
|
for i in addresses:
|
|
|
|
seenTime, stream, services, ip, port = i
|
|
|
|
decodedIP = protocol.checkIPAddress(ip)
|
|
|
|
if stream not in state.streamsInWhichIAmParticipating:
|
|
|
|
continue
|
|
|
|
#print "maybe adding %s in stream %i to knownnodes (%i)" % (decodedIP, stream, len(knownnodes.knownNodes[stream]))
|
2017-05-27 19:09:21 +02:00
|
|
|
if decodedIP is not False and seenTime > time.time() - BMProto.addressAlive:
|
2017-05-25 23:04:33 +02:00
|
|
|
peer = state.Peer(decodedIP, port)
|
|
|
|
if peer in knownnodes.knownNodes[stream] and knownnodes.knownNodes[stream][peer] > seenTime:
|
|
|
|
continue
|
2017-05-29 00:24:07 +02:00
|
|
|
if len(knownnodes.knownNodes[stream]) < 20000:
|
|
|
|
with knownnodes.knownNodesLock:
|
|
|
|
knownnodes.knownNodes[stream][peer] = seenTime
|
|
|
|
#knownnodes.knownNodes[stream][peer] = seenTime
|
|
|
|
#AddrUploadQueue().put((stream, peer))
|
2017-05-24 16:51:49 +02:00
|
|
|
return True
|
2017-04-16 18:27:15 +02:00
|
|
|
|
2017-05-25 14:59:18 +02:00
|
|
|
def bm_command_portcheck(self):
|
|
|
|
portCheckerQueue.put(state.Peer(self.destination, self.peerNode.port))
|
|
|
|
return True
|
|
|
|
|
2017-04-04 10:46:01 +02:00
|
|
|
def bm_command_ping(self):
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.CreatePacket('pong'))
|
2017-05-24 16:51:49 +02:00
|
|
|
return True
|
2017-04-16 18:27:15 +02:00
|
|
|
|
2017-04-04 10:46:01 +02:00
|
|
|
def bm_command_pong(self):
|
2017-04-16 18:27:15 +02:00
|
|
|
# nothing really
|
2017-05-24 16:51:49 +02:00
|
|
|
return True
|
2017-04-04 10:46:01 +02:00
|
|
|
|
2017-03-20 18:32:26 +01:00
|
|
|
def bm_command_verack(self):
|
2017-03-11 11:12:08 +01:00
|
|
|
self.verackReceived = True
|
2017-04-16 18:27:15 +02:00
|
|
|
if self.verackSent:
|
|
|
|
if self.isSSL:
|
|
|
|
self.set_state("tls_init", self.payloadLength)
|
2017-05-24 16:51:49 +02:00
|
|
|
self.bm_proto_reset()
|
|
|
|
return False
|
2017-04-16 18:27:15 +02:00
|
|
|
else:
|
2017-05-24 16:51:49 +02:00
|
|
|
self.set_connection_fully_established()
|
|
|
|
return True
|
|
|
|
return True
|
2017-03-11 11:12:08 +01:00
|
|
|
|
2017-03-20 18:32:26 +01:00
|
|
|
def bm_command_version(self):
|
2017-04-16 18:27:15 +02:00
|
|
|
#self.remoteProtocolVersion, self.services, self.timestamp, padding1, self.myExternalIP, padding2, self.remoteNodeIncomingPort = protocol.VersionPacket.unpack(self.payload[:protocol.VersionPacket.size])
|
|
|
|
self.remoteProtocolVersion, self.services, self.timestamp, self.sockNode, self.peerNode, self.nonce, self.userAgent, self.streams = self.decode_payload_content("IQQiiQlslv")
|
2017-05-27 19:09:21 +02:00
|
|
|
self.nonce = struct.pack('>Q', self.nonce)
|
2017-04-16 18:27:15 +02:00
|
|
|
self.timeOffset = self.timestamp - int(time.time())
|
2017-05-27 22:30:30 +02:00
|
|
|
logger.debug("remoteProtocolVersion: %i", self.remoteProtocolVersion)
|
|
|
|
logger.debug("services: %08X", self.services)
|
|
|
|
logger.debug("time offset: %i", self.timestamp - int(time.time()))
|
|
|
|
logger.debug("my external IP: %s", self.sockNode.host)
|
|
|
|
logger.debug("remote node incoming port: %i", self.peerNode.port)
|
|
|
|
logger.debug("user agent: %s", self.userAgent)
|
|
|
|
logger.debug("streams: [%s]", ",".join(map(str,self.streams)))
|
2017-04-04 10:46:01 +02:00
|
|
|
if not self.peerValidityChecks():
|
|
|
|
# TODO ABORT
|
|
|
|
return True
|
2017-05-25 14:59:18 +02:00
|
|
|
#shared.connectedHostsList[self.destination] = self.streams[0]
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.CreatePacket('verack'))
|
2017-04-04 10:46:01 +02:00
|
|
|
self.verackSent = True
|
2017-05-24 16:51:49 +02:00
|
|
|
if not self.isOutbound:
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.assembleVersionMessage(self.destination.host, self.destination.port, network.connectionpool.BMConnectionPool().streams, True))
|
2017-05-29 00:24:07 +02:00
|
|
|
#print "%s:%i: Sending version" % (self.destination.host, self.destination.port)
|
2017-04-04 10:46:01 +02:00
|
|
|
if ((self.services & protocol.NODE_SSL == protocol.NODE_SSL) and
|
|
|
|
protocol.haveSSL(not self.isOutbound)):
|
|
|
|
self.isSSL = True
|
|
|
|
if self.verackReceived:
|
|
|
|
if self.isSSL:
|
|
|
|
self.set_state("tls_init", self.payloadLength)
|
2017-05-24 16:51:49 +02:00
|
|
|
self.bm_proto_reset()
|
|
|
|
return False
|
2017-04-04 10:46:01 +02:00
|
|
|
else:
|
2017-05-24 16:51:49 +02:00
|
|
|
self.set_connection_fully_established()
|
|
|
|
return True
|
|
|
|
return True
|
2017-04-04 10:46:01 +02:00
|
|
|
|
|
|
|
def peerValidityChecks(self):
|
|
|
|
if self.remoteProtocolVersion < 3:
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.assembleErrorMessage(fatal=2,
|
2017-04-16 18:27:15 +02:00
|
|
|
errorText="Your is using an old protocol. Closing connection."))
|
2017-04-04 10:46:01 +02:00
|
|
|
logger.debug ('Closing connection to old protocol version %s, node: %s',
|
2017-05-27 19:09:21 +02:00
|
|
|
str(self.remoteProtocolVersion), str(self.destination))
|
2017-04-04 10:46:01 +02:00
|
|
|
return False
|
2017-05-27 19:09:21 +02:00
|
|
|
if self.timeOffset > BMProto.maxTimeOffset:
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.assembleErrorMessage(fatal=2,
|
2017-04-16 18:27:15 +02:00
|
|
|
errorText="Your time is too far in the future compared to mine. Closing connection."))
|
2017-04-04 10:46:01 +02:00
|
|
|
logger.info("%s's time is too far in the future (%s seconds). Closing connection to it.",
|
2017-05-27 19:09:21 +02:00
|
|
|
self.destination, self.timeOffset)
|
2017-04-04 10:46:01 +02:00
|
|
|
shared.timeOffsetWrongCount += 1
|
|
|
|
return False
|
2017-05-27 19:09:21 +02:00
|
|
|
elif self.timeOffset < -BMProto.maxTimeOffset:
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.assembleErrorMessage(fatal=2,
|
2017-04-16 18:27:15 +02:00
|
|
|
errorText="Your time is too far in the past compared to mine. Closing connection."))
|
2017-04-04 10:46:01 +02:00
|
|
|
logger.info("%s's time is too far in the past (timeOffset %s seconds). Closing connection to it.",
|
2017-05-27 19:09:21 +02:00
|
|
|
self.destination, self.timeOffset)
|
2017-04-04 10:46:01 +02:00
|
|
|
shared.timeOffsetWrongCount += 1
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
shared.timeOffsetWrongCount = 0
|
|
|
|
if len(self.streams) == 0:
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.assembleErrorMessage(fatal=2,
|
2017-04-16 18:27:15 +02:00
|
|
|
errorText="We don't have shared stream interests. Closing connection."))
|
2017-04-04 10:46:01 +02:00
|
|
|
logger.debug ('Closed connection to %s because there is no overlapping interest in streams.',
|
2017-05-27 19:09:21 +02:00
|
|
|
str(self.destination))
|
2017-04-04 10:46:01 +02:00
|
|
|
return False
|
2017-05-24 16:51:49 +02:00
|
|
|
if self.destination in network.connectionpool.BMConnectionPool().inboundConnections:
|
|
|
|
try:
|
|
|
|
if not protocol.checkSocksIP(self.destination.host):
|
2017-05-25 23:04:33 +02:00
|
|
|
self.writeQueue.put(protocol.assembleErrorMessage(fatal=2,
|
2017-05-24 16:51:49 +02:00
|
|
|
errorText="Too many connections from your IP. Closing connection."))
|
|
|
|
logger.debug ('Closed connection to %s because we are already connected to that IP.',
|
2017-05-27 19:09:21 +02:00
|
|
|
str(self.destination))
|
2017-05-24 16:51:49 +02:00
|
|
|
return False
|
|
|
|
except:
|
|
|
|
pass
|
2017-05-27 19:09:21 +02:00
|
|
|
if self.nonce == protocol.eightBytesOfRandomDataUsedToDetectConnectionsToSelf:
|
|
|
|
self.writeQueue.put(protocol.assembleErrorMessage(fatal=2,
|
|
|
|
errorText="I'm connected to myself. Closing connection."))
|
|
|
|
logger.debug ("Closed connection to %s because I'm connected to myself.",
|
|
|
|
str(self.destination))
|
2017-05-29 14:41:02 +02:00
|
|
|
return False
|
2017-05-27 19:09:21 +02:00
|
|
|
|
2017-03-20 18:32:26 +01:00
|
|
|
return True
|
2017-03-11 11:12:08 +01:00
|
|
|
|
2017-05-27 19:09:21 +02:00
|
|
|
@staticmethod
|
|
|
|
def assembleAddr(peerList):
|
|
|
|
if type(peerList) is state.Peer:
|
|
|
|
peerList = (peerList)
|
|
|
|
# TODO handle max length, now it's done by upper layers
|
|
|
|
payload = addresses.encodeVarint(len(peerList))
|
|
|
|
for address in peerList:
|
|
|
|
stream, peer, timestamp = address
|
|
|
|
payload += struct.pack(
|
|
|
|
'>Q', timestamp) # 64-bit time
|
|
|
|
payload += struct.pack('>I', stream)
|
|
|
|
payload += struct.pack(
|
|
|
|
'>q', 1) # service bit flags offered by this node
|
|
|
|
payload += protocol.encodeHost(peer.host)
|
|
|
|
payload += struct.pack('>H', peer.port) # remote port
|
|
|
|
return protocol.CreatePacket('addr', payload)
|
2017-04-04 10:46:01 +02:00
|
|
|
|
2017-05-24 16:51:49 +02:00
|
|
|
def close(self, reason=None):
|
2017-05-25 23:04:33 +02:00
|
|
|
self.set_state("close")
|
2017-05-29 00:24:07 +02:00
|
|
|
if reason is None:
|
|
|
|
#logger.debug("%s:%i: closing, %s", self.destination.host, self.destination.port, ''.join(traceback.format_stack()))
|
|
|
|
logger.debug("%s:%i: closing", self.destination.host, self.destination.port)
|
|
|
|
#traceback.print_stack()
|
|
|
|
else:
|
|
|
|
logger.debug("%s:%i: closing, %s", self.destination.host, self.destination.port, reason)
|
2017-05-27 19:09:21 +02:00
|
|
|
AdvancedDispatcher.close(self)
|