733 lines
29 KiB
Python
Executable File
733 lines
29 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
'''
|
|
Onionr - P2P Microblogging Platform & Social network.
|
|
|
|
This file contains both the OnionrCommunicate class for communcating with peers
|
|
and code to operate as a daemon, getting commands from the command queue database (see core.Core.daemonQueue)
|
|
'''
|
|
'''
|
|
This program is free software: you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation, either version 3 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
'''
|
|
import sqlite3, requests, hmac, hashlib, time, sys, os, math, logger, urllib.parse, base64, binascii, random, json, threading
|
|
import core, onionrutils, onionrcrypto, netcontroller, onionrproofs, btc, config, onionrplugins as plugins
|
|
|
|
class OnionrCommunicate:
|
|
def __init__(self, debug, developmentMode):
|
|
'''
|
|
OnionrCommunicate
|
|
|
|
This class handles communication with nodes in the Onionr network.
|
|
'''
|
|
|
|
self._core = core.Core()
|
|
self._utils = onionrutils.OnionrUtils(self._core)
|
|
self._crypto = onionrcrypto.OnionrCrypto(self._core)
|
|
self._netController = netcontroller.NetController(0) # arg is the HS port but not needed rn in this file
|
|
|
|
self.newHashes = {} # use this to not keep hashes around too long if we cant get their data
|
|
self.keepNewHash = 12
|
|
self.ignoredHashes = []
|
|
|
|
self.highFailureAmount = 7
|
|
|
|
self.communicatorThreads = 0
|
|
self.maxThreads = 75
|
|
self.processBlocksThreads = 0
|
|
self.lookupBlocksThreads = 0
|
|
|
|
self.blocksProcessing = [] # list of blocks currently processing, to avoid trying a block twice at once in 2 seperate threads
|
|
self.peerStatus = {} # network actions (active requests) for peers used mainly to prevent conflicting actions in threads
|
|
'''
|
|
logger.info('Starting Bitcoin Node... with Tor socks port:' + str(sys.argv[2]), timestamp=True)
|
|
try:
|
|
self.bitcoin = btc.OnionrBTC(torP=int(sys.argv[2]))
|
|
except _gdbm.error:
|
|
pass
|
|
logger.info('Bitcoin Node started, on block: ' + self.bitcoin.node.getBlockHash(self.bitcoin.node.getLastBlockHeight()), timestamp=True)
|
|
'''
|
|
#except:
|
|
#logger.fatal('Failed to start Bitcoin Node, exiting...')
|
|
#exit(1)
|
|
|
|
blockProcessTimer = 0
|
|
blockProcessAmount = 20
|
|
highFailureTimer = 0
|
|
highFailureRate = 10
|
|
heartBeatTimer = 0
|
|
heartBeatRate = 10
|
|
pexTimer = 120 # How often we should check for new peers
|
|
pexCount = 0
|
|
logger.debug('Communicator debugging enabled.')
|
|
torID = open('data/hs/hostname').read()
|
|
|
|
apiRunningCheckRate = 10
|
|
apiRunningCheckCount = 0
|
|
|
|
self.peerData = {} # Session data for peers (recent reachability, speed, etc)
|
|
|
|
if os.path.exists(self._core.queueDB):
|
|
self._core.clearDaemonQueue()
|
|
|
|
# Loads in and starts the enabled plugins
|
|
plugins.reload()
|
|
|
|
while True:
|
|
command = self._core.daemonQueue()
|
|
# Process blocks based on a timer
|
|
blockProcessTimer += 1
|
|
heartBeatTimer += 1
|
|
pexCount += 1
|
|
if highFailureTimer == highFailureRate:
|
|
highFailureTimer = 0
|
|
for i in self.peerData:
|
|
if self.peerData[i]['failCount'] >= self.highFailureAmount:
|
|
self.peerData[i]['failCount'] -= 1
|
|
if pexTimer == pexCount:
|
|
pT1 = threading.Thread(target=self.getNewPeers, name="pT1")
|
|
pT1.start()
|
|
pT2 = threading.Thread(target=self.getNewPeers, name="pT2")
|
|
pT2.start()
|
|
pexCount = 0 # TODO: do not reset timer if low peer count
|
|
if heartBeatRate == heartBeatTimer:
|
|
logger.debug('Communicator heartbeat')
|
|
heartBeatTimer = 0
|
|
if blockProcessTimer == blockProcessAmount:
|
|
lT1 = threading.Thread(target=self.lookupBlocks, name="lt1", args=(isThread=True,))
|
|
lT2 = threading.Thread(target=self.lookupBlocks, name="lt2", args=(isThread=True,))
|
|
lT3 = threading.Thread(target=self.lookupBlocks, name="lt3", args=(isThread=True,))
|
|
lT4 = threading.Thread(target=self.lookupBlocks, name="lt4", args=(isThread=True,))
|
|
pbT1 = threading.Thread(target=self.processBlocks, name='pbT1', args=(isThread=True,))
|
|
pbT2 = threading.Thread(target=self.processBlocks, name='pbT2', args=(isThread=True,))
|
|
pbT3 = threading.Thread(target=self.processBlocks, name='pbT3', args=(isThread=True,))
|
|
pbT4 = threading.Thread(target=self.processBlocks, name='pbT4', args=(isThread=True,))
|
|
if (self.maxThreads - 8) >= threading.active_count():
|
|
lT1.start()
|
|
lT2.start()
|
|
lT3.start()
|
|
lT4.start()
|
|
pbT1.start()
|
|
pbT2.start()
|
|
pbT3.start()
|
|
pbT4.start()
|
|
blockProcessTimer = 0
|
|
else:
|
|
logger.debug(threading.active_count())
|
|
logger.debug('Too many threads.')
|
|
if command != False:
|
|
if command[0] == 'shutdown':
|
|
logger.info('Daemon received exit command.', timestamp=True)
|
|
break
|
|
elif command[0] == 'announceNode':
|
|
announceAttempts = 3
|
|
announceAttemptCount = 0
|
|
announceVal = False
|
|
logger.info('Announcing node to ' + command[1], timestamp=True)
|
|
while not announceVal:
|
|
announceAttemptCount += 1
|
|
announceVal = self.performGet('announce', command[1], data=self._core.hsAdder.replace('\n', ''), skipHighFailureAddress=True)
|
|
logger.info(announceVal)
|
|
if announceAttemptCount >= announceAttempts:
|
|
logger.warn('Unable to announce to ' + command[1])
|
|
break
|
|
elif command[0] == 'runCheck':
|
|
logger.info('Status check; looks good.')
|
|
open('data/.runcheck', 'w+').close()
|
|
elif command[0] == 'event':
|
|
# todo
|
|
pass
|
|
elif command[0] == 'checkCallbacks':
|
|
try:
|
|
data = json.loads(command[1])
|
|
|
|
logger.info('Checking for callbacks with connection %s...' % data['id'])
|
|
|
|
self.check_callbacks(data, config.get('dc_execcallbacks', True))
|
|
|
|
events.event('incoming_direct_connection', data = {'callback' : True, 'communicator' : self, 'data' : data})
|
|
except Exception as e:
|
|
logger.error('Failed to interpret callbacks for checking', e)
|
|
elif command[0] == 'incomingDirectConnection':
|
|
try:
|
|
data = json.loads(command[1])
|
|
|
|
logger.info('Handling incoming connection %s...' % data['id'])
|
|
|
|
self.incoming_direct_connection(data)
|
|
|
|
events.event('incoming_direct_connection', data = {'callback' : False, 'communicator' : self, 'data' : data})
|
|
except Exception as e:
|
|
logger.error('Failed to handle callbacks for checking', e)
|
|
|
|
apiRunningCheckCount += 1
|
|
|
|
# check if local API is up
|
|
if apiRunningCheckCount > apiRunningCheckRate:
|
|
if self._core._utils.localCommand('ping') != 'pong':
|
|
for i in range(4):
|
|
if self._utils.localCommand('ping') == 'pong':
|
|
apiRunningCheckCount = 0
|
|
break # break for loop
|
|
time.sleep(1)
|
|
else:
|
|
# This executes if the api is NOT detected to be running
|
|
logger.error('Daemon detected API crash (or otherwise unable to reach API after long time), stopping...')
|
|
break # break main daemon loop
|
|
apiRunningCheckCount = 0
|
|
|
|
time.sleep(1)
|
|
|
|
self._netController.killTor()
|
|
return
|
|
|
|
future_callbacks = {}
|
|
connection_handlers = {}
|
|
id_peer_cache = {}
|
|
|
|
def get_connection_handlers(self, name = None):
|
|
'''
|
|
Returns a list of callback handlers by name, or, if name is None, it returns all handlers.
|
|
'''
|
|
|
|
if name is None:
|
|
return self.connection_handlers
|
|
elif name in self.connection_handlers:
|
|
return self.connection_handlers[name]
|
|
else:
|
|
return list()
|
|
|
|
def add_connection_handler(self, name, handler):
|
|
'''
|
|
Adds a function to be called when an connection that is NOT a callback is received.
|
|
Takes in the name of the communication type and the handler as input
|
|
'''
|
|
|
|
if not name in self.connection_handlers:
|
|
self.connection_handlers[name] = list()
|
|
|
|
self.connection_handlers[name].append(handler)
|
|
|
|
return
|
|
|
|
def remove_connection_handler(self, name, handler = None):
|
|
'''
|
|
Removes a connection handler if specified, or removes all by name
|
|
'''
|
|
|
|
if handler is None:
|
|
if name in self.connection_handlers:
|
|
self.connection_handlers[name].remove(handler)
|
|
elif name in self.connection_handlers:
|
|
del self.connection_handlers[name]
|
|
|
|
return
|
|
|
|
|
|
def set_callback(self, identifier, callback):
|
|
'''
|
|
(Over)writes a callback by communication identifier
|
|
'''
|
|
|
|
if not callback is None:
|
|
self.future_callbacks[identifier] = callback
|
|
return True
|
|
|
|
return False
|
|
|
|
def unset_callback(self, identifier):
|
|
'''
|
|
Unsets a callback by communication identifier, if set
|
|
'''
|
|
|
|
if identifier in future_callbacks:
|
|
del self.future_callbacks[identifier]
|
|
return True
|
|
|
|
return False
|
|
|
|
def get_callback(self, identifier):
|
|
'''
|
|
Returns a callback by communication identifier if set, or None
|
|
'''
|
|
|
|
if identifier in self.future_callbacks:
|
|
return self.future_callbacks[id]
|
|
|
|
return None
|
|
|
|
def direct_connect(self, peer, data = None, callback = None, log = True):
|
|
'''
|
|
Communicates something directly with the client
|
|
|
|
- `peer` should obviously be the peer id to request.
|
|
- `data` should be a dict (NOT str), with the parameter "type"
|
|
ex. {'type': 'sendMessage', 'content': 'hey, this is a dm'}
|
|
In that dict, the key 'token' must NEVER be set. If it is, it will
|
|
be overwritten.
|
|
- if `callback` is set to a function, it will call that function
|
|
back if/when the client the request is sent to decides to respond.
|
|
Do NOT depend on a response, because users can configure their
|
|
clients not to respond to this type of request.
|
|
- `log` is set to True by default-- what this does is log the
|
|
request for debug purposes. Should be False for sensitive actions.
|
|
'''
|
|
|
|
# TODO: Timing attack prevention
|
|
try:
|
|
# does not need to be secure random, only used for keeping track of async responses
|
|
# Actually, on second thought, it does need to be secure random. Otherwise, if it is predictable, someone could trigger arbitrary callbacks that have been saved on the local node, wrecking all kinds of havoc. Better just to keep it secure random.
|
|
identifier = self._utils.token(32)
|
|
if 'id' in data:
|
|
identifier = data['id']
|
|
|
|
if not identifier in id_peer_cache:
|
|
id_peer_cache[identifier] = peer
|
|
|
|
if type(data) == str:
|
|
# if someone inputs a string instead of a dict, it will assume it's the type
|
|
data = {'type' : data}
|
|
|
|
data['id'] = identifier
|
|
data['token'] = '' # later put PoW stuff here or whatever is needed
|
|
data_str = json.dumps(data)
|
|
|
|
events.event('outgoing_direct_connection', data = {'callback' : True, 'communicator' : self, 'data' : data, 'id' : identifier, 'token' : token, 'peer' : peer, 'callback' : callback, 'log' : log})
|
|
|
|
logger.debug('Direct connection (identifier: "%s"): %s' + (identifier, data_str))
|
|
try:
|
|
self.performGet('directMessage', peer, data_str)
|
|
except:
|
|
logger.warn('Failed to connect to peer: "%s".' % str(peer))
|
|
return False
|
|
|
|
if not callback is None:
|
|
self.set_callback(identifier, callback)
|
|
|
|
return True
|
|
except Exception as e:
|
|
logger.warn('Unknown error, failed to execute direct connect (peer: "%s").' % str(peer), e)
|
|
|
|
return False
|
|
|
|
def direct_connect_response(self, identifier, data, peer = None, callback = None, log = True):
|
|
'''
|
|
Responds to a previous connection. Hostname will be pulled from id_peer_cache if not specified in `peer` parameter.
|
|
|
|
If yet another callback is requested, it can be put in the `callback` parameter.
|
|
'''
|
|
|
|
if config.get('dc_response', True):
|
|
data['id'] = identifier
|
|
data['sender'] = open('data/hs/hostname').read()
|
|
data['callback'] = True
|
|
|
|
if (origin is None) and (identifier in id_peer_cache):
|
|
origin = id_peer_cache[identifier]
|
|
|
|
if not identifier in id_peer_cache:
|
|
id_peer_cache[identifier] = peer
|
|
|
|
if origin is None:
|
|
logger.warn('Failed to identify peer for connection %s' % str(identifier))
|
|
return False
|
|
else:
|
|
return self.direct_connect(peer, data = data, callback = callback, log = log)
|
|
else:
|
|
logger.warn('Node tried to respond to direct connection id %s, but it was rejected due to `dc_response` restriction.' % str(identifier))
|
|
return False
|
|
|
|
|
|
def check_callbacks(self, data, execute = True, remove = True):
|
|
'''
|
|
Check if a callback is set, and if so, execute it
|
|
'''
|
|
|
|
try:
|
|
if type(data) is str:
|
|
data = json.loads(data)
|
|
|
|
if 'id' in data: # TODO: prevent enumeration, require extra PoW
|
|
identifier = data['id']
|
|
|
|
if identifier in self.future_callbacks:
|
|
if execute:
|
|
self.get_callback(identifier)(data)
|
|
logger.debug('Request callback "%s" executed.' % str(identifier))
|
|
if remove:
|
|
self.unset_callback(identifier)
|
|
|
|
return True
|
|
|
|
logger.warn('Unable to find request callback for ID "%s".' % str(identifier))
|
|
else:
|
|
logger.warn('Unable to identify callback request, `id` parameter missing: %s' % json.dumps(data))
|
|
except Exception as e:
|
|
logger.warn('Unknown error, failed to execute direct connection callback (peer: "%s").' % str(peer), e)
|
|
|
|
return False
|
|
|
|
def incoming_direct_connection(self, data):
|
|
'''
|
|
This code is run whenever there is a new incoming connection.
|
|
'''
|
|
|
|
if 'type' in data and data['type'] in self.connection_handlers:
|
|
for handler in self.get_connection_handlers(name):
|
|
handler(data)
|
|
|
|
return
|
|
|
|
def getNewPeers(self):
|
|
'''
|
|
Get new peers and ed25519 keys
|
|
'''
|
|
|
|
peersCheck = 1 # Amount of peers to ask for new peers + keys
|
|
peersChecked = 0
|
|
peerList = list(self._core.listAdders()) # random ordered list of peers
|
|
newKeys = []
|
|
newAdders = []
|
|
if len(peerList) > 0:
|
|
maxN = len(peerList) - 1
|
|
else:
|
|
peersCheck = 0
|
|
maxN = 0
|
|
|
|
if len(peerList) > peersCheck:
|
|
peersCheck = len(peerList)
|
|
|
|
while peersCheck > peersChecked:
|
|
#i = secrets.randbelow(maxN) # cant use prior to 3.6
|
|
i = random.randint(0, maxN)
|
|
|
|
try:
|
|
if self.peerStatusTaken(peerList[i], 'pex') or self.peerStatusTaken(peerList[i], 'kex'):
|
|
continue
|
|
except IndexError:
|
|
pass
|
|
|
|
logger.info('Using ' + peerList[i] + ' to find new peers', timestamp=True)
|
|
|
|
try:
|
|
newAdders = self.performGet('pex', peerList[i], skipHighFailureAddress=True)
|
|
logger.debug('Attempting to merge address: ')
|
|
logger.debug(newAdders)
|
|
self._utils.mergeAdders(newAdders)
|
|
except requests.exceptions.ConnectionError:
|
|
logger.info(peerList[i] + ' connection failed', timestamp=True)
|
|
continue
|
|
else:
|
|
try:
|
|
logger.info('Using ' + peerList[i] + ' to find new keys')
|
|
newKeys = self.performGet('kex', peerList[i], skipHighFailureAddress=True)
|
|
logger.debug('Attempting to merge pubkey: ')
|
|
logger.debug(newKeys)
|
|
# TODO: Require keys to come with POW token (very large amount of POW)
|
|
self._utils.mergeKeys(newKeys)
|
|
except requests.exceptions.ConnectionError:
|
|
logger.info(peerList[i] + ' connection failed', timestamp=True)
|
|
continue
|
|
else:
|
|
peersChecked += 1
|
|
return
|
|
|
|
def lookupBlocks(self, isThread=False):
|
|
'''
|
|
Lookup blocks and merge new ones
|
|
'''
|
|
if isThread:
|
|
self.lookupBlocksThreads += 1
|
|
peerList = self._core.listAdders()
|
|
blocks = ''
|
|
|
|
for i in peerList:
|
|
if self.peerStatusTaken(i, 'getBlockHashes') or self.peerStatusTaken(i, 'getDBHash'):
|
|
continue
|
|
try:
|
|
if self.peerData[i]['failCount'] >= self.highFailureAmount:
|
|
continue
|
|
except KeyError:
|
|
pass
|
|
|
|
lastDB = self._core.getAddressInfo(i, 'DBHash')
|
|
|
|
if lastDB == None:
|
|
logger.debug('Fetching hash from ' + str(i) + ', no previous known.')
|
|
else:
|
|
logger.debug('Fetching hash from ' + str(i) + ', ' + str(lastDB) + ' last known')
|
|
|
|
currentDB = self.performGet('getDBHash', i)
|
|
|
|
if currentDB != False:
|
|
logger.debug(i + " hash db (from request): " + currentDB)
|
|
else:
|
|
logger.warn("Error getting hash db status for " + i)
|
|
|
|
if currentDB != False:
|
|
if lastDB != currentDB:
|
|
logger.debug('Fetching hash from ' + i + ' - ' + currentDB + ' current hash.')
|
|
try:
|
|
blocks += self.performGet('getBlockHashes', i)
|
|
except TypeError:
|
|
logger.warn('Failed to get data hash from ' + i)
|
|
self.peerData[i]['failCount'] -= 1
|
|
if self._utils.validateHash(currentDB):
|
|
self._core.setAddressInfo(i, "DBHash", currentDB)
|
|
|
|
if len(blocks.strip()) != 0:
|
|
pass
|
|
#logger.debug('BLOCKS:' + blocks)
|
|
|
|
blockList = blocks.split('\n')
|
|
|
|
for i in blockList:
|
|
if len(i.strip()) == 0:
|
|
continue
|
|
if self._utils.hasBlock(i):
|
|
continue
|
|
if i in self.ignoredHashes:
|
|
continue
|
|
|
|
#logger.debug('Exchanged block (blockList): ' + i)
|
|
if not self._utils.validateHash(i):
|
|
# skip hash if it isn't valid
|
|
logger.warn('Hash ' + i + ' is not valid')
|
|
continue
|
|
else:
|
|
self.newHashes[i] = 0
|
|
logger.debug('Adding ' + i + ' to hash database...')
|
|
self._core.addToBlockDB(i)
|
|
self.lookupBlocksThreads -= 1
|
|
return
|
|
|
|
def processBlocks(self, isThread=False):
|
|
'''
|
|
Work with the block database and download any missing blocks
|
|
|
|
This is meant to be called from the communicator daemon on its timer.
|
|
'''
|
|
if isThread:
|
|
self.processBlocksThreads += 1
|
|
for i in self._core.getBlockList(unsaved=True).split("\n"):
|
|
if i != "":
|
|
if i in self.blocksProcessing or i in self.ignoredHashes:
|
|
#logger.debug('already processing ' + i)
|
|
continue
|
|
else:
|
|
self.blocksProcessing.append(i)
|
|
try:
|
|
self.newHashes[i]
|
|
except KeyError:
|
|
self.newHashes[i] = 0
|
|
|
|
# check if a new hash has been around too long, delete it from database and add it to ignore list
|
|
if self.newHashes[i] >= self.keepNewHash:
|
|
logger.warn('Ignoring block ' + i + ' because it took to long to get valid data.')
|
|
del self.newHashes[i]
|
|
self._core.removeBlock(i)
|
|
self.ignoredHashes.append(i)
|
|
continue
|
|
|
|
self.newHashes[i] += 1
|
|
logger.warn('UNSAVED BLOCK: ' + i)
|
|
data = self.downloadBlock(i)
|
|
|
|
# if block was successfull gotten (hash already verified)
|
|
if data:
|
|
del self.newHashes[i] # remove from probation list
|
|
|
|
# deal with block metadata
|
|
blockContent = self._core.getData(i)
|
|
try:
|
|
#blockMetadata = json.loads(self._core.getData(i)).split('}')[0] + '}'
|
|
blockMetadata = self._core.getData(i).split(b'}')[0]
|
|
try:
|
|
blockMetadata = blockMetadata.decode()
|
|
except AttributeError:
|
|
pass
|
|
|
|
blockMetadata = json.loads(blockMetadata + '}')
|
|
|
|
try:
|
|
blockMetadata['sig']
|
|
blockMetadata['id']
|
|
except KeyError:
|
|
pass
|
|
|
|
else:
|
|
creator = self._utils.getPeerByHashId(blockMetadata['id'])
|
|
try:
|
|
creator = creator.decode()
|
|
except AttributeError:
|
|
pass
|
|
|
|
if self._core._crypto.edVerify(blockContent.split(b'}')[1], creator, blockMetadata['sig'], encodedData=True):
|
|
self._core.updateBlockInfo(i, 'sig', 'true')
|
|
else:
|
|
self._core.updateBlockInfo(i, 'sig', 'false')
|
|
try:
|
|
logger.info('Block type is ' + blockMetadata['type'])
|
|
self._core.updateBlockInfo(i, 'dataType', blockMetadata['type'])
|
|
self.removeBlockFromProcessingList(i)
|
|
self.removeBlockFromProcessingList(i)
|
|
except KeyError:
|
|
logger.warn('Block has no type')
|
|
pass
|
|
except json.decoder.JSONDecodeError:
|
|
logger.warn('Could not decode block metadata')
|
|
self.removeBlockFromProcessingList(i)
|
|
self.processBlocksThreads -= 1
|
|
return
|
|
|
|
def removeBlockFromProcessingList(self, block):
|
|
try:
|
|
self.blocksProcessing.remove(block)
|
|
except ValueError:
|
|
return False
|
|
else:
|
|
return True
|
|
|
|
def downloadBlock(self, hash, peerTries=3):
|
|
'''
|
|
Download a block from random order of peers
|
|
'''
|
|
|
|
retVal = False
|
|
peerList = self._core.listAdders()
|
|
blocks = ''
|
|
peerTryCount = 0
|
|
|
|
for i in peerList:
|
|
try:
|
|
if self.peerData[i]['failCount'] >= self.highFailureAmount:
|
|
continue
|
|
except KeyError:
|
|
pass
|
|
if peerTryCount >= peerTries:
|
|
break
|
|
|
|
hasher = hashlib.sha3_256()
|
|
data = self.performGet('getData', i, hash, skipHighFailureAddress=True)
|
|
|
|
if data == False or len(data) > 10000000 or data == '':
|
|
peerTryCount += 1
|
|
continue
|
|
|
|
try:
|
|
data = base64.b64decode(data)
|
|
except binascii.Error:
|
|
data = b''
|
|
|
|
hasher.update(data)
|
|
digest = hasher.hexdigest()
|
|
|
|
if type(digest) is bytes:
|
|
digest = digest.decode()
|
|
|
|
if digest == hash.strip():
|
|
self._core.setData(data)
|
|
logger.info('Successfully obtained data for ' + hash, timestamp=True)
|
|
retVal = True
|
|
break
|
|
'''
|
|
if data.startswith(b'-txt-'):
|
|
self._core.setBlockType(hash, 'txt')
|
|
if len(data) < 120:
|
|
logger.debug('Block text:\n' + data.decode())
|
|
'''
|
|
else:
|
|
logger.warn("Failed to validate " + hash + " " + " hash calculated was " + digest)
|
|
peerTryCount += 1
|
|
|
|
return retVal
|
|
|
|
def urlencode(self, data):
|
|
'''
|
|
URL encodes the data
|
|
'''
|
|
return urllib.parse.quote_plus(data)
|
|
|
|
def performGet(self, action, peer, data=None, skipHighFailureAddress=False, peerType='tor', selfCheck=True):
|
|
'''
|
|
Performs a request to a peer through Tor or i2p (currently only Tor)
|
|
'''
|
|
|
|
if not peer.endswith('.onion') and not peer.endswith('.onion/'):
|
|
raise PeerError('Currently only Tor .onion peers are supported. You must manually specify .onion')
|
|
|
|
if len(self._core.hsAdder.strip()) == 0:
|
|
raise Exception("Could not perform self address check in performGet due to not knowing our address")
|
|
if selfCheck:
|
|
if peer.replace('/', '') == self._core.hsAdder:
|
|
logger.warn('Tried to performget to own hidden service, but selfCheck was not set to false')
|
|
return
|
|
|
|
# Store peer in peerData dictionary (non permanent)
|
|
if not peer in self.peerData:
|
|
self.peerData[peer] = {'connectCount': 0, 'failCount': 0, 'lastConnectTime': math.floor(time.time())}
|
|
socksPort = sys.argv[2]
|
|
'''We use socks5h to use tor as DNS'''
|
|
proxies = {'http': 'socks5://127.0.0.1:' + str(socksPort), 'https': 'socks5://127.0.0.1:' + str(socksPort)}
|
|
headers = {'user-agent': 'PyOnionr'}
|
|
url = 'http://' + peer + '/public/?action=' + self.urlencode(action)
|
|
if data != None:
|
|
url = url + '&data=' + self.urlencode(data)
|
|
try:
|
|
if skipHighFailureAddress and self.peerData[peer]['failCount'] > self.highFailureAmount:
|
|
retData = False
|
|
logger.debug('Skipping ' + peer + ' because of high failure rate')
|
|
else:
|
|
self.peerStatus[peer] = action
|
|
logger.debug('Contacting ' + peer + ' on port ' + socksPort)
|
|
r = requests.get(url, headers=headers, proxies=proxies, timeout=(15, 30))
|
|
retData = r.text
|
|
except requests.exceptions.RequestException as e:
|
|
logger.warn(action + " failed with peer " + peer + ": " + str(e))
|
|
retData = False
|
|
|
|
if not retData:
|
|
self.peerData[peer]['failCount'] += 1
|
|
else:
|
|
self.peerData[peer]['connectCount'] += 1
|
|
self.peerData[peer]['failCount'] -= 1
|
|
self.peerData[peer]['lastConnectTime'] = math.floor(time.time())
|
|
return retData
|
|
|
|
def peerStatusTaken(self, peer, status):
|
|
'''
|
|
Returns if a peer is currently performing a specified action
|
|
'''
|
|
try:
|
|
if self.peerStatus[peer] == status:
|
|
return True
|
|
except KeyError:
|
|
pass
|
|
return False
|
|
|
|
shouldRun = False
|
|
debug = True
|
|
developmentMode = False
|
|
if config.get('devmode', True):
|
|
developmentMode = True
|
|
try:
|
|
if sys.argv[1] == 'run':
|
|
shouldRun = True
|
|
except IndexError:
|
|
pass
|
|
if shouldRun:
|
|
try:
|
|
OnionrCommunicate(debug, developmentMode)
|
|
except KeyboardInterrupt:
|
|
sys.exit(1)
|
|
pass
|