merge crypto
This commit is contained in:
commit
56ca0abc63
@ -198,6 +198,8 @@ class API:
|
|||||||
resp = Response(self._utils.getBlockDBHash())
|
resp = Response(self._utils.getBlockDBHash())
|
||||||
elif action == 'getBlockHashes':
|
elif action == 'getBlockHashes':
|
||||||
resp = Response(self._core.getBlockList())
|
resp = Response(self._core.getBlockList())
|
||||||
|
elif action == 'directMessage':
|
||||||
|
resp = Response(self._core.handle_direct_connection(data))
|
||||||
elif action == 'announce':
|
elif action == 'announce':
|
||||||
if data != '':
|
if data != '':
|
||||||
# TODO: require POW for this
|
# TODO: require POW for this
|
||||||
|
@ -119,13 +119,13 @@ class OnionrCommunicate:
|
|||||||
blockProcessTimer = 0
|
blockProcessTimer = 0
|
||||||
if command != False:
|
if command != False:
|
||||||
if command[0] == 'shutdown':
|
if command[0] == 'shutdown':
|
||||||
logger.info('Daemon recieved exit command.', timestamp=True)
|
logger.info('Daemon received exit command.', timestamp=True)
|
||||||
break
|
break
|
||||||
elif command[0] == 'announceNode':
|
elif command[0] == 'announceNode':
|
||||||
announceAttempts = 3
|
announceAttempts = 3
|
||||||
announceAttemptCount = 0
|
announceAttemptCount = 0
|
||||||
announceVal = False
|
announceVal = False
|
||||||
logger.info('Announcing our node to ' + command[1], timestamp=True)
|
logger.info('Announcing node to ' + command[1], timestamp=True)
|
||||||
while not announceVal:
|
while not announceVal:
|
||||||
announceAttemptCount += 1
|
announceAttemptCount += 1
|
||||||
announceVal = self.performGet('announce', command[1], data=self._core.hsAdder.replace('\n', ''), skipHighFailureAddress=True)
|
announceVal = self.performGet('announce', command[1], data=self._core.hsAdder.replace('\n', ''), skipHighFailureAddress=True)
|
||||||
@ -136,6 +136,31 @@ class OnionrCommunicate:
|
|||||||
elif command[0] == 'runCheck':
|
elif command[0] == 'runCheck':
|
||||||
logger.info('Status check; looks good.')
|
logger.info('Status check; looks good.')
|
||||||
open('data/.runcheck', 'w+').close()
|
open('data/.runcheck', 'w+').close()
|
||||||
|
elif command[0] == 'event':
|
||||||
|
# todo
|
||||||
|
pass
|
||||||
|
elif command[0] == 'checkCallbacks':
|
||||||
|
try:
|
||||||
|
data = json.loads(command[1])
|
||||||
|
|
||||||
|
logger.info('Checking for callbacks with connection %s...' % data['id'])
|
||||||
|
|
||||||
|
self.check_callbacks(data, config.get('dc_execcallbacks', True))
|
||||||
|
|
||||||
|
events.event('incoming_direct_connection', data = {'callback' : True, 'communicator' : self, 'data' : data})
|
||||||
|
except Exception as e:
|
||||||
|
logger.error('Failed to interpret callbacks for checking', e)
|
||||||
|
elif command[0] == 'incomingDirectConnection':
|
||||||
|
try:
|
||||||
|
data = json.loads(command[1])
|
||||||
|
|
||||||
|
logger.info('Handling incoming connection %s...' % data['id'])
|
||||||
|
|
||||||
|
self.incoming_direct_connection(data)
|
||||||
|
|
||||||
|
events.event('incoming_direct_connection', data = {'callback' : False, 'communicator' : self, 'data' : data})
|
||||||
|
except Exception as e:
|
||||||
|
logger.error('Failed to handle callbacks for checking', e)
|
||||||
|
|
||||||
apiRunningCheckCount += 1
|
apiRunningCheckCount += 1
|
||||||
|
|
||||||
@ -149,7 +174,7 @@ class OnionrCommunicate:
|
|||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
else:
|
else:
|
||||||
# This executes if the api is NOT detected to be running
|
# This executes if the api is NOT detected to be running
|
||||||
logger.error('Daemon detected API crash (or otherwise unable to reach API after long time, stopping)')
|
logger.error('Daemon detected API crash (or otherwise unable to reach API after long time), stopping...')
|
||||||
break # break main daemon loop
|
break # break main daemon loop
|
||||||
apiRunningCheckCount = 0
|
apiRunningCheckCount = 0
|
||||||
|
|
||||||
@ -158,10 +183,208 @@ class OnionrCommunicate:
|
|||||||
self._netController.killTor()
|
self._netController.killTor()
|
||||||
return
|
return
|
||||||
|
|
||||||
|
future_callbacks = {}
|
||||||
|
connection_handlers = {}
|
||||||
|
id_peer_cache = {}
|
||||||
|
|
||||||
|
def get_connection_handlers(self, name = None):
|
||||||
|
'''
|
||||||
|
Returns a list of callback handlers by name, or, if name is None, it returns all handlers.
|
||||||
|
'''
|
||||||
|
|
||||||
|
if name is None:
|
||||||
|
return self.connection_handlers
|
||||||
|
elif name in self.connection_handlers:
|
||||||
|
return self.connection_handlers[name]
|
||||||
|
else
|
||||||
|
return list()
|
||||||
|
|
||||||
|
def add_connection_handler(self, name, handler):
|
||||||
|
'''
|
||||||
|
Adds a function to be called when an connection that is NOT a callback is received.
|
||||||
|
Takes in the name of the communication type and the handler as input
|
||||||
|
'''
|
||||||
|
|
||||||
|
if not name in self.connection_handlers:
|
||||||
|
self.connection_handlers[name] = list()
|
||||||
|
|
||||||
|
self.connection_handlers[name].append(handler)
|
||||||
|
|
||||||
|
return
|
||||||
|
|
||||||
|
def remove_connection_handler(self, name, handler = None):
|
||||||
|
'''
|
||||||
|
Removes a connection handler if specified, or removes all by name
|
||||||
|
'''
|
||||||
|
|
||||||
|
if handler is None:
|
||||||
|
if name in self.connection_handlers:
|
||||||
|
self.connection_handlers[name].remove(handler)
|
||||||
|
elif name in self.connection_handlers:
|
||||||
|
del self.connection_handlers[name]
|
||||||
|
|
||||||
|
return
|
||||||
|
|
||||||
|
|
||||||
|
def set_callback(self, identifier, callback):
|
||||||
|
'''
|
||||||
|
(Over)writes a callback by communication identifier
|
||||||
|
'''
|
||||||
|
|
||||||
|
if not callback is None:
|
||||||
|
self.future_callbacks[identifier] = callback
|
||||||
|
return True
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
def unset_callback(self, identifier):
|
||||||
|
'''
|
||||||
|
Unsets a callback by communication identifier, if set
|
||||||
|
'''
|
||||||
|
|
||||||
|
if identifier in future_callbacks:
|
||||||
|
del self.future_callbacks[identifier]
|
||||||
|
return True
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
def get_callback(self, identifier):
|
||||||
|
'''
|
||||||
|
Returns a callback by communication identifier if set, or None
|
||||||
|
'''
|
||||||
|
|
||||||
|
if identifier in self.future_callbacks:
|
||||||
|
return self.future_callbacks[id]
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
def direct_connect(self, peer, data = None, callback = None, log = True):
|
||||||
|
'''
|
||||||
|
Communicates something directly with the client
|
||||||
|
|
||||||
|
- `peer` should obviously be the peer id to request.
|
||||||
|
- `data` should be a dict (NOT str), with the parameter "type"
|
||||||
|
ex. {'type': 'sendMessage', 'content': 'hey, this is a dm'}
|
||||||
|
In that dict, the key 'token' must NEVER be set. If it is, it will
|
||||||
|
be overwritten.
|
||||||
|
- if `callback` is set to a function, it will call that function
|
||||||
|
back if/when the client the request is sent to decides to respond.
|
||||||
|
Do NOT depend on a response, because users can configure their
|
||||||
|
clients not to respond to this type of request.
|
||||||
|
- `log` is set to True by default-- what this does is log the
|
||||||
|
request for debug purposes. Should be False for sensitive actions.
|
||||||
|
'''
|
||||||
|
|
||||||
|
# TODO: Timing attack prevention
|
||||||
|
try:
|
||||||
|
# does not need to be secure random, only used for keeping track of async responses
|
||||||
|
# Actually, on second thought, it does need to be secure random. Otherwise, if it is predictable, someone could trigger arbitrary callbacks that have been saved on the local node, wrecking all kinds of havoc. Better just to keep it secure random.
|
||||||
|
identifier = self._utils.token(32)
|
||||||
|
if 'id' in data:
|
||||||
|
identifier = data['id']
|
||||||
|
|
||||||
|
if not identifier in id_peer_cache:
|
||||||
|
id_peer_cache[identifier] = peer
|
||||||
|
|
||||||
|
if type(data) == str:
|
||||||
|
# if someone inputs a string instead of a dict, it will assume it's the type
|
||||||
|
data = {'type' : data}
|
||||||
|
|
||||||
|
data['id'] = identifier
|
||||||
|
data['token'] = '' # later put PoW stuff here or whatever is needed
|
||||||
|
data_str = json.dumps(data)
|
||||||
|
|
||||||
|
events.event('outgoing_direct_connection', data = {'callback' : True, 'communicator' : self, 'data' : data, 'id' : identifier, 'token' : token, 'peer' : peer, 'callback' : callback, 'log' : log})
|
||||||
|
|
||||||
|
logger.debug('Direct connection (identifier: "%s"): %s' + (identifier, data_str))
|
||||||
|
try:
|
||||||
|
self.performGet('directMessage', peer, data_str)
|
||||||
|
except:
|
||||||
|
logger.warn('Failed to connect to peer: "%s".' % str(peer))
|
||||||
|
return False
|
||||||
|
|
||||||
|
if not callback is None:
|
||||||
|
self.set_callback(identifier, callback)
|
||||||
|
|
||||||
|
return True
|
||||||
|
except Exception as e:
|
||||||
|
logger.warn('Unknown error, failed to execute direct connect (peer: "%s").' % str(peer), e)
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
def direct_connect_response(self, identifier, data, peer = None, callback = None, log = True):
|
||||||
|
'''
|
||||||
|
Responds to a previous connection. Hostname will be pulled from id_peer_cache if not specified in `peer` parameter.
|
||||||
|
|
||||||
|
If yet another callback is requested, it can be put in the `callback` parameter.
|
||||||
|
'''
|
||||||
|
|
||||||
|
if config.get('dc_response', True):
|
||||||
|
data['id'] = identifier
|
||||||
|
data['sender'] = open('data/hs/hostname').read()
|
||||||
|
data['callback'] = True
|
||||||
|
|
||||||
|
if (origin is None) and (identifier in id_peer_cache):
|
||||||
|
origin = id_peer_cache[identifier]
|
||||||
|
|
||||||
|
if not identifier in id_peer_cache:
|
||||||
|
id_peer_cache[identifier] = peer
|
||||||
|
|
||||||
|
if origin is None:
|
||||||
|
logger.warn('Failed to identify peer for connection %s' % str(identifier))
|
||||||
|
return False
|
||||||
|
else:
|
||||||
|
return self.direct_connect(peer, data = data, callback = callback, log = log)
|
||||||
|
else:
|
||||||
|
logger.warn('Node tried to respond to direct connection id %s, but it was rejected due to `dc_response` restriction.' % str(identifier))
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def check_callbacks(self, data, execute = True, remove = True):
|
||||||
|
'''
|
||||||
|
Check if a callback is set, and if so, execute it
|
||||||
|
'''
|
||||||
|
|
||||||
|
try:
|
||||||
|
if type(data) is str:
|
||||||
|
data = json.loads(data)
|
||||||
|
|
||||||
|
if 'id' in data: # TODO: prevent enumeration, require extra PoW
|
||||||
|
identifier = data['id']
|
||||||
|
|
||||||
|
if identifier in self.future_callbacks:
|
||||||
|
if execute:
|
||||||
|
self.get_callback(identifier)(data)
|
||||||
|
logger.debug('Request callback "%s" executed.' % str(identifier))
|
||||||
|
if remove:
|
||||||
|
self.unset_callback(identifier)
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
|
logger.warn('Unable to find request callback for ID "%s".' % str(identifier))
|
||||||
|
else:
|
||||||
|
logger.warn('Unable to identify callback request, `id` parameter missing: %s' % json.dumps(data))
|
||||||
|
except Exception as e:
|
||||||
|
logger.warn('Unknown error, failed to execute direct connection callback (peer: "%s").' % str(peer), e)
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
def incoming_direct_connection(self, data):
|
||||||
|
'''
|
||||||
|
This code is run whenever there is a new incoming connection.
|
||||||
|
'''
|
||||||
|
|
||||||
|
if 'type' in data and data['type'] in self.connection_handlers:
|
||||||
|
for handler in self.get_connection_handlers(name):
|
||||||
|
handler(data)
|
||||||
|
|
||||||
|
return
|
||||||
|
|
||||||
def getNewPeers(self):
|
def getNewPeers(self):
|
||||||
'''
|
'''
|
||||||
Get new peers and ed25519 keys
|
Get new peers and ed25519 keys
|
||||||
'''
|
'''
|
||||||
|
|
||||||
peersCheck = 5 # Amount of peers to ask for new peers + keys
|
peersCheck = 5 # Amount of peers to ask for new peers + keys
|
||||||
peersChecked = 0
|
peersChecked = 0
|
||||||
peerList = list(self._core.listAdders()) # random ordered list of peers
|
peerList = list(self._core.listAdders()) # random ordered list of peers
|
||||||
@ -187,6 +410,7 @@ class OnionrCommunicate:
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
logger.info('Using ' + peerList[i] + ' to find new peers', timestamp=True)
|
logger.info('Using ' + peerList[i] + ' to find new peers', timestamp=True)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
newAdders = self.performGet('pex', peerList[i], skipHighFailureAddress=True)
|
newAdders = self.performGet('pex', peerList[i], skipHighFailureAddress=True)
|
||||||
logger.debug('Attempting to merge address: ')
|
logger.debug('Attempting to merge address: ')
|
||||||
@ -214,8 +438,10 @@ class OnionrCommunicate:
|
|||||||
'''
|
'''
|
||||||
Lookup blocks and merge new ones
|
Lookup blocks and merge new ones
|
||||||
'''
|
'''
|
||||||
|
|
||||||
peerList = self._core.listAdders()
|
peerList = self._core.listAdders()
|
||||||
blocks = ''
|
blocks = ''
|
||||||
|
|
||||||
for i in peerList:
|
for i in peerList:
|
||||||
if self.peerStatusTaken(i, 'getBlockHashes') or self.peerStatusTaken(i, 'getDBHash'):
|
if self.peerStatusTaken(i, 'getBlockHashes') or self.peerStatusTaken(i, 'getDBHash'):
|
||||||
continue
|
continue
|
||||||
@ -224,16 +450,21 @@ class OnionrCommunicate:
|
|||||||
continue
|
continue
|
||||||
except KeyError:
|
except KeyError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
lastDB = self._core.getAddressInfo(i, 'DBHash')
|
lastDB = self._core.getAddressInfo(i, 'DBHash')
|
||||||
|
|
||||||
if lastDB == None:
|
if lastDB == None:
|
||||||
logger.debug('Fetching hash from ' + i + ' No previous known.')
|
logger.debug('Fetching hash from ' + str(i) + ', no previous known.')
|
||||||
else:
|
else:
|
||||||
logger.debug('Fetching hash from ' + str(i) + ', ' + lastDB + ' last known')
|
logger.debug('Fetching hash from ' + str(i) + ', ' + str(lastDB) + ' last known')
|
||||||
|
|
||||||
currentDB = self.performGet('getDBHash', i)
|
currentDB = self.performGet('getDBHash', i)
|
||||||
|
|
||||||
if currentDB != False:
|
if currentDB != False:
|
||||||
logger.debug(i + " hash db (from request): " + currentDB)
|
logger.debug(i + " hash db (from request): " + currentDB)
|
||||||
else:
|
else:
|
||||||
logger.warn("Error getting hash db status for " + i)
|
logger.warn("Error getting hash db status for " + i)
|
||||||
|
|
||||||
if currentDB != False:
|
if currentDB != False:
|
||||||
if lastDB != currentDB:
|
if lastDB != currentDB:
|
||||||
logger.debug('Fetching hash from ' + i + ' - ' + currentDB + ' current hash.')
|
logger.debug('Fetching hash from ' + i + ' - ' + currentDB + ' current hash.')
|
||||||
@ -244,10 +475,13 @@ class OnionrCommunicate:
|
|||||||
self.peerData[i]['failCount'] -= 1
|
self.peerData[i]['failCount'] -= 1
|
||||||
if self._utils.validateHash(currentDB):
|
if self._utils.validateHash(currentDB):
|
||||||
self._core.setAddressInfo(i, "DBHash", currentDB)
|
self._core.setAddressInfo(i, "DBHash", currentDB)
|
||||||
|
|
||||||
if len(blocks.strip()) != 0:
|
if len(blocks.strip()) != 0:
|
||||||
pass
|
pass
|
||||||
#logger.debug('BLOCKS:' + blocks)
|
#logger.debug('BLOCKS:' + blocks)
|
||||||
|
|
||||||
blockList = blocks.split('\n')
|
blockList = blocks.split('\n')
|
||||||
|
|
||||||
for i in blockList:
|
for i in blockList:
|
||||||
if len(i.strip()) == 0:
|
if len(i.strip()) == 0:
|
||||||
continue
|
continue
|
||||||
@ -255,6 +489,7 @@ class OnionrCommunicate:
|
|||||||
continue
|
continue
|
||||||
if i in self.ignoredHashes:
|
if i in self.ignoredHashes:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
#logger.debug('Exchanged block (blockList): ' + i)
|
#logger.debug('Exchanged block (blockList): ' + i)
|
||||||
if not self._utils.validateHash(i):
|
if not self._utils.validateHash(i):
|
||||||
# skip hash if it isn't valid
|
# skip hash if it isn't valid
|
||||||
@ -284,6 +519,7 @@ class OnionrCommunicate:
|
|||||||
self.newHashes[i]
|
self.newHashes[i]
|
||||||
except KeyError:
|
except KeyError:
|
||||||
self.newHashes[i] = 0
|
self.newHashes[i] = 0
|
||||||
|
|
||||||
# check if a new hash has been around too long, delete it from database and add it to ignore list
|
# check if a new hash has been around too long, delete it from database and add it to ignore list
|
||||||
if self.newHashes[i] >= self.keepNewHash:
|
if self.newHashes[i] >= self.keepNewHash:
|
||||||
logger.warn('Ignoring block ' + i + ' because it took to long to get valid data.')
|
logger.warn('Ignoring block ' + i + ' because it took to long to get valid data.')
|
||||||
@ -291,6 +527,7 @@ class OnionrCommunicate:
|
|||||||
self._core.removeBlock(i)
|
self._core.removeBlock(i)
|
||||||
self.ignoredHashes.append(i)
|
self.ignoredHashes.append(i)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
self.newHashes[i] += 1
|
self.newHashes[i] += 1
|
||||||
logger.warn('UNSAVED BLOCK: ' + i)
|
logger.warn('UNSAVED BLOCK: ' + i)
|
||||||
data = self.downloadBlock(i)
|
data = self.downloadBlock(i)
|
||||||
@ -308,18 +545,22 @@ class OnionrCommunicate:
|
|||||||
blockMetadata = blockMetadata.decode()
|
blockMetadata = blockMetadata.decode()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
blockMetadata = json.loads(blockMetadata + '}')
|
blockMetadata = json.loads(blockMetadata + '}')
|
||||||
|
|
||||||
try:
|
try:
|
||||||
blockMetadata['sig']
|
blockMetadata['sig']
|
||||||
blockMetadata['id']
|
blockMetadata['id']
|
||||||
except KeyError:
|
except KeyError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
else:
|
else:
|
||||||
creator = self._utils.getPeerByHashId(blockMetadata['id'])
|
creator = self._utils.getPeerByHashId(blockMetadata['id'])
|
||||||
try:
|
try:
|
||||||
creator = creator.decode()
|
creator = creator.decode()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
if self._core._crypto.edVerify(blockContent.split(b'}')[1], creator, blockMetadata['sig'], encodedData=True):
|
if self._core._crypto.edVerify(blockContent.split(b'}')[1], creator, blockMetadata['sig'], encodedData=True):
|
||||||
self._core.updateBlockInfo(i, 'sig', 'true')
|
self._core.updateBlockInfo(i, 'sig', 'true')
|
||||||
else:
|
else:
|
||||||
@ -340,28 +581,36 @@ class OnionrCommunicate:
|
|||||||
'''
|
'''
|
||||||
Download a block from random order of peers
|
Download a block from random order of peers
|
||||||
'''
|
'''
|
||||||
|
|
||||||
retVal = False
|
retVal = False
|
||||||
peerList = self._core.listAdders()
|
peerList = self._core.listAdders()
|
||||||
blocks = ''
|
blocks = ''
|
||||||
peerTryCount = 0
|
peerTryCount = 0
|
||||||
|
|
||||||
for i in peerList:
|
for i in peerList:
|
||||||
if self.peerData[i]['failCount'] >= self.highFailureAmount:
|
if self.peerData[i]['failCount'] >= self.highFailureAmount:
|
||||||
continue
|
continue
|
||||||
if peerTryCount >= peerTries:
|
if peerTryCount >= peerTries:
|
||||||
break
|
break
|
||||||
|
|
||||||
hasher = hashlib.sha3_256()
|
hasher = hashlib.sha3_256()
|
||||||
data = self.performGet('getData', i, hash, skipHighFailureAddress=True)
|
data = self.performGet('getData', i, hash, skipHighFailureAddress=True)
|
||||||
|
|
||||||
if data == False or len(data) > 10000000 or data == '':
|
if data == False or len(data) > 10000000 or data == '':
|
||||||
peerTryCount += 1
|
peerTryCount += 1
|
||||||
continue
|
continue
|
||||||
|
|
||||||
try:
|
try:
|
||||||
data = base64.b64decode(data)
|
data = base64.b64decode(data)
|
||||||
except binascii.Error:
|
except binascii.Error:
|
||||||
data = b''
|
data = b''
|
||||||
|
|
||||||
hasher.update(data)
|
hasher.update(data)
|
||||||
digest = hasher.hexdigest()
|
digest = hasher.hexdigest()
|
||||||
|
|
||||||
if type(digest) is bytes:
|
if type(digest) is bytes:
|
||||||
digest = digest.decode()
|
digest = digest.decode()
|
||||||
|
|
||||||
if digest == hash.strip():
|
if digest == hash.strip():
|
||||||
self._core.setData(data)
|
self._core.setData(data)
|
||||||
logger.info('Successfully obtained data for ' + hash, timestamp=True)
|
logger.info('Successfully obtained data for ' + hash, timestamp=True)
|
||||||
|
@ -17,10 +17,9 @@
|
|||||||
You should have received a copy of the GNU General Public License
|
You should have received a copy of the GNU General Public License
|
||||||
along with this program. If not, see <https://www.gnu.org/licenses/>.
|
along with this program. If not, see <https://www.gnu.org/licenses/>.
|
||||||
'''
|
'''
|
||||||
import sqlite3, os, sys, time, math, base64, tarfile, getpass, simplecrypt, hashlib, nacl, logger, json
|
import sqlite3, os, sys, time, math, base64, tarfile, getpass, simplecrypt, hashlib, nacl, logger, json, netcontroller
|
||||||
#from Crypto.Cipher import AES
|
#from Crypto.Cipher import AES
|
||||||
#from Crypto import Random
|
#from Crypto import Random
|
||||||
import netcontroller
|
|
||||||
|
|
||||||
import onionrutils, onionrcrypto, btc, onionrevents as events
|
import onionrutils, onionrcrypto, btc, onionrevents as events
|
||||||
|
|
||||||
@ -547,7 +546,26 @@ class Core:
|
|||||||
conn.close()
|
conn.close()
|
||||||
return
|
return
|
||||||
|
|
||||||
def getBlockList(self, unsaved=False):
|
def handle_direct_connection(self, data):
|
||||||
|
'''
|
||||||
|
Handles direct messages
|
||||||
|
'''
|
||||||
|
try:
|
||||||
|
data = json.loads(data)
|
||||||
|
|
||||||
|
# TODO: Determine the sender, verify, etc
|
||||||
|
if ('callback' in data) and (data['callback'] is True):
|
||||||
|
# then this is a response to the message we sent earlier
|
||||||
|
self.daemonQueueAdd('checkCallbacks', json.dumps(data))
|
||||||
|
else:
|
||||||
|
# then we should handle it and respond accordingly
|
||||||
|
self.daemonQueueAdd('incomingDirectConnection', json.dumps(data))
|
||||||
|
except Exception as e:
|
||||||
|
logger.warn('Failed to handle incoming direct message: %s' % str(e))
|
||||||
|
|
||||||
|
return
|
||||||
|
|
||||||
|
def getBlockList(self, unsaved = False):
|
||||||
'''
|
'''
|
||||||
Get list of our blocks
|
Get list of our blocks
|
||||||
'''
|
'''
|
||||||
|
@ -59,8 +59,11 @@ class Onionr:
|
|||||||
if not data_exists:
|
if not data_exists:
|
||||||
os.mkdir('data/')
|
os.mkdir('data/')
|
||||||
|
|
||||||
exists = os.path.exists(config.get_config_file())
|
if os.path.exists('static-data/default_config.json'):
|
||||||
config.set_config({'devmode': True, 'log': {'file': {'output': True, 'path': 'data/output.log'}, 'console': {'output': True, 'color': True}}}) # this is the default config, it will be overwritten if a config file already exists. Else, it saves it
|
config.set_config(json.loads(open('static-data/default_config.json').read())) # this is the default config, it will be overwritten if a config file already exists. Else, it saves it
|
||||||
|
else:
|
||||||
|
# the default config file doesn't exist, try hardcoded config
|
||||||
|
config.set_config({'devmode': True, 'log': {'file': {'output': True, 'path': 'data/output.log'}, 'console': {'output': True, 'color': True}}})
|
||||||
if not exists:
|
if not exists:
|
||||||
config.save()
|
config.save()
|
||||||
config.reload() # this will read the configuration file into memory
|
config.reload() # this will read the configuration file into memory
|
||||||
@ -454,7 +457,7 @@ class Onionr:
|
|||||||
|
|
||||||
os.makedirs(plugins.get_plugins_folder(plugin_name))
|
os.makedirs(plugins.get_plugins_folder(plugin_name))
|
||||||
with open(plugins.get_plugins_folder(plugin_name) + '/main.py', 'a') as main:
|
with open(plugins.get_plugins_folder(plugin_name) + '/main.py', 'a') as main:
|
||||||
main.write(open('static-data/default_plugin.txt').read().replace('$user', os.getlogin()).replace('$date', datetime.datetime.now().strftime('%Y-%m-%d')))
|
main.write(open('static-data/default_plugin.py').read().replace('$user', os.getlogin()).replace('$date', datetime.datetime.now().strftime('%Y-%m-%d')))
|
||||||
|
|
||||||
logger.info('Enabling plugin "%s"...' % plugin_name)
|
logger.info('Enabling plugin "%s"...' % plugin_name)
|
||||||
plugins.enable(plugin_name, self)
|
plugins.enable(plugin_name, self)
|
||||||
|
@ -19,15 +19,16 @@
|
|||||||
'''
|
'''
|
||||||
|
|
||||||
import config, logger, onionrplugins as plugins, onionrpluginapi as pluginapi
|
import config, logger, onionrplugins as plugins, onionrpluginapi as pluginapi
|
||||||
|
from threading import Thread
|
||||||
|
|
||||||
def get_pluginapi(onionr, data):
|
def get_pluginapi(onionr, data):
|
||||||
return pluginapi.pluginapi(onionr, data)
|
return pluginapi.pluginapi(onionr, data)
|
||||||
|
|
||||||
def event(event_name, data = {}, onionr = None):
|
def __event_caller(event_name, data = {}, onionr = None):
|
||||||
'''
|
'''
|
||||||
Calls an event on all plugins (if defined)
|
DO NOT call this function, this is for threading code only.
|
||||||
|
Instead, call onionrevents.event
|
||||||
'''
|
'''
|
||||||
|
|
||||||
for plugin in plugins.get_enabled_plugins():
|
for plugin in plugins.get_enabled_plugins():
|
||||||
try:
|
try:
|
||||||
call(plugins.get_plugin(plugin), event_name, data, get_pluginapi(onionr, data))
|
call(plugins.get_plugin(plugin), event_name, data, get_pluginapi(onionr, data))
|
||||||
@ -38,6 +39,19 @@ def event(event_name, data = {}, onionr = None):
|
|||||||
logger.warn('Event \"' + event_name + '\" failed for plugin \"' + plugin + '\".')
|
logger.warn('Event \"' + event_name + '\" failed for plugin \"' + plugin + '\".')
|
||||||
logger.debug(str(e))
|
logger.debug(str(e))
|
||||||
|
|
||||||
|
|
||||||
|
def event(event_name, data = {}, onionr = None, threaded = True):
|
||||||
|
'''
|
||||||
|
Calls an event on all plugins (if defined)
|
||||||
|
'''
|
||||||
|
|
||||||
|
if threaded:
|
||||||
|
thread = Thread(target = __event_caller, args = (event_name, data, onionr))
|
||||||
|
thread.start()
|
||||||
|
return thread
|
||||||
|
else:
|
||||||
|
__event_caller(event_name, data, onionr)
|
||||||
|
|
||||||
def call(plugin, event_name, data = None, pluginapi = None):
|
def call(plugin, event_name, data = None, pluginapi = None):
|
||||||
'''
|
'''
|
||||||
Calls an event on a plugin if one is defined
|
Calls an event on a plugin if one is defined
|
||||||
|
@ -421,3 +421,6 @@ class OnionrUtils:
|
|||||||
return False
|
return False
|
||||||
except:
|
except:
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
def token(self, size = 32):
|
||||||
|
return binascii.hexlify(os.urandom(size))
|
||||||
|
16
onionr/static-data/default_config.json
Normal file
16
onionr/static-data/default_config.json
Normal file
@ -0,0 +1,16 @@
|
|||||||
|
{
|
||||||
|
"devmode": true,
|
||||||
|
"dc_response": true,
|
||||||
|
|
||||||
|
"log": {
|
||||||
|
"file": {
|
||||||
|
"output": true,
|
||||||
|
"path": "data/output.log"
|
||||||
|
},
|
||||||
|
|
||||||
|
"console": {
|
||||||
|
"output": true,
|
||||||
|
"color": true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
@ -171,15 +171,19 @@ class OnionrTests(unittest.TestCase):
|
|||||||
if not plugins.exists('test'):
|
if not plugins.exists('test'):
|
||||||
os.makedirs(plugins.get_plugins_folder('test'))
|
os.makedirs(plugins.get_plugins_folder('test'))
|
||||||
with open(plugins.get_plugins_folder('test') + '/main.py', 'a') as main:
|
with open(plugins.get_plugins_folder('test') + '/main.py', 'a') as main:
|
||||||
main.write("print('Running')\n\ndef on_test(pluginapi, data = None):\n print('received test event!')\n return True\n\ndef on_start(pluginapi, data = None):\n print('start event called')\n\ndef on_stop(pluginapi, data = None):\n print('stop event called')\n\ndef on_enable(pluginapi, data = None):\n print('enable event called')\n\ndef on_disable(pluginapi, data = None):\n print('disable event called')\n")
|
main.write("print('Running')\n\ndef on_test(pluginapi, data = None):\n print('received test event!')\n print('thread test started...')\n import time\n time.sleep(1)\n \n return True\n\ndef on_start(pluginapi, data = None):\n print('start event called')\n\ndef on_stop(pluginapi, data = None):\n print('stop event called')\n\ndef on_enable(pluginapi, data = None):\n print('enable event called')\n\ndef on_disable(pluginapi, data = None):\n print('disable event called')\n")
|
||||||
plugins.enable('test')
|
plugins.enable('test')
|
||||||
|
|
||||||
|
|
||||||
plugins.start('test')
|
plugins.start('test')
|
||||||
if not events.call(plugins.get_plugin('test'), 'test'):
|
if not events.call(plugins.get_plugin('test'), 'enable'):
|
||||||
self.assertTrue(False)
|
self.assertTrue(False)
|
||||||
|
|
||||||
events.event('test', data = {'tests': self})
|
logger.debug('preparing to start thread', timestamp = False)
|
||||||
|
thread = events.event('test', data = {'tests': self})
|
||||||
|
logger.debug('thread running...', timestamp = False)
|
||||||
|
thread.join()
|
||||||
|
logger.debug('thread finished.', timestamp = False)
|
||||||
|
|
||||||
self.assertTrue(True)
|
self.assertTrue(True)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user