def get_current_txouts(self): real_subsidy = self.quarkd_work.value['subsidy'] if self.quarkd_work.value['payee_amount'] >= 0: real_pay = real_subsidy - self.quarkd_work.value['payee_amount'] return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.quarkd_work.value['bits'].target, real_pay, self.net) else: return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.quarkd_work.value['bits'].target, real_subsidy, self.net)
def get_current_txouts(self): #i = 210240 real_subsidy = self.dashd_work.value['subsidy'] #while i <= self.dashd_work.value['height']: #real_subsidy = real_subsidy*92.9/100 #i = i + 210240 if self.dashd_work.value['height'] > 158000 + ((576 * 30) * 17): real_pay = (real_subsidy) * 40 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 15): real_pay = (real_subsidy) * 42.5 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 13): real_pay = (real_subsidy) * 45 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 11): real_pay = (real_subsidy) * 47.5 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 9): real_pay = (real_subsidy) * 50 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 7): real_pay = (real_subsidy) * 52.5 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 6): real_pay = (real_subsidy) * 55 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 5): real_pay = (real_subsidy) * 57.5 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 4): real_pay = (real_subsidy) * 60 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 3): real_pay = (real_subsidy) * 62.5 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 2): real_pay = (real_subsidy) * 65 / 100 elif self.dashd_work.value['height'] > 158000 + ((576 * 30) * 1): real_pay = (real_subsidy) * 70 / 100 else: real_pay = (real_subsidy) * 75 / 100 return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.dashd_work.value['bits'].target, real_pay, self.net)
def get_current_txouts(self): #i = 210240 real_subsidy = self.dashd_work.value['subsidy'] #while i <= self.dashd_work.value['height']: #real_subsidy = real_subsidy*92.9/100 #i = i + 210240 if self.dashd_work.value['height'] > 158000+((576*30)* 17): real_pay = (real_subsidy)*40/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 15): real_pay = (real_subsidy)*42.5/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 13): real_pay = (real_subsidy)*45/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 11): real_pay = (real_subsidy)*47.5/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 9): real_pay = (real_subsidy)*50/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 7): real_pay = (real_subsidy)*52.5/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 6): real_pay = (real_subsidy)*55/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 5): real_pay = (real_subsidy)*57.5/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 4): real_pay = (real_subsidy)*60/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 3): real_pay = (real_subsidy)*62.5/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 2): real_pay = (real_subsidy)*65/100 elif self.dashd_work.value['height'] > 158000+((576*30)* 1): real_pay = (real_subsidy)*70/100 else: real_pay = (real_subsidy)*75/100 return p2pool_data.get_expected_payouts(self.tracker, self.best_share_var.value, self.dashd_work.value['bits'].target, real_pay, self.net)
def get_current_txouts(self): return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.bitcoind_work.value["bits"].target, self.bitcoind_work.value["subsidy"], self.net, )
def get_current_txouts(self): #i = 210240 real_subsidy = self.futurocoind_work.value['subsidy'] #while i <= self.futurocoind_work.value['height']: #real_subsidy = real_subsidy*92.9/100 #i = i + 210240 # Use value from getblocktemplate's result. if self.futurocoind_work.value['payment_amount'] >= 0: real_pay = real_subsidy - self.futurocoind_work.value[ 'payment_amount'] return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.futurocoind_work.value['bits'].target, real_pay, self.net) if self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 17): real_pay = (real_subsidy) * 40 / 100 elif self.futurocoind_work.value['height'] > 158000 + ( (576 * 30) * 15): real_pay = (real_subsidy) * 42.5 / 100 elif self.futurocoind_work.value['height'] > 158000 + ( (576 * 30) * 13): real_pay = (real_subsidy) * 45 / 100 elif self.futurocoind_work.value['height'] > 158000 + ( (576 * 30) * 11): real_pay = (real_subsidy) * 47.5 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 9): real_pay = (real_subsidy) * 50 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 7): real_pay = (real_subsidy) * 52.5 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 6): real_pay = (real_subsidy) * 55 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 5): real_pay = (real_subsidy) * 57.5 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 4): real_pay = (real_subsidy) * 60 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 3): real_pay = (real_subsidy) * 62.5 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 2): real_pay = (real_subsidy) * 65 / 100 elif self.futurocoind_work.value['height'] > 158000 + ((576 * 30) * 1): real_pay = (real_subsidy) * 70 / 100 else: real_pay = (real_subsidy) * 75 / 100 return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.futurocoind_work.value['bits'].target, real_pay, self.net)
def get_current_txouts(self): return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.dcrd_work.value['bits'].target, self.dcrd_work.value['subsidy'], self.net)
def get_current_txouts(self): return p2pool_data.get_expected_payouts(self.tracker, self.best_share_var.value, self.ezoncoind_work.value['bits'].target, self.ezoncoind_work.value['subsidy'], self.net)
def main(args, net, datadir_path, merged_urls, worker_endpoint): try: print 'p2pool (version %s)' % (p2pool.__version__,) print traffic_happened = variable.Event() @defer.inlineCallbacks def connect_p2p(): # connect to bitcoind over bitcoin-p2p print '''Testing bitcoind P2P connection to '%s:%s'...''' % (args.bitcoind_address, args.bitcoind_p2p_port) factory = bitcoin_p2p.ClientFactory(net.PARENT) reactor.connectTCP(args.bitcoind_address, args.bitcoind_p2p_port, factory) yield factory.getProtocol() # waits until handshake is successful print ' ...success!' print defer.returnValue(factory) if args.testnet: # establish p2p connection first if testnet so bitcoind can work without connections factory = yield connect_p2p() # connect to bitcoind over JSON-RPC and do initial getmemorypool url = '%s://%s:%i/' % ('https' if args.bitcoind_rpc_ssl else 'http', args.bitcoind_address, args.bitcoind_rpc_port) print '''Testing bitcoind RPC connection to '%s' with username '%s'...''' % (url, args.bitcoind_rpc_username) bitcoind = jsonrpc.Proxy(url, dict(Authorization='Basic ' + base64.b64encode(args.bitcoind_rpc_username + ':' + args.bitcoind_rpc_password)), timeout=30) @deferral.retry('Error while checking Bitcoin connection:', 1) @defer.inlineCallbacks def check(): if not (yield net.PARENT.RPC_CHECK(bitcoind)): print >>sys.stderr, " Check failed! Make sure that you're connected to the right bitcoind with --bitcoind-rpc-port!" raise deferral.RetrySilentlyException() if not net.VERSION_CHECK((yield bitcoind.rpc_getinfo())['version']): print >>sys.stderr, ' Bitcoin version too old! Upgrade to 0.6.4 or newer!' raise deferral.RetrySilentlyException() yield check() temp_work = yield getwork(bitcoind) if not args.testnet: factory = yield connect_p2p() block_height_var = variable.Variable(None) @defer.inlineCallbacks def poll_height(): block_height_var.set((yield deferral.retry('Error while calling getblockcount:')(bitcoind.rpc_getblockcount)())) yield poll_height() task.LoopingCall(poll_height).start(60*60) bitcoind_warning_var = variable.Variable(None) @defer.inlineCallbacks def poll_warnings(): errors = (yield deferral.retry('Error while calling getmininginfo:')(bitcoind.rpc_getmininginfo)())['errors'] bitcoind_warning_var.set(errors if errors != '' else None) yield poll_warnings() task.LoopingCall(poll_warnings).start(20*60) print ' ...success!' print ' Current block hash: %x' % (temp_work['previous_block'],) print ' Current block height: %i' % (block_height_var.value,) print print 'Determining payout address...' if args.pubkey_hash is None: address_path = os.path.join(datadir_path, 'cached_payout_address') if os.path.exists(address_path): with open(address_path, 'rb') as f: address = f.read().strip('\r\n') print ' Loaded cached address: %s...' % (address,) else: address = None if address is not None: res = yield deferral.retry('Error validating cached address:', 5)(lambda: bitcoind.rpc_validateaddress(address))() if not res['isvalid'] or not res['ismine']: print ' Cached address is either invalid or not controlled by local bitcoind!' address = None if address is None: print ' Getting payout address from bitcoind...' address = yield deferral.retry('Error getting payout address from bitcoind:', 5)(lambda: bitcoind.rpc_getaccountaddress('p2pool'))() with open(address_path, 'wb') as f: f.write(address) my_pubkey_hash = bitcoin_data.address_to_pubkey_hash(address, net.PARENT) else: my_pubkey_hash = args.pubkey_hash print ' ...success! Payout address:', bitcoin_data.pubkey_hash_to_address(my_pubkey_hash, net.PARENT) print my_share_hashes = set() my_doa_share_hashes = set() tracker = p2pool_data.OkayTracker(net, my_share_hashes, my_doa_share_hashes) shared_share_hashes = set() ss = p2pool_data.ShareStore(os.path.join(datadir_path, 'shares.'), net) known_verified = set() print "Loading shares..." for i, (mode, contents) in enumerate(ss.get_shares()): if mode == 'share': if contents.hash in tracker.items: continue shared_share_hashes.add(contents.hash) contents.time_seen = 0 tracker.add(contents) if len(tracker.items) % 1000 == 0 and tracker.items: print " %i" % (len(tracker.items),) elif mode == 'verified_hash': known_verified.add(contents) else: raise AssertionError() print " ...inserting %i verified shares..." % (len(known_verified),) for h in known_verified: if h not in tracker.items: ss.forget_verified_share(h) continue tracker.verified.add(tracker.items[h]) print " ...done loading %i shares!" % (len(tracker.items),) print tracker.removed.watch(lambda share: ss.forget_share(share.hash)) tracker.verified.removed.watch(lambda share: ss.forget_verified_share(share.hash)) tracker.removed.watch(lambda share: shared_share_hashes.discard(share.hash)) print 'Initializing work...' # BITCOIND WORK bitcoind_work = variable.Variable((yield getwork(bitcoind))) @defer.inlineCallbacks def work_poller(): while True: flag = factory.new_block.get_deferred() try: bitcoind_work.set((yield getwork(bitcoind, bitcoind_work.value['use_getblocktemplate']))) except: log.err() yield defer.DeferredList([flag, deferral.sleep(15)], fireOnOneCallback=True) work_poller() # PEER WORK best_block_header = variable.Variable(None) def handle_header(new_header): # check that header matches current target if not (net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(new_header)) <= bitcoind_work.value['bits'].target): return bitcoind_best_block = bitcoind_work.value['previous_block'] if (best_block_header.value is None or ( new_header['previous_block'] == bitcoind_best_block and bitcoin_data.hash256(bitcoin_data.block_header_type.pack(best_block_header.value)) == bitcoind_best_block ) # new is child of current and previous is current or ( bitcoin_data.hash256(bitcoin_data.block_header_type.pack(new_header)) == bitcoind_best_block and best_block_header.value['previous_block'] != bitcoind_best_block )): # new is current and previous is not a child of current best_block_header.set(new_header) @defer.inlineCallbacks def poll_header(): handle_header((yield factory.conn.value.get_block_header(bitcoind_work.value['previous_block']))) bitcoind_work.changed.watch(lambda _: poll_header()) yield deferral.retry('Error while requesting best block header:')(poll_header)() # BEST SHARE get_height_rel_highest = yield height_tracker.get_height_rel_highest_func(bitcoind, factory, lambda: bitcoind_work.value['previous_block'], net) best_share_var = variable.Variable(None) desired_var = variable.Variable(None) def set_best_share(): best, desired = tracker.think(get_height_rel_highest, bitcoind_work.value['previous_block'], bitcoind_work.value['bits']) best_share_var.set(best) desired_var.set(desired) bitcoind_work.changed.watch(lambda _: set_best_share()) set_best_share() print ' ...success!' print # setup p2p logic and join p2pool network class Node(p2p.Node): def handle_shares(self, shares, peer): if len(shares) > 5: print 'Processing %i shares from %s...' % (len(shares), '%s:%i' % peer.addr if peer is not None else None) new_count = 0 for share in shares: if share.hash in tracker.items: #print 'Got duplicate share, ignoring. Hash: %s' % (p2pool_data.format_hash(share.hash),) continue new_count += 1 #print 'Received share %s from %r' % (p2pool_data.format_hash(share.hash), share.peer.addr if share.peer is not None else None) tracker.add(share) if new_count: set_best_share() if len(shares) > 5: print '... done processing %i shares. New: %i Have: %i/~%i' % (len(shares), new_count, len(tracker.items), 2*net.CHAIN_LENGTH) @defer.inlineCallbacks def handle_share_hashes(self, hashes, peer): new_hashes = [x for x in hashes if x not in tracker.items] if not new_hashes: return try: shares = yield peer.get_shares( hashes=new_hashes, parents=0, stops=[], ) except: log.err(None, 'in handle_share_hashes:') else: self.handle_shares(shares, peer) def handle_get_shares(self, hashes, parents, stops, peer): parents = min(parents, 1000//len(hashes)) stops = set(stops) shares = [] for share_hash in hashes: for share in tracker.get_chain(share_hash, min(parents + 1, tracker.get_height(share_hash))): if share.hash in stops: break shares.append(share) print 'Sending %i shares to %s:%i' % (len(shares), peer.addr[0], peer.addr[1]) return shares def handle_bestblock(self, header, peer): if net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(header)) > header['bits'].target: raise p2p.PeerMisbehavingError('received block header fails PoW test') handle_header(header) @deferral.retry('Error submitting primary block: (will retry)', 10, 10) def submit_block_p2p(block): if factory.conn.value is None: print >>sys.stderr, 'No bitcoind connection when block submittal attempted! %s%064x' % (net.PARENT.BLOCK_EXPLORER_URL_PREFIX, bitcoin_data.hash256(bitcoin_data.block_header_type.pack(block['header']))) raise deferral.RetrySilentlyException() factory.conn.value.send_block(block=block) @deferral.retry('Error submitting block: (will retry)', 10, 10) @defer.inlineCallbacks def submit_block_rpc(block, ignore_failure): if bitcoind_work.value['use_getblocktemplate']: result = yield bitcoind.rpc_submitblock(bitcoin_data.block_type.pack(block).encode('hex')) success = result is None else: result = yield bitcoind.rpc_getmemorypool(bitcoin_data.block_type.pack(block).encode('hex')) success = result success_expected = net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(block['header'])) <= block['header']['bits'].target if (not success and success_expected and not ignore_failure) or (success and not success_expected): print >>sys.stderr, 'Block submittal result: %s (%r) Expected: %s' % (success, result, success_expected) def submit_block(block, ignore_failure): submit_block_p2p(block) submit_block_rpc(block, ignore_failure) @tracker.verified.added.watch def _(share): if share.pow_hash <= share.header['bits'].target: submit_block(share.as_block(tracker), ignore_failure=True) print print 'GOT BLOCK FROM PEER! Passing to bitcoind! %s bitcoin: %s%064x' % (p2pool_data.format_hash(share.hash), net.PARENT.BLOCK_EXPLORER_URL_PREFIX, share.header_hash) print def spread(): if (get_height_rel_highest(share.header['previous_block']) > -5 or bitcoind_work.value['previous_block'] in [share.header['previous_block'], share.header_hash]): broadcast_share(share.hash) spread() reactor.callLater(5, spread) # so get_height_rel_highest can update print 'Joining p2pool network using port %i...' % (args.p2pool_port,) @defer.inlineCallbacks def parse(x): if ':' in x: ip, port = x.split(':') defer.returnValue(((yield reactor.resolve(ip)), int(port))) else: defer.returnValue(((yield reactor.resolve(x)), net.P2P_PORT)) addrs = {} if os.path.exists(os.path.join(datadir_path, 'addrs')): try: with open(os.path.join(datadir_path, 'addrs'), 'rb') as f: addrs.update(dict((tuple(k), v) for k, v in json.loads(f.read()))) except: print >>sys.stderr, 'error parsing addrs' for addr_df in map(parse, net.BOOTSTRAP_ADDRS): try: addr = yield addr_df if addr not in addrs: addrs[addr] = (0, time.time(), time.time()) except: log.err() connect_addrs = set() for addr_df in map(parse, args.p2pool_nodes): try: connect_addrs.add((yield addr_df)) except: log.err() p2p_node = Node( best_share_hash_func=lambda: best_share_var.value, port=args.p2pool_port, net=net, addr_store=addrs, connect_addrs=connect_addrs, max_incoming_conns=args.p2pool_conns, traffic_happened=traffic_happened, ) p2p_node.start() def save_addrs(): with open(os.path.join(datadir_path, 'addrs'), 'wb') as f: f.write(json.dumps(p2p_node.addr_store.items())) task.LoopingCall(save_addrs).start(60) @best_block_header.changed.watch def _(header): for peer in p2p_node.peers.itervalues(): peer.send_bestblock(header=header) @defer.inlineCallbacks def broadcast_share(share_hash): shares = [] for share in tracker.get_chain(share_hash, min(5, tracker.get_height(share_hash))): if share.hash in shared_share_hashes: break shared_share_hashes.add(share.hash) shares.append(share) for peer in list(p2p_node.peers.itervalues()): yield peer.sendShares([share for share in shares if share.peer is not peer]) # send share when the chain changes to their chain best_share_var.changed.watch(broadcast_share) def save_shares(): for share in tracker.get_chain(best_share_var.value, min(tracker.get_height(best_share_var.value), 2*net.CHAIN_LENGTH)): ss.add_share(share) if share.hash in tracker.verified.items: ss.add_verified_hash(share.hash) task.LoopingCall(save_shares).start(60) @apply @defer.inlineCallbacks def download_shares(): while True: desired = yield desired_var.get_when_satisfies(lambda val: len(val) != 0) peer2, share_hash = random.choice(desired) if len(p2p_node.peers) == 0: yield deferral.sleep(1) continue peer = random.choice(p2p_node.peers.values()) print 'Requesting parent share %s from %s' % (p2pool_data.format_hash(share_hash), '%s:%i' % peer.addr) try: shares = yield peer.get_shares( hashes=[share_hash], parents=500, stops=[], ) except: log.err(None, 'in download_shares:') continue if not shares: yield deferral.sleep(1) # sleep so we don't keep rerequesting the same share nobody has continue p2p_node.handle_shares(shares, peer) print ' ...success!' print if args.upnp: @defer.inlineCallbacks def upnp_thread(): while True: try: is_lan, lan_ip = yield ipdiscover.get_local_ip() if is_lan: pm = yield portmapper.get_port_mapper() yield pm._upnp.add_port_mapping(lan_ip, args.p2pool_port, args.p2pool_port, 'p2pool', 'TCP') except defer.TimeoutError: pass except: if p2pool.DEBUG: log.err(None, 'UPnP error:') yield deferral.sleep(random.expovariate(1/120)) upnp_thread() # start listening for workers with a JSON-RPC server print 'Listening for workers on %r port %i...' % (worker_endpoint[0], worker_endpoint[1]) get_current_txouts = lambda: p2pool_data.get_expected_payouts(tracker, best_share_var.value, bitcoind_work.value['bits'].target, bitcoind_work.value['subsidy'], net) wb = work.WorkerBridge(my_pubkey_hash, net, args.donation_percentage, bitcoind_work, best_block_header, merged_urls, best_share_var, tracker, my_share_hashes, my_doa_share_hashes, args.worker_fee, p2p_node, submit_block, set_best_share, broadcast_share, block_height_var) web_root = web.get_web_root(tracker, bitcoind_work, get_current_txouts, datadir_path, net, wb.get_stale_counts, my_pubkey_hash, wb.local_rate_monitor, args.worker_fee, p2p_node, wb.my_share_hashes, wb.pseudoshare_received, wb.share_received, best_share_var, bitcoind_warning_var, traffic_happened) worker_interface.WorkerInterface(wb).attach_to(web_root, get_handler=lambda request: request.redirect('/static/')) deferral.retry('Error binding to worker port:', traceback=False)(reactor.listenTCP)(worker_endpoint[1], server.Site(web_root), interface=worker_endpoint[0]) with open(os.path.join(os.path.join(datadir_path, 'ready_flag')), 'wb') as f: pass print ' ...success!' print # done! print 'Started successfully!' print 'Go to http://127.0.0.1:%i/ to view graphs and statistics!' % (worker_endpoint[1],) if args.donation_percentage > 0.51: print '''Donating %.1f%% of work towards P2Pool's development. Thanks for the tip!''' % (args.donation_percentage,) elif args.donation_percentage < 0.49: print '''Donating %.1f%% of work towards P2Pool's development. Please donate to encourage further development of P2Pool!''' % (args.donation_percentage,) else: print '''Donating %.1f%% of work towards P2Pool's development. Thank you!''' % (args.donation_percentage,) print 'You can increase this amount with --give-author argument! (or decrease it, if you must)' print if hasattr(signal, 'SIGALRM'): signal.signal(signal.SIGALRM, lambda signum, frame: reactor.callFromThread( sys.stderr.write, 'Watchdog timer went off at:\n' + ''.join(traceback.format_stack()) )) signal.siginterrupt(signal.SIGALRM, False) task.LoopingCall(signal.alarm, 30).start(1) if args.irc_announce: from twisted.words.protocols import irc class IRCClient(irc.IRCClient): nickname = 'p2pool%02i' % (random.randrange(100),) channel = net.ANNOUNCE_CHANNEL def lineReceived(self, line): if p2pool.DEBUG: print repr(line) irc.IRCClient.lineReceived(self, line) def signedOn(self): irc.IRCClient.signedOn(self) self.factory.resetDelay() self.join(self.channel) @defer.inlineCallbacks def new_share(share): if share.pow_hash <= share.header['bits'].target and abs(share.timestamp - time.time()) < 10*60: yield deferral.sleep(random.expovariate(1/60)) message = '\x02%s BLOCK FOUND by %s! %s%064x' % (net.NAME.upper(), bitcoin_data.script2_to_address(share.new_script, net.PARENT), net.PARENT.BLOCK_EXPLORER_URL_PREFIX, share.header_hash) if all('%x' % (share.header_hash,) not in old_message for old_message in self.recent_messages): self.say(self.channel, message) self._remember_message(message) self.watch_id = tracker.verified.added.watch(new_share) self.recent_messages = [] def _remember_message(self, message): self.recent_messages.append(message) while len(self.recent_messages) > 100: self.recent_messages.pop(0) def privmsg(self, user, channel, message): if channel == self.channel: self._remember_message(message) def connectionLost(self, reason): tracker.verified.added.unwatch(self.watch_id) print 'IRC connection lost:', reason.getErrorMessage() class IRCClientFactory(protocol.ReconnectingClientFactory): protocol = IRCClient reactor.connectTCP("irc.freenode.net", 6667, IRCClientFactory()) @defer.inlineCallbacks def status_thread(): last_str = None last_time = 0 while True: yield deferral.sleep(3) try: height = tracker.get_height(best_share_var.value) this_str = 'P2Pool: %i shares in chain (%i verified/%i total) Peers: %i (%i incoming)' % ( height, len(tracker.verified.items), len(tracker.items), len(p2p_node.peers), sum(1 for peer in p2p_node.peers.itervalues() if peer.incoming), ) + (' FDs: %i R/%i W' % (len(reactor.getReaders()), len(reactor.getWriters())) if p2pool.DEBUG else '') datums, dt = wb.local_rate_monitor.get_datums_in_last() my_att_s = sum(datum['work']/dt for datum in datums) this_str += '\n Local: %sH/s in last %s Local dead on arrival: %s Expected time to share: %s' % ( math.format(int(my_att_s)), math.format_dt(dt), math.format_binomial_conf(sum(1 for datum in datums if datum['dead']), len(datums), 0.95), math.format_dt(2**256 / tracker.items[best_share_var.value].max_target / my_att_s) if my_att_s and best_share_var.value else '???', ) if height > 2: (stale_orphan_shares, stale_doa_shares), shares, _ = wb.get_stale_counts() stale_prop = p2pool_data.get_average_stale_prop(tracker, best_share_var.value, min(60*60//net.SHARE_PERIOD, height)) real_att_s = p2pool_data.get_pool_attempts_per_second(tracker, best_share_var.value, min(height - 1, 60*60//net.SHARE_PERIOD)) / (1 - stale_prop) this_str += '\n Shares: %i (%i orphan, %i dead) Stale rate: %s Efficiency: %s Current payout: %.4f %s' % ( shares, stale_orphan_shares, stale_doa_shares, math.format_binomial_conf(stale_orphan_shares + stale_doa_shares, shares, 0.95), math.format_binomial_conf(stale_orphan_shares + stale_doa_shares, shares, 0.95, lambda x: (1 - x)/(1 - stale_prop)), get_current_txouts().get(bitcoin_data.pubkey_hash_to_script2(my_pubkey_hash), 0)*1e-8, net.PARENT.SYMBOL, ) this_str += '\n Pool: %sH/s Stale rate: %.1f%% Expected time to block: %s' % ( math.format(int(real_att_s)), 100*stale_prop, math.format_dt(2**256 / bitcoind_work.value['bits'].target / real_att_s), ) for warning in p2pool_data.get_warnings(tracker, best_share_var.value, net, bitcoind_warning_var.value, bitcoind_work.value): print >>sys.stderr, '#'*40 print >>sys.stderr, '>>> Warning: ' + warning print >>sys.stderr, '#'*40 if this_str != last_str or time.time() > last_time + 15: print this_str last_str = this_str last_time = time.time() except: log.err() status_thread() except: reactor.stop() log.err(None, 'Fatal error:')
def main(args, net, datadir_path, merged_urls, worker_endpoint): try: print 'p2pool (version %s)' % (p2pool.__version__, ) print traffic_happened = variable.Event() @defer.inlineCallbacks def connect_p2p(): # connect to bitcoind over bitcoin-p2p print '''Testing bitcoind P2P connection to '%s:%s'...''' % ( args.bitcoind_address, args.bitcoind_p2p_port) factory = bitcoin_p2p.ClientFactory(net.PARENT) reactor.connectTCP(args.bitcoind_address, args.bitcoind_p2p_port, factory) yield factory.getProtocol() # waits until handshake is successful print ' ...success!' print defer.returnValue(factory) if args.testnet: # establish p2p connection first if testnet so bitcoind can work without connections factory = yield connect_p2p() # connect to bitcoind over JSON-RPC and do initial getmemorypool url = '%s://%s:%i/' % ('https' if args.bitcoind_rpc_ssl else 'http', args.bitcoind_address, args.bitcoind_rpc_port) print '''Testing bitcoind RPC connection to '%s' with username '%s'...''' % ( url, args.bitcoind_rpc_username) bitcoind = jsonrpc.Proxy( url, dict(Authorization='Basic ' + base64.b64encode(args.bitcoind_rpc_username + ':' + args.bitcoind_rpc_password)), timeout=30) @deferral.retry('Error while checking Bitcoin connection:', 1) @defer.inlineCallbacks def check(): if not (yield net.PARENT.RPC_CHECK(bitcoind)): print >> sys.stderr, " Check failed! Make sure that you're connected to the right bitcoind with --bitcoind-rpc-port!" raise deferral.RetrySilentlyException() if not net.VERSION_CHECK( (yield bitcoind.rpc_getinfo())['version']): print >> sys.stderr, ' Bitcoin version too old! Upgrade to 0.6.4 or newer!' raise deferral.RetrySilentlyException() yield check() temp_work = yield getwork(bitcoind) if not args.testnet: factory = yield connect_p2p() block_height_var = variable.Variable(None) @defer.inlineCallbacks def poll_height(): block_height_var.set( (yield deferral.retry('Error while calling getblockcount:')( bitcoind.rpc_getblockcount)())) yield poll_height() task.LoopingCall(poll_height).start(60 * 60) bitcoind_warning_var = variable.Variable(None) @defer.inlineCallbacks def poll_warnings(): errors = (yield deferral.retry('Error while calling getmininginfo:')( bitcoind.rpc_getmininginfo)())['errors'] bitcoind_warning_var.set(errors if errors != '' else None) yield poll_warnings() task.LoopingCall(poll_warnings).start(20 * 60) print ' ...success!' print ' Current block hash: %x' % (temp_work['previous_block'], ) print ' Current block height: %i' % (block_height_var.value, ) print print 'Determining payout address...' if args.pubkey_hash is None: address_path = os.path.join(datadir_path, 'cached_payout_address') if os.path.exists(address_path): with open(address_path, 'rb') as f: address = f.read().strip('\r\n') print ' Loaded cached address: %s...' % (address, ) else: address = None if address is not None: res = yield deferral.retry( 'Error validating cached address:', 5)(lambda: bitcoind.rpc_validateaddress(address))() if not res['isvalid'] or not res['ismine']: print ' Cached address is either invalid or not controlled by local bitcoind!' address = None if address is None: print ' Getting payout address from bitcoind...' address = yield deferral.retry( 'Error getting payout address from bitcoind:', 5)(lambda: bitcoind.rpc_getaccountaddress('p2pool'))() with open(address_path, 'wb') as f: f.write(address) my_pubkey_hash = bitcoin_data.address_to_pubkey_hash( address, net.PARENT) else: my_pubkey_hash = args.pubkey_hash print ' ...success! Payout address:', bitcoin_data.pubkey_hash_to_address( my_pubkey_hash, net.PARENT) print my_share_hashes = set() my_doa_share_hashes = set() tracker = p2pool_data.OkayTracker(net, my_share_hashes, my_doa_share_hashes) shared_share_hashes = set() ss = p2pool_data.ShareStore(os.path.join(datadir_path, 'shares.'), net) known_verified = set() print "Loading shares..." for i, (mode, contents) in enumerate(ss.get_shares()): if mode == 'share': if contents.hash in tracker.items: continue shared_share_hashes.add(contents.hash) contents.time_seen = 0 tracker.add(contents) if len(tracker.items) % 1000 == 0 and tracker.items: print " %i" % (len(tracker.items), ) elif mode == 'verified_hash': known_verified.add(contents) else: raise AssertionError() print " ...inserting %i verified shares..." % ( len(known_verified), ) for h in known_verified: if h not in tracker.items: ss.forget_verified_share(h) continue tracker.verified.add(tracker.items[h]) print " ...done loading %i shares!" % (len(tracker.items), ) print tracker.removed.watch(lambda share: ss.forget_share(share.hash)) tracker.verified.removed.watch( lambda share: ss.forget_verified_share(share.hash)) tracker.removed.watch( lambda share: shared_share_hashes.discard(share.hash)) print 'Initializing work...' # BITCOIND WORK bitcoind_work = variable.Variable((yield getwork(bitcoind))) @defer.inlineCallbacks def work_poller(): while True: flag = factory.new_block.get_deferred() try: bitcoind_work.set( (yield getwork(bitcoind, bitcoind_work.value['use_getblocktemplate']))) except: log.err() yield defer.DeferredList([flag, deferral.sleep(15)], fireOnOneCallback=True) work_poller() # PEER WORK best_block_header = variable.Variable(None) def handle_header(new_header): # check that header matches current target if not (net.PARENT.POW_FUNC( bitcoin_data.block_header_type.pack(new_header)) <= bitcoind_work.value['bits'].target): return bitcoind_best_block = bitcoind_work.value['previous_block'] if (best_block_header.value is None or (new_header['previous_block'] == bitcoind_best_block and bitcoin_data.hash256( bitcoin_data.block_header_type.pack( best_block_header.value)) == bitcoind_best_block ) # new is child of current and previous is current or (bitcoin_data.hash256( bitcoin_data.block_header_type.pack(new_header)) == bitcoind_best_block and best_block_header.value['previous_block'] != bitcoind_best_block) ): # new is current and previous is not a child of current best_block_header.set(new_header) @defer.inlineCallbacks def poll_header(): handle_header((yield factory.conn.value.get_block_header( bitcoind_work.value['previous_block']))) bitcoind_work.changed.watch(lambda _: poll_header()) yield deferral.retry('Error while requesting best block header:')( poll_header)() # BEST SHARE get_height_rel_highest = yield height_tracker.get_height_rel_highest_func( bitcoind, factory, lambda: bitcoind_work.value['previous_block'], net) best_share_var = variable.Variable(None) desired_var = variable.Variable(None) def set_best_share(): best, desired = tracker.think( get_height_rel_highest, bitcoind_work.value['previous_block'], bitcoind_work.value['bits']) best_share_var.set(best) desired_var.set(desired) bitcoind_work.changed.watch(lambda _: set_best_share()) set_best_share() print ' ...success!' print # setup p2p logic and join p2pool network class Node(p2p.Node): def handle_shares(self, shares, peer): if len(shares) > 5: print 'Processing %i shares from %s...' % ( len(shares), '%s:%i' % peer.addr if peer is not None else None) new_count = 0 for share in shares: if share.hash in tracker.items: #print 'Got duplicate share, ignoring. Hash: %s' % (p2pool_data.format_hash(share.hash),) continue new_count += 1 #print 'Received share %s from %r' % (p2pool_data.format_hash(share.hash), share.peer.addr if share.peer is not None else None) tracker.add(share) if new_count: set_best_share() if len(shares) > 5: print '... done processing %i shares. New: %i Have: %i/~%i' % ( len(shares), new_count, len( tracker.items), 2 * net.CHAIN_LENGTH) @defer.inlineCallbacks def handle_share_hashes(self, hashes, peer): new_hashes = [x for x in hashes if x not in tracker.items] if not new_hashes: return try: shares = yield peer.get_shares( hashes=new_hashes, parents=0, stops=[], ) except: log.err(None, 'in handle_share_hashes:') else: self.handle_shares(shares, peer) def handle_get_shares(self, hashes, parents, stops, peer): parents = min(parents, 1000 // len(hashes)) stops = set(stops) shares = [] for share_hash in hashes: for share in tracker.get_chain( share_hash, min(parents + 1, tracker.get_height(share_hash))): if share.hash in stops: break shares.append(share) print 'Sending %i shares to %s:%i' % ( len(shares), peer.addr[0], peer.addr[1]) return shares def handle_bestblock(self, header, peer): if net.PARENT.POW_FUNC( bitcoin_data.block_header_type.pack( header)) > header['bits'].target: raise p2p.PeerMisbehavingError( 'received block header fails PoW test') handle_header(header) @deferral.retry('Error submitting primary block: (will retry)', 10, 10) def submit_block_p2p(block): if factory.conn.value is None: print >> sys.stderr, 'No bitcoind connection when block submittal attempted! %s%064x' % ( net.PARENT.BLOCK_EXPLORER_URL_PREFIX, bitcoin_data.hash256( bitcoin_data.block_header_type.pack(block['header']))) raise deferral.RetrySilentlyException() factory.conn.value.send_block(block=block) @deferral.retry('Error submitting block: (will retry)', 10, 10) @defer.inlineCallbacks def submit_block_rpc(block, ignore_failure): if bitcoind_work.value['use_getblocktemplate']: result = yield bitcoind.rpc_submitblock( bitcoin_data.block_type.pack(block).encode('hex')) success = result is None else: result = yield bitcoind.rpc_getmemorypool( bitcoin_data.block_type.pack(block).encode('hex')) success = result success_expected = net.PARENT.POW_FUNC( bitcoin_data.block_header_type.pack( block['header'])) <= block['header']['bits'].target if (not success and success_expected and not ignore_failure) or (success and not success_expected): print >> sys.stderr, 'Block submittal result: %s (%r) Expected: %s' % ( success, result, success_expected) def submit_block(block, ignore_failure): submit_block_p2p(block) submit_block_rpc(block, ignore_failure) @tracker.verified.added.watch def _(share): if share.pow_hash <= share.header['bits'].target: submit_block(share.as_block(tracker), ignore_failure=True) print print 'GOT BLOCK FROM PEER! Passing to bitcoind! %s bitcoin: %s%064x' % ( p2pool_data.format_hash(share.hash), net.PARENT.BLOCK_EXPLORER_URL_PREFIX, share.header_hash) print def spread(): if (get_height_rel_highest(share.header['previous_block']) > -5 or bitcoind_work.value['previous_block'] in [ share.header['previous_block'], share.header_hash ]): broadcast_share(share.hash) spread() reactor.callLater( 5, spread) # so get_height_rel_highest can update print 'Joining p2pool network using port %i...' % (args.p2pool_port, ) @defer.inlineCallbacks def parse(x): if ':' in x: ip, port = x.split(':') defer.returnValue(((yield reactor.resolve(ip)), int(port))) else: defer.returnValue(((yield reactor.resolve(x)), net.P2P_PORT)) addrs = {} if os.path.exists(os.path.join(datadir_path, 'addrs')): try: with open(os.path.join(datadir_path, 'addrs'), 'rb') as f: addrs.update( dict((tuple(k), v) for k, v in json.loads(f.read()))) except: print >> sys.stderr, 'error parsing addrs' for addr_df in map(parse, net.BOOTSTRAP_ADDRS): try: addr = yield addr_df if addr not in addrs: addrs[addr] = (0, time.time(), time.time()) except: log.err() connect_addrs = set() for addr_df in map(parse, args.p2pool_nodes): try: connect_addrs.add((yield addr_df)) except: log.err() p2p_node = Node( best_share_hash_func=lambda: best_share_var.value, port=args.p2pool_port, net=net, addr_store=addrs, connect_addrs=connect_addrs, max_incoming_conns=args.p2pool_conns, traffic_happened=traffic_happened, ) p2p_node.start() def save_addrs(): with open(os.path.join(datadir_path, 'addrs'), 'wb') as f: f.write(json.dumps(p2p_node.addr_store.items())) task.LoopingCall(save_addrs).start(60) @best_block_header.changed.watch def _(header): for peer in p2p_node.peers.itervalues(): peer.send_bestblock(header=header) @defer.inlineCallbacks def broadcast_share(share_hash): shares = [] for share in tracker.get_chain( share_hash, min(5, tracker.get_height(share_hash))): if share.hash in shared_share_hashes: break shared_share_hashes.add(share.hash) shares.append(share) for peer in list(p2p_node.peers.itervalues()): yield peer.sendShares( [share for share in shares if share.peer is not peer]) # send share when the chain changes to their chain best_share_var.changed.watch(broadcast_share) def save_shares(): for share in tracker.get_chain( best_share_var.value, min(tracker.get_height(best_share_var.value), 2 * net.CHAIN_LENGTH)): ss.add_share(share) if share.hash in tracker.verified.items: ss.add_verified_hash(share.hash) task.LoopingCall(save_shares).start(60) @apply @defer.inlineCallbacks def download_shares(): while True: desired = yield desired_var.get_when_satisfies( lambda val: len(val) != 0) peer2, share_hash = random.choice(desired) if len(p2p_node.peers) == 0: yield deferral.sleep(1) continue peer = random.choice(p2p_node.peers.values()) print 'Requesting parent share %s from %s' % ( p2pool_data.format_hash(share_hash), '%s:%i' % peer.addr) try: shares = yield peer.get_shares( hashes=[share_hash], parents=500, stops=[], ) except: log.err(None, 'in download_shares:') continue if not shares: yield deferral.sleep( 1 ) # sleep so we don't keep rerequesting the same share nobody has continue p2p_node.handle_shares(shares, peer) print ' ...success!' print if args.upnp: @defer.inlineCallbacks def upnp_thread(): while True: try: is_lan, lan_ip = yield ipdiscover.get_local_ip() if is_lan: pm = yield portmapper.get_port_mapper() yield pm._upnp.add_port_mapping( lan_ip, args.p2pool_port, args.p2pool_port, 'p2pool', 'TCP') except defer.TimeoutError: pass except: if p2pool.DEBUG: log.err(None, 'UPnP error:') yield deferral.sleep(random.expovariate(1 / 120)) upnp_thread() # start listening for workers with a JSON-RPC server print 'Listening for workers on %r port %i...' % (worker_endpoint[0], worker_endpoint[1]) get_current_txouts = lambda: p2pool_data.get_expected_payouts( tracker, best_share_var.value, bitcoind_work.value['bits'].target, bitcoind_work.value['subsidy'], net) wb = work.WorkerBridge(my_pubkey_hash, net, args.donation_percentage, bitcoind_work, best_block_header, merged_urls, best_share_var, tracker, my_share_hashes, my_doa_share_hashes, args.worker_fee, p2p_node, submit_block, set_best_share, broadcast_share, block_height_var) web_root = web.get_web_root( tracker, bitcoind_work, get_current_txouts, datadir_path, net, wb.get_stale_counts, my_pubkey_hash, wb.local_rate_monitor, args.worker_fee, p2p_node, wb.my_share_hashes, wb.pseudoshare_received, wb.share_received, best_share_var, bitcoind_warning_var, traffic_happened) worker_interface.WorkerInterface(wb).attach_to( web_root, get_handler=lambda request: request.redirect('/static/')) deferral.retry('Error binding to worker port:', traceback=False)( reactor.listenTCP)(worker_endpoint[1], server.Site(web_root), interface=worker_endpoint[0]) with open(os.path.join(os.path.join(datadir_path, 'ready_flag')), 'wb') as f: pass print ' ...success!' print # done! print 'Started successfully!' print 'Go to http://127.0.0.1:%i/ to view graphs and statistics!' % ( worker_endpoint[1], ) if args.donation_percentage > 0.51: print '''Donating %.1f%% of work towards P2Pool's development. Thanks for the tip!''' % ( args.donation_percentage, ) elif args.donation_percentage < 0.49: print '''Donating %.1f%% of work towards P2Pool's development. Please donate to encourage further development of P2Pool!''' % ( args.donation_percentage, ) else: print '''Donating %.1f%% of work towards P2Pool's development. Thank you!''' % ( args.donation_percentage, ) print 'You can increase this amount with --give-author argument! (or decrease it, if you must)' print if hasattr(signal, 'SIGALRM'): signal.signal( signal.SIGALRM, lambda signum, frame: reactor.callFromThread( sys.stderr.write, 'Watchdog timer went off at:\n' + ''. join(traceback.format_stack()))) signal.siginterrupt(signal.SIGALRM, False) task.LoopingCall(signal.alarm, 30).start(1) if args.irc_announce: from twisted.words.protocols import irc class IRCClient(irc.IRCClient): nickname = 'p2pool%02i' % (random.randrange(100), ) channel = net.ANNOUNCE_CHANNEL def lineReceived(self, line): if p2pool.DEBUG: print repr(line) irc.IRCClient.lineReceived(self, line) def signedOn(self): irc.IRCClient.signedOn(self) self.factory.resetDelay() self.join(self.channel) @defer.inlineCallbacks def new_share(share): if share.pow_hash <= share.header[ 'bits'].target and abs(share.timestamp - time.time()) < 10 * 60: yield deferral.sleep(random.expovariate(1 / 60)) message = '\x02%s BLOCK FOUND by %s! %s%064x' % ( net.NAME.upper(), bitcoin_data.script2_to_address( share.new_script, net.PARENT), net.PARENT.BLOCK_EXPLORER_URL_PREFIX, share.header_hash) if all('%x' % (share.header_hash, ) not in old_message for old_message in self.recent_messages): self.say(self.channel, message) self._remember_message(message) self.watch_id = tracker.verified.added.watch(new_share) self.recent_messages = [] def _remember_message(self, message): self.recent_messages.append(message) while len(self.recent_messages) > 100: self.recent_messages.pop(0) def privmsg(self, user, channel, message): if channel == self.channel: self._remember_message(message) def connectionLost(self, reason): tracker.verified.added.unwatch(self.watch_id) print 'IRC connection lost:', reason.getErrorMessage() class IRCClientFactory(protocol.ReconnectingClientFactory): protocol = IRCClient reactor.connectTCP("irc.freenode.net", 6667, IRCClientFactory()) @defer.inlineCallbacks def status_thread(): last_str = None last_time = 0 while True: yield deferral.sleep(3) try: height = tracker.get_height(best_share_var.value) this_str = 'P2Pool: %i shares in chain (%i verified/%i total) Peers: %i (%i incoming)' % ( height, len(tracker.verified.items), len(tracker.items), len(p2p_node.peers), sum(1 for peer in p2p_node.peers.itervalues() if peer.incoming), ) + (' FDs: %i R/%i W' % (len(reactor.getReaders()), len(reactor.getWriters())) if p2pool.DEBUG else '') datums, dt = wb.local_rate_monitor.get_datums_in_last() my_att_s = sum(datum['work'] / dt for datum in datums) this_str += '\n Local: %sH/s in last %s Local dead on arrival: %s Expected time to share: %s' % ( math.format(int(my_att_s)), math.format_dt(dt), math.format_binomial_conf( sum(1 for datum in datums if datum['dead']), len(datums), 0.95), math.format_dt(2**256 / tracker.items[ best_share_var.value].max_target / my_att_s) if my_att_s and best_share_var.value else '???', ) if height > 2: (stale_orphan_shares, stale_doa_shares), shares, _ = wb.get_stale_counts() stale_prop = p2pool_data.get_average_stale_prop( tracker, best_share_var.value, min(60 * 60 // net.SHARE_PERIOD, height)) real_att_s = p2pool_data.get_pool_attempts_per_second( tracker, best_share_var.value, min(height - 1, 60 * 60 // net.SHARE_PERIOD)) / (1 - stale_prop) this_str += '\n Shares: %i (%i orphan, %i dead) Stale rate: %s Efficiency: %s Current payout: %.4f %s' % ( shares, stale_orphan_shares, stale_doa_shares, math.format_binomial_conf( stale_orphan_shares + stale_doa_shares, shares, 0.95), math.format_binomial_conf( stale_orphan_shares + stale_doa_shares, shares, 0.95, lambda x: (1 - x) / (1 - stale_prop)), get_current_txouts().get( bitcoin_data.pubkey_hash_to_script2( my_pubkey_hash), 0) * 1e-8, net.PARENT.SYMBOL, ) this_str += '\n Pool: %sH/s Stale rate: %.1f%% Expected time to block: %s' % ( math.format(int(real_att_s)), 100 * stale_prop, math.format_dt( 2**256 / bitcoind_work.value['bits'].target / real_att_s), ) for warning in p2pool_data.get_warnings( tracker, best_share_var.value, net, bitcoind_warning_var.value, bitcoind_work.value): print >> sys.stderr, '#' * 40 print >> sys.stderr, '>>> Warning: ' + warning print >> sys.stderr, '#' * 40 if this_str != last_str or time.time() > last_time + 15: print this_str last_str = this_str last_time = time.time() except: log.err() status_thread() except: reactor.stop() log.err(None, 'Fatal error:')
def get_current_txouts(self): real_subsidy = self.eternityd_work.value['subsidy'] real_pay = (real_subsidy) / 2 return p2pool_data.get_expected_payouts( self.tracker, self.best_share_var.value, self.eternityd_work.value['bits'].target, real_pay, self.net)