diff --git a/bittensor/_logging/__init__.py b/bittensor/_logging/__init__.py index 6958a97d4a..13636849ed 100644 --- a/bittensor/_logging/__init__.py +++ b/bittensor/_logging/__init__.py @@ -304,25 +304,12 @@ def rpc_log( synapse = synapse ) - - @classmethod - def create_receptor_log( cls, endpoint: 'bittensor.Endpoint' ): - """ Debug logging for the connection between endpoints - """ - logger.debug( 'endpoint', receptor=True, action = '' + 'Connect'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) ) - @classmethod def update_receptor_log( cls, endpoint: 'bittensor.Endpoint' ): """ Debug logging for updating the connection with endpoint """ logger.debug( 'endpoint', receptor=True, action = '' + 'Update'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) ) - @classmethod - def destroy_receptor_log( cls, endpoint: 'bittensor.Endpoint' ): - """ Debug logging for destroying connection with endpoint - """ - logger.debug( 'endpoint', receptor=True, action = '' + 'Destroy'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) ) - @classmethod def success( cls, prefix:str, sufix:str ): """ Success logging diff --git a/bittensor/_neuron/text/core_server/__init__.py b/bittensor/_neuron/text/core_server/__init__.py index 2440d9edc9..1e4d56d4e2 100644 --- a/bittensor/_neuron/text/core_server/__init__.py +++ b/bittensor/_neuron/text/core_server/__init__.py @@ -214,7 +214,7 @@ def run( # Load/Create our bittensor wallet. self.wallet.reregister(subtensor=self.subtensor, netuid = self.config.netuid) - self.metagraph.load().sync(netuid = self.config.netuid, subtensor=self.subtensor).save() + self.metagraph.sync(netuid = self.config.netuid, subtensor=self.subtensor).save() # Create our optimizer. optimizer = torch.optim.SGD( diff --git a/bittensor/_neuron/text/core_validator/__init__.py b/bittensor/_neuron/text/core_validator/__init__.py index 1e5136669e..a02b94a1a0 100644 --- a/bittensor/_neuron/text/core_validator/__init__.py +++ b/bittensor/_neuron/text/core_validator/__init__.py @@ -844,7 +844,7 @@ def add_args( cls, parser ): parser.add_argument('--nucleus.dropout', type=float, help='the dropout value', default=0.2) parser.add_argument('--nucleus.importance', type=float, help='hyperparameter for the importance loss', default=3) parser.add_argument('--nucleus.noise_multiplier', type=float, help='Standard deviation multipler on weights', default=2 ) - parser.add_argument('--nucleus.no_dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False ) + parser.add_argument('--nucleus.dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False ) parser.add_argument('--nucleus.scaling_law_power', type=float, help='Power for modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1) parser.add_argument('--nucleus.synergy_scaling_law_power', type=float, help='Power for synergy modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1) parser.add_argument('--nucleus.logits_divergence', type=float, help=' the divergence value for logit anomaly detection (default value: -1, pulling from subtensor directly)', default=-1) @@ -985,10 +985,11 @@ def forward( timeout=bittensor.__blocktime__ ) - if self.config.nucleus.no_dendrite_backward: - query_responses = [[syn.detach().to(self.device) for syn in res] for res in query_responses] - return_ops = [ops.detach().to(self.device) for ops in return_ops] - times = [t.detach().to(self.device) for t in times] + if not self.config.nucleus.dendrite_backward: + query_responses = [[syn.detach() for syn in res] for res in query_responses] + return_ops = [ops.detach() for ops in return_ops] + times = [t.detach() for t in times] + # Send responses to device. This is required to ensure we move the responses # Onto the correct device. diff --git a/bittensor/_receptor/receptor_pool_impl.py b/bittensor/_receptor/receptor_pool_impl.py index db76bb3c5a..04ae83c368 100644 --- a/bittensor/_receptor/receptor_pool_impl.py +++ b/bittensor/_receptor/receptor_pool_impl.py @@ -338,7 +338,6 @@ def _destroy_receptors_over_max_allowed( self ): if receptor_to_remove != None: try: - bittensor.logging.destroy_receptor_log(receptor_to_remove.endpoint) self.receptors[ receptor_to_remove.endpoint.hotkey ].close() del self.receptors[ receptor_to_remove.endpoint.hotkey ] except KeyError: @@ -370,7 +369,6 @@ def _get_or_create_receptor_for_endpoint( self, endpoint: 'bittensor.Endpoint' ) # ---- Or: Create a new receptor ---- else: - bittensor.logging.create_receptor_log( endpoint ) receptor = bittensor.receptor ( endpoint = endpoint, wallet = self.wallet, diff --git a/bittensor/utils/weight_utils.py b/bittensor/utils/weight_utils.py index 3a66baa2dc..c89d6a9c34 100644 --- a/bittensor/utils/weight_utils.py +++ b/bittensor/utils/weight_utils.py @@ -138,4 +138,4 @@ def convert_weights_and_uids_for_emit( uids: torch.LongTensor, weights: torch.Fl weight_vals.append( uint16_val ) weight_uids.append( uid_i ) - return weight_uids, weight_vals + return weight_uids, weight_vals \ No newline at end of file