diff --git a/bittensor/_logging/__init__.py b/bittensor/_logging/__init__.py
index 6958a97d4a..13636849ed 100644
--- a/bittensor/_logging/__init__.py
+++ b/bittensor/_logging/__init__.py
@@ -304,25 +304,12 @@ def rpc_log(
synapse = synapse
)
-
- @classmethod
- def create_receptor_log( cls, endpoint: 'bittensor.Endpoint' ):
- """ Debug logging for the connection between endpoints
- """
- logger.debug( 'endpoint', receptor=True, action = '' + 'Connect'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) )
-
@classmethod
def update_receptor_log( cls, endpoint: 'bittensor.Endpoint' ):
""" Debug logging for updating the connection with endpoint
"""
logger.debug( 'endpoint', receptor=True, action = '' + 'Update'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) )
- @classmethod
- def destroy_receptor_log( cls, endpoint: 'bittensor.Endpoint' ):
- """ Debug logging for destroying connection with endpoint
- """
- logger.debug( 'endpoint', receptor=True, action = '' + 'Destroy'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) )
-
@classmethod
def success( cls, prefix:str, sufix:str ):
""" Success logging
diff --git a/bittensor/_neuron/text/core_server/__init__.py b/bittensor/_neuron/text/core_server/__init__.py
index 2440d9edc9..1e4d56d4e2 100644
--- a/bittensor/_neuron/text/core_server/__init__.py
+++ b/bittensor/_neuron/text/core_server/__init__.py
@@ -214,7 +214,7 @@ def run(
# Load/Create our bittensor wallet.
self.wallet.reregister(subtensor=self.subtensor, netuid = self.config.netuid)
- self.metagraph.load().sync(netuid = self.config.netuid, subtensor=self.subtensor).save()
+ self.metagraph.sync(netuid = self.config.netuid, subtensor=self.subtensor).save()
# Create our optimizer.
optimizer = torch.optim.SGD(
diff --git a/bittensor/_neuron/text/core_validator/__init__.py b/bittensor/_neuron/text/core_validator/__init__.py
index 1e5136669e..a02b94a1a0 100644
--- a/bittensor/_neuron/text/core_validator/__init__.py
+++ b/bittensor/_neuron/text/core_validator/__init__.py
@@ -844,7 +844,7 @@ def add_args( cls, parser ):
parser.add_argument('--nucleus.dropout', type=float, help='the dropout value', default=0.2)
parser.add_argument('--nucleus.importance', type=float, help='hyperparameter for the importance loss', default=3)
parser.add_argument('--nucleus.noise_multiplier', type=float, help='Standard deviation multipler on weights', default=2 )
- parser.add_argument('--nucleus.no_dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False )
+ parser.add_argument('--nucleus.dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False )
parser.add_argument('--nucleus.scaling_law_power', type=float, help='Power for modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1)
parser.add_argument('--nucleus.synergy_scaling_law_power', type=float, help='Power for synergy modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1)
parser.add_argument('--nucleus.logits_divergence', type=float, help=' the divergence value for logit anomaly detection (default value: -1, pulling from subtensor directly)', default=-1)
@@ -985,10 +985,11 @@ def forward(
timeout=bittensor.__blocktime__
)
- if self.config.nucleus.no_dendrite_backward:
- query_responses = [[syn.detach().to(self.device) for syn in res] for res in query_responses]
- return_ops = [ops.detach().to(self.device) for ops in return_ops]
- times = [t.detach().to(self.device) for t in times]
+ if not self.config.nucleus.dendrite_backward:
+ query_responses = [[syn.detach() for syn in res] for res in query_responses]
+ return_ops = [ops.detach() for ops in return_ops]
+ times = [t.detach() for t in times]
+
# Send responses to device. This is required to ensure we move the responses
# Onto the correct device.
diff --git a/bittensor/_receptor/receptor_pool_impl.py b/bittensor/_receptor/receptor_pool_impl.py
index db76bb3c5a..04ae83c368 100644
--- a/bittensor/_receptor/receptor_pool_impl.py
+++ b/bittensor/_receptor/receptor_pool_impl.py
@@ -338,7 +338,6 @@ def _destroy_receptors_over_max_allowed( self ):
if receptor_to_remove != None:
try:
- bittensor.logging.destroy_receptor_log(receptor_to_remove.endpoint)
self.receptors[ receptor_to_remove.endpoint.hotkey ].close()
del self.receptors[ receptor_to_remove.endpoint.hotkey ]
except KeyError:
@@ -370,7 +369,6 @@ def _get_or_create_receptor_for_endpoint( self, endpoint: 'bittensor.Endpoint' )
# ---- Or: Create a new receptor ----
else:
- bittensor.logging.create_receptor_log( endpoint )
receptor = bittensor.receptor (
endpoint = endpoint,
wallet = self.wallet,
diff --git a/bittensor/utils/weight_utils.py b/bittensor/utils/weight_utils.py
index 3a66baa2dc..c89d6a9c34 100644
--- a/bittensor/utils/weight_utils.py
+++ b/bittensor/utils/weight_utils.py
@@ -138,4 +138,4 @@ def convert_weights_and_uids_for_emit( uids: torch.LongTensor, weights: torch.Fl
weight_vals.append( uint16_val )
weight_uids.append( uid_i )
- return weight_uids, weight_vals
+ return weight_uids, weight_vals
\ No newline at end of file