Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
13 changes: 0 additions & 13 deletions bittensor/_logging/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -304,25 +304,12 @@ def rpc_log(
synapse = synapse
)


@classmethod
def create_receptor_log( cls, endpoint: 'bittensor.Endpoint' ):
""" Debug logging for the connection between endpoints
"""
logger.debug( 'endpoint', receptor=True, action = '<green>' + 'Connect'.center(16) + '</green>', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) )

@classmethod
def update_receptor_log( cls, endpoint: 'bittensor.Endpoint' ):
""" Debug logging for updating the connection with endpoint
"""
logger.debug( 'endpoint', receptor=True, action = '<blue>' + 'Update'.center(16) + '</blue>', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) )

@classmethod
def destroy_receptor_log( cls, endpoint: 'bittensor.Endpoint' ):
""" Debug logging for destroying connection with endpoint
"""
logger.debug( 'endpoint', receptor=True, action = '<red>' + 'Destroy'.center(16) + '</red>', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) )

@classmethod
def success( cls, prefix:str, sufix:str ):
""" Success logging
Expand Down
2 changes: 1 addition & 1 deletion bittensor/_neuron/text/core_server/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -214,7 +214,7 @@ def run(

# Load/Create our bittensor wallet.
self.wallet.reregister(subtensor=self.subtensor, netuid = self.config.netuid)
self.metagraph.load().sync(netuid = self.config.netuid, subtensor=self.subtensor).save()
self.metagraph.sync(netuid = self.config.netuid, subtensor=self.subtensor).save()

# Create our optimizer.
optimizer = torch.optim.SGD(
Expand Down
11 changes: 6 additions & 5 deletions bittensor/_neuron/text/core_validator/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -844,7 +844,7 @@ def add_args( cls, parser ):
parser.add_argument('--nucleus.dropout', type=float, help='the dropout value', default=0.2)
parser.add_argument('--nucleus.importance', type=float, help='hyperparameter for the importance loss', default=3)
parser.add_argument('--nucleus.noise_multiplier', type=float, help='Standard deviation multipler on weights', default=2 )
parser.add_argument('--nucleus.no_dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False )
parser.add_argument('--nucleus.dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False )
parser.add_argument('--nucleus.scaling_law_power', type=float, help='Power for modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1)
parser.add_argument('--nucleus.synergy_scaling_law_power', type=float, help='Power for synergy modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1)
parser.add_argument('--nucleus.logits_divergence', type=float, help=' the divergence value for logit anomaly detection (default value: -1, pulling from subtensor directly)', default=-1)
Expand Down Expand Up @@ -985,10 +985,11 @@ def forward(
timeout=bittensor.__blocktime__
)

if self.config.nucleus.no_dendrite_backward:
query_responses = [[syn.detach().to(self.device) for syn in res] for res in query_responses]
return_ops = [ops.detach().to(self.device) for ops in return_ops]
times = [t.detach().to(self.device) for t in times]
if not self.config.nucleus.dendrite_backward:
query_responses = [[syn.detach() for syn in res] for res in query_responses]
return_ops = [ops.detach() for ops in return_ops]
times = [t.detach() for t in times]


# Send responses to device. This is required to ensure we move the responses
# Onto the correct device.
Expand Down
2 changes: 0 additions & 2 deletions bittensor/_receptor/receptor_pool_impl.py
Original file line number Diff line number Diff line change
Expand Up @@ -338,7 +338,6 @@ def _destroy_receptors_over_max_allowed( self ):

if receptor_to_remove != None:
try:
bittensor.logging.destroy_receptor_log(receptor_to_remove.endpoint)
self.receptors[ receptor_to_remove.endpoint.hotkey ].close()
del self.receptors[ receptor_to_remove.endpoint.hotkey ]
except KeyError:
Expand Down Expand Up @@ -370,7 +369,6 @@ def _get_or_create_receptor_for_endpoint( self, endpoint: 'bittensor.Endpoint' )

# ---- Or: Create a new receptor ----
else:
bittensor.logging.create_receptor_log( endpoint )
receptor = bittensor.receptor (
endpoint = endpoint,
wallet = self.wallet,
Expand Down
2 changes: 1 addition & 1 deletion bittensor/utils/weight_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -138,4 +138,4 @@ def convert_weights_and_uids_for_emit( uids: torch.LongTensor, weights: torch.Fl
weight_vals.append( uint16_val )
weight_uids.append( uid_i )

return weight_uids, weight_vals
return weight_uids, weight_vals