From 09ae97f934b1aad849e4bbaa1478974c55cb835c Mon Sep 17 00:00:00 2001 From: isabella618033 Date: Tue, 21 Mar 2023 13:48:15 -0400 Subject: [PATCH 1/5] remove metagraph.load --- bittensor/_neuron/text/core_server/__init__.py | 2 +- bittensor/utils/weight_utils.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/bittensor/_neuron/text/core_server/__init__.py b/bittensor/_neuron/text/core_server/__init__.py index 2440d9edc9..1e4d56d4e2 100644 --- a/bittensor/_neuron/text/core_server/__init__.py +++ b/bittensor/_neuron/text/core_server/__init__.py @@ -214,7 +214,7 @@ def run( # Load/Create our bittensor wallet. self.wallet.reregister(subtensor=self.subtensor, netuid = self.config.netuid) - self.metagraph.load().sync(netuid = self.config.netuid, subtensor=self.subtensor).save() + self.metagraph.sync(netuid = self.config.netuid, subtensor=self.subtensor).save() # Create our optimizer. optimizer = torch.optim.SGD( diff --git a/bittensor/utils/weight_utils.py b/bittensor/utils/weight_utils.py index 3a66baa2dc..c89d6a9c34 100644 --- a/bittensor/utils/weight_utils.py +++ b/bittensor/utils/weight_utils.py @@ -138,4 +138,4 @@ def convert_weights_and_uids_for_emit( uids: torch.LongTensor, weights: torch.Fl weight_vals.append( uint16_val ) weight_uids.append( uid_i ) - return weight_uids, weight_vals + return weight_uids, weight_vals \ No newline at end of file From 370e97c195aa97e9df48430e9e7333e5cf2a8156 Mon Sep 17 00:00:00 2001 From: isabella618033 Date: Tue, 21 Mar 2023 13:50:00 -0400 Subject: [PATCH 2/5] remove destroy receptor log --- bittensor/_logging/__init__.py | 6 ------ bittensor/_receptor/receptor_pool_impl.py | 1 - 2 files changed, 7 deletions(-) diff --git a/bittensor/_logging/__init__.py b/bittensor/_logging/__init__.py index 6958a97d4a..1b57fcc392 100644 --- a/bittensor/_logging/__init__.py +++ b/bittensor/_logging/__init__.py @@ -317,12 +317,6 @@ def update_receptor_log( cls, endpoint: 'bittensor.Endpoint' ): """ logger.debug( 'endpoint', receptor=True, action = '' + 'Update'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) ) - @classmethod - def destroy_receptor_log( cls, endpoint: 'bittensor.Endpoint' ): - """ Debug logging for destroying connection with endpoint - """ - logger.debug( 'endpoint', receptor=True, action = '' + 'Destroy'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) ) - @classmethod def success( cls, prefix:str, sufix:str ): """ Success logging diff --git a/bittensor/_receptor/receptor_pool_impl.py b/bittensor/_receptor/receptor_pool_impl.py index db76bb3c5a..be0e952b17 100644 --- a/bittensor/_receptor/receptor_pool_impl.py +++ b/bittensor/_receptor/receptor_pool_impl.py @@ -338,7 +338,6 @@ def _destroy_receptors_over_max_allowed( self ): if receptor_to_remove != None: try: - bittensor.logging.destroy_receptor_log(receptor_to_remove.endpoint) self.receptors[ receptor_to_remove.endpoint.hotkey ].close() del self.receptors[ receptor_to_remove.endpoint.hotkey ] except KeyError: From 6895aff9f0616a7e0d2e6405d2797b01fd8f209a Mon Sep 17 00:00:00 2001 From: isabella618033 Date: Tue, 21 Mar 2023 13:54:04 -0400 Subject: [PATCH 3/5] remove create receptor log --- bittensor/_logging/__init__.py | 7 ------- bittensor/_receptor/receptor_pool_impl.py | 1 - 2 files changed, 8 deletions(-) diff --git a/bittensor/_logging/__init__.py b/bittensor/_logging/__init__.py index 1b57fcc392..13636849ed 100644 --- a/bittensor/_logging/__init__.py +++ b/bittensor/_logging/__init__.py @@ -304,13 +304,6 @@ def rpc_log( synapse = synapse ) - - @classmethod - def create_receptor_log( cls, endpoint: 'bittensor.Endpoint' ): - """ Debug logging for the connection between endpoints - """ - logger.debug( 'endpoint', receptor=True, action = '' + 'Connect'.center(16) + '', uid=str(endpoint.uid).center(4), hotkey=endpoint.hotkey, coldkey=endpoint.coldkey, ip_str=endpoint.ip_str().center(27) ) - @classmethod def update_receptor_log( cls, endpoint: 'bittensor.Endpoint' ): """ Debug logging for updating the connection with endpoint diff --git a/bittensor/_receptor/receptor_pool_impl.py b/bittensor/_receptor/receptor_pool_impl.py index be0e952b17..04ae83c368 100644 --- a/bittensor/_receptor/receptor_pool_impl.py +++ b/bittensor/_receptor/receptor_pool_impl.py @@ -369,7 +369,6 @@ def _get_or_create_receptor_for_endpoint( self, endpoint: 'bittensor.Endpoint' ) # ---- Or: Create a new receptor ---- else: - bittensor.logging.create_receptor_log( endpoint ) receptor = bittensor.receptor ( endpoint = endpoint, wallet = self.wallet, From 8e35e9959bb846c04c1fb042203f4f66dbdb40cc Mon Sep 17 00:00:00 2001 From: isabella618033 Date: Tue, 21 Mar 2023 14:27:06 -0400 Subject: [PATCH 4/5] validator dendrite backward default to false --- bittensor/_neuron/text/core_validator/__init__.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/bittensor/_neuron/text/core_validator/__init__.py b/bittensor/_neuron/text/core_validator/__init__.py index 1e5136669e..c7327176c2 100644 --- a/bittensor/_neuron/text/core_validator/__init__.py +++ b/bittensor/_neuron/text/core_validator/__init__.py @@ -844,7 +844,7 @@ def add_args( cls, parser ): parser.add_argument('--nucleus.dropout', type=float, help='the dropout value', default=0.2) parser.add_argument('--nucleus.importance', type=float, help='hyperparameter for the importance loss', default=3) parser.add_argument('--nucleus.noise_multiplier', type=float, help='Standard deviation multipler on weights', default=2 ) - parser.add_argument('--nucleus.no_dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False ) + parser.add_argument('--nucleus.dendrite_backward', action='store_true', help='Pass backward request to the server side or not', default=False ) parser.add_argument('--nucleus.scaling_law_power', type=float, help='Power for modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1) parser.add_argument('--nucleus.synergy_scaling_law_power', type=float, help='Power for synergy modified scaling law, powered down to improve dynamic range, e.g. 3 → 6 nats for 0.5. (default value: -1, pulling from subtensor directly)', default=-1) parser.add_argument('--nucleus.logits_divergence', type=float, help=' the divergence value for logit anomaly detection (default value: -1, pulling from subtensor directly)', default=-1) @@ -985,7 +985,7 @@ def forward( timeout=bittensor.__blocktime__ ) - if self.config.nucleus.no_dendrite_backward: + if not self.config.nucleus.dendrite_backward: query_responses = [[syn.detach().to(self.device) for syn in res] for res in query_responses] return_ops = [ops.detach().to(self.device) for ops in return_ops] times = [t.detach().to(self.device) for t in times] From cde7d2f01da8abc0ff2f36e493897abfebe5be29 Mon Sep 17 00:00:00 2001 From: isabella618033 Date: Tue, 21 Mar 2023 16:14:23 -0400 Subject: [PATCH 5/5] fix --- bittensor/_neuron/text/core_validator/__init__.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/bittensor/_neuron/text/core_validator/__init__.py b/bittensor/_neuron/text/core_validator/__init__.py index c7327176c2..a02b94a1a0 100644 --- a/bittensor/_neuron/text/core_validator/__init__.py +++ b/bittensor/_neuron/text/core_validator/__init__.py @@ -986,9 +986,10 @@ def forward( ) if not self.config.nucleus.dendrite_backward: - query_responses = [[syn.detach().to(self.device) for syn in res] for res in query_responses] - return_ops = [ops.detach().to(self.device) for ops in return_ops] - times = [t.detach().to(self.device) for t in times] + query_responses = [[syn.detach() for syn in res] for res in query_responses] + return_ops = [ops.detach() for ops in return_ops] + times = [t.detach() for t in times] + # Send responses to device. This is required to ensure we move the responses # Onto the correct device.