From 8dc1c72a0b9d4d7573d59672f2948f63d0af6ae9 Mon Sep 17 00:00:00 2001 From: Eugene Date: Fri, 29 Sep 2023 10:04:29 -0700 Subject: [PATCH 1/2] merge master into revolution --- .circleci/config.yml | 3 +- CHANGELOG.md | 33 ++++-- README.md | 6 +- bittensor/commands/list.py | 1 - bittensor/extrinsics/prometheus.py | 1 + bittensor/extrinsics/staking.py | 1 + bittensor/extrinsics/transfer.py | 1 + neurons/text/prompting/miners/BTLM/README.md | 117 +++++++++++++++++++ setup.py | 5 +- tests/integration_tests/test_cli.py | 32 ++--- 10 files changed, 160 insertions(+), 40 deletions(-) create mode 100644 neurons/text/prompting/miners/BTLM/README.md diff --git a/.circleci/config.yml b/.circleci/config.yml index db17d1104a..34f1050f6d 100644 --- a/.circleci/config.yml +++ b/.circleci/config.yml @@ -2,6 +2,7 @@ version: 2.1 orbs: python: circleci/python@2.1.1 + python-lib: dialogue/python-lib@0.1.55 # coveralls: coveralls/coveralls@1.0.6 jobs: @@ -242,7 +243,7 @@ workflows: pr-requirements: jobs: - black: - python-version: "3.9.13" + python-version: "3.8.12" - build-and-test: matrix: parameters: diff --git a/CHANGELOG.md b/CHANGELOG.md index 44f54aa42c..9b10736b03 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,27 @@ # Changelog +## 5.3.4 / 2023-08-16 + +# What's Changed +* Removes miniupnpc by @ifrit98 (completely unused and requires a sudo install) +* Fixes blacklist vpermit_required by @inquinim e80d3d5 +* Add try/except and timeout to version checking with exception handles by @ifrit98 a6a89fd +* Further updates CONTRIBUTING.md and DEVELOPMENT_WORKFLOW.md by @gitphantomman 3fefdbb +* Adds automatic compatibility checks to circleci for all major python3 supported versions. add checks by @ifrit98 #1484 + +**Full Changelog**: https://github.com/opentensor/bittensor/compare/v5.3.3...v5.3.4 + + +## 5.3.3 / 2023-07-26 + +## What's Changed +* Remove datasets requirement by @camfairchild in 2eabf0002b01 +* Relax bittensor-* requirements by @camfairchild in da9300ba5b2 + + +**Full Changelog**: https://github.com/opentensor/bittensor/compare/v5.3.2...v5.3.3 + + ## 5.3.2 / 2023-07-25 ## What's Changed @@ -15,16 +37,6 @@ ## 5.3.1 / 2023-07-06 -<<<<<<< HEAD -## What's Changed -* bump bittensor-wallet req, update cryptography security req by @@ifrit98 in [91d13b0](https://github.com/opentensor/bittensor/commit/91d13b0fa711621cbf823708d4368b1b387e42c4) -* Fixes Discord Link Issue #1442 by @camfairchild in [54d6248](https://github.com/opentensor/bittensor/commit/54d62487d4cb59e0b5edcd53acdca013108d155b) -* move mocks to bittensor_wallet package by @camfairchild in https://github.com/opentensor/bittensor/pull/1441 -* Bump bittensor-wallet version to 0.0.4 - -**Full Changelog**: https://github.com/opentensor/bittensor/compare/v5.3.0...v5.3.1 - -======= ## What's Changed * bump bittensor-wallet req, update cryptography security req by @@ifrit98 in [91d13b0](https://github.com/opentensor/bittensor/commit/91d13b0fa711621cbf823708d4368b1b387e42c4) * Fixes Discord Link Issue #1442 by @camfairchild in [54d6248](https://github.com/opentensor/bittensor/commit/54d62487d4cb59e0b5edcd53acdca013108d155b) @@ -32,7 +44,6 @@ * Bump bittensor-wallet version to 0.0.4 **Full Changelog**: https://github.com/opentensor/bittensor/compare/v5.3.0...v5.3.1 ->>>>>>> 18b8a7cdc5c00cf88dc2600c0e448a640df99caa ## 5.3.0 / 2023-07-04 diff --git a/README.md b/README.md index eb24ee8b1e..78af6c19e6 100644 --- a/README.md +++ b/README.md @@ -1,7 +1,7 @@
# **Bittensor** -[![Discord Chat](https://img.shields.io/discord/308323056592486420.svg)](https://discord.gg/realbittensor) +[![Discord Chat](https://img.shields.io/discord/308323056592486420.svg)](https://discord.gg/bittensor) [![PyPI version](https://badge.fury.io/py/bittensor.svg)](https://badge.fury.io/py/bittensor) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) @@ -9,7 +9,7 @@ ### Internet-scale Neural Networks -[Discord](https://discord.gg/bittensor) • [Network](https://taostats.io/) • [Research](https://drive.google.com/file/d/1VnsobL6lIAAqcA1_Tbm8AYIQscfJV4KU) +[Discord](https://discord.gg/bittensor) • [Network](https://taostats.io/) • [Research](https://bittensor.com/whitepaper)
@@ -17,7 +17,7 @@ Bittensor is a mining network, similar to Bitcoin, that includes built-in incent The token based mechanism under which the miners are incentivized ensures that they are constantly driven to make their knowledge output more useful, in terms of speed, intelligence and diversity. The value generated by the network is distributed directly to the individuals producing that value, without intermediarias. Anyone can participate in this endeavour, extract value from the network, and govern Bittensor. The network is open to all participants, and no individual or group has full control over what is learned, who can profit from it, or who can access it. -To learn more about Bittensor, please read our [paper](https://drive.google.com/file/d/1VnsobL6lIAAqcA1_Tbm8AYIQscfJV4KU/view). +To learn more about Bittensor, please read our [paper](https://bittensor.com/whitepaper). # Install There are three ways to install Bittensor diff --git a/bittensor/commands/list.py b/bittensor/commands/list.py index 97182b3e89..209ddf979d 100644 --- a/bittensor/commands/list.py +++ b/bittensor/commands/list.py @@ -86,6 +86,5 @@ def check_config(config: "bittensor.config"): @staticmethod def add_args(parser: argparse.ArgumentParser): list_parser = parser.add_parser("list", help="""List wallets""") - bittensor.wallet.add_args(list_parser) bittensor.subtensor.add_args(list_parser) diff --git a/bittensor/extrinsics/prometheus.py b/bittensor/extrinsics/prometheus.py index 55a3224852..bc80d12f2d 100644 --- a/bittensor/extrinsics/prometheus.py +++ b/bittensor/extrinsics/prometheus.py @@ -22,6 +22,7 @@ import bittensor.utils.networking as net + def prometheus_extrinsic( subtensor: "bittensor.subtensor", wallet: "bittensor.wallet", diff --git a/bittensor/extrinsics/staking.py b/bittensor/extrinsics/staking.py index 926717270c..eb0acfe9dc 100644 --- a/bittensor/extrinsics/staking.py +++ b/bittensor/extrinsics/staking.py @@ -23,6 +23,7 @@ from bittensor.utils.balance import Balance + def add_stake_extrinsic( subtensor: "bittensor.subtensor", wallet: "bittensor.wallet", diff --git a/bittensor/extrinsics/transfer.py b/bittensor/extrinsics/transfer.py index 86fcce70a5..cd251316eb 100644 --- a/bittensor/extrinsics/transfer.py +++ b/bittensor/extrinsics/transfer.py @@ -24,6 +24,7 @@ from ..utils import is_valid_bittensor_address_or_public_key + def transfer_extrinsic( subtensor: "bittensor.subtensor", wallet: "bittensor.wallet", diff --git a/neurons/text/prompting/miners/BTLM/README.md b/neurons/text/prompting/miners/BTLM/README.md new file mode 100644 index 0000000000..71828ca82c --- /dev/null +++ b/neurons/text/prompting/miners/BTLM/README.md @@ -0,0 +1,117 @@ +## Bittensor LM (BTLM) Miner +Bittensor LM 1.3B Language Model +This code is for running the very small Bittenso Language Model created by Cerebras. + +# Example Usage +``` +python3 neurons/text/prompting/miners/cerebras/neuron.py +``` + +# Full Usage +``` +usage: neuron.py [-h] [--cerebras.device CEREBRAS.DEVICE] [--cerebras.max_length CEREBRAS.MAX_LENGTH] [--cerebras.do_sample] + [--cerebras.no_repeat_ngram_size CEREBRAS.NO_REPEAT_NGRAM_SIZE] + [--netuid NETUID] [--neuron.name NEURON.NAME] [--neuron.blocks_per_epoch NEURON.BLOCKS_PER_EPOCH] [--neuron.no_set_weights] + [--neuron.max_batch_size NEURON.MAX_BATCH_SIZE] [--neuron.max_sequence_len NEURON.MAX_SEQUENCE_LEN] + [--neuron.blacklist.hotkeys [NEURON.BLACKLIST.HOTKEYS ...]] [--neuron.blacklist.allow_non_registered] + [--neuron.blacklist.default_stake NEURON.BLACKLIST.DEFAULT_STAKE] [--neuron.default_priority NEURON.DEFAULT_PRIORITY] + [--wallet.name WALLET.NAME] [--wallet.hotkey WALLET.HOTKEY] [--wallet.path WALLET.PATH] [--wallet._mock] + [--wallet.reregister WALLET.REREGISTER] [--axon.priority.max_workers AXON.PRIORITY.MAX_WORKERS] + [--axon.priority.maxsize AXON.PRIORITY.MAXSIZE] [--axon.port AXON.PORT] [--axon.ip AXON.IP] + [--axon.external_port AXON.EXTERNAL_PORT] [--axon.external_ip AXON.EXTERNAL_IP] [--axon.max_workers AXON.MAX_WORKERS] + [--axon.maximum_concurrent_rpcs AXON.MAXIMUM_CONCURRENT_RPCS] [--subtensor.network SUBTENSOR.NETWORK] + [--subtensor.chain_endpoint SUBTENSOR.CHAIN_ENDPOINT] [--subtensor._mock] + [--subtensor.register.num_processes SUBTENSOR.REGISTER.NUM_PROCESSES] + [--subtensor.register.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL] [--subtensor.register.no_output_in_place] + [--subtensor.register.verbose] [--subtensor.register.cuda.use_cuda] [--subtensor.register.cuda.no_cuda] + [--subtensor.register.cuda.dev_id SUBTENSOR.REGISTER.CUDA.DEV_ID [SUBTENSOR.REGISTER.CUDA.DEV_ID ...]] + [--subtensor.register.cuda.TPB SUBTENSOR.REGISTER.CUDA.TPB] [--logging.debug] [--logging.trace] [--logging.record_log] + [--logging.logging_dir LOGGING.LOGGING_DIR] [--metagraph._mock] [--config CONFIG] [--strict] + +optional arguments: + -h, --help show this help message and exit + --cerebras.device CEREBRAS.DEVICE + Device to load model + --cerebras.max_length CEREBRAS.MAX_LENGTH + The maximum length (in tokens) of the generated text. + --cerebras.do_sample Whether to use sampling or not (if not, uses greedy decoding). + --cerebras.no_repeat_ngram_size CEREBRAS.NO_REPEAT_NGRAM_SIZE + The size of the n-grams to avoid repeating in the generated text. + --cerebras.model_size {1.3B,2.7B,6.7B,13B} + Model size to use. + --netuid NETUID Subnet netuid + --neuron.name NEURON.NAME + Trials for this miner go in miner.root / (wallet_cold - wallet_hot) / miner.name + --neuron.blocks_per_epoch NEURON.BLOCKS_PER_EPOCH + Blocks until the miner sets weights on chain + --neuron.no_set_weights + If True, the model does not set weights. + --neuron.max_batch_size NEURON.MAX_BATCH_SIZE + The maximum batch size for forward requests. + --neuron.max_sequence_len NEURON.MAX_SEQUENCE_LEN + The maximum sequence length for forward requests. + --neuron.blacklist.hotkeys [NEURON.BLACKLIST.HOTKEYS ...] + To blacklist certain hotkeys + --neuron.blacklist.allow_non_registered + If True, the miner will allow non-registered hotkeys to mine. + --neuron.blacklist.default_stake NEURON.BLACKLIST.DEFAULT_STAKE + Set default stake for miners. + --neuron.default_priority NEURON.DEFAULT_PRIORITY + Set default priority for miners. + --wallet.name WALLET.NAME + The name of the wallet to unlock for running bittensor (name mock is reserved for mocking this wallet) + --wallet.hotkey WALLET.HOTKEY + The name of wallet's hotkey. + --wallet.path WALLET.PATH + The path to your bittensor wallets + --wallet._mock To turn on wallet mocking for testing purposes. + --wallet.reregister WALLET.REREGISTER + Whether to reregister the wallet if it is not already registered. + --axon.priority.max_workers AXON.PRIORITY.MAX_WORKERS + maximum number of threads in thread pool + --axon.priority.maxsize AXON.PRIORITY.MAXSIZE + maximum size of tasks in priority queue + --axon.port AXON.PORT + The local port this axon endpoint is bound to. i.e. 8091 + --axon.ip AXON.IP The local ip this axon binds to. ie. [::] + --axon.external_port AXON.EXTERNAL_PORT + The public port this axon broadcasts to the network. i.e. 8091 + --axon.external_ip AXON.EXTERNAL_IP + The external ip this axon broadcasts to the network to. ie. [::] + --axon.max_workers AXON.MAX_WORKERS + The maximum number connection handler threads working simultaneously on this endpoint. The grpc server distributes + new worker threads to service requests up to this number. + --axon.maximum_concurrent_rpcs AXON.MAXIMUM_CONCURRENT_RPCS + Maximum number of allowed active connections + --subtensor.network SUBTENSOR.NETWORK + The subtensor network flag. The likely choices are: -- finney (main network) -- local (local running network) -- + mock (creates a mock connection (for testing)) If this option is set it overloads subtensor.chain_endpoint with an + entry point node from that network. + --subtensor.chain_endpoint SUBTENSOR.CHAIN_ENDPOINT + The subtensor endpoint flag. If set, overrides the --network flag. + --subtensor._mock To turn on subtensor mocking for testing purposes. + --subtensor.register.num_processes SUBTENSOR.REGISTER.NUM_PROCESSES, -n SUBTENSOR.REGISTER.NUM_PROCESSES + Number of processors to use for registration + --subtensor.register.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL, --subtensor.register.cuda.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL, --cuda.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL, -u SUBTENSOR.REGISTER.UPDATE_INTERVAL + The number of nonces to process before checking for next block during registration + --subtensor.register.no_output_in_place, --no_output_in_place + Whether to not ouput the registration statistics in-place. Set flag to disable output in-place. + --subtensor.register.verbose + Whether to ouput the registration statistics verbosely. + --subtensor.register.cuda.use_cuda, --cuda, --cuda.use_cuda + Set flag to use CUDA to register. + --subtensor.register.cuda.no_cuda, --no_cuda, --cuda.no_cuda + Set flag to not use CUDA for registration + --subtensor.register.cuda.dev_id SUBTENSOR.REGISTER.CUDA.DEV_ID [SUBTENSOR.REGISTER.CUDA.DEV_ID ...], --cuda.dev_id SUBTENSOR.REGISTER.CUDA.DEV_ID [SUBTENSOR.REGISTER.CUDA.DEV_ID ...] + Set the CUDA device id(s). Goes by the order of speed. (i.e. 0 is the fastest). + --subtensor.register.cuda.TPB SUBTENSOR.REGISTER.CUDA.TPB, --cuda.TPB SUBTENSOR.REGISTER.CUDA.TPB + Set the number of Threads Per Block for CUDA. + --logging.debug Turn on bittensor debugging information + --logging.trace Turn on bittensor trace level information + --logging.record_log Turns on logging to file. + --logging.logging_dir LOGGING.LOGGING_DIR + Logging default root directory. + --metagraph._mock To turn on metagraph mocking for testing purposes. + --config CONFIG If set, defaults are overridden by passed file. + --strict If flagged, config will check that only exact arguemnts have been set. + ``` \ No newline at end of file diff --git a/setup.py b/setup.py index 0a84a8b825..6a3707d106 100644 --- a/setup.py +++ b/setup.py @@ -77,7 +77,9 @@ def read_requirements(path): license="MIT", python_requires=">=3.8", install_requires=requirements, - extras_require={"dev": extra_requirements_dev}, + extras_require={ + "dev": extra_requirements_dev, + }, scripts=["bin/btcli"], classifiers=[ "Development Status :: 3 - Alpha", @@ -89,7 +91,6 @@ def read_requirements(path): "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", - "Programming Language :: Python :: 3.11", "Topic :: Scientific/Engineering", "Topic :: Scientific/Engineering :: Mathematics", "Topic :: Scientific/Engineering :: Artificial Intelligence", diff --git a/tests/integration_tests/test_cli.py b/tests/integration_tests/test_cli.py index 49cf805f99..5db68fe23b 100644 --- a/tests/integration_tests/test_cli.py +++ b/tests/integration_tests/test_cli.py @@ -803,8 +803,7 @@ def test_stake_with_specific_hotkeys(self, _): ) success, err = _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -879,8 +878,7 @@ def test_stake_with_all_hotkeys(self, _): # Set the coldkey balance success, err = _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -978,8 +976,7 @@ def test_stake_with_exclude_hotkeys_from_all(self, _): # Set the coldkey balance _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -1093,8 +1090,7 @@ def test_stake_with_multiple_hotkeys_max_stake(self, _): ) _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -1192,8 +1188,7 @@ def test_stake_with_multiple_hotkeys_max_stake_not_enough_balance(self, _): ) _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -1286,8 +1281,7 @@ def test_stake_with_single_hotkey_max_stake(self, _): ) _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -1374,8 +1368,7 @@ def test_stake_with_single_hotkey_max_stake_not_enough_balance(self, _): ) _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -1469,8 +1462,7 @@ def test_stake_with_single_hotkey_max_stake_enough_stake(self, _): ) success, err = _subtensor_mock.force_set_balance( - ss58_address=mock_coldkey_kp.ss58_address, - balance=mock_balance.rao, + ss58_address=mock_coldkey_kp.ss58_address, balance=mock_balance.rao ) cli = bittensor.cli(config) @@ -1630,9 +1622,7 @@ def test_delegate_stake(self, _): ) # Make the first wallet a delegate - success = _subtensor_mock.nominate( - wallet=mock_wallets[0], - ) + success = _subtensor_mock.nominate(wallet=mock_wallets[0]) self.assertTrue(success) cli = bittensor.cli(config) @@ -1718,9 +1708,7 @@ def test_undelegate_stake(self, _): ) # Make the first wallet a delegate - success = _subtensor_mock.nominate( - wallet=mock_wallets[0], - ) + success = _subtensor_mock.nominate(wallet=mock_wallets[0]) self.assertTrue(success) # Stake to the delegate From aabff76a3a26efb5332837521d4f2d6f6ab3451f Mon Sep 17 00:00:00 2001 From: Eugene Date: Fri, 29 Sep 2023 10:06:34 -0700 Subject: [PATCH 2/2] remove depreicated content --- neurons/text/prompting/miners/BTLM/README.md | 117 ------------------- 1 file changed, 117 deletions(-) delete mode 100644 neurons/text/prompting/miners/BTLM/README.md diff --git a/neurons/text/prompting/miners/BTLM/README.md b/neurons/text/prompting/miners/BTLM/README.md deleted file mode 100644 index 71828ca82c..0000000000 --- a/neurons/text/prompting/miners/BTLM/README.md +++ /dev/null @@ -1,117 +0,0 @@ -## Bittensor LM (BTLM) Miner -Bittensor LM 1.3B Language Model -This code is for running the very small Bittenso Language Model created by Cerebras. - -# Example Usage -``` -python3 neurons/text/prompting/miners/cerebras/neuron.py -``` - -# Full Usage -``` -usage: neuron.py [-h] [--cerebras.device CEREBRAS.DEVICE] [--cerebras.max_length CEREBRAS.MAX_LENGTH] [--cerebras.do_sample] - [--cerebras.no_repeat_ngram_size CEREBRAS.NO_REPEAT_NGRAM_SIZE] - [--netuid NETUID] [--neuron.name NEURON.NAME] [--neuron.blocks_per_epoch NEURON.BLOCKS_PER_EPOCH] [--neuron.no_set_weights] - [--neuron.max_batch_size NEURON.MAX_BATCH_SIZE] [--neuron.max_sequence_len NEURON.MAX_SEQUENCE_LEN] - [--neuron.blacklist.hotkeys [NEURON.BLACKLIST.HOTKEYS ...]] [--neuron.blacklist.allow_non_registered] - [--neuron.blacklist.default_stake NEURON.BLACKLIST.DEFAULT_STAKE] [--neuron.default_priority NEURON.DEFAULT_PRIORITY] - [--wallet.name WALLET.NAME] [--wallet.hotkey WALLET.HOTKEY] [--wallet.path WALLET.PATH] [--wallet._mock] - [--wallet.reregister WALLET.REREGISTER] [--axon.priority.max_workers AXON.PRIORITY.MAX_WORKERS] - [--axon.priority.maxsize AXON.PRIORITY.MAXSIZE] [--axon.port AXON.PORT] [--axon.ip AXON.IP] - [--axon.external_port AXON.EXTERNAL_PORT] [--axon.external_ip AXON.EXTERNAL_IP] [--axon.max_workers AXON.MAX_WORKERS] - [--axon.maximum_concurrent_rpcs AXON.MAXIMUM_CONCURRENT_RPCS] [--subtensor.network SUBTENSOR.NETWORK] - [--subtensor.chain_endpoint SUBTENSOR.CHAIN_ENDPOINT] [--subtensor._mock] - [--subtensor.register.num_processes SUBTENSOR.REGISTER.NUM_PROCESSES] - [--subtensor.register.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL] [--subtensor.register.no_output_in_place] - [--subtensor.register.verbose] [--subtensor.register.cuda.use_cuda] [--subtensor.register.cuda.no_cuda] - [--subtensor.register.cuda.dev_id SUBTENSOR.REGISTER.CUDA.DEV_ID [SUBTENSOR.REGISTER.CUDA.DEV_ID ...]] - [--subtensor.register.cuda.TPB SUBTENSOR.REGISTER.CUDA.TPB] [--logging.debug] [--logging.trace] [--logging.record_log] - [--logging.logging_dir LOGGING.LOGGING_DIR] [--metagraph._mock] [--config CONFIG] [--strict] - -optional arguments: - -h, --help show this help message and exit - --cerebras.device CEREBRAS.DEVICE - Device to load model - --cerebras.max_length CEREBRAS.MAX_LENGTH - The maximum length (in tokens) of the generated text. - --cerebras.do_sample Whether to use sampling or not (if not, uses greedy decoding). - --cerebras.no_repeat_ngram_size CEREBRAS.NO_REPEAT_NGRAM_SIZE - The size of the n-grams to avoid repeating in the generated text. - --cerebras.model_size {1.3B,2.7B,6.7B,13B} - Model size to use. - --netuid NETUID Subnet netuid - --neuron.name NEURON.NAME - Trials for this miner go in miner.root / (wallet_cold - wallet_hot) / miner.name - --neuron.blocks_per_epoch NEURON.BLOCKS_PER_EPOCH - Blocks until the miner sets weights on chain - --neuron.no_set_weights - If True, the model does not set weights. - --neuron.max_batch_size NEURON.MAX_BATCH_SIZE - The maximum batch size for forward requests. - --neuron.max_sequence_len NEURON.MAX_SEQUENCE_LEN - The maximum sequence length for forward requests. - --neuron.blacklist.hotkeys [NEURON.BLACKLIST.HOTKEYS ...] - To blacklist certain hotkeys - --neuron.blacklist.allow_non_registered - If True, the miner will allow non-registered hotkeys to mine. - --neuron.blacklist.default_stake NEURON.BLACKLIST.DEFAULT_STAKE - Set default stake for miners. - --neuron.default_priority NEURON.DEFAULT_PRIORITY - Set default priority for miners. - --wallet.name WALLET.NAME - The name of the wallet to unlock for running bittensor (name mock is reserved for mocking this wallet) - --wallet.hotkey WALLET.HOTKEY - The name of wallet's hotkey. - --wallet.path WALLET.PATH - The path to your bittensor wallets - --wallet._mock To turn on wallet mocking for testing purposes. - --wallet.reregister WALLET.REREGISTER - Whether to reregister the wallet if it is not already registered. - --axon.priority.max_workers AXON.PRIORITY.MAX_WORKERS - maximum number of threads in thread pool - --axon.priority.maxsize AXON.PRIORITY.MAXSIZE - maximum size of tasks in priority queue - --axon.port AXON.PORT - The local port this axon endpoint is bound to. i.e. 8091 - --axon.ip AXON.IP The local ip this axon binds to. ie. [::] - --axon.external_port AXON.EXTERNAL_PORT - The public port this axon broadcasts to the network. i.e. 8091 - --axon.external_ip AXON.EXTERNAL_IP - The external ip this axon broadcasts to the network to. ie. [::] - --axon.max_workers AXON.MAX_WORKERS - The maximum number connection handler threads working simultaneously on this endpoint. The grpc server distributes - new worker threads to service requests up to this number. - --axon.maximum_concurrent_rpcs AXON.MAXIMUM_CONCURRENT_RPCS - Maximum number of allowed active connections - --subtensor.network SUBTENSOR.NETWORK - The subtensor network flag. The likely choices are: -- finney (main network) -- local (local running network) -- - mock (creates a mock connection (for testing)) If this option is set it overloads subtensor.chain_endpoint with an - entry point node from that network. - --subtensor.chain_endpoint SUBTENSOR.CHAIN_ENDPOINT - The subtensor endpoint flag. If set, overrides the --network flag. - --subtensor._mock To turn on subtensor mocking for testing purposes. - --subtensor.register.num_processes SUBTENSOR.REGISTER.NUM_PROCESSES, -n SUBTENSOR.REGISTER.NUM_PROCESSES - Number of processors to use for registration - --subtensor.register.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL, --subtensor.register.cuda.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL, --cuda.update_interval SUBTENSOR.REGISTER.UPDATE_INTERVAL, -u SUBTENSOR.REGISTER.UPDATE_INTERVAL - The number of nonces to process before checking for next block during registration - --subtensor.register.no_output_in_place, --no_output_in_place - Whether to not ouput the registration statistics in-place. Set flag to disable output in-place. - --subtensor.register.verbose - Whether to ouput the registration statistics verbosely. - --subtensor.register.cuda.use_cuda, --cuda, --cuda.use_cuda - Set flag to use CUDA to register. - --subtensor.register.cuda.no_cuda, --no_cuda, --cuda.no_cuda - Set flag to not use CUDA for registration - --subtensor.register.cuda.dev_id SUBTENSOR.REGISTER.CUDA.DEV_ID [SUBTENSOR.REGISTER.CUDA.DEV_ID ...], --cuda.dev_id SUBTENSOR.REGISTER.CUDA.DEV_ID [SUBTENSOR.REGISTER.CUDA.DEV_ID ...] - Set the CUDA device id(s). Goes by the order of speed. (i.e. 0 is the fastest). - --subtensor.register.cuda.TPB SUBTENSOR.REGISTER.CUDA.TPB, --cuda.TPB SUBTENSOR.REGISTER.CUDA.TPB - Set the number of Threads Per Block for CUDA. - --logging.debug Turn on bittensor debugging information - --logging.trace Turn on bittensor trace level information - --logging.record_log Turns on logging to file. - --logging.logging_dir LOGGING.LOGGING_DIR - Logging default root directory. - --metagraph._mock To turn on metagraph mocking for testing purposes. - --config CONFIG If set, defaults are overridden by passed file. - --strict If flagged, config will check that only exact arguemnts have been set. - ``` \ No newline at end of file