From dc813c426e5aa2bce7d8d64b1055399ca7239dfa Mon Sep 17 00:00:00 2001 From: Aleksandr Borzunov Date: Wed, 30 Nov 2022 04:48:58 +0000 Subject: [PATCH] Add `src/petals/cli` --- src/petals/cli/__init__.py | 0 src/petals/cli/config.json | 20 + src/petals/cli/convert_model.py | 93 + src/petals/cli/deploy_server.sh | 79 + src/petals/cli/inference_one_block.py | 53 + .../cli/local_server_config_example.cfg | 5 + .../cli/remote_server_config_example.cfg | 6 + src/petals/cli/run_local_servers.sh | 109 + src/petals/cli/run_remote_servers.sh | 110 + src/petals/cli/run_server.py | 146 ++ src/petals/cli/speed_test.py | 1941 +++++++++++++++++ 11 files changed, 2562 insertions(+) create mode 100644 src/petals/cli/__init__.py create mode 100644 src/petals/cli/config.json create mode 100644 src/petals/cli/convert_model.py create mode 100644 src/petals/cli/deploy_server.sh create mode 100644 src/petals/cli/inference_one_block.py create mode 100644 src/petals/cli/local_server_config_example.cfg create mode 100644 src/petals/cli/remote_server_config_example.cfg create mode 100644 src/petals/cli/run_local_servers.sh create mode 100644 src/petals/cli/run_remote_servers.sh create mode 100644 src/petals/cli/run_server.py create mode 100755 src/petals/cli/speed_test.py diff --git a/src/petals/cli/__init__.py b/src/petals/cli/__init__.py new file mode 100644 index 0000000..e69de29 diff --git a/src/petals/cli/config.json b/src/petals/cli/config.json new file mode 100644 index 0000000..ca7ffbb --- /dev/null +++ b/src/petals/cli/config.json @@ -0,0 +1,20 @@ +{ + "apply_residual_connection_post_layernorm": false, + "attention_dropout": 0.0, + "attention_softmax_in_fp32": true, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_dropout": 0.0, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "masked_softmax_fusion": true, + "model_type": "bloom", + "n_embed": 14336, + "n_layer": 70, + "num_attention_heads": 112, + "pretraining_tp": 4, + "slow_but_exact": false, + "transformers_version": "4.20.0.dev0", + "use_cache": true, + "vocab_size": 250880 +} \ No newline at end of file diff --git a/src/petals/cli/convert_model.py b/src/petals/cli/convert_model.py new file mode 100644 index 0000000..9174847 --- /dev/null +++ b/src/petals/cli/convert_model.py @@ -0,0 +1,93 @@ +import argparse +import os + +import psutil +import torch.backends.quantized +import torch.nn as nn +import transformers +from hivemind.utils.logging import get_logger, use_hivemind_log_handler +from huggingface_hub import Repository +from tqdm.auto import tqdm + +from petals.client import BloomModel +from petals.bloom.from_pretrained import BLOCK_BRANCH_PREFIX, CLIENT_BRANCH +from petals.client import DistributedBloomConfig + +use_hivemind_log_handler("in_root_logger") +logger = get_logger(__file__) + +DTYPE_MAP = dict(bfloat16=torch.bfloat16, float16=torch.float16, float32=torch.float32, auto="auto") + + +if __name__ == "__main__": + parser = argparse.ArgumentParser(description="Load bloom layers and convert to 8-bit using torch quantization.") + + parser.add_argument("--model", type=str, default="bigscience/bloom-6b3", help="Model name for from_pretrained") + parser.add_argument("--revision", type=str, default=None, help="Optional commit id from HF hub") + parser.add_argument("--torch_dtype", type=str, default="auto", help="Load initial model in this dtype") + parser.add_argument("--output_path", type=str, default="./converted_model", help="Track output repo to this folder") + parser.add_argument("--output_repo", type=str, default="bigscience/test-bloomd", help="Push to this HF hub repo") + parser.add_argument("--client_branch", type=str, default=CLIENT_BRANCH, help="Save client version to this branch") + parser.add_argument( + "--block_branch_prefix", type=str, default=BLOCK_BRANCH_PREFIX, help="Save blocks to branches with this prefix" + ) + parser.add_argument( + "--commit_message", type=str, default="push-o-matic", help="Use this commit message for all parts" + ) + parser.add_argument("--use_auth_token", type=str, default=None, help="auth token for from_pretrained") + parser.add_argument("--resize_token_embeddings", type=int, default=None, help="change the vocabulary size") + args = parser.parse_args() + + free_ram_gb = psutil.virtual_memory().available / 2**30 + if args.model == "bigscience/bloom" and free_ram_gb < 400: + logger.warning(f"ACHTUNG! converting bloom-176b will use up 350-400GB RAM, you have {free_ram_gb:.3f} free") + + assert args.torch_dtype in DTYPE_MAP, f"torch_dtype must be one of {list(DTYPE_MAP.keys())}" + if os.path.exists(args.output_path) and ( + len(os.listdir(args.output_path)) != 0 or not os.path.isdir(args.output_path) + ): + raise FileExistsError(f"Output path {args.output_path} already exists and is not an empty directory") + + logger.info(f"Loading source model {args.model} (this may take a few minutes)") + config = DistributedBloomConfig.from_pretrained( + args.model, use_auth_token=args.use_auth_token, revision=args.revision + ) + config.dht_prefix = args.output_repo + + model = BloomModel.from_pretrained( + args.model, use_auth_token=args.use_auth_token, revision=args.revision, torch_dtype=DTYPE_MAP[args.torch_dtype] + ) + if args.resize_token_embeddings: + logger.info(f"Resizing token embeddings, new size = {args.resize_token_embeddings}") + model.resize_token_embeddings(args.resize_token_embeddings) + config.vocab_size = args.resize_token_embeddings + + tokenizer = transformers.AutoTokenizer.from_pretrained( + args.model, use_auth_token=args.use_auth_token, revision=args.revision + ) + os.makedirs(args.output_path, exist_ok=True) + + repo = Repository(args.output_path, clone_from=args.output_repo, use_auth_token=args.use_auth_token) + repo.git_pull() + + transformer_blocks = model.h + logger.info( + f"Saving transformer blocks to {args.output_repo}@{args.block_branch_prefix}0" + f" - {args.output_repo}@{args.block_branch_prefix}{len(transformer_blocks)}" + ) + for i, block in enumerate(tqdm(transformer_blocks)): + repo.git_checkout(args.client_branch, create_branch_ok=True) + with repo.commit( + commit_message=args.commit_message, branch=args.block_branch_prefix + str(i), track_large_files=True + ): + torch.save(block.state_dict(), "./pytorch_model.bin") + + logger.info(f"Saving client-side modules to {args.output_repo}@{args.client_branch}") + repo.git_checkout(args.client_branch, create_branch_ok=True) + with repo.commit(commit_message=args.commit_message, branch=args.client_branch, track_large_files=True): + model.h = nn.ModuleList() + model.save_pretrained(".") + tokenizer.save_pretrained(".") + config.save_pretrained(".") + + logger.info(f"Converted {args.model} and pushed to {args.output_repo}") diff --git a/src/petals/cli/deploy_server.sh b/src/petals/cli/deploy_server.sh new file mode 100644 index 0000000..6a87312 --- /dev/null +++ b/src/petals/cli/deploy_server.sh @@ -0,0 +1,79 @@ +#!/usr/bin/env bash + +################# +# Parse options # +################# + +instructions() { + echo "Usage: $0 [-m] [-i] [ -d ] [ -p ] [ -b ] [-a] [-t]" >&2 + echo " -m: model name" + echo " -i: initial peer" + echo " -d: device" >&2 + echo " -p: server identity path" >&2 + echo " -b: block_ids" >&2 + echo " -a: host maddrs" >&2 + echo " -t: whether to run local tests" >&2 + exit 1 +} + +if [ ! $# -ge 8 ]; then + instructions +fi + +while getopts ":m:i:d:p:b:a:t:" option; do + case $option in + m) MODEL_NAME=${OPTARG} + ;; + i) INITIAL_PEER=${OPTARG} + ;; + d) DEVICE=${OPTARG} + ;; + p) SERVER_ID_PATH=${OPTARG} + ;; + b) BLOCK_IDS=${OPTARG} + ;; + a) HOST_MADDR=${OPTARG} # TODO: allow several maddrs + ;; + t) RUN_LOCAL_TESTS=true + ;; + \?) instructions + ;; + esac +done + + +echo "==========" +echo "= Config =" +echo "==========" +echo "Model name: ${MODEL_NAME}" +echo "Initial peer: ${INITIAL_PEER}" +echo "Device: ${DEVICE}" +echo "Server name: ${SERVER_ID_PATH}" +echo "Server address: ${HOST_MADDR}" +echo "Bloom blocks: ${BLOCK_IDS}" + + +########################### +# Install or activate env # +########################### + +# TODO fix bug with self calling +source ~/miniconda3/etc/profile.d/conda.sh +if conda env list | grep ".*bloom-demo.*" >/dev/null 2>/dev/null; then + conda activate bloom-demo +else + conda create -y --name bloom-demo python=3.8.12 pip + conda activate bloom-demo + + conda install -y -c conda-forge cudatoolkit-dev==11.3.1 cudatoolkit==11.3.1 cudnn==8.2.1.32 + pip install -i https://pypi.org/simple torch==1.12.0+cu113 -f https://download.pytorch.org/whl/torch_stable.html + pip install -i https://pypi.org/simple . + pip install -i https://test.pypi.org/simple/ bitsandbytes-cuda113 +fi + +############## +# Run server # +############## + +python -m petals.cli.run_server --converted_model_name_or_path ${MODEL_NAME} --device ${DEVICE} --initial_peer ${INITIAL_PEER} \ + --block_indices ${BLOCK_IDS} --compression UNIFORM_8BIT --identity_path ${SERVER_ID_PATH} --host_maddrs ${HOST_MADDR} --load_in_8bit &> ${SERVER_ID_PATH}.log diff --git a/src/petals/cli/inference_one_block.py b/src/petals/cli/inference_one_block.py new file mode 100644 index 0000000..a5e5a00 --- /dev/null +++ b/src/petals/cli/inference_one_block.py @@ -0,0 +1,53 @@ +import argparse + +import torch +from hivemind.utils.logging import get_logger, use_hivemind_log_handler +from tqdm.auto import trange + +from petals.bloom.block import BloomBlock +from petals.bloom.model import BloomConfig +from petals.bloom.ops import build_alibi_tensor + +use_hivemind_log_handler("in_root_logger") +logger = get_logger(__file__) + +logger.warning("inference_one_block will soon be deprecated in favour of tests!") + + +def print_device_info(device=None): + """Prints device stats. Code from https://stackoverflow.com/a/53374933/12891528""" + device = torch.device(device or ("cuda" if torch.cuda.is_available() else "cpu")) + logger.info(f"Using device: {device}") + + # Additional Info when using cuda + if device.type == "cuda": + logger.info(torch.cuda.get_device_name(0)) + logger.info(f"Memory Usage:") + logger.info(f"Allocated: {round(torch.cuda.memory_allocated(0) / 1024 ** 3, 1)} GB") + logger.info(f"Cached: {round(torch.cuda.memory_cached(0) / 1024 ** 3, 1)} GB") + + +if __name__ == "__main__": + parser = argparse.ArgumentParser(description="Run a single bloom block locally on dummy data") + parser.add_argument("--config", required=True, type=str, help="Path to a config json file") + parser.add_argument("--state_dict", default=None, type=str, help="Optional path to saved block state dict") + parser.add_argument("--layer_index", default=0, type=int, help="Optional path to saved block state dict") + parser.add_argument("--num_steps", default=500, type=int, help="How many inference steps to run") + parser.add_argument("--device", default=None, type=str, help="Run inference on this device") + args = parser.parse_args() + + if args.device is None: + args.device = "cuda" if torch.cuda.is_available() else "cpu" + + config = BloomConfig.from_json_file(args.config) + block = BloomBlock(config, args.layer_index).to(args.device) + + cache = None + + for i in trange(args.num_steps): + dummy_input = torch.randn(1, 1, config.hidden_size, device=args.device) + alibi = build_alibi_tensor(i + 1, config.num_attention_heads).to(args.device) + with torch.no_grad(): + outputs, cache = block.forward(dummy_input, alibi=alibi, use_cache=True, layer_past=cache) + + print_device_info(args.device) diff --git a/src/petals/cli/local_server_config_example.cfg b/src/petals/cli/local_server_config_example.cfg new file mode 100644 index 0000000..8cbfe45 --- /dev/null +++ b/src/petals/cli/local_server_config_example.cfg @@ -0,0 +1,5 @@ +device=cpu +block_ids=2:3 +id_path=./server.id +maddr=/ip4/127.0.0.1/tcp/30000 +# diff --git a/src/petals/cli/remote_server_config_example.cfg b/src/petals/cli/remote_server_config_example.cfg new file mode 100644 index 0000000..54df7af --- /dev/null +++ b/src/petals/cli/remote_server_config_example.cfg @@ -0,0 +1,6 @@ +name=bloom-peer-0.bloom.net +device=cpu +block_ids=1:3 +id_path=./server.id +maddr=/ip4/0.0.0.0/tcp/30000 +# \ No newline at end of file diff --git a/src/petals/cli/run_local_servers.sh b/src/petals/cli/run_local_servers.sh new file mode 100644 index 0000000..0e449cb --- /dev/null +++ b/src/petals/cli/run_local_servers.sh @@ -0,0 +1,109 @@ +# !/usr/bin/env bash + +################# +# Parse options # +################# + +instructions() { + echo "Usage: $0 [-n] [-c]" >&2 + echo " -n: number of servers to run" >&2 + echo " -c: path to the server configs" >&2 + exit 1 +} + +if [ $# != 4 ]; then + instructions +fi + +while getopts ":n:c:t:" option; do + case $option in + n) NUM_SERVERS=${OPTARG} + ;; + c) CONFIG_PATH=${OPTARG} + ;; + \?) instructions + ;; + esac +done + + +########################### +# Install or activate env # +########################### + +source ~/miniconda3/etc/profile.d/conda.sh +if conda env list | grep ".*bloom-demo.*" >/dev/null 2>/dev/null; then + conda activate bloom-demo +else + conda create -y --name bloom-demo python=3.8.12 pip + conda activate bloom-demo + + conda install -y -c conda-forge cudatoolkit-dev==11.3.1 cudatoolkit==11.3.1 cudnn==8.2.1.32 + pip install -i https://pypi.org/simple torch==1.12.0+cu113 -f https://download.pytorch.org/whl/torch_stable.html + pip install -i https://pypi.org/simple -r . + pip install -i https://test.pypi.org/simple/ bitsandbytes-cuda113 +fi + + +####################### +# Create Initial peer # +####################### + +hivemind-dht &> tmp.out & +sleep 5 +INITIAL_PEER=$(python -c "with open('tmp.out') as f: print(f.readlines()[1].split()[-1])" ) +echo "Initial peer: ${INITIAL_PEER}" + + +############################## +# Initialize the config file # +############################## + +typeset -A cfg +cfg=( # set default values in config array + [device]="cpu" + [block_ids]="1:2" + [id_path]="server.id" + [maddr]="/ip4/127.0.0.1/tcp/30000" +) + +############### +# Run servers # +############### + +for SERVER_ID in $(seq 0 $(( $NUM_SERVERS - 1 )) ) +do + ############### + # Read config # + ############### + + while read line + do + if echo $line | grep -F = &>/dev/null + then + varname=$(echo "$line" | cut -d '=' -f 1) + cfg[$varname]=$(echo "$line" | cut -d '=' -f 2-) + fi + done < ${CONFIG_PATH}/server_${SERVER_ID}.cfg + + echo "=== Server #${SERVER_ID} ===" + echo "Server ID: ${cfg[id_path]}" + echo "Device: ${cfg[device]}" + echo "Bloom block ids: ${cfg[block_ids]}" + echo "Host maddr: ${cfg[maddr]}" + echo "" + + ############## + # Run server # + ############## + + tmux new-session -d -s "Server_${SERVER_ID}" bash cli/deploy_server.sh -m "bigscience/test-bloomd" -i ${INITIAL_PEER} -d ${cfg[device]} -p ${cfg[id_path]} -b ${cfg[block_ids]} -a ${cfg[maddr]} +done + +##################### +# Kill initial peer # +##################### + +sleep 10 +pkill -f hivemind-dht # TODO: kill only particular pids of hivemind-dht +rm tmp.out \ No newline at end of file diff --git a/src/petals/cli/run_remote_servers.sh b/src/petals/cli/run_remote_servers.sh new file mode 100644 index 0000000..e3f30de --- /dev/null +++ b/src/petals/cli/run_remote_servers.sh @@ -0,0 +1,110 @@ +# !/usr/bin/env bash + +SSH_KEY_PATH="~/.ssh/" + +################# +# Parse options # +################# + +instructions() { + echo "Usage: $0 [-u] [-n] [-c]" >&2 + echo " -u: username" >&2 + echo " -n: number of servers to run" >&2 + echo " -c: path to the server configs" >&2 + exit 1 +} + +if [ $# != 6 ]; then + instructions +fi + +while getopts ":u:n:c:" option; do + case $option in + u) USERNAME=${OPTARG} + ;; + n) NUM_SERVERS=${OPTARG} + ;; + c) CONFIG_PATH=${OPTARG} + ;; + \?) instructions + ;; + esac +done + + +########################### +# Install or activate env # +########################### + +source ~/miniconda3/etc/profile.d/conda.sh +if conda env list | grep ".*bloom-demo.*" >/dev/null 2>/dev/null; then + conda activate bloom-demo +else + conda create -y --name bloom-demo python=3.8.12 pip + conda activate bloom-demo + + conda install -y -c conda-forge cudatoolkit-dev==11.3.1 cudatoolkit==11.3.1 cudnn==8.2.1.32 + pip install -i https://pypi.org/simple torch==1.12.0+cu113 -f https://download.pytorch.org/whl/torch_stable.html + pip install -i https://pypi.org/simple -r . +fi + + +####################### +# Create Initial peer # +####################### + +hivemind-dht &> tmp.out & + +sleep 5 +INITIAL_PEER=$(python -c "with open('tmp.out') as f: print(f.readlines()[1].split()[-2])" ) +rm tmp.out +echo "Initial peer: ${INITIAL_PEER}" + + +############################## +# Initialize the config file # +############################## + +typeset -A cfg +cfg=( # set default values in config array + [name]="" + [device]="cpu" + [block_ids]="1:2" + [id_path]="server.id" + [maddr]="/ip4/0.0.0.0/tcp/30000" +) + +############### +# Run servers # +############### + +for SERVER_ID in $(seq 0 $(( $NUM_SERVERS - 1 )) ) +do + ############### + # Read config # + ############### + + while read line + do + if echo $line | grep -F = &>/dev/null + then + varname=$(echo "$line" | cut -d '=' -f 1) + cfg[$varname]=$(echo "$line" | cut -d '=' -f 2-) + fi + done < ${CONFIG_PATH}/server_${SERVER_ID}.cfg + + SERVER_NAME="${USERNAME}@${cfg[name]}" + echo "=== Server #${SERVER_ID} ===" + echo "Server name ${SERVER_NAME}" + echo "Server ID: ${cfg[id_path]}" + echo "Device: ${cfg[device]}" + echo "Bloom block ids: ${cfg[block_ids]}" + echo "Host maddr: ${cfg[maddr]}" + echo "=================" + + ############## + # Run server # + ############## + + ssh -i ${SSH_KEY_PATH} ${SERVER_NAME} "tmux new-session -d -s 'Server_${SERVER_ID}' 'cd bloom-demo && bash cli/deploy_server.sh -i ${INITIAL_PEER} -d ${cfg[device]} -p ${cfg[id_path]} -b ${cfg[block_ids]} -a ${cfg[maddr]}'" +done \ No newline at end of file diff --git a/src/petals/cli/run_server.py b/src/petals/cli/run_server.py new file mode 100644 index 0000000..e7b147e --- /dev/null +++ b/src/petals/cli/run_server.py @@ -0,0 +1,146 @@ +import argparse + +import configargparse +from hivemind.proto.runtime_pb2 import CompressionType +from hivemind.utils.limits import increase_file_limit +from hivemind.utils.logging import get_logger, use_hivemind_log_handler +from humanfriendly import parse_size + +from petals.constants import PUBLIC_INITIAL_PEERS +from petals.server.server import Server + +use_hivemind_log_handler("in_root_logger") +logger = get_logger(__file__) + + +def main(): + # fmt:off + parser = configargparse.ArgParser(default_config_files=["config.yml"], + formatter_class=argparse.ArgumentDefaultsHelpFormatter) + parser.add('-c', '--config', required=False, is_config_file=True, help='config file path') + + group = parser.add_mutually_exclusive_group(required=True) + group.add_argument('--converted_model_name_or_path', type=str, default=None, + help="path or name of a pretrained model, converted with cli/convert_model.py") + group.add_argument('model', nargs='?', type=str, help="same as --converted_model_name_or_path") + + parser.add_argument('--num_blocks', type=int, default=None, help="The number of blocks to serve") + parser.add_argument('--block_indices', type=str, default=None, help="Specific block indices to serve") + parser.add_argument('--prefix', type=str, default=None, help="Announce all blocks with this prefix. By default," + "use the same name as in the converted model.") + parser.add_argument('--host_maddrs', nargs='+', default=['/ip4/0.0.0.0/tcp/0', '/ip6/::/tcp/0'], required=False, + help='Multiaddrs to listen for external connections from other peers. Default: all IPv4/IPv6 interfaces, a random free TCP port') + parser.add_argument('--announce_maddrs', nargs='+', default=None, required=False, + help='Visible multiaddrs the host announces for external connections from other peers') + + parser.add_argument('--compression', type=str, default='NONE', required=False, help='Tensor compression communication') + + parser.add_argument('--num_handlers', type=int, default=8, required=False, + help='server will use this many processes to handle incoming requests') + parser.add_argument('--min_batch_size', type=int, default=1, + help='Minimum required batch size for all operations (in total tokens)') + parser.add_argument('--max_batch_size', type=int, default=2048, + help='The total number of tokens in the same batch will not exceed this value') + parser.add_argument('--prefetch_batches', type=int, default=1, required=False, + help='Pre-form this many subsequent batches while GPU is processing the current one') + parser.add_argument('--sender_threads', type=int, default=1, required=False, + help='Use this many threads to pass results/exceptions from Runtime to Pools') + parser.add_argument('--inference_max_length', type=int, default=2048, + help='Maximum total sequence length permitted per inference, defaults to 16384 tokens') + parser.add_argument('--cache_dir', type=str, default=None, + help='Path to a directory in which a downloaded pretrained model configuration should be cached if the standard cache should not be used.') + parser.add_argument('--device', type=str, default=None, required=False, + help='all blocks will use this device in torch notation; default: cuda if available else cpu') + parser.add_argument("--torch_dtype", type=str, default="auto", + help="Use this dtype to store block weights and do computations. " + "By default, respect the dtypes in the pre-trained state dict.") + parser.add_argument('--attn_cache_size', type=str, default=None, + help='The size of GPU memory allocated for storing past attention keys/values between inference' + ' steps; examples: 500MB or 1.2GB or 1073741824 (bytes); be warned: 1KB != 1KiB') + parser.add_argument('--alloc_timeout', type=float, default=60, + help='If the cache is full, the server will wait for this number of seconds hoping that some memory will be freed ' + 'before rejecting the request') + parser.add_argument('--revision', type=str, default='main', + help="The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a git-based system for storing models" + "and other artifacts on huggingface.co, so `revision` can be any identifier allowed by git.") + + parser.add_argument('--throughput', + type=lambda value: value if value in ['auto', 'eval'] else float(value), + default='auto', + help='Expected server throughput (a float measured in RPS). ' + 'If set to "auto" (default), the script evaluates network and compute throughput ' + 'on the first run and uses these estimates for future runs. ' + 'If set to "eval", the script re-evaluates the throughput and overrides the cache.') + parser.add_argument('--update_period', type=float, required=False, default=30, + help='Server will report blocks to DHT once in this many seconds') + parser.add_argument('--expiration', type=float, required=False, default=None, + help='DHT entries will expire after this many seconds') + parser.add_argument('--request_timeout', type=float, required=False, default=3 * 60, + help='Timeout for the whole rpc_forward/rpc_backward/rpc_forward_stream/rpc_backward_stream request') + parser.add_argument('--session_timeout', type=float, required=False, default=30 * 60, + help='Timeout for the whole inference session') + parser.add_argument('--step_timeout', type=float, required=False, default=5 * 60, + help="Timeout for waiting the next step's inputs inside an inference session") + + group = parser.add_mutually_exclusive_group() + group.add_argument('--initial_peers', type=str, nargs='*', required=False, default=PUBLIC_INITIAL_PEERS, + help='Multiaddrs of one or more DHT peers from the target swarm. Default: connects to the public swarm') + group.add_argument('--new_swarm', action='store_true', + help='Start a new private swarm (i.e., do not connect to any initial peers)') + + parser.add_argument('--increase_file_limit', action='store_true', + help='On *nix, this will increase the max number of processes ' + 'a server can spawn before hitting "Too many open files"; Use at your own risk.') + parser.add_argument('--stats_report_interval', type=int, required=False, + help='Interval between two reports of batch processing performance statistics') + + parser.add_argument('--custom_module_path', type=str, required=False, + help='Path of a file with custom nn.modules, wrapped into special decorator') + parser.add_argument('--identity_path', type=str, required=False, help='Path to identity file to be used in P2P') + + parser.add_argument("--balance_quality", type=float, default=0.75, + help="Rebalance the swarm if its throughput is worse than this share of the optimal " + "throughput. Use 0.0 to disable rebalancing, values > 1.0 to force rebalancing " + "on each check for debugging purposes.") + parser.add_argument("--mean_balance_check_period", type=float, default=60, + help="Check the swarm's balance every N seconds (and rebalance it if necessary)") + + parser.add_argument("--use_auth_token", type=str, default=None, help="auth token for from_pretrained") + parser.add_argument('--load_in_8bit', action='store_true', help='Convert the loaded model into mixed-8bit quantized model.') + + # fmt:on + args = vars(parser.parse_args()) + args.pop("config", None) + + args["converted_model_name_or_path"] = args.pop("model") or args["converted_model_name_or_path"] + + if args.pop("increase_file_limit"): + increase_file_limit() + + compression_type = args.pop("compression").upper() + compression = getattr(CompressionType, compression_type) + + attn_cache_size = args.pop("attn_cache_size") + if attn_cache_size is not None: + attn_cache_size = parse_size(attn_cache_size) + assert isinstance( + attn_cache_size, (int, type(None)) + ), "unrecognized value for attention_cache_bytes, examples: 1.5GB or 1500MB or 1572864000 (bytes)" + + if args.pop("new_swarm"): + args["initial_peers"] = [] + + use_auth_token = args.pop("use_auth_token") + args["use_auth_token"] = True if use_auth_token in ("True", "true", "") else use_auth_token + + server = Server(**args, compression=compression, attn_cache_size=attn_cache_size) + try: + server.run() + except KeyboardInterrupt: + logger.info("Caught KeyboardInterrupt, shutting down") + finally: + server.shutdown() + + +if __name__ == "__main__": + main() diff --git a/src/petals/cli/speed_test.py b/src/petals/cli/speed_test.py new file mode 100755 index 0000000..d3342c3 --- /dev/null +++ b/src/petals/cli/speed_test.py @@ -0,0 +1,1941 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# Copyright 2012 Matt Martz +# All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); you may +# not use this file except in compliance with the License. You may obtain +# a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT +# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the +# License for the specific language governing permissions and limitations +# under the License. + +import csv +import datetime +import errno +import math +import os +import platform +import re +import signal +import socket +import sys +import threading +import timeit +import xml.parsers.expat + +try: + import gzip + + GZIP_BASE = gzip.GzipFile +except ImportError: + gzip = None + GZIP_BASE = object + +__version__ = "2.1.4b1" + + +class FakeShutdownEvent(object): + """Class to fake a threading.Event.isSet so that users of this module + are not required to register their own threading.Event() + """ + + @staticmethod + def isSet(): + "Dummy method to always return false" "" + return False + + is_set = isSet + + +# Some global variables we use +DEBUG = False +_GLOBAL_DEFAULT_TIMEOUT = object() +PY25PLUS = sys.version_info[:2] >= (2, 5) +PY26PLUS = sys.version_info[:2] >= (2, 6) +PY32PLUS = sys.version_info[:2] >= (3, 2) +PY310PLUS = sys.version_info[:2] >= (3, 10) + +# Begin import game to handle Python 2 and Python 3 +try: + import json +except ImportError: + try: + import simplejson as json + except ImportError: + json = None + +try: + import xml.etree.ElementTree as ET + + try: + from xml.etree.ElementTree import _Element as ET_Element + except ImportError: + pass +except ImportError: + from xml.dom import minidom as DOM + from xml.parsers.expat import ExpatError + + ET = None + +try: + from urllib2 import ( + AbstractHTTPHandler, + HTTPDefaultErrorHandler, + HTTPError, + HTTPErrorProcessor, + HTTPRedirectHandler, + OpenerDirector, + ProxyHandler, + Request, + URLError, + urlopen, + ) +except ImportError: + from urllib.request import ( + AbstractHTTPHandler, + HTTPDefaultErrorHandler, + HTTPError, + HTTPErrorProcessor, + HTTPRedirectHandler, + OpenerDirector, + ProxyHandler, + Request, + URLError, + urlopen, + ) + +try: + from httplib import BadStatusLine, HTTPConnection +except ImportError: + from http.client import BadStatusLine, HTTPConnection + +try: + from httplib import HTTPSConnection +except ImportError: + try: + from http.client import HTTPSConnection + except ImportError: + HTTPSConnection = None + +try: + from httplib import FakeSocket +except ImportError: + FakeSocket = None + +try: + from Queue import Queue +except ImportError: + from queue import Queue + +try: + from urlparse import urlparse +except ImportError: + from urllib.parse import urlparse + +try: + from urlparse import parse_qs +except ImportError: + try: + from urllib.parse import parse_qs + except ImportError: + from cgi import parse_qs + +try: + from hashlib import md5 +except ImportError: + from md5 import md5 + +try: + from argparse import SUPPRESS as ARG_SUPPRESS, ArgumentParser as ArgParser + + PARSER_TYPE_INT = int + PARSER_TYPE_STR = str + PARSER_TYPE_FLOAT = float +except ImportError: + from optparse import SUPPRESS_HELP as ARG_SUPPRESS, OptionParser as ArgParser + + PARSER_TYPE_INT = "int" + PARSER_TYPE_STR = "string" + PARSER_TYPE_FLOAT = "float" + +try: + from cStringIO import StringIO + + BytesIO = None +except ImportError: + try: + from StringIO import StringIO + + BytesIO = None + except ImportError: + from io import BytesIO, StringIO + +try: + import __builtin__ +except ImportError: + import builtins + from io import FileIO, TextIOWrapper + + class _Py3Utf8Output(TextIOWrapper): + """UTF-8 encoded wrapper around stdout for py3, to override + ASCII stdout + """ + + def __init__(self, f, **kwargs): + buf = FileIO(f.fileno(), "w") + super(_Py3Utf8Output, self).__init__(buf, encoding="utf8", errors="strict") + + def write(self, s): + super(_Py3Utf8Output, self).write(s) + self.flush() + + _py3_print = getattr(builtins, "print") + try: + _py3_utf8_stdout = _Py3Utf8Output(sys.stdout) + _py3_utf8_stderr = _Py3Utf8Output(sys.stderr) + except OSError: + # sys.stdout/sys.stderr is not a compatible stdout/stderr object + # just use it and hope things go ok + _py3_utf8_stdout = sys.stdout + _py3_utf8_stderr = sys.stderr + + def to_utf8(v): + """No-op encode to utf-8 for py3""" + return v + + def print_(*args, **kwargs): + """Wrapper function for py3 to print, with a utf-8 encoded stdout""" + if kwargs.get("file") == sys.stderr: + kwargs["file"] = _py3_utf8_stderr + else: + kwargs["file"] = kwargs.get("file", _py3_utf8_stdout) + _py3_print(*args, **kwargs) + +else: + del __builtin__ + + def to_utf8(v): + """Encode value to utf-8 if possible for py2""" + try: + return v.encode("utf8", "strict") + except AttributeError: + return v + + def print_(*args, **kwargs): + """The new-style print function for Python 2.4 and 2.5. + + Taken from https://pypi.python.org/pypi/six/ + + Modified to set encoding to UTF-8 always, and to flush after write + """ + fp = kwargs.pop("file", sys.stdout) + if fp is None: + return + + def write(data): + if not isinstance(data, basestring): + data = str(data) + # If the file has an encoding, encode unicode with it. + encoding = "utf8" # Always trust UTF-8 for output + if isinstance(fp, file) and isinstance(data, unicode) and encoding is not None: + errors = getattr(fp, "errors", None) + if errors is None: + errors = "strict" + data = data.encode(encoding, errors) + fp.write(data) + fp.flush() + + want_unicode = False + sep = kwargs.pop("sep", None) + if sep is not None: + if isinstance(sep, unicode): + want_unicode = True + elif not isinstance(sep, str): + raise TypeError("sep must be None or a string") + end = kwargs.pop("end", None) + if end is not None: + if isinstance(end, unicode): + want_unicode = True + elif not isinstance(end, str): + raise TypeError("end must be None or a string") + if kwargs: + raise TypeError("invalid keyword arguments to print()") + if not want_unicode: + for arg in args: + if isinstance(arg, unicode): + want_unicode = True + break + if want_unicode: + newline = unicode("\n") + space = unicode(" ") + else: + newline = "\n" + space = " " + if sep is None: + sep = space + if end is None: + end = newline + for i, arg in enumerate(args): + if i: + write(sep) + write(arg) + write(end) + + +# Exception "constants" to support Python 2 through Python 3 +try: + import ssl + + try: + CERT_ERROR = (ssl.CertificateError,) + except AttributeError: + CERT_ERROR = tuple() + + HTTP_ERRORS = (HTTPError, URLError, socket.error, ssl.SSLError, BadStatusLine) + CERT_ERROR +except ImportError: + ssl = None + HTTP_ERRORS = (HTTPError, URLError, socket.error, BadStatusLine) + +if PY32PLUS: + etree_iter = ET.Element.iter +elif PY25PLUS: + etree_iter = ET_Element.getiterator + +if PY26PLUS: + thread_is_alive = threading.Thread.is_alive +else: + thread_is_alive = threading.Thread.isAlive + + +def event_is_set(event): + try: + return event.is_set() + except AttributeError: + return event.isSet() + + +class SpeedtestException(Exception): + """Base exception for this module""" + + +class SpeedtestCLIError(SpeedtestException): + """Generic exception for raising errors during CLI operation""" + + +class SpeedtestHTTPError(SpeedtestException): + """Base HTTP exception for this module""" + + +class SpeedtestConfigError(SpeedtestException): + """Configuration XML is invalid""" + + +class SpeedtestServersError(SpeedtestException): + """Servers XML is invalid""" + + +class ConfigRetrievalError(SpeedtestHTTPError): + """Could not retrieve config.php""" + + +class ServersRetrievalError(SpeedtestHTTPError): + """Could not retrieve speedtest-servers.php""" + + +class InvalidServerIDType(SpeedtestException): + """Server ID used for filtering was not an integer""" + + +class NoMatchedServers(SpeedtestException): + """No servers matched when filtering""" + + +class SpeedtestMiniConnectFailure(SpeedtestException): + """Could not connect to the provided speedtest mini server""" + + +class InvalidSpeedtestMiniServer(SpeedtestException): + """Server provided as a speedtest mini server does not actually appear + to be a speedtest mini server + """ + + +class ShareResultsConnectFailure(SpeedtestException): + """Could not connect to speedtest.net API to POST results""" + + +class ShareResultsSubmitFailure(SpeedtestException): + """Unable to successfully POST results to speedtest.net API after + connection + """ + + +class SpeedtestUploadTimeout(SpeedtestException): + """testlength configuration reached during upload + Used to ensure the upload halts when no additional data should be sent + """ + + +class SpeedtestBestServerFailure(SpeedtestException): + """Unable to determine best server""" + + +class SpeedtestMissingBestServer(SpeedtestException): + """get_best_server not called or not able to determine best server""" + + +def create_connection(address, timeout=_GLOBAL_DEFAULT_TIMEOUT, source_address=None): + """Connect to *address* and return the socket object. + + Convenience function. Connect to *address* (a 2-tuple ``(host, + port)``) and return the socket object. Passing the optional + *timeout* parameter will set the timeout on the socket instance + before attempting to connect. If no *timeout* is supplied, the + global default timeout setting returned by :func:`getdefaulttimeout` + is used. If *source_address* is set it must be a tuple of (host, port) + for the socket to bind as a source address before making the connection. + An host of '' or port 0 tells the OS to use the default. + + Largely vendored from Python 2.7, modified to work with Python 2.4 + """ + + host, port = address + err = None + for res in socket.getaddrinfo(host, port, 0, socket.SOCK_STREAM): + af, socktype, proto, canonname, sa = res + sock = None + try: + sock = socket.socket(af, socktype, proto) + if timeout is not _GLOBAL_DEFAULT_TIMEOUT: + sock.settimeout(float(timeout)) + if source_address: + sock.bind(source_address) + sock.connect(sa) + return sock + + except socket.error: + err = get_exception() + if sock is not None: + sock.close() + + if err is not None: + raise err + else: + raise socket.error("getaddrinfo returns an empty list") + + +class SpeedtestHTTPConnection(HTTPConnection): + """Custom HTTPConnection to support source_address across + Python 2.4 - Python 3 + """ + + def __init__(self, *args, **kwargs): + source_address = kwargs.pop("source_address", None) + timeout = kwargs.pop("timeout", 10) + + self._tunnel_host = None + + HTTPConnection.__init__(self, *args, **kwargs) + + self.source_address = source_address + self.timeout = timeout + + def connect(self): + """Connect to the host and port specified in __init__.""" + try: + self.sock = socket.create_connection((self.host, self.port), self.timeout, self.source_address) + except (AttributeError, TypeError): + self.sock = create_connection((self.host, self.port), self.timeout, self.source_address) + + if self._tunnel_host: + self._tunnel() + + +if HTTPSConnection: + + class SpeedtestHTTPSConnection(HTTPSConnection): + """Custom HTTPSConnection to support source_address across + Python 2.4 - Python 3 + """ + + default_port = 443 + + def __init__(self, *args, **kwargs): + source_address = kwargs.pop("source_address", None) + timeout = kwargs.pop("timeout", 10) + + self._tunnel_host = None + + HTTPSConnection.__init__(self, *args, **kwargs) + + self.timeout = timeout + self.source_address = source_address + + def connect(self): + "Connect to a host on a given (SSL) port." + try: + self.sock = socket.create_connection((self.host, self.port), self.timeout, self.source_address) + except (AttributeError, TypeError): + self.sock = create_connection((self.host, self.port), self.timeout, self.source_address) + + if self._tunnel_host: + self._tunnel() + + if ssl: + try: + kwargs = {} + if hasattr(ssl, "SSLContext"): + if self._tunnel_host: + kwargs["server_hostname"] = self._tunnel_host + else: + kwargs["server_hostname"] = self.host + self.sock = self._context.wrap_socket(self.sock, **kwargs) + except AttributeError: + self.sock = ssl.wrap_socket(self.sock) + try: + self.sock.server_hostname = self.host + except AttributeError: + pass + elif FakeSocket: + # Python 2.4/2.5 support + try: + self.sock = FakeSocket(self.sock, socket.ssl(self.sock)) + except AttributeError: + raise SpeedtestException("This version of Python does not support HTTPS/SSL " "functionality") + else: + raise SpeedtestException("This version of Python does not support HTTPS/SSL " "functionality") + + +def _build_connection(connection, source_address, timeout, context=None): + """Cross Python 2.4 - Python 3 callable to build an ``HTTPConnection`` or + ``HTTPSConnection`` with the args we need + + Called from ``http(s)_open`` methods of ``SpeedtestHTTPHandler`` or + ``SpeedtestHTTPSHandler`` + """ + + def inner(host, **kwargs): + kwargs.update({"source_address": source_address, "timeout": timeout}) + if context: + kwargs["context"] = context + return connection(host, **kwargs) + + return inner + + +class SpeedtestHTTPHandler(AbstractHTTPHandler): + """Custom ``HTTPHandler`` that can build a ``HTTPConnection`` with the + args we need for ``source_address`` and ``timeout`` + """ + + def __init__(self, debuglevel=0, source_address=None, timeout=10): + AbstractHTTPHandler.__init__(self, debuglevel) + self.source_address = source_address + self.timeout = timeout + + def http_open(self, req): + return self.do_open(_build_connection(SpeedtestHTTPConnection, self.source_address, self.timeout), req) + + http_request = AbstractHTTPHandler.do_request_ + + +class SpeedtestHTTPSHandler(AbstractHTTPHandler): + """Custom ``HTTPSHandler`` that can build a ``HTTPSConnection`` with the + args we need for ``source_address`` and ``timeout`` + """ + + def __init__(self, debuglevel=0, context=None, source_address=None, timeout=10): + AbstractHTTPHandler.__init__(self, debuglevel) + self._context = context + self.source_address = source_address + self.timeout = timeout + + def https_open(self, req): + return self.do_open( + _build_connection( + SpeedtestHTTPSConnection, + self.source_address, + self.timeout, + context=self._context, + ), + req, + ) + + https_request = AbstractHTTPHandler.do_request_ + + +def build_opener(source_address=None, timeout=10): + """Function similar to ``urllib2.build_opener`` that will build + an ``OpenerDirector`` with the explicit handlers we want, + ``source_address`` for binding, ``timeout`` and our custom + `User-Agent` + """ + + printer("Timeout set to %d" % timeout, debug=True) + + if source_address: + source_address_tuple = (source_address, 0) + printer("Binding to source address: %r" % (source_address_tuple,), debug=True) + else: + source_address_tuple = None + + handlers = [ + ProxyHandler(), + SpeedtestHTTPHandler(source_address=source_address_tuple, timeout=timeout), + SpeedtestHTTPSHandler(source_address=source_address_tuple, timeout=timeout), + HTTPDefaultErrorHandler(), + HTTPRedirectHandler(), + HTTPErrorProcessor(), + ] + + opener = OpenerDirector() + opener.addheaders = [("User-agent", build_user_agent())] + + for handler in handlers: + opener.add_handler(handler) + + return opener + + +class GzipDecodedResponse(GZIP_BASE): + """A file-like object to decode a response encoded with the gzip + method, as described in RFC 1952. + + Largely copied from ``xmlrpclib``/``xmlrpc.client`` and modified + to work for py2.4-py3 + """ + + def __init__(self, response): + # response doesn't support tell() and read(), required by + # GzipFile + if not gzip: + raise SpeedtestHTTPError("HTTP response body is gzip encoded, " "but gzip support is not available") + IO = BytesIO or StringIO + self.io = IO() + while 1: + chunk = response.read(1024) + if len(chunk) == 0: + break + self.io.write(chunk) + self.io.seek(0) + gzip.GzipFile.__init__(self, mode="rb", fileobj=self.io) + + def close(self): + try: + gzip.GzipFile.close(self) + finally: + self.io.close() + + +def get_exception(): + """Helper function to work with py2.4-py3 for getting the current + exception in a try/except block + """ + return sys.exc_info()[1] + + +def distance(origin, destination): + """Determine distance between 2 sets of [lat,lon] in km""" + + lat1, lon1 = origin + lat2, lon2 = destination + radius = 6371 # km + + dlat = math.radians(lat2 - lat1) + dlon = math.radians(lon2 - lon1) + a = math.sin(dlat / 2) * math.sin(dlat / 2) + math.cos(math.radians(lat1)) * math.cos( + math.radians(lat2) + ) * math.sin(dlon / 2) * math.sin(dlon / 2) + c = 2 * math.atan2(math.sqrt(a), math.sqrt(1 - a)) + d = radius * c + + return d + + +def build_user_agent(): + """Build a Mozilla/5.0 compatible User-Agent string""" + + ua_tuple = ( + "Mozilla/5.0", + "(%s; U; %s; en-us)" % (platform.platform(), platform.architecture()[0]), + "Python/%s" % platform.python_version(), + "(KHTML, like Gecko)", + "speedtest-cli/%s" % __version__, + ) + user_agent = " ".join(ua_tuple) + printer("User-Agent: %s" % user_agent, debug=True) + return user_agent + + +def build_request(url, data=None, headers=None, bump="0", secure=False): + """Build a urllib2 request object + + This function automatically adds a User-Agent header to all requests + + """ + + if not headers: + headers = {} + + if url[0] == ":": + scheme = ("http", "https")[bool(secure)] + schemed_url = "%s%s" % (scheme, url) + else: + schemed_url = url + + if "?" in url: + delim = "&" + else: + delim = "?" + + # WHO YOU GONNA CALL? CACHE BUSTERS! + final_url = "%s%sx=%s.%s" % (schemed_url, delim, int(timeit.time.time() * 1000), bump) + + headers.update( + { + "Cache-Control": "no-cache", + } + ) + + printer("%s %s" % (("GET", "POST")[bool(data)], final_url), debug=True) + + return Request(final_url, data=data, headers=headers) + + +def catch_request(request, opener=None): + """Helper function to catch common exceptions encountered when + establishing a connection with a HTTP/HTTPS request + + """ + + if opener: + _open = opener.open + else: + _open = urlopen + + try: + uh = _open(request) + if request.get_full_url() != uh.geturl(): + printer("Redirected to %s" % uh.geturl(), debug=True) + return uh, False + except HTTP_ERRORS: + e = get_exception() + return None, e + + +def get_response_stream(response): + """Helper function to return either a Gzip reader if + ``Content-Encoding`` is ``gzip`` otherwise the response itself + + """ + + try: + getheader = response.headers.getheader + except AttributeError: + getheader = response.getheader + + if getheader("content-encoding") == "gzip": + return GzipDecodedResponse(response) + + return response + + +def get_attributes_by_tag_name(dom, tag_name): + """Retrieve an attribute from an XML document and return it in a + consistent format + + Only used with xml.dom.minidom, which is likely only to be used + with python versions older than 2.5 + """ + elem = dom.getElementsByTagName(tag_name)[0] + return dict(list(elem.attributes.items())) + + +def print_dots(shutdown_event): + """Built in callback function used by Thread classes for printing + status + """ + + def inner(current, total, start=False, end=False): + if event_is_set(shutdown_event): + return + + sys.stdout.write(".") + if current + 1 == total and end is True: + sys.stdout.write("\n") + sys.stdout.flush() + + return inner + + +def do_nothing(*args, **kwargs): + pass + + +class HTTPDownloader(threading.Thread): + """Thread class for retrieving a URL""" + + def __init__(self, i, request, start, timeout, opener=None, shutdown_event=None): + threading.Thread.__init__(self) + self.request = request + self.result = [0] + self.starttime = start + self.timeout = timeout + self.i = i + if opener: + self._opener = opener.open + else: + self._opener = urlopen + + if shutdown_event: + self._shutdown_event = shutdown_event + else: + self._shutdown_event = FakeShutdownEvent() + + def run(self): + try: + if (timeit.default_timer() - self.starttime) <= self.timeout: + f = self._opener(self.request) + while ( + not event_is_set(self._shutdown_event) and (timeit.default_timer() - self.starttime) <= self.timeout + ): + self.result.append(len(f.read(10240))) + if self.result[-1] == 0: + break + f.close() + except IOError: + pass + except HTTP_ERRORS: + pass + + +class HTTPUploaderData(object): + """File like object to improve cutting off the upload once the timeout + has been reached + """ + + def __init__(self, length, start, timeout, shutdown_event=None): + self.length = length + self.start = start + self.timeout = timeout + + if shutdown_event: + self._shutdown_event = shutdown_event + else: + self._shutdown_event = FakeShutdownEvent() + + self._data = None + + self.total = [0] + + def pre_allocate(self): + chars = "0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ" + multiplier = int(round(int(self.length) / 36.0)) + IO = BytesIO or StringIO + try: + self._data = IO(("content1=%s" % (chars * multiplier)[0 : int(self.length) - 9]).encode()) + except MemoryError: + raise SpeedtestCLIError("Insufficient memory to pre-allocate upload data. Please " "use --no-pre-allocate") + + @property + def data(self): + if not self._data: + self.pre_allocate() + return self._data + + def read(self, n=10240): + if (timeit.default_timer() - self.start) <= self.timeout and not event_is_set(self._shutdown_event): + chunk = self.data.read(n) + self.total.append(len(chunk)) + return chunk + else: + raise SpeedtestUploadTimeout() + + def __len__(self): + return self.length + + +class HTTPUploader(threading.Thread): + """Thread class for putting a URL""" + + def __init__(self, i, request, start, size, timeout, opener=None, shutdown_event=None): + threading.Thread.__init__(self) + self.request = request + self.request.data.start = self.starttime = start + self.size = size + self.result = 0 + self.timeout = timeout + self.i = i + + if opener: + self._opener = opener.open + else: + self._opener = urlopen + + if shutdown_event: + self._shutdown_event = shutdown_event + else: + self._shutdown_event = FakeShutdownEvent() + + def run(self): + request = self.request + try: + if (timeit.default_timer() - self.starttime) <= self.timeout and not event_is_set(self._shutdown_event): + try: + f = self._opener(request) + except TypeError: + # PY24 expects a string or buffer + # This also causes issues with Ctrl-C, but we will concede + # for the moment that Ctrl-C on PY24 isn't immediate + request = build_request(self.request.get_full_url(), data=request.data.read(self.size)) + f = self._opener(request) + f.read(11) + f.close() + self.result = sum(self.request.data.total) + else: + self.result = 0 + except (IOError, SpeedtestUploadTimeout): + self.result = sum(self.request.data.total) + except HTTP_ERRORS: + self.result = 0 + + +class SpeedtestResults(object): + """Class for holding the results of a speedtest, including: + + Download speed + Upload speed + Ping/Latency to test server + Data about server that the test was run against + + Additionally this class can return a result data as a dictionary or CSV, + as well as submit a POST of the result data to the speedtest.net API + to get a share results image link. + """ + + def __init__(self, download=0, upload=0, ping=0, server=None, client=None, opener=None, secure=False): + self.download = download + self.upload = upload + self.ping = ping + if server is None: + self.server = {} + else: + self.server = server + self.client = client or {} + + self._share = None + self.timestamp = "%sZ" % datetime.datetime.utcnow().isoformat() + self.bytes_received = 0 + self.bytes_sent = 0 + + if opener: + self._opener = opener + else: + self._opener = build_opener() + + self._secure = secure + + def __repr__(self): + return repr(self.dict()) + + def share(self): + """POST data to the speedtest.net API to obtain a share results + link + """ + + if self._share: + return self._share + + download = int(round(self.download / 1000.0, 0)) + ping = int(round(self.ping, 0)) + upload = int(round(self.upload / 1000.0, 0)) + + # Build the request to send results back to speedtest.net + # We use a list instead of a dict because the API expects parameters + # in a certain order + api_data = [ + "recommendedserverid=%s" % self.server["id"], + "ping=%s" % ping, + "screenresolution=", + "promo=", + "download=%s" % download, + "screendpi=", + "upload=%s" % upload, + "testmethod=http", + "hash=%s" % md5(("%s-%s-%s-%s" % (ping, upload, download, "297aae72")).encode()).hexdigest(), + "touchscreen=none", + "startmode=pingselect", + "accuracy=1", + "bytesreceived=%s" % self.bytes_received, + "bytessent=%s" % self.bytes_sent, + "serverid=%s" % self.server["id"], + ] + + headers = {"Referer": "http://c.speedtest.net/flash/speedtest.swf"} + request = build_request( + "://www.speedtest.net/api/api.php", data="&".join(api_data).encode(), headers=headers, secure=self._secure + ) + f, e = catch_request(request, opener=self._opener) + if e: + raise ShareResultsConnectFailure(e) + + response = f.read() + code = f.code + f.close() + + if int(code) != 200: + raise ShareResultsSubmitFailure("Could not submit results to " "speedtest.net") + + qsargs = parse_qs(response.decode()) + resultid = qsargs.get("resultid") + if not resultid or len(resultid) != 1: + raise ShareResultsSubmitFailure("Could not submit results to " "speedtest.net") + + self._share = "http://www.speedtest.net/result/%s.png" % resultid[0] + + return self._share + + def dict(self): + """Return dictionary of result data""" + + return { + "download": self.download, + "upload": self.upload, + "ping": self.ping, + "server": self.server, + "timestamp": self.timestamp, + "bytes_sent": self.bytes_sent, + "bytes_received": self.bytes_received, + "share": self._share, + "client": self.client, + } + + @staticmethod + def csv_header(delimiter=","): + """Return CSV Headers""" + + row = [ + "Server ID", + "Sponsor", + "Server Name", + "Timestamp", + "Distance", + "Ping", + "Download", + "Upload", + "Share", + "IP Address", + ] + out = StringIO() + writer = csv.writer(out, delimiter=delimiter, lineterminator="") + writer.writerow([to_utf8(v) for v in row]) + return out.getvalue() + + def csv(self, delimiter=","): + """Return data in CSV format""" + + data = self.dict() + out = StringIO() + writer = csv.writer(out, delimiter=delimiter, lineterminator="") + row = [ + data["server"]["id"], + data["server"]["sponsor"], + data["server"]["name"], + data["timestamp"], + data["server"]["d"], + data["ping"], + data["download"], + data["upload"], + self._share or "", + self.client["ip"], + ] + writer.writerow([to_utf8(v) for v in row]) + return out.getvalue() + + def json(self, pretty=False): + """Return data in JSON format""" + + kwargs = {} + if pretty: + kwargs.update({"indent": 4, "sort_keys": True}) + return json.dumps(self.dict(), **kwargs) + + +class Speedtest(object): + """Class for performing standard speedtest.net testing operations""" + + def __init__(self, config=None, source_address=None, timeout=10, secure=False, shutdown_event=None): + self.config = {} + + self._source_address = source_address + self._timeout = timeout + self._opener = build_opener(source_address, timeout) + + self._secure = secure + + if shutdown_event: + self._shutdown_event = shutdown_event + else: + self._shutdown_event = FakeShutdownEvent() + + self.get_config() + if config is not None: + self.config.update(config) + + self.servers = {} + self.closest = [] + self._best = {} + + self.results = SpeedtestResults( + client=self.config["client"], + opener=self._opener, + secure=secure, + ) + + @property + def best(self): + if not self._best: + self.get_best_server() + return self._best + + def get_config(self): + """Download the speedtest.net configuration and return only the data + we are interested in + """ + + headers = {} + if gzip: + headers["Accept-Encoding"] = "gzip" + request = build_request("://www.speedtest.net/speedtest-config.php", headers=headers, secure=self._secure) + uh, e = catch_request(request, opener=self._opener) + if e: + raise ConfigRetrievalError(e) + configxml_list = [] + + stream = get_response_stream(uh) + + while 1: + try: + configxml_list.append(stream.read(1024)) + except (OSError, EOFError): + raise ConfigRetrievalError(get_exception()) + if len(configxml_list[-1]) == 0: + break + stream.close() + uh.close() + + if int(uh.code) != 200: + return None + + configxml = "".encode().join(configxml_list) + + printer("Config XML:\n%s" % configxml, debug=True) + + try: + try: + root = ET.fromstring(configxml) + except ET.ParseError: + e = get_exception() + raise SpeedtestConfigError("Malformed speedtest.net configuration: %s" % e) + server_config = root.find("server-config").attrib + download = root.find("download").attrib + upload = root.find("upload").attrib + # times = root.find('times').attrib + client = root.find("client").attrib + + except AttributeError: + try: + root = DOM.parseString(configxml) + except ExpatError: + e = get_exception() + raise SpeedtestConfigError("Malformed speedtest.net configuration: %s" % e) + server_config = get_attributes_by_tag_name(root, "server-config") + download = get_attributes_by_tag_name(root, "download") + upload = get_attributes_by_tag_name(root, "upload") + # times = get_attributes_by_tag_name(root, 'times') + client = get_attributes_by_tag_name(root, "client") + + ignore_servers = [int(i) for i in server_config["ignoreids"].split(",") if i] + + ratio = int(upload["ratio"]) + upload_max = int(upload["maxchunkcount"]) + up_sizes = [32768, 65536, 131072, 262144, 524288, 1048576, 7340032] + sizes = {"upload": up_sizes[ratio - 1 :], "download": [350, 500, 750, 1000, 1500, 2000, 2500, 3000, 3500, 4000]} + + size_count = len(sizes["upload"]) + + upload_count = int(math.ceil(upload_max / size_count)) + + counts = {"upload": upload_count, "download": int(download["threadsperurl"])} + + threads = {"upload": int(upload["threads"]), "download": int(server_config["threadcount"]) * 2} + + length = {"upload": int(upload["testlength"]), "download": int(download["testlength"])} + + self.config.update( + { + "client": client, + "ignore_servers": ignore_servers, + "sizes": sizes, + "counts": counts, + "threads": threads, + "length": length, + "upload_max": upload_count * size_count, + } + ) + + try: + self.lat_lon = (float(client["lat"]), float(client["lon"])) + except ValueError: + raise SpeedtestConfigError("Unknown location: lat=%r lon=%r" % (client.get("lat"), client.get("lon"))) + + printer("Config:\n%r" % self.config, debug=True) + + return self.config + + def get_servers(self, servers=None, exclude=None): + """Retrieve a the list of speedtest.net servers, optionally filtered + to servers matching those specified in the ``servers`` argument + """ + if servers is None: + servers = [] + + if exclude is None: + exclude = [] + + self.servers.clear() + + for server_list in (servers, exclude): + for i, s in enumerate(server_list): + try: + server_list[i] = int(s) + except ValueError: + raise InvalidServerIDType("%s is an invalid server type, must be int" % s) + + urls = [ + "://www.speedtest.net/speedtest-servers-static.php", + "http://c.speedtest.net/speedtest-servers-static.php", + "://www.speedtest.net/speedtest-servers.php", + "http://c.speedtest.net/speedtest-servers.php", + ] + + headers = {} + if gzip: + headers["Accept-Encoding"] = "gzip" + + errors = [] + for url in urls: + try: + request = build_request( + "%s?threads=%s" % (url, self.config["threads"]["download"]), headers=headers, secure=self._secure + ) + uh, e = catch_request(request, opener=self._opener) + if e: + errors.append("%s" % e) + raise ServersRetrievalError() + + stream = get_response_stream(uh) + + serversxml_list = [] + while 1: + try: + serversxml_list.append(stream.read(1024)) + except (OSError, EOFError): + raise ServersRetrievalError(get_exception()) + if len(serversxml_list[-1]) == 0: + break + + stream.close() + uh.close() + + if int(uh.code) != 200: + raise ServersRetrievalError() + + serversxml = "".encode().join(serversxml_list) + + printer("Servers XML:\n%s" % serversxml, debug=True) + + try: + try: + try: + root = ET.fromstring(serversxml) + except ET.ParseError: + e = get_exception() + raise SpeedtestServersError("Malformed speedtest.net server list: %s" % e) + elements = etree_iter(root, "server") + except AttributeError: + try: + root = DOM.parseString(serversxml) + except ExpatError: + e = get_exception() + raise SpeedtestServersError("Malformed speedtest.net server list: %s" % e) + elements = root.getElementsByTagName("server") + except (SyntaxError, xml.parsers.expat.ExpatError): + raise ServersRetrievalError() + + for server in elements: + try: + attrib = server.attrib + except AttributeError: + attrib = dict(list(server.attributes.items())) + + if servers and int(attrib.get("id")) not in servers: + continue + + if int(attrib.get("id")) in self.config["ignore_servers"] or int(attrib.get("id")) in exclude: + continue + + try: + d = distance(self.lat_lon, (float(attrib.get("lat")), float(attrib.get("lon")))) + except Exception: + continue + + attrib["d"] = d + + try: + self.servers[d].append(attrib) + except KeyError: + self.servers[d] = [attrib] + + break + + except ServersRetrievalError: + continue + + if (servers or exclude) and not self.servers: + raise NoMatchedServers() + + return self.servers + + def set_mini_server(self, server): + """Instead of querying for a list of servers, set a link to a + speedtest mini server + """ + + urlparts = urlparse(server) + + name, ext = os.path.splitext(urlparts[2]) + if ext: + url = os.path.dirname(server) + else: + url = server + + request = build_request(url) + uh, e = catch_request(request, opener=self._opener) + if e: + raise SpeedtestMiniConnectFailure("Failed to connect to %s" % server) + else: + text = uh.read() + uh.close() + + extension = re.findall('upload_?[Ee]xtension: "([^"]+)"', text.decode()) + if not extension: + for ext in ["php", "asp", "aspx", "jsp"]: + try: + f = self._opener.open("%s/speedtest/upload.%s" % (url, ext)) + except Exception: + pass + else: + data = f.read().strip().decode() + if f.code == 200 and len(data.splitlines()) == 1 and re.match("size=[0-9]", data): + extension = [ext] + break + if not urlparts or not extension: + raise InvalidSpeedtestMiniServer("Invalid Speedtest Mini Server: " "%s" % server) + + self.servers = [ + { + "sponsor": "Speedtest Mini", + "name": urlparts[1], + "d": 0, + "url": "%s/speedtest/upload.%s" % (url.rstrip("/"), extension[0]), + "latency": 0, + "id": 0, + } + ] + + return self.servers + + def get_closest_servers(self, limit=5): + """Limit servers to the closest speedtest.net servers based on + geographic distance + """ + + if not self.servers: + self.get_servers() + + for d in sorted(self.servers.keys()): + for s in self.servers[d]: + self.closest.append(s) + if len(self.closest) == limit: + break + else: + continue + break + + printer("Closest Servers:\n%r" % self.closest, debug=True) + return self.closest + + def get_best_server(self, servers=None): + """Perform a speedtest.net "ping" to determine which speedtest.net + server has the lowest latency + """ + + if not servers: + if not self.closest: + servers = self.get_closest_servers() + servers = self.closest + + if self._source_address: + source_address_tuple = (self._source_address, 0) + else: + source_address_tuple = None + + user_agent = build_user_agent() + + results = {} + for server in servers: + cum = [] + url = os.path.dirname(server["url"]) + stamp = int(timeit.time.time() * 1000) + latency_url = "%s/latency.txt?x=%s" % (url, stamp) + for i in range(0, 3): + this_latency_url = "%s.%s" % (latency_url, i) + printer("%s %s" % ("GET", this_latency_url), debug=True) + urlparts = urlparse(latency_url) + try: + if urlparts[0] == "https": + h = SpeedtestHTTPSConnection(urlparts[1], source_address=source_address_tuple) + else: + h = SpeedtestHTTPConnection(urlparts[1], source_address=source_address_tuple) + headers = {"User-Agent": user_agent} + path = "%s?%s" % (urlparts[2], urlparts[4]) + start = timeit.default_timer() + h.request("GET", path, headers=headers) + r = h.getresponse() + total = timeit.default_timer() - start + except HTTP_ERRORS: + e = get_exception() + printer("ERROR: %r" % e, debug=True) + cum.append(3600) + continue + + text = r.read(9) + if int(r.status) == 200 and text == "test=test".encode(): + cum.append(total) + else: + cum.append(3600) + h.close() + + avg = round((sum(cum) / 6) * 1000.0, 3) + results[avg] = server + + try: + fastest = sorted(results.keys())[0] + except IndexError: + raise SpeedtestBestServerFailure("Unable to connect to servers to " "test latency.") + best = results[fastest] + best["latency"] = fastest + + self.results.ping = fastest + self.results.server = best + + self._best.update(best) + printer("Best Server:\n%r" % best, debug=True) + return best + + def download(self, callback=do_nothing, threads=None): + """Test download speed against speedtest.net + + A ``threads`` value of ``None`` will fall back to those dictated + by the speedtest.net configuration + """ + + urls = [] + for size in self.config["sizes"]["download"]: + for _ in range(0, self.config["counts"]["download"]): + urls.append("%s/random%sx%s.jpg" % (os.path.dirname(self.best["url"]), size, size)) + + request_count = len(urls) + requests = [] + for i, url in enumerate(urls): + requests.append(build_request(url, bump=i, secure=self._secure)) + + max_threads = threads or self.config["threads"]["download"] + in_flight = {"threads": 0} + + def producer(q, requests, request_count): + for i, request in enumerate(requests): + thread = HTTPDownloader( + i, + request, + start, + self.config["length"]["download"], + opener=self._opener, + shutdown_event=self._shutdown_event, + ) + while in_flight["threads"] >= max_threads: + timeit.time.sleep(0.001) + thread.start() + q.put(thread, True) + in_flight["threads"] += 1 + callback(i, request_count, start=True) + + finished = [] + + def consumer(q, request_count): + _is_alive = thread_is_alive + while len(finished) < request_count: + thread = q.get(True) + while _is_alive(thread): + thread.join(timeout=0.001) + in_flight["threads"] -= 1 + finished.append(sum(thread.result)) + callback(thread.i, request_count, end=True) + + q = Queue(max_threads) + prod_thread = threading.Thread(target=producer, args=(q, requests, request_count)) + cons_thread = threading.Thread(target=consumer, args=(q, request_count)) + start = timeit.default_timer() + prod_thread.start() + cons_thread.start() + _is_alive = thread_is_alive + while _is_alive(prod_thread): + prod_thread.join(timeout=0.001) + while _is_alive(cons_thread): + cons_thread.join(timeout=0.001) + + stop = timeit.default_timer() + self.results.bytes_received = sum(finished) + self.results.download = (self.results.bytes_received / (stop - start)) * 8.0 + if self.results.download > 100000: + self.config["threads"]["upload"] = 8 + return self.results.download + + def upload(self, callback=do_nothing, pre_allocate=True, threads=None): + """Test upload speed against speedtest.net + + A ``threads`` value of ``None`` will fall back to those dictated + by the speedtest.net configuration + """ + + sizes = [] + + for size in self.config["sizes"]["upload"]: + for _ in range(0, self.config["counts"]["upload"]): + sizes.append(size) + + # request_count = len(sizes) + request_count = self.config["upload_max"] + + requests = [] + for i, size in enumerate(sizes): + # We set ``0`` for ``start`` and handle setting the actual + # ``start`` in ``HTTPUploader`` to get better measurements + data = HTTPUploaderData(size, 0, self.config["length"]["upload"], shutdown_event=self._shutdown_event) + if pre_allocate: + data.pre_allocate() + + headers = {"Content-length": size} + requests.append((build_request(self.best["url"], data, secure=self._secure, headers=headers), size)) + + max_threads = threads or self.config["threads"]["upload"] + in_flight = {"threads": 0} + + def producer(q, requests, request_count): + for i, request in enumerate(requests[:request_count]): + thread = HTTPUploader( + i, + request[0], + start, + request[1], + self.config["length"]["upload"], + opener=self._opener, + shutdown_event=self._shutdown_event, + ) + while in_flight["threads"] >= max_threads: + timeit.time.sleep(0.001) + thread.start() + q.put(thread, True) + in_flight["threads"] += 1 + callback(i, request_count, start=True) + + finished = [] + + def consumer(q, request_count): + _is_alive = thread_is_alive + while len(finished) < request_count: + thread = q.get(True) + while _is_alive(thread): + thread.join(timeout=0.001) + in_flight["threads"] -= 1 + finished.append(thread.result) + callback(thread.i, request_count, end=True) + + q = Queue(threads or self.config["threads"]["upload"]) + prod_thread = threading.Thread(target=producer, args=(q, requests, request_count)) + cons_thread = threading.Thread(target=consumer, args=(q, request_count)) + start = timeit.default_timer() + prod_thread.start() + cons_thread.start() + _is_alive = thread_is_alive + while _is_alive(prod_thread): + prod_thread.join(timeout=0.1) + while _is_alive(cons_thread): + cons_thread.join(timeout=0.1) + + stop = timeit.default_timer() + self.results.bytes_sent = sum(finished) + self.results.upload = (self.results.bytes_sent / (stop - start)) * 8.0 + return self.results.upload + + +def ctrl_c(shutdown_event): + """Catch Ctrl-C key sequence and set a SHUTDOWN_EVENT for our threaded + operations + """ + + def inner(signum, frame): + shutdown_event.set() + printer("\nCancelling...", error=True) + sys.exit(0) + + return inner + + +def version(): + """Print the version""" + + printer("speedtest-cli %s" % __version__) + printer("Python %s" % sys.version.replace("\n", "")) + sys.exit(0) + + +def csv_header(delimiter=","): + """Print the CSV Headers""" + + printer(SpeedtestResults.csv_header(delimiter=delimiter)) + sys.exit(0) + + +def parse_args(): + """Function to handle building and parsing of command line arguments""" + description = ( + "Command line interface for testing internet bandwidth using " + "speedtest.net.\n" + "------------------------------------------------------------" + "--------------\n" + "https://github.com/sivel/speedtest-cli" + ) + + parser = ArgParser(description=description) + # Give optparse.OptionParser an `add_argument` method for + # compatibility with argparse.ArgumentParser + try: + parser.add_argument = parser.add_option + except AttributeError: + pass + parser.add_argument( + "--no-download", + dest="download", + default=True, + action="store_const", + const=False, + help="Do not perform download test", + ) + parser.add_argument( + "--no-upload", dest="upload", default=True, action="store_const", const=False, help="Do not perform upload test" + ) + parser.add_argument( + "--single", + default=False, + action="store_true", + help="Only use a single connection instead of " "multiple. This simulates a typical file " "transfer.", + ) + parser.add_argument( + "--bytes", + dest="units", + action="store_const", + const=("byte", 8), + default=("bit", 1), + help="Display values in bytes instead of bits. Does " + "not affect the image generated by --share, nor " + "output from --json or --csv", + ) + parser.add_argument( + "--share", + action="store_true", + help="Generate and provide a URL to the speedtest.net " "share results image, not displayed with --csv", + ) + parser.add_argument( + "--simple", action="store_true", default=False, help="Suppress verbose output, only show basic " "information" + ) + parser.add_argument( + "--csv", + action="store_true", + default=False, + help="Suppress verbose output, only show basic " + "information in CSV format. Speeds listed in " + "bit/s and not affected by --bytes", + ) + parser.add_argument( + "--csv-delimiter", + default=",", + type=PARSER_TYPE_STR, + help="Single character delimiter to use in CSV " 'output. Default ","', + ) + parser.add_argument("--csv-header", action="store_true", default=False, help="Print CSV headers") + parser.add_argument( + "--json", + action="store_true", + default=False, + help="Suppress verbose output, only show basic " + "information in JSON format. Speeds listed in " + "bit/s and not affected by --bytes", + ) + parser.add_argument( + "--list", action="store_true", help="Display a list of speedtest.net servers " "sorted by distance" + ) + parser.add_argument( + "--server", + type=PARSER_TYPE_INT, + action="append", + help="Specify a server ID to test against. Can be " "supplied multiple times", + ) + parser.add_argument( + "--exclude", + type=PARSER_TYPE_INT, + action="append", + help="Exclude a server from selection. Can be " "supplied multiple times", + ) + parser.add_argument("--mini", help="URL of the Speedtest Mini server") + parser.add_argument("--source", help="Source IP address to bind to") + parser.add_argument("--timeout", default=10, type=PARSER_TYPE_FLOAT, help="HTTP timeout in seconds. Default 10") + parser.add_argument( + "--secure", + action="store_true", + help="Use HTTPS instead of HTTP when communicating " "with speedtest.net operated servers", + ) + parser.add_argument( + "--no-pre-allocate", + dest="pre_allocate", + action="store_const", + default=True, + const=False, + help="Do not pre allocate upload data. Pre allocation " + "is enabled by default to improve upload " + "performance. To support systems with " + "insufficient memory, use this option to avoid a " + "MemoryError", + ) + parser.add_argument("--version", action="store_true", help="Show the version number and exit") + parser.add_argument("--debug", action="store_true", help=ARG_SUPPRESS, default=ARG_SUPPRESS) + + options = parser.parse_args() + if isinstance(options, tuple): + args = options[0] + else: + args = options + return args + + +def validate_optional_args(args): + """Check if an argument was provided that depends on a module that may + not be part of the Python standard library. + + If such an argument is supplied, and the module does not exist, exit + with an error stating which module is missing. + """ + optional_args = { + "json": ("json/simplejson python module", json), + "secure": ("SSL support", HTTPSConnection), + } + + for arg, info in optional_args.items(): + if getattr(args, arg, False) and info[1] is None: + raise SystemExit("%s is not installed. --%s is " "unavailable" % (info[0], arg)) + + +def printer(string, quiet=False, debug=False, error=False, **kwargs): + """Helper function print a string with various features""" + + if debug and not DEBUG: + return + + if debug: + if sys.stdout.isatty(): + out = "\033[1;30mDEBUG: %s\033[0m" % string + else: + out = "DEBUG: %s" % string + else: + out = string + + if error: + kwargs["file"] = sys.stderr + + if not quiet: + print_(out, **kwargs) + + +def shell(): + """Run the full speedtest.net test""" + + global DEBUG + shutdown_event = threading.Event() + + signal.signal(signal.SIGINT, ctrl_c(shutdown_event)) + + args = parse_args() + + # Print the version and exit + if args.version: + version() + + if not args.download and not args.upload: + raise SpeedtestCLIError("Cannot supply both --no-download and " "--no-upload") + + if len(args.csv_delimiter) != 1: + raise SpeedtestCLIError("--csv-delimiter must be a single character") + + if args.csv_header: + csv_header(args.csv_delimiter) + + validate_optional_args(args) + + debug = getattr(args, "debug", False) + if debug == "SUPPRESSHELP": + debug = False + if debug: + DEBUG = True + + if args.simple or args.csv or args.json: + quiet = True + else: + quiet = False + + if args.csv or args.json: + machine_format = True + else: + machine_format = False + + # Don't set a callback if we are running quietly + if quiet or debug: + callback = do_nothing + else: + callback = print_dots(shutdown_event) + + printer("Retrieving speedtest.net configuration...", quiet) + try: + speedtest = Speedtest(source_address=args.source, timeout=args.timeout, secure=args.secure) + except (ConfigRetrievalError,) + HTTP_ERRORS: + printer("Cannot retrieve speedtest configuration", error=True) + raise SpeedtestCLIError(get_exception()) + + if args.list: + try: + speedtest.get_servers() + except (ServersRetrievalError,) + HTTP_ERRORS: + printer("Cannot retrieve speedtest server list", error=True) + raise SpeedtestCLIError(get_exception()) + + for _, servers in sorted(speedtest.servers.items()): + for server in servers: + line = "%(id)5s) %(sponsor)s (%(name)s, %(country)s) " "[%(d)0.2f km]" % server + try: + printer(line) + except IOError: + e = get_exception() + if e.errno != errno.EPIPE: + raise + sys.exit(0) + + printer("Testing from %(isp)s (%(ip)s)..." % speedtest.config["client"], quiet) + + if not args.mini: + printer("Retrieving speedtest.net server list...", quiet) + try: + speedtest.get_servers(servers=args.server, exclude=args.exclude) + except NoMatchedServers: + raise SpeedtestCLIError("No matched servers: %s" % ", ".join("%s" % s for s in args.server)) + except (ServersRetrievalError,) + HTTP_ERRORS: + printer("Cannot retrieve speedtest server list", error=True) + raise SpeedtestCLIError(get_exception()) + except InvalidServerIDType: + raise SpeedtestCLIError( + "%s is an invalid server type, must " "be an int" % ", ".join("%s" % s for s in args.server) + ) + + if args.server and len(args.server) == 1: + printer("Retrieving information for the selected server...", quiet) + else: + printer("Selecting best server based on ping...", quiet) + speedtest.get_best_server() + elif args.mini: + speedtest.get_best_server(speedtest.set_mini_server(args.mini)) + + results = speedtest.results + + printer("Hosted by %(sponsor)s (%(name)s) [%(d)0.2f km]: " "%(latency)s ms" % results.server, quiet) + + if args.download: + printer("Testing download speed", quiet, end=("", "\n")[bool(debug)]) + speedtest.download(callback=callback, threads=(None, 1)[args.single]) + printer("Download: %0.2f M%s/s" % ((results.download / 1000.0 / 1000.0) / args.units[1], args.units[0]), quiet) + else: + printer("Skipping download test", quiet) + + if args.upload: + printer("Testing upload speed", quiet, end=("", "\n")[bool(debug)]) + speedtest.upload(callback=callback, pre_allocate=args.pre_allocate, threads=(None, 1)[args.single]) + printer("Upload: %0.2f M%s/s" % ((results.upload / 1000.0 / 1000.0) / args.units[1], args.units[0]), quiet) + else: + printer("Skipping upload test", quiet) + + printer("Results:\n%r" % results.dict(), debug=True) + + if not args.simple and args.share: + results.share() + + if args.simple: + printer( + "Ping: %s ms\nDownload: %0.2f M%s/s\nUpload: %0.2f M%s/s" + % ( + results.ping, + (results.download / 1000.0 / 1000.0) / args.units[1], + args.units[0], + (results.upload / 1000.0 / 1000.0) / args.units[1], + args.units[0], + ) + ) + elif args.csv: + printer(results.csv(delimiter=args.csv_delimiter)) + elif args.json: + printer(results.json()) + + if args.share and not machine_format: + printer("Share results: %s" % results.share()) + + +def main(): + try: + shell() + except KeyboardInterrupt: + printer("\nCancelling...", error=True) + except (SpeedtestException, SystemExit): + e = get_exception() + # Ignore a successful exit, or argparse exit + if getattr(e, "code", 1) not in (0, 2): + msg = "%s" % e + if not msg: + msg = "%r" % e + raise SystemExit("ERROR: %s" % msg) + + +if __name__ == "__main__": + main()