From fe3b8d6e66a62ecc49074df0caea9b0f1f2018a7 Mon Sep 17 00:00:00 2001 From: Aleksandr Borzunov Date: Sun, 6 Aug 2023 22:01:55 +0000 Subject: [PATCH] Append .amd to reported version --- src/petals/server/handler.py | 2 +- src/petals/server/server.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/src/petals/server/handler.py b/src/petals/server/handler.py index d3776de..637e2ee 100644 --- a/src/petals/server/handler.py +++ b/src/petals/server/handler.py @@ -608,7 +608,7 @@ class TransformerConnectionHandler(ConnectionHandler): backend = self.module_backends[request.uid] if request.uid else next(iter(self.module_backends.values())) result = { - "version": petals.__version__, + "version": petals.__version__ + ".amd", "dht_client_mode": self.dht.client_mode, CACHE_TOKENS_AVAILABLE: backend.memory_cache.bytes_left // max(backend.cache_bytes_per_token.values()), } diff --git a/src/petals/server/server.py b/src/petals/server/server.py index a505fac..7f0de41 100644 --- a/src/petals/server/server.py +++ b/src/petals/server/server.py @@ -238,7 +238,7 @@ class Server: self.server_info = ServerInfo( state=ServerState.JOINING, public_name=public_name, - version=petals.__version__, + version=petals.__version__ + ".amd", adapters=tuple(adapters), torch_dtype=str(torch_dtype).replace("torch.", ""), quant_type=quant_type.name.lower(),