From d7a97aa0b60fcf2a861ccd2ffb40e34cdf2c7650 Mon Sep 17 00:00:00 2001 From: Nicolas Patry Date: Mon, 15 May 2023 15:14:17 +0200 Subject: [PATCH] Removing dead variables. (#327) # What does this PR do? Fixes # (issue) ## Before submitting - [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case). - [ ] Did you read the [contributor guideline](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md#start-contributing-pull-requests), Pull Request section? - [ ] Was this discussed/approved via a Github issue or the [forum](https://discuss.huggingface.co/)? Please add a link to it if that's the case. - [ ] Did you make sure to update the documentation with your changes? Here are the [documentation guidelines](https://github.com/huggingface/transformers/tree/main/docs), and [here are tips on formatting docstrings](https://github.com/huggingface/transformers/tree/main/docs#writing-source-documentation). - [ ] Did you write any new necessary tests? ## Who can review? Anyone in the community is free to review the PR once the tests have passed. Feel free to tag members/contributors who may be interested in your PR. --- server/text_generation_server/models/bloom.py | 1 - server/text_generation_server/models/flash_llama.py | 3 --- server/text_generation_server/models/flash_neox.py | 2 -- server/text_generation_server/models/flash_santacoder.py | 3 --- server/text_generation_server/models/galactica.py | 1 - server/text_generation_server/models/gpt_neox.py | 1 - server/text_generation_server/models/opt.py | 1 - server/text_generation_server/models/t5.py | 1 - 8 files changed, 13 deletions(-) diff --git a/server/text_generation_server/models/bloom.py b/server/text_generation_server/models/bloom.py index f6a6903..ed95929 100644 --- a/server/text_generation_server/models/bloom.py +++ b/server/text_generation_server/models/bloom.py @@ -72,7 +72,6 @@ class BLOOMSharded(BLOOM): quantize: Optional[str] = None, ): self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16 diff --git a/server/text_generation_server/models/flash_llama.py b/server/text_generation_server/models/flash_llama.py index 0b63f90..aa0b448 100644 --- a/server/text_generation_server/models/flash_llama.py +++ b/server/text_generation_server/models/flash_llama.py @@ -29,7 +29,6 @@ tracer = trace.get_tracer(__name__) class FlashLlama(FlashCausalLM): def __init__(self, model_id: str, revision: Optional[str] = None, quantize=False): - self.past_pad = None if torch.cuda.is_available(): device = torch.device("cuda") dtype = torch.float16 @@ -150,9 +149,7 @@ class FlashLlamaSharded(FlashLlama): revision: Optional[str] = None, quantize: Optional[str] = None, ): - self.past_pad = None self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16 diff --git a/server/text_generation_server/models/flash_neox.py b/server/text_generation_server/models/flash_neox.py index 168c919..fc741f5 100644 --- a/server/text_generation_server/models/flash_neox.py +++ b/server/text_generation_server/models/flash_neox.py @@ -33,9 +33,7 @@ class FlashNeoXSharded(FlashNeoX): def __init__( self, model_id: str, revision: Optional[str] = None, quantize: bool = False ): - self.past_pad = None self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16 diff --git a/server/text_generation_server/models/flash_santacoder.py b/server/text_generation_server/models/flash_santacoder.py index 51a8998..f810bb0 100644 --- a/server/text_generation_server/models/flash_santacoder.py +++ b/server/text_generation_server/models/flash_santacoder.py @@ -28,7 +28,6 @@ tracer = trace.get_tracer(__name__) class FlashSantacoder(FlashCausalLM): def __init__(self, model_id: str, revision: Optional[str] = None, quantize=False): - self.past_pad = None if torch.cuda.is_available(): device = torch.device("cuda") dtype = torch.float16 @@ -173,9 +172,7 @@ class FlashSantacoderSharded(FlashSantacoder): def __init__( self, model_id: str, revision: Optional[str] = None, quantize: bool = False ): - self.past_pad = None self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16 diff --git a/server/text_generation_server/models/galactica.py b/server/text_generation_server/models/galactica.py index c6dd4c3..a011125 100644 --- a/server/text_generation_server/models/galactica.py +++ b/server/text_generation_server/models/galactica.py @@ -199,7 +199,6 @@ class GalacticaSharded(Galactica): quantize: Optional[str] = None, ): self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16 diff --git a/server/text_generation_server/models/gpt_neox.py b/server/text_generation_server/models/gpt_neox.py index 215bb2b..3e8557b 100644 --- a/server/text_generation_server/models/gpt_neox.py +++ b/server/text_generation_server/models/gpt_neox.py @@ -38,7 +38,6 @@ class GPTNeoxSharded(CausalLM): quantize: Optional[str] = None, ): self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16 diff --git a/server/text_generation_server/models/opt.py b/server/text_generation_server/models/opt.py index 8d856b1..c83c335 100644 --- a/server/text_generation_server/models/opt.py +++ b/server/text_generation_server/models/opt.py @@ -51,7 +51,6 @@ class OPTSharded(OPT): self, model_id: str, revision: Optional[str] = None, quantize: bool = False ): self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16 diff --git a/server/text_generation_server/models/t5.py b/server/text_generation_server/models/t5.py index b5e7710..6fe77ca 100644 --- a/server/text_generation_server/models/t5.py +++ b/server/text_generation_server/models/t5.py @@ -38,7 +38,6 @@ class T5Sharded(Seq2SeqLM): quantize: Optional[str] = None, ): self.process_group, rank, world_size = initialize_torch_distributed() - self.master = rank == 0 if torch.cuda.is_available(): device = torch.device(f"cuda:{rank}") dtype = torch.float16