From 5dc1221c6557df3e7fdf09d4ecf1ffbdf4b3b42b Mon Sep 17 00:00:00 2001 From: lllyasviel Date: Tue, 31 Oct 2023 14:41:18 -0700 Subject: [PATCH] maintain gpt --- fooocus_version.py | 2 +- .../prompt_expansion/fooocus_expansion/positive.txt | 3 --- modules/expansion.py | 12 +++++++----- 3 files changed, 8 insertions(+), 9 deletions(-) diff --git a/fooocus_version.py b/fooocus_version.py index f7c3cbd2..cb54aeb5 100644 --- a/fooocus_version.py +++ b/fooocus_version.py @@ -1 +1 @@ -version = '2.1.770' +version = '2.1.771' diff --git a/models/prompt_expansion/fooocus_expansion/positive.txt b/models/prompt_expansion/fooocus_expansion/positive.txt index 4e35c773..a7131449 100644 --- a/models/prompt_expansion/fooocus_expansion/positive.txt +++ b/models/prompt_expansion/fooocus_expansion/positive.txt @@ -18,8 +18,6 @@ aesthetic affectionate affirmed affirming -aged -ageless agile aimed airy @@ -654,7 +652,6 @@ outfitted outstanding overjoyed pacified -painterly paired palpable panoramic diff --git a/modules/expansion.py b/modules/expansion.py index a8f3d9c8..2478410c 100644 --- a/modules/expansion.py +++ b/modules/expansion.py @@ -31,18 +31,20 @@ class FooocusExpansion: positive_words = open(os.path.join(fooocus_expansion_path, 'positive.txt'), encoding='utf-8').read().splitlines() + positive_words = [x.lower() for x in positive_words if x != ''] # new_content = '\n'.join(sorted(list(set(positive_words)))) - # t198 = self.tokenizer('\n', return_tensors="np") - # t11 = self.tokenizer(',', return_tensors="np") - # positive_ids = [11, 198, self.tokenizer.eos_token_id] + # eos = self.tokenizer.eos_token_id - positive_ids = [11] + symbols = '-+,.;?!!!' self.bad_words_ids = [] for k, v in self.tokenizer.vocab.items(): - if k.replace('Ġ', '') not in positive_words and v not in positive_ids: + if k.replace('Ġ', '').lower() not in positive_words and k not in symbols: self.bad_words_ids.append([v]) + else: + # print(k) + pass self.model = AutoModelForCausalLM.from_pretrained(fooocus_expansion_path) self.model.eval()