Skip to content

Commit dfdace3

Browse files
committed
add disable parameter for pre-tokenizer
1 parent acd2bda commit dfdace3

8 files changed

Lines changed: 21 additions & 10 deletions

File tree

FlagEmbedding/inference/embedder/decoder_only/base.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -224,7 +224,8 @@ def encode_single_device(
224224

225225
# tokenize without padding to get the correct length
226226
all_inputs = []
227-
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize'):
227+
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize',
228+
disable=len(sentences) < 256):
228229
sentences_batch = sentences[start_index:start_index + batch_size]
229230
inputs_batch = self.tokenizer(
230231
sentences_batch,

FlagEmbedding/inference/embedder/decoder_only/icl.py

Lines changed: 7 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -178,9 +178,12 @@ def stop_self_query_pool(self):
178178
if self.query_pool is not None:
179179
self.stop_multi_process_pool(self.query_pool)
180180
self.query_pool = None
181-
self.model.to('cpu')
181+
try:
182+
self.model.to('cpu')
183+
torch.cuda.empty_cache()
184+
except:
185+
pass
182186
gc.collect()
183-
torch.cuda.empty_cache()
184187

185188
def encode_queries(
186189
self,
@@ -483,7 +486,8 @@ def encode_single_device(
483486

484487
# tokenize without padding to get the correct length
485488
all_inputs = []
486-
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize'):
489+
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize',
490+
disable=len(sentences) < 256):
487491
sentences_batch = sentences[start_index:start_index + batch_size]
488492
inputs_batch = self.tokenizer(
489493
sentences_batch,

FlagEmbedding/inference/embedder/encoder_only/base.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -205,7 +205,8 @@ def encode_single_device(
205205

206206
# tokenize without padding to get the correct length
207207
all_inputs = []
208-
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize'):
208+
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize',
209+
disable=len(sentences) < 256):
209210
sentences_batch = sentences[start_index:start_index + batch_size]
210211
inputs_batch = self.tokenizer(
211212
sentences_batch,

FlagEmbedding/inference/embedder/encoder_only/m3.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -369,7 +369,8 @@ def _process_colbert_vecs(colbert_vecs: np.ndarray, attention_mask: list):
369369

370370
# tokenize without padding to get the correct length
371371
all_inputs = []
372-
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize'):
372+
for start_index in trange(0, len(sentences), batch_size, desc='pre tokenize',
373+
disable=len(sentences) < 256):
373374
sentences_batch = sentences[start_index:start_index + batch_size]
374375
inputs_batch = self.tokenizer(
375376
sentences_batch,

FlagEmbedding/inference/reranker/decoder_only/base.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -309,7 +309,8 @@ def compute_score_single_gpu(
309309
# tokenize without padding to get the correct length
310310
all_queries_inputs = []
311311
all_passages_inputs = []
312-
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize"):
312+
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize",
313+
disable=len(sentence_pairs) < 128):
313314
sentences_batch = sentence_pairs[start_index:start_index + batch_size]
314315
queries = [s[0] for s in sentences_batch]
315316
passages = [s[1] for s in sentences_batch]

FlagEmbedding/inference/reranker/decoder_only/layerwise.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -191,7 +191,8 @@ def compute_score_single_gpu(
191191
# tokenize without padding to get the correct length
192192
all_queries_inputs = []
193193
all_passages_inputs = []
194-
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize"):
194+
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize",
195+
disable=len(sentence_pairs) < 128):
195196
sentences_batch = sentence_pairs[start_index:start_index + batch_size]
196197
queries = [s[0] for s in sentences_batch]
197198
passages = [s[1] for s in sentences_batch]

FlagEmbedding/inference/reranker/decoder_only/lightweight.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -262,7 +262,8 @@ def compute_score_single_gpu(
262262
# tokenize without padding to get the correct length
263263
all_queries_inputs = []
264264
all_passages_inputs = []
265-
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize"):
265+
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize",
266+
disable=len(sentence_pairs) < 128):
266267
sentences_batch = sentence_pairs[start_index:start_index + batch_size]
267268
queries = [s[0] for s in sentences_batch]
268269
passages = [s[1] for s in sentences_batch]

FlagEmbedding/inference/reranker/encoder_only/base.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -121,7 +121,8 @@ def compute_score_single_gpu(
121121

122122
# tokenize without padding to get the correct length
123123
all_inputs = []
124-
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize"):
124+
for start_index in trange(0, len(sentence_pairs), batch_size, desc="pre tokenize",
125+
disable=len(sentence_pairs) < 128):
125126
sentences_batch = sentence_pairs[start_index:start_index + batch_size]
126127
queries = [s[0] for s in sentences_batch]
127128
passages = [s[1] for s in sentences_batch]

0 commit comments

Comments
 (0)