Skip to content

Commit 64f0f52

Browse files
authored
chore: pre download the tokenizer instead of mid operation (#191)
This gpt2 tokenizer is used to estimate the tokens a query takes up. Signed-off-by: Anupam Kumar <[email protected]>
1 parent 24d4b86 commit 64f0f52

File tree

1 file changed

+9
-2
lines changed

1 file changed

+9
-2
lines changed

context_chat_backend/controller.py

Lines changed: 9 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -52,8 +52,15 @@
5252
logger = logging.getLogger('ccb.controller')
5353

5454
models_to_fetch = {
55-
"https://huggingface.co/Ralriki/multilingual-e5-large-instruct-GGUF/resolve/8738f8d3d8f311808479ecd5756607e24c6ca811/multilingual-e5-large-instruct-q6_k.gguf": { # noqa: E501
56-
"save_path": os.path.join(persistent_storage(), 'model_files', "multilingual-e5-large-instruct-q6_k.gguf")
55+
# embedding model
56+
'https://huggingface.co/Ralriki/multilingual-e5-large-instruct-GGUF/resolve/8738f8d3d8f311808479ecd5756607e24c6ca811/multilingual-e5-large-instruct-q6_k.gguf': { # noqa: E501
57+
'save_path': os.path.join(persistent_storage(), 'model_files', 'multilingual-e5-large-instruct-q6_k.gguf')
58+
},
59+
# tokenizer model for estimating token count of queries
60+
'gpt2': {
61+
'cache_dir': os.path.join(persistent_storage(), 'model_files/hub'),
62+
'allow_patterns': ['config.json', 'merges.txt', 'tokenizer.json', 'tokenizer_config.json', 'vocab.json'],
63+
'revision': '607a30d783dfa663caf39e06633721c8d4cfcd7e',
5764
}
5865
}
5966
app_enabled = Event()

0 commit comments

Comments
 (0)