From e9c948317152756f0d3cb7a5d8f480a8f8850ed1 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Fri, 3 May 2024 08:10:44 -0700 Subject: [PATCH] Improve the logging messages while loading models --- modules/models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/models.py b/modules/models.py index c89bcae4..687af8ba 100644 --- a/modules/models.py +++ b/modules/models.py @@ -107,10 +107,10 @@ def load_model(model_name, loader=None): elif loader in ['llama.cpp', 'llamacpp_HF']: shared.settings['truncation_length'] = shared.args.n_ctx + logger.info(f"Loaded \"{model_name}\" in {(time.time()-t0):.2f} seconds.") logger.info(f"LOADER: \"{loader}\"") logger.info(f"TRUNCATION LENGTH: {shared.settings['truncation_length']}") logger.info(f"INSTRUCTION TEMPLATE: \"{metadata['instruction_template']}\"") - logger.info(f"Loaded the model in {(time.time()-t0):.2f} seconds.") return model, tokenizer