From 582485bc6396540b3c20533a3e659f82cce8583b Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Gabriel=20Mart=C3=ADn=20Bl=C3=A1zquez?= Date: Tue, 19 Dec 2023 17:07:10 +0100 Subject: [PATCH] Apply suggestions from code review Co-authored-by: Alvaro Bartolome --- examples/pipeline-llamacpp-and-openai-process.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/examples/pipeline-llamacpp-and-openai-process.py b/examples/pipeline-llamacpp-and-openai-process.py index 0507c624db..df5f8cf11f 100644 --- a/examples/pipeline-llamacpp-and-openai-process.py +++ b/examples/pipeline-llamacpp-and-openai-process.py @@ -34,7 +34,7 @@ def load_llama_cpp_llm(task: "Task") -> "LLM": from distilabel.llm import LlamaCppLLM llama = Llama( - model_path="notus-7b-v1.Q4_0.gguf", n_gpu_layers=10, n_ctx=1024, verbose=True + model_path="", n_gpu_layers=10, n_ctx=1024, verbose=False ) return LlamaCppLLM( model=llama, task=task, max_new_tokens=512, prompt_format="zephyr" @@ -70,7 +70,7 @@ def load_openai_llm(task: "Task") -> "LLM": dataset = pipeline.generate( dataset, # type: ignore num_generations=2, - batch_size=2, + batch_size=1, enable_checkpoints=True, display_progress_bar=False, )