From 9bc2063e785c29a67e9dfef34de5c3ba63aa87d6 Mon Sep 17 00:00:00 2001 From: Croos3r Date: Mon, 1 Jul 2024 23:44:17 +0200 Subject: [PATCH] feat: added new data about models and split name and parameters subtitle Took 16 minutes --- src/texts.ts | 26 +++++++++++++++++++------- 1 file changed, 19 insertions(+), 7 deletions(-) diff --git a/src/texts.ts b/src/texts.ts index bcd1d97..b18640d 100644 --- a/src/texts.ts +++ b/src/texts.ts @@ -1,30 +1,42 @@ -export const models: Array<{ title: string; description: string }> = [ +export const models: Array<{ name: string; description: string; subtitle: string; type: string }> = [ { - title: "Nous Hermes 2 Pro (Llama 3 8B)", + name: "Nous Hermes 2 Pro", + subtitle: "Llama 3 8B", + type: "fast", description: "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.", }, { - title: "NeuralBeagle (7B)", + name: "NeuralBeagle", + subtitle: "7B", + type: "fast", description: "NeuralBeagle14-7B is a DPO fine-tune of mlabonne/Beagle14-7B using the argilla/distilabel-intel-orca-dpo-pairs preference dataset.", }, { - title: "Mixtral (8x7B MOE)", + name: "Mixtral", + subtitle: "8x7B MOE", + type: "smart", description: "Mixtral 8x7B, a high-quality sparse mixture of experts model (SMoE) with open weights.", }, { - title: "Nous Hermes 2 (34B)", + name: "Nous Hermes 2 ", + subtitle: "34B", + type: "smart", description: "Nous Hermes 2 Yi 34B was trained on 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape.", }, { - title: "Llama 3 Instruct (70B)", + name: "Llama 3 Instruct", + subtitle: "70B", + type: "genius, slow", description: "Llama 3 instruction-tuned models are fine-tuned and optimized for dialogue/chat use cases and outperform many of the available open-source chat models on common benchmarks.", }, { - title: "DeepSeek Coder (6.7B)", + name: "DeepSeek Coder", + subtitle: "6.7B", + type: "developer", description: "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.", },