From ea48eded22299bf1394ab75a102d9f91bc8594eb Mon Sep 17 00:00:00 2001 From: majorli Date: Mon, 25 Mar 2024 15:56:21 +0800 Subject: [PATCH] unify Llama 2 name to be official name Signed-off-by: majorli --- README.md | 6 +++--- RELEASE.md | 4 ++-- nlp/llm/llama2-7b/megatron-deepspeed/README.md | 4 ++-- nlp/llm/llama2-7b_reward_sft/deepspeed/README.md | 2 +- nlp/llm/llama2-7b_sft/megatron-deepspeed/README.md | 4 ++-- 5 files changed, 10 insertions(+), 10 deletions(-) diff --git a/README.md b/README.md index f5becce73..95ad94365 100644 --- a/README.md +++ b/README.md @@ -429,9 +429,9 @@ DeepSparkHub甄选上百个应用算法和模型,覆盖AI和通用计算各领 -------- | ------ | ---- [ChatGLM-6B](nlp/llm/chatglm-6b/deepspeed/README.md) | PyTorch (DeepSpeed) | ADGEN & chatglm-6b [LLaMA-7B](nlp/llm/llama-7b/colossalai/README.md) | PyTorch (Colossal-AI) | llama-7b-hf -[LLaMA2-7B](nlp/llm/llama2-7b/megatron-deepspeed/README.md) | PyTorch (Megatron-DeepSpeed) | Bookcorpus -[LLaMA2-7B Reward Model Finetuning](nlp/llm/llama2-7b_reward_sft/deepspeed/README.md) | PyTorch (DeepSpeed) | Dahoas/rm-static -[LLaMA2-7B SFT](nlp/llm/llama2-7b_sft/megatron-deepspeed/README.md) | PyTorch (Megatron-DeepSpeed) | gpt_small-117M +[Llama-2-7B](nlp/llm/llama2-7b/megatron-deepspeed/README.md) | PyTorch (Megatron-DeepSpeed) | Bookcorpus +[Llama-2-7B Reward Model Finetuning](nlp/llm/llama2-7b_reward_sft/deepspeed/README.md) | PyTorch (DeepSpeed) | Dahoas/rm-static +[Llama-2-7B SFT](nlp/llm/llama2-7b_sft/megatron-deepspeed/README.md) | PyTorch (Megatron-DeepSpeed) | gpt_small-117M #### Text Correction diff --git a/RELEASE.md b/RELEASE.md index 66404c30a..9a1b0fd1d 100644 --- a/RELEASE.md +++ b/RELEASE.md @@ -28,8 +28,8 @@ * 新增了基于分布式训练框架的大语言模型训练示例。 - * Megatron-DeepSpeed LLaMA2-7B SFT - * DeepSpeed LLaMa2-7B Reward Model Finetuning + * Megatron-DeepSpeed Llama-2-7B SFT + * DeepSpeed Llama-2-7B Reward Model Finetuning ### 问题修复 diff --git a/nlp/llm/llama2-7b/megatron-deepspeed/README.md b/nlp/llm/llama2-7b/megatron-deepspeed/README.md index 60718f5bd..1ae436c93 100644 --- a/nlp/llm/llama2-7b/megatron-deepspeed/README.md +++ b/nlp/llm/llama2-7b/megatron-deepspeed/README.md @@ -1,8 +1,8 @@ -# Megatron-DeepSpeed LLaMA2-7B +# Megatron-DeepSpeed Llama-2-7B ## Model description -LLaMA2 is a large language model released by Meta in 2023, with parameters ranging from 7B to 70B. Compared to LLaMA, the training corpus of LLaMA2 is 40% longer, and the context length has been upgraded from 2048 to 4096, allowing for understanding and generating longer texts. +Llama 2 is a large language model released by Meta in 2023, with parameters ranging from 7B to 70B. Compared to LLaMA, the training corpus of Llama 2 is 40% longer, and the context length has been upgraded from 2048 to 4096, allowing for understanding and generating longer texts. ## Step 1: Installation diff --git a/nlp/llm/llama2-7b_reward_sft/deepspeed/README.md b/nlp/llm/llama2-7b_reward_sft/deepspeed/README.md index f72a66f3b..e1f1d66ed 100644 --- a/nlp/llm/llama2-7b_reward_sft/deepspeed/README.md +++ b/nlp/llm/llama2-7b_reward_sft/deepspeed/README.md @@ -1,4 +1,4 @@ -# LLaMa2 7B Reward Model Finetuning +# DeepSpeed Llama-2-7B Reward Model Finetuning ## Model description LLaMA2 is a large language model released by Meta in 2023, with parameters ranging from 7B to 70B. Compared to LLaMA, the training corpus of LLaMA2 is 40% longer, and the context length has been upgraded from 2048 to 4096, allowing for understanding and generating longer texts. diff --git a/nlp/llm/llama2-7b_sft/megatron-deepspeed/README.md b/nlp/llm/llama2-7b_sft/megatron-deepspeed/README.md index 62a52f657..17ec826ea 100644 --- a/nlp/llm/llama2-7b_sft/megatron-deepspeed/README.md +++ b/nlp/llm/llama2-7b_sft/megatron-deepspeed/README.md @@ -1,8 +1,8 @@ -# Megatron-DeepSpeed LLaMA2-7B SFT +# Megatron-DeepSpeed Llama-2-7B SFT ## Model description -LLaMA2 is a large language model released by Meta in 2023, with parameters ranging from 7B to 70B. Compared to LLaMA, the training corpus of LLaMA2 is 40% longer, and the context length has been upgraded from 2048 to 4096, allowing for understanding and generating longer texts. +Llama 2 is a large language model released by Meta in 2023, with parameters ranging from 7B to 70B. Compared to LLaMA, the training corpus of Llama 2 is 40% longer, and the context length has been upgraded from 2048 to 4096, allowing for understanding and generating longer texts. ## Step 1: Installation -- Gitee