LlaMa-2-7B&LlaMa-2-7B-chat

模型介绍

LlaMa 2

LlaMa 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B pretrained model. Links to other models can be found in the index at the bottom.

LlaMa-2-7b-chat

LlaMa-2-7b

启动方式

  1. 创建工作空间,按需选择GPU资源

  2. 挂载公有镜像pytorch: official-torch2.0-cu1117。选择预训练模型: llama-2-7b和llama-2-7b-chat

  3. 创建工作空间

  1. 创建完成后,打开vscode或jupyter,在终端中输入以下命令,可能需要学术网站加速

cd /ark-contexts/imported_models

sudo git clone https://github.com/facebookresearch/llama.git

cd llama

sudo pip install -e .
  1. 使用模型

LlaMa-2-7b

torchrun --nproc_per_node 1 example_text_completion.py \
--ckpt_dir ../llama-2-7b/Llama-2-7b/ \
--tokenizer_path ../llama-2-7b/Llama-2-7b/tokenizer.model \
--max_seq_len 128 --max_batch_size 4

LlaMa-2-7b-chat

torchrun --nproc_per_node 1 example_chat_completion.py \
--ckpt_dir ../llama-2-7b-chat/Llama-2-7b-chat/ \
--tokenizer_path ../llama-2-7b-chat/Llama-2-7b-chat/tokenizer.model \
--max_seq_len 512 --max_batch_size 4

Last updated