update instruct_config and set all random seed to 42
This commit is contained in:
parent
dba2e2d680
commit
f0d41f937b
|
@ -16,7 +16,8 @@ model:
|
|||
shared_input_output_embedding: False
|
||||
train:
|
||||
train_batch_size: 2
|
||||
num_training_steps: 40000
|
||||
# 1B token for 1 epoch, 5epoch
|
||||
num_training_steps: 20000
|
||||
num_warmup_steps: 500
|
||||
initializer_range: 1.0e-2
|
||||
lr: 2.0e-4
|
||||
|
@ -29,5 +30,5 @@ train:
|
|||
log_interval: 50
|
||||
eval_interval: 500
|
||||
save_interval: 1000
|
||||
work_dir: "data/saved_ckpt/7B"
|
||||
work_dir: "data/saved_ckpt/7B_instruction"
|
||||
project_name: "Llama Instruction"
|
||||
|
|
|
@ -2,7 +2,7 @@
|
|||
Author: LiangSong(sl12160010@gmail.com)
|
||||
Date: 2023-04-24 20:05:21
|
||||
LastEditors: LiangSong(sl12160010@gmail.com)
|
||||
LastEditTime: 2023-05-03 10:23:41
|
||||
LastEditTime: 2023-05-04 08:42:58
|
||||
FilePath: /Open-Llama/dataset/dataset.py
|
||||
Description:
|
||||
|
||||
|
@ -12,9 +12,10 @@ import math
|
|||
import torch
|
||||
import random
|
||||
from glob import glob
|
||||
from datasets import load_dataset, interleave_datasets
|
||||
from datasets import load_dataset
|
||||
|
||||
|
||||
random.seed(42)
|
||||
def pretrain_transform(batch):
|
||||
# wudao preprocess
|
||||
if "title" in batch and "content" in batch:
|
||||
|
|
|
@ -2,13 +2,12 @@
|
|||
Author: LiangSong(sl12160010@gmail.com)
|
||||
Date: 2023-04-24 20:05:21
|
||||
LastEditors: LiangSong(sl12160010@gmail.com)
|
||||
LastEditTime: 2023-05-02 23:55:37
|
||||
LastEditTime: 2023-05-04 08:41:37
|
||||
FilePath: /Open-Llama/solver/trainer.py
|
||||
Description:
|
||||
|
||||
Copyright (c) 2023 by LiangSong(sl12160010@gmail.com), All Rights Reserved.
|
||||
"""
|
||||
import os
|
||||
import time
|
||||
import wandb
|
||||
import torch
|
||||
|
|
|
@ -2,8 +2,8 @@
|
|||
Author: LiangSong(sl12160010@gmail.com)
|
||||
Date: 2023-03-24 20:49:03
|
||||
LastEditors: LiangSong(sl12160010@gmail.com)
|
||||
LastEditTime: 2023-04-05 22:40:29
|
||||
FilePath: /Open-Llama/dataset/train_tokenizer.py
|
||||
LastEditTime: 2023-05-04 08:42:21
|
||||
FilePath: /Open-Llama/utils/train_tokenizer.py
|
||||
Description:
|
||||
|
||||
Copyright (c) 2023 by LiangSong(sl12160010@gmail.com), All Rights Reserved.
|
||||
|
@ -11,6 +11,9 @@ Copyright (c) 2023 by LiangSong(sl12160010@gmail.com), All Rights Reserved.
|
|||
import random
|
||||
from dataset.data_iter import DataIter, create_shard_kwargs
|
||||
|
||||
|
||||
random.seed(42)
|
||||
|
||||
wudao_patterns = [
|
||||
"data/pretrain_data/part-wudao-*.jsonl.zst",
|
||||
]
|
||||
|
|
Loading…
Reference in New Issue
Block a user