| from dataclasses import dataclass, field |
| from typing import Optional |
|
|
| from transformers import TrainingArguments |
|
|
|
|
| @dataclass |
| class ModelArgs: |
| |
| model_name_or_path: str = field( |
| default="meta-llama/Llama-2-7b-chat-hf", |
| metadata={ |
| "help": "Path to pretrained model or model identifier from huggingface.co/models" |
| }, |
| ) |
| super_tokenizer_name_or_path: str = field( |
| default="/share/ninglu_shao/code/PluginTransformer/data/outputs/90k_0104+8-longalpaca_0106/super_tokenizer", |
| metadata={ |
| "help": "Path to pretrained model or model identifier from huggingface.co/models" |
| }, |
| ) |
| |
| super_tokenizer_num_hidden_layers: int = field( |
| default=8, |
| metadata={"help": "Encoder model's layer num."}, |
| ) |
| is_model_frozen: bool = field( |
| default=True, |
| metadata={"help": "Freeze or not decoder model."}, |
| ) |
| use_flash_attention_2: bool = field( |
| default=True, |
| metadata={"help": "Use flash attention 2 or not."}, |
| ) |
| dtype: str = field( |
| default="bf16", |
| ) |
| device_map: Optional[str] = field( |
| default=None, |
| ) |
|
|
|
|
| @dataclass |
| class DataArgs: |
| |
| dataset_list: str = field( |
| default="wikipedia", |
| metadata={"help": "Path of dataset"}, |
| ) |
| dataset_save_dir: str = field( |
| default="/share/ninglu_shao/data/PluginTransformer", |
| metadata={"help": "The path to save dataset."}, |
| ) |
|
|
| def __post_init__(self): |
| self.dataset_list = [dataset.strip() for dataset in self.dataset_list.split(",")] |
|
|
|
|
| @dataclass |
| class TrainingArgs(TrainingArguments): |
| |
| output_dir: str = field( |
| default="outputs/test_4", |
| metadata={ |
| "help": "The output directory where the model predictions and checkpoints will be written." |
| }, |
| ) |
| overwrite_output_dir: bool = field( |
| default=False, |
| metadata={"help": "If True, overwrite the content of the output directory."}, |
| ) |
| |
| learning_rate: float = field( |
| default=1e-4, |
| metadata={"help": "The initial learning rate for optimizer."}, |
| ) |
| warmup_ratio: float = field( |
| default=0.1, |
| metadata={"help": "The ratio of warmup steps for optimizer."}, |
| ) |
| num_train_epochs: float = field( |
| default=1, |
| metadata={"help": "Total number of training epochs to perform."}, |
| ) |
| per_device_train_batch_size: int = field( |
| default=8, |
| metadata={"help": "The batch size per GPU/TPU core/CPU for training."}, |
| ) |
| |
| dataloader_num_workers: int = field( |
| default=32, |
| metadata={"help": "Number of subprocesses to use for data loading."}, |
| ) |
| remove_unused_columns: bool = field( |
| default=False, |
| metadata={ |
| "help": "Whether or not to automatically remove the columns unused by the model forward method." |
| }, |
| ) |
| |
| save_strategy: str = field( |
| default="steps", |
| metadata={"help": "The checkpoint save strategy to adopt during training."}, |
| ) |
| save_steps: int = field( |
| default=500, |
| metadata={"help": "Saving frequency according to saving strategy"}, |
| ) |
| save_total_limit: int = field( |
| default=None, |
| metadata={"help": "How many checkpoints to keep in the output_dir."}, |
| ) |
| logging_steps: int = field( |
| default=10, |
| metadata={"help": "Logging frequency according to logging strategy."}, |
| ) |
| |
| fp16: bool = field( |
| default=False, |
| metadata={ |
| "help": "Whether to use fp16 16-bit (mixed) precision training instead of 32-bit training." |
| }, |
| ) |
| bf16: bool = field( |
| default=True, |
| metadata={ |
| "help": "Whether to use bf16 16-bit (mixed) precision training instead of 32-bit training." |
| }, |
| ) |
| |
| |
| |
| |
| |
| |
|
|
|
|
| @dataclass |
| class GenerationArgs: |
| do_sample: bool = field( |
| default=False, |
| metadata={"help": "Sample when decoding?"}, |
| ) |
| num_return_sequences: int = field( |
| default=1, |
| metadata={"help": "How many sequences to generate?"}, |
| ) |
| max_length: int = field( |
| default=1024, |
| metadata={"help": "Maximum length."}, |
| ) |