File size: 547 Bytes
30131af
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
from transformers.models.llama.configuration_llama import \
    LlamaConfig as LlamaConfigOriginal


class LlamaConfig(LlamaConfigOriginal):
    def __init__(
        self,
        use_xpos=False,
        position_interpolation_scale=1,
        ntk_alpha=None,
        transformer_engine=None,
        **kwargs
    ):
        self.use_xpos = use_xpos
        self.position_interpolation_scale = position_interpolation_scale
        self.transformer_engine = transformer_engine
        self.ntk_alpha = ntk_alpha
        super().__init__(**kwargs)