File size: 696 Bytes
ed221d6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
from transformers import PretrainedConfig


class Erbb1MlpConfig(PretrainedConfig):
    model_type = "erbb1_mlp"

    def __init__(
        self,
        d_in: int = 768,
        d_hidden: int = 1024,
        n_layers: int = 6,
        dropout: float = 0.1,
        layer_norm_eps: float | None = 1e-12,
        dataset_mean: float | None = None,
        dataset_std: float | None = None,
        **kwargs,
    ):
        self.d_in = d_in
        self.d_hidden = d_hidden
        self.n_layers = n_layers
        self.dropout = dropout
        self.layer_norm_eps = layer_norm_eps
        self.dataset_mean = dataset_mean
        self.dataset_std = dataset_std
        super().__init__(**kwargs)