|
4 | 4 | # https://huggingface.co/mosaicml/mpt-7b/blob/main/configuration_mpt.py
|
5 | 5 | """A HuggingFace-style model configuration."""
|
6 | 6 | import warnings
|
7 |
| -from typing import Any, Dict, Optional, Union |
| 7 | +from typing import Any, Optional, Union |
8 | 8 |
|
9 | 9 | from transformers import PretrainedConfig
|
10 | 10 |
|
11 |
| -attn_config_defaults: Dict = { |
| 11 | +attn_config_defaults: dict = { |
12 | 12 | 'attn_type': 'multihead_attention',
|
13 | 13 | 'attn_pdrop': 0.0,
|
14 | 14 | 'attn_impl': 'triton',
|
|
20 | 20 | 'alibi': False,
|
21 | 21 | 'alibi_bias_max': 8
|
22 | 22 | }
|
23 |
| -ffn_config_defaults: Dict = {'ffn_type': 'mptmlp'} |
24 |
| -init_config_defaults: Dict = { |
| 23 | +ffn_config_defaults: dict = {'ffn_type': 'mptmlp'} |
| 24 | +init_config_defaults: dict = { |
25 | 25 | 'name': 'kaiming_normal_',
|
26 | 26 | 'fan_mode': 'fan_in',
|
27 | 27 | 'init_nonlinearity': 'relu',
|
@@ -52,15 +52,15 @@ def __init__(self,
|
52 | 52 | resid_pdrop: float = 0.0,
|
53 | 53 | emb_pdrop: float = 0.0,
|
54 | 54 | learned_pos_emb: bool = True,
|
55 |
| - attn_config: Dict = attn_config_defaults, |
56 |
| - ffn_config: Dict = ffn_config_defaults, |
| 55 | + attn_config: dict = attn_config_defaults, |
| 56 | + ffn_config: dict = ffn_config_defaults, |
57 | 57 | init_device: str = 'cpu',
|
58 | 58 | logit_scale: Optional[Union[float, str]] = None,
|
59 | 59 | no_bias: bool = False,
|
60 | 60 | embedding_fraction: float = 1.0,
|
61 | 61 | norm_type: str = 'low_precision_layernorm',
|
62 | 62 | use_cache: bool = False,
|
63 |
| - init_config: Dict = init_config_defaults, |
| 63 | + init_config: dict = init_config_defaults, |
64 | 64 | fc_type: str = 'torch',
|
65 | 65 | verbose: Optional[int] = None,
|
66 | 66 | **kwargs: Any):
|
@@ -102,8 +102,8 @@ def __init__(self,
|
102 | 102 | self._validate_config()
|
103 | 103 |
|
104 | 104 | def _set_config_defaults(
|
105 |
| - self, config: Dict[str, Any], |
106 |
| - config_defaults: Dict[str, Any]) -> Dict[str, Any]: |
| 105 | + self, config: dict[str, Any], |
| 106 | + config_defaults: dict[str, Any]) -> dict[str, Any]: |
107 | 107 | for (k, v) in config_defaults.items():
|
108 | 108 | if k not in config:
|
109 | 109 | config[k] = v
|
|
0 commit comments