88from ais_bench .benchmark .models .local_models .base import BaseModel
99from ais_bench .benchmark .models import APITemplateParser
1010from ais_bench .benchmark .registry import MODELS
11-
12- from mindspore import Tensor , Model
13- from mindformers import MindFormerConfig , build_context
14- from mindformers .models import build_network
15- from mindformers .core .parallel_config import build_parallel_config
16- from mindformers .utils .load_checkpoint_utils import get_load_path_after_hf_convert
17- from mindformers .trainer .utils import transform_and_load_checkpoint
18-
11+ try :
12+ from mindspore import Tensor , Model
13+ from mindformers import MindFormerConfig , build_context
14+ from mindformers .models import build_network
15+ from mindformers .core .parallel_config import build_parallel_config
16+ from mindformers .utils .load_checkpoint_utils import get_load_path_after_hf_convert
17+ from mindformers .trainer .utils import transform_and_load_checkpoint
18+ except ModuleNotFoundError as _mf_err :
19+ MIND_FORMERS_INSTALL_GUIDE = (
20+ "MindSpore/MindFormers is not available in the current environment. "
21+ "To use `MindFormerModel`, install MindSpore and MindFormers and ensure any required "
22+ "environment variables for your device are configured.\n \n "
23+ "Quick steps (examples):\n "
24+ " 1) Follow the official MindSpore installation guide for your platform: "
25+ "https://www.mindspore.cn/install/en.\n "
26+ " 2) Install MindFormers: `pip install mindformers` (or follow its docs); "
27+ "alternatively, add a local MindFormers checkout to your PYTHONPATH, e.g.:\n "
28+ " PYTHONPATH=/path/to/mindformers:$PYTHONPATH\n "
29+ )
30+ raise ModuleNotFoundError (MIND_FORMERS_INSTALL_GUIDE ) from _mf_err
1931
2032
2133class MultiTokenEOSCriteria (transformers .StoppingCriteria ):
@@ -61,7 +73,7 @@ def __init__(self,
6173 path : str ,
6274 checkpoint : Optional [str ] = None ,
6375 yaml_cfg_file : Optional [str ] = None ,
64- batch_size : int = 1 ,
76+ build_batch_size : int = 1 ,
6577 max_seq_len : int = 2048 ,
6678 tokenizer_path : Optional [str ] = None ,
6779 tokenizer_kwargs : dict = dict (),
@@ -79,7 +91,7 @@ def __init__(self,
7991 max_seq_len = max_seq_len ,
8092 tokenizer_only = tokenizer_only ,
8193 meta_template = meta_template )
82- self .batch_size = batch_size
94+ self .batch_size = build_batch_size
8395 self .pad_token_id = pad_token_id
8496 self .pretrained_model_path = path
8597 if mode not in ['none' , 'mid' ]:
0 commit comments