You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
Similar error by running "python app.py" and caught this error: "HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': 'checkpoints/llama-7b_hf/'. Use repo_type argument if needed."
Similar error by running "python app.py" and caught this error: "HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': 'checkpoints/llama-7b_hf/'. Use repo_type argument if needed."
I fixed it by replace app.py:359 from llama_path = "checkpoints/llama-7b_hf/" to llama_path = "checkpoints/llama-7b-hf/"
执行命名如下
torchrun --nproc_per_node=8 /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py
--lm_path /home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf
--tokenizer_path /home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf
--pretrained_path /home/jn/th/work/Multimodal-GPT/checkpoints/OpenFlamingo-9B/checkpoint.pt
--run_name train-my-gpt4
--learning_rate 1e-5
--lr_scheduler cosine
--batch_size 1
--tuning_config /home/jn/th/work/Multimodal-GPT/configs/lora_config.py
--dataset_config /home/jn/th/work/Multimodal-GPT/configs/dataset_config.py
--report_to_wandb
如下错误:
WARNING:torch.distributed.run:
Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init tokenizer
init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.init tokenizer
init tokenizer
init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use
repo_type
argument if needed." ││ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type
argument if needed.WARNING:torch.distributed.elastic.multiprocessing.api:Sending process 23886 closing signal SIGTERM
ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 1) local_rank: 0 (pid: 23882) of binary: /opt/conda/bin/python
Traceback (most recent call last):
File "/opt/conda/bin/torchrun", line 8, in
sys.exit(main())
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/elastic/multiprocessing/errors/init.py", line 346, in wrapper
return f(*args, **kwargs)
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/run.py", line 762, in main
run(args)
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/run.py", line 756, in run
)(*cmd_args)
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/launcher/api.py", line 132, in call
return launch_agent(self._config, self._entrypoint, list(args))
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/launcher/api.py", line 248, in launch_agent
failures=result.failures,
torch.distributed.elastic.multiprocessing.errors.ChildFailedError:
/home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py FAILED
Failures:
[1]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 1 (local_rank: 1)
exitcode : 1 (pid: 23883)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[2]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 2 (local_rank: 2)
exitcode : 1 (pid: 23884)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[3]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 3 (local_rank: 3)
exitcode : 1 (pid: 23885)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[4]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 5 (local_rank: 5)
exitcode : 1 (pid: 23887)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[5]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 6 (local_rank: 6)
exitcode : 1 (pid: 23888)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[6]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 7 (local_rank: 7)
exitcode : 1 (pid: 23889)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
Root Cause (first observed failure):
[0]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 0 (local_rank: 0)
exitcode : 1 (pid: 23882)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
The text was updated successfully, but these errors were encountered: