| from typing import TYPE_CHECKING | |
| from transformers.utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available | |
| _import_structure = { | |
| "configuration_codify": ["CODIFY_PRETRAINED_CONFIG_ARCHIVE_MAP", "CodifyConfig", "CodifyOnnxConfig"], | |
| } | |
| try: | |
| if not is_tokenizers_available(): | |
| raise OptionalDependencyNotAvailable() | |
| except OptionalDependencyNotAvailable: | |
| pass | |
| else: | |
| _import_structure["tokenization_codify_fast"] = ["CodifyTokenizerFast"] | |
| try: | |
| if not is_torch_available(): | |
| raise OptionalDependencyNotAvailable() | |
| except OptionalDependencyNotAvailable: | |
| pass | |
| else: | |
| _import_structure["modeling_codify"] = [ | |
| "CODIFY_PRETRAINED_MODEL_ARCHIVE_LIST", | |
| "CodifyForCausalLM", | |
| "CodifyModel", | |
| "CodifyPreTrainedModel", | |
| ] | |
| if TYPE_CHECKING: | |
| from .configuration_codify import CODIFY_PRETRAINED_CONFIG_ARCHIVE_MAP, CodifyConfig, CodifyOnnxConfig | |
| try: | |
| if not is_tokenizers_available(): | |
| raise OptionalDependencyNotAvailable() | |
| except OptionalDependencyNotAvailable: | |
| pass | |
| else: | |
| from .tokenization_codify_fast import CodifyTokenizerFast | |
| try: | |
| if not is_torch_available(): | |
| raise OptionalDependencyNotAvailable() | |
| except OptionalDependencyNotAvailable: | |
| pass | |
| else: | |
| from .modeling_codify import ( | |
| CODIFY_PRETRAINED_MODEL_ARCHIVE_LIST, | |
| CodifyForCausalLM, | |
| CodifyModel, | |
| CodifyPreTrainedModel, | |
| ) | |
| else: | |
| import sys | |
| sys.modules[__name__] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__) | |