| 
39 | 39 | # so that mypy, pylint or other static linters can recognize them,  | 
40 | 40 | # given that they are not exported using `__all__` in this file.  | 
41 | 41 | from .utils import is_bitsandbytes_available as is_bitsandbytes_available  | 
42 |  | -from .utils import is_flax_available as is_flax_available  | 
43 |  | -from .utils import is_keras_nlp_available as is_keras_nlp_available  | 
44 | 42 | from .utils import is_scipy_available as is_scipy_available  | 
45 | 43 | from .utils import is_sentencepiece_available as is_sentencepiece_available  | 
46 | 44 | from .utils import is_speech_available as is_speech_available  | 
47 |  | -from .utils import is_tensorflow_text_available as is_tensorflow_text_available  | 
48 | 45 | from .utils import is_timm_available as is_timm_available  | 
49 | 46 | from .utils import is_tokenizers_available as is_tokenizers_available  | 
50 | 47 | from .utils import is_torch_available as is_torch_available  | 
 | 
223 | 220 |         "is_bitsandbytes_available",  | 
224 | 221 |         "is_datasets_available",  | 
225 | 222 |         "is_faiss_available",  | 
226 |  | -        "is_flax_available",  | 
227 |  | -        "is_keras_nlp_available",  | 
228 | 223 |         "is_matplotlib_available",  | 
229 | 224 |         "is_phonemizer_available",  | 
230 | 225 |         "is_psutil_available",  | 
 | 
236 | 231 |         "is_sentencepiece_available",  | 
237 | 232 |         "is_sklearn_available",  | 
238 | 233 |         "is_speech_available",  | 
239 |  | -        "is_tensorflow_text_available",  | 
240 | 234 |         "is_timm_available",  | 
241 | 235 |         "is_tokenizers_available",  | 
242 | 236 |         "is_torch_available",  | 
 | 
573 | 567 |     from .generation import EpsilonLogitsWarper as EpsilonLogitsWarper  | 
574 | 568 |     from .generation import EtaLogitsWarper as EtaLogitsWarper  | 
575 | 569 |     from .generation import ExponentialDecayLengthPenalty as ExponentialDecayLengthPenalty  | 
576 |  | -    from .generation import FlaxForcedBOSTokenLogitsProcessor as FlaxForcedBOSTokenLogitsProcessor  | 
577 |  | -    from .generation import FlaxForcedEOSTokenLogitsProcessor as FlaxForcedEOSTokenLogitsProcessor  | 
578 |  | -    from .generation import FlaxForceTokensLogitsProcessor as FlaxForceTokensLogitsProcessor  | 
579 |  | -    from .generation import FlaxGenerationMixin as FlaxGenerationMixin  | 
580 |  | -    from .generation import FlaxLogitsProcessor as FlaxLogitsProcessor  | 
581 |  | -    from .generation import FlaxLogitsProcessorList as FlaxLogitsProcessorList  | 
582 |  | -    from .generation import FlaxLogitsWarper as FlaxLogitsWarper  | 
583 |  | -    from .generation import FlaxMinLengthLogitsProcessor as FlaxMinLengthLogitsProcessor  | 
584 |  | -    from .generation import FlaxSuppressTokensAtBeginLogitsProcessor as FlaxSuppressTokensAtBeginLogitsProcessor  | 
585 |  | -    from .generation import FlaxSuppressTokensLogitsProcessor as FlaxSuppressTokensLogitsProcessor  | 
586 |  | -    from .generation import FlaxTemperatureLogitsWarper as FlaxTemperatureLogitsWarper  | 
587 |  | -    from .generation import FlaxTopKLogitsWarper as FlaxTopKLogitsWarper  | 
588 |  | -    from .generation import FlaxTopPLogitsWarper as FlaxTopPLogitsWarper  | 
589 |  | -    from .generation import FlaxWhisperTimeStampLogitsProcessor as FlaxWhisperTimeStampLogitsProcessor  | 
590 | 570 |     from .generation import ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor  | 
591 | 571 |     from .generation import ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor  | 
592 | 572 |     from .generation import GenerationConfig as GenerationConfig  | 
 | 
648 | 628 |     from .integrations import is_wandb_available as is_wandb_available  | 
649 | 629 |     from .integrations.executorch import TorchExportableModuleWithStaticCache as TorchExportableModuleWithStaticCache  | 
650 | 630 |     from .integrations.executorch import convert_and_export_with_cache as convert_and_export_with_cache  | 
651 |  | -    from .keras_callbacks import KerasMetricCallback as KerasMetricCallback  | 
652 |  | -    from .keras_callbacks import PushToHubCallback as PushToHubCallback  | 
653 | 631 |     from .masking_utils import AttentionMaskInterface as AttentionMaskInterface  | 
654 | 632 |     from .model_debugging_utils import model_addition_debugger_context as model_addition_debugger_context  | 
655 | 633 | 
 
  | 
656 | 634 |     # Model Cards  | 
657 | 635 |     from .modelcard import ModelCard as ModelCard  | 
658 |  | -    from .modeling_flax_utils import FlaxPreTrainedModel as FlaxPreTrainedModel  | 
659 | 636 |     from .modeling_layers import GradientCheckpointingLayer as GradientCheckpointingLayer  | 
660 | 637 |     from .modeling_rope_utils import ROPE_INIT_FUNCTIONS as ROPE_INIT_FUNCTIONS  | 
661 | 638 |     from .modeling_rope_utils import dynamic_rope_update as dynamic_rope_update  | 
662 |  | - | 
663 | 639 |     from .modeling_utils import AttentionInterface as AttentionInterface  | 
664 | 640 |     from .modeling_utils import PreTrainedModel as PreTrainedModel  | 
665 | 641 |     from .models import *  | 
 | 
0 commit comments