From 152569e7482df006e7956388199490e7c86bdfbe Mon Sep 17 00:00:00 2001 From: Cyril Vallez Date: Tue, 1 Oct 2024 15:47:41 +0200 Subject: [PATCH] Update __all__ --- src/transformers/models/glm/configuration_glm.py | 3 +++ src/transformers/models/glm/modeling_glm.py | 3 +++ 2 files changed, 6 insertions(+) diff --git a/src/transformers/models/glm/configuration_glm.py b/src/transformers/models/glm/configuration_glm.py index eb2fe38f1117c8..93e48edf59537d 100644 --- a/src/transformers/models/glm/configuration_glm.py +++ b/src/transformers/models/glm/configuration_glm.py @@ -141,3 +141,6 @@ def __init__( tie_word_embeddings=tie_word_embeddings, **kwargs, ) + + +__all__ = ["GlmConfig"] diff --git a/src/transformers/models/glm/modeling_glm.py b/src/transformers/models/glm/modeling_glm.py index a72597a31fdcf3..5a3c501032fdfa 100644 --- a/src/transformers/models/glm/modeling_glm.py +++ b/src/transformers/models/glm/modeling_glm.py @@ -1386,3 +1386,6 @@ def forward( hidden_states=outputs.hidden_states, attentions=outputs.attentions, ) + + +__all__ = ["GlmModel", "GlmForCausalLM", "GlmForSequenceClassification", "GlmForTokenClassification"]