chore(deps): update all dependencies #296
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
==2.6.4
->==2.7.0
==2.16.3
->==2.18.1
==2023.12.25
->==2024.4.16
v1.25.0
->v1.26.0
==0.15.2
->==0.19.1
==4.39.3
->==4.40.0
Release Notes
pydantic/pydantic (pydantic)
v2.7.0
Compare Source
GitHub release
The code released in v2.7.0 is practically identical to that of v2.7.0b1.
What's Changed
Packaging
pyproject.toml
sections by @Viicos in #8899pydantic-core
tov2.18.1
by @sydney-runkle in #9211jiter
v0.2.0
by @samuelcolvin in pydantic/pydantic-core#1250New Features
FieldInfo.description
by @Viicos in #6563with_config
decorator to comply with typing spec by @Viicos in #8611ByteSize.human_readable
by @jks15satoshi in #8706Secret
base type by @conradogarciaberrotaran in #8519Sphinx
inventories for cross references in docs by @Viicos in #8682deprecated
fields by @Viicos in #8237field_serializer('*')
by @ornariece in #9001model_config
is defined as a model property by @alexeyt101 in #9004create_model()
to supporttyping.Annotated
as input by @wannieman98 in #8947ClickhouseDsn
support by @solidguy7 in #9062re.Pattern[str]
topattern
field by @jag-k in #9053serialize_as_any
runtime setting by @sydney-runkle in #8830typing.Self
by @Youssefares in #9023context
to serialization by @ornariece in #8965Finalized in v2.7.0, rather than v2.7.0b1:
warnings
parameter for serialization utilities to allow raising a warning by @Lance-Drane in #9166Changes
model_construct
behavior withextra
by @sydney-runkle in #8807RootModel
subclasses by @sydney-runkle in #8857PEP570
syntax by @Viicos in #8940enum
andtype
to the JSON schema for single item literals by @dmontagu in #8944update_json_schema
internal function by @sydney-runkle in #9125Performance
enum
validator improvements by @samuelcolvin in #9045enum
validation and serialization to Rust by @samuelcolvin in #9064aarch64
(Note: SIMD on x86 will be implemented in a future release) by @samuelcolvin in in pydantic/jiter#65Cow<str>
fromjiter
by @davidhewitt in pydantic/pydantic-core#1231Fixes
Sequence
s by @sydney-runkle in #8614__qualname__
by @anci3ntr0ck in #8642__pydantic_extra__
annotation being a string or inherited by @alexmojaki in #8659NameEmail
by @Holi0317 in #8650BaseModel
by @bluenote10 in #8651mypy
plugin andno_strict_optional = True
by @dmontagu in #8666ByteSize
errortype
change by @sydney-runkle in #8681__pydantic_config__
ignored for TypeDict by @13sin in #8734pytest v8.0.0
due topytest.warns()
starting to work insidepytest.raises()
by @mgorny in #8678is_valid_field
from 1.x formypy
plugin by @DanielNoord in #8738mypy
strict equality flag by @dmontagu in #8799FieldInfo.__repr_args__
by @sydney-runkle in #8801BaseModel
type annotations to be resolvable bytyping.get_type_hints
by @devmonkey22 in #7680AliasGenerator
by @sydney-runkle in #8810date
->datetime
timezone assumption fix by @sydney-runkle in #8823ast.Str
by @Viicos in #8837deprecated
decorators by @Viicos in #8877NameEmail
if name includes an email address by @NeevCohen in #8860TypeAdapter
's typing compatible with special forms by @adriangb in #8923enum
s by @dmontagu in #8920model_json_schema
usage by @sydney-runkle in #8928mypy
plugin by @dmontagu in #9008PlainSerializer
usage with std type constructor by @sydney-runkle in #9031Model.__getattr__()
by @NeevCohen in #9082ClassVar
forward ref inherited from parent class by @alexmojaki in #9097True
by @andresliszt in #8977deque
when passed toSequence[blah blah blah]
by @sydney-runkle in #9128model_post_init
by @Viicos in #9134model_construct
withvalidation_alias
by @ornariece in #9144Literal
null
types by @bruno-f-cruz in #9135New Contributors
pydantic/pydantic-core (pydantic_core)
v2.18.1
: 2024-04-11Compare Source
What's Changed
coerce_numbers_to_str option
inStringSchema
by @NeevCohen in https://github.com/pydantic/pydantic-core/pull/1262str
->int
by @samuelcolvin in https://github.com/pydantic/pydantic-core/pull/1266New Contributors
Full Changelog: pydantic/pydantic-core@v2.18.0...v2.18.1
v2.18.0
: 2024-04-02Compare Source
What's Changed
ser_json_inf_nan
inference by @sydney-runkle in https://github.com/pydantic/pydantic-core/pull/12510.2.0
by @samuelcolvin in https://github.com/pydantic/pydantic-core/pull/1250New Contributors
Full Changelog: pydantic/pydantic-core@v2.17.0...v2.18.0
v2.17.0
Compare Source
What's Changed
Packaging
smallvec
from 1.11.2 to 1.13.1 by @dependabot in https://github.com/pydantic/pydantic-core/pull/1181regex
from 1.10.2 to 1.10.3 by @dependabot in https://github.com/pydantic/pydantic-core/pull/1180uuid
from 1.6.1 to 1.7.0 by @dependabot in https://github.com/pydantic/pydantic-core/pull/1179serde
from 1.0.195 to 1.0.196 by @dependabot in https://github.com/pydantic/pydantic-core/pull/1177serde_json
from 1.0.109 to 1.0.114 by @dependabot in https://github.com/pydantic/pydantic-core/pull/1211ahash
from 0.8.7 to 0.8.10 by @dependabot in https://github.com/pydantic/pydantic-core/pull/1210strum_macros
from 0.25.3 to 0.26.1 by @dependabot in https://github.com/pydantic/pydantic-core/pull/1208PyO3
0.21 beta by @davidhewitt in https://github.com/pydantic/pydantic-core/pull/1222pyodide
to 0.25.0 by @samdobson in https://github.com/pydantic/pydantic-core/pull/1199speedate
by @sydney-runkle in https://github.com/pydantic/pydantic-core/pull/1244Fixes
__dict__
changes during iteration by @alexmojaki in https://github.com/pydantic/pydantic-core/pull/1196TzInfo
equality check based on offset by @13sin in https://github.com/pydantic/pydantic-core/pull/1197BigInt
fromstr
by @sydney-runkle in https://github.com/pydantic/pydantic-core/pull/1204_pydantic_core.pyi
by @Viicos in https://github.com/pydantic/pydantic-core/pull/1217Input
trait'a
and'py
lifetimes by @davidhewitt in https://github.com/pydantic/pydantic-core/pull/1227computed_field
is excluded by @sydney-runkle in https://github.com/pydantic/pydantic-core/pull/1228Input
forstr
by @davidhewitt in https://github.com/pydantic/pydantic-core/pull/1229with_new_extra
by @davidhewitt in https://github.com/pydantic/pydantic-core/pull/1233Input
trait to have singleas_python
cast for python inputs by @davidhewitt in https://github.com/pydantic/pydantic-core/pull/1241__pydantic_extra__
isNone
, even ifextra='allow'
by @sydney-runkle in https://github.com/pydantic/pydantic-core/pull/1236Performance
New Features
context
to serialization by @ornariece in https://github.com/pydantic/pydantic-core/pull/1215serialize_as_any
runtime flag support by @sydney-runkle in https://github.com/pydantic/pydantic-core/pull/1194Changes
speedate
change: Serialize duration to hour minute second, instead of just seconds by @kakilangit in https://github.com/pydantic/speedate/pull/50New Contributors
Full Changelog: pydantic/pydantic-core@v2.16.3...v2.17.0
mrabarnett/mrab-regex (regex)
v2024.4.16
Compare Source
slackapi/slack-github-action (slackapi/slack-github-action)
v1.26.0
: Slack Send V1.26.0Compare Source
What's Changed
This release provides an escape hatch for sending the JSON content of a payload file exactly as is, without replacing any templated variables!
Previously a payload file was parsed and templated variables were replaced with values from
github.context
andgithub.env
. Any undefined variables were replaced with???
in this process, which might have caused questions.That remains the default behavior, but now the JSON contents of a payload file can be sent exactly as written by setting the
payload-file-path-parsed
input tofalse
:With this change, the contents of the example
payload-slack-content.json
will be sent to a webhook URL exactly as is!Recent commits
Enhancements
Documentation
Maintenance
Dependencies
New Contributors
Full Changelog: slackapi/slack-github-action@v1.25.0...v1.26.0
huggingface/tokenizers (tokenizers)
v0.19.1
Compare Source
What's Changed
ignore_merges
by @ArthurZucker in https://github.com/huggingface/tokenizers/pull/1504Full Changelog: huggingface/tokenizers@v0.19.0...v0.19.1
v0.19.0
Compare Source
What's Changed
remove black
] And use ruff by @ArthurZucker in https://github.com/huggingface/tokenizers/pull/1436AddedVocabulary
. by @eaplatanios in https://github.com/huggingface/tokenizers/pull/1443Full Changelog: huggingface/tokenizers@v0.15.2...v0.19.0
huggingface/transformers (transformers)
v4.40.0
: : Llama 3, Idefics 2, Recurrent Gemma, Jamba, DBRX, OLMo, Qwen2MoE, Grounding DinoCompare Source
New model additions
Llama 3
Llama 3 is supported in this release through the Llama 2 architecture and some fixes in the
tokenizers
library.Idefics2
The Idefics2 model was created by the Hugging Face M4 team and authored by Léo Tronchon, Hugo Laurencon, Victor Sanh. The accompanying blog post can be found here.
Idefics2 is an open multimodal model that accepts arbitrary sequences of image and text inputs and produces text outputs. The model can answer questions about images, describe visual content, create stories grounded on multiple images, or simply behave as a pure language model without visual inputs. It improves upon IDEFICS-1, notably on document understanding, OCR, or visual reasoning. Idefics2 is lightweight (8 billion parameters) and treats images in their native aspect ratio and resolution, which allows for varying inference efficiency.
Recurrent Gemma
Recurrent Gemma architecture. Taken from the original paper.
The Recurrent Gemma model was proposed in RecurrentGemma: Moving Past Transformers for Efficient Open Language Models by the Griffin, RLHF and Gemma Teams of Google.
The abstract from the paper is the following:
We introduce RecurrentGemma, an open language model which uses Google’s novel Griffin architecture. Griffin combines linear recurrences with local attention to achieve excellent performance on language. It has a fixed-sized state, which reduces memory use and enables efficient inference on long sequences. We provide a pre-trained model with 2B non-embedding parameters, and an instruction tuned variant. Both models achieve comparable performance to Gemma-2B despite being trained on fewer tokens.
Jamba
Jamba is a pretrained, mixture-of-experts (MoE) generative text model, with 12B active parameters and an overall of 52B parameters across all experts. It supports a 256K context length, and can fit up to 140K tokens on a single 80GB GPU.
As depicted in the diagram below, Jamba’s architecture features a blocks-and-layers approach that allows Jamba to successfully integrate Transformer and Mamba architectures altogether. Each Jamba block contains either an attention or a Mamba layer, followed by a multi-layer perceptron (MLP), producing an overall ratio of one Transformer layer out of every eight total layers.
Jamba introduces the first
HybridCache
object that allows it to natively support assisted generation, contrastive search, speculative decoding, beam search and all of the awesome features from thegenerate
API!DBRX
DBRX is a transformer-based decoder-only large language model (LLM) that was trained using next-token prediction. It uses a fine-grained mixture-of-experts (MoE) architecture with 132B total parameters of which 36B parameters are active on any input.
It was pre-trained on 12T tokens of text and code data. Compared to other open MoE models like Mixtral-8x7B and Grok-1, DBRX is fine-grained, meaning it uses a larger number of smaller experts. DBRX has 16 experts and chooses 4, while Mixtral-8x7B and Grok-1 have 8 experts and choose 2.
This provides 65x more possible combinations of experts and the authors found that this improves model quality. DBRX uses rotary position encodings (RoPE), gated linear units (GLU), and grouped query attention (GQA).
OLMo
The OLMo model was proposed in OLMo: Accelerating the Science of Language Models by Dirk Groeneveld, Iz Beltagy, Pete Walsh, Akshita Bhagia, Rodney Kinney, Oyvind Tafjord, Ananya Harsh Jha, Hamish Ivison, Ian Magnusson, Yizhong Wang, Shane Arora, David Atkinson, Russell Authur, Khyathi Raghavi Chandu, Arman Cohan, Jennifer Dumas, Yanai Elazar, Yuling Gu, Jack Hessel, Tushar Khot, William Merrill, Jacob Morrison, Niklas Muennighoff, Aakanksha Naik, Crystal Nam, Matthew E. Peters, Valentina Pyatkin, Abhilasha Ravichander, Dustin Schwenk, Saurabh Shah, Will Smith, Emma Strubell, Nishant Subramani, Mitchell Wortsman, Pradeep Dasigi, Nathan Lambert, Kyle Richardson, Luke Zettlemoyer, Jesse Dodge, Kyle Lo, Luca Soldaini, Noah A. Smith, Hannaneh Hajishirzi.
OLMo is a series of Open Language Models designed to enable the science of language models. The OLMo models are trained on the Dolma dataset. We release all code, checkpoints, logs (coming soon), and details involved in training these models.
Qwen2MoE
Qwen2MoE is the new model series of large language models from the Qwen team. Previously, we released the Qwen series, including Qwen-72B, Qwen-1.8B, Qwen-VL, Qwen-Audio, etc.
Model Details
Qwen2MoE is a language model series including decoder language models of different model sizes. For each size, we release the base language model and the aligned chat model. Qwen2MoE has the following architectural choices:
Qwen2MoE is based on the Transformer architecture with SwiGLU activation, attention QKV bias, group query attention, mixture of sliding window attention and full attention, etc. Additionally, we have an improved tokenizer adaptive to multiple natural languages and codes.
Qwen2MoE employs Mixture of Experts (MoE) architecture, where the models are upcycled from dense language models. For instance, Qwen1.5-MoE-A2.7B is upcycled from Qwen-1.8B. It has 14.3B parameters in total and 2.7B activated parameters during runtime, while it achieves comparable performance with Qwen1.5-7B, with only 25% of the training resources.
Grounding Dino
Taken from the original paper.
The Grounding DINO model was proposed in Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection by Shilong Liu, Zhaoyang Zeng, Tianhe Ren, Feng Li, Hao Zhang, Jie Yang, Chunyuan Li, Jianwei Yang, Hang Su, Jun Zhu, Lei Zhang. Grounding DINO extends a closed-set object detection model with a text encoder, enabling open-set object detection. The model achieves remarkable results, such as 52.5 AP on COCO zero-shot.
Static pretrained maps
Static pretrained maps have been removed from the library's internals and are currently deprecated. These used to reflect all the available checkpoints for a given architecture on the Hugging Face Hub, but their presence does not make sense in light of the huge growth of checkpoint shared by the community.
With the objective of lowering the bar of model contributions and reviewing, we first start by removing legacy objects such as this one which do not serve a purpose.
Notable improvements
Processors improvements
Processors are ungoing changes in order to uniformize them and make them clearer to use.
SDPA
Push to Hub for pipelines
Pipelines can now be pushed to Hub using a convenient
push_to_hub
method.push_to_hub
to pipeline by @not-lain in #29172Flash Attention 2 for more models (M2M100, NLLB, GPT2, MusicGen) !
Thanks to the community contribution, Flash Attention 2 has been integrated for more architectures
Improvements and bugfixes
-
andthe
from custom_tools.md by @windsonsea in #29767-OO
mode fordocstring_decorator
by @matthid in #29689Latest PyTorch + TensorFlow [dev]
by @ydshieh in #29764LlavaNext
] Fix llava next unsafe imports by @ArthurZucker in #29773set_seed
by @muellerzr in #29778torch_dtype
in the run_mlm example by @jla524 in #29776bos token
to Blip generations by @zucchini-nlp in #29642quality
] update quality check to make sure we check imports 😈 by @ArthurZucker in #29771vocab_size
by @fxmarty in #29389AssistedCandidateGenerator
by @gante in #29787cleanup
] vestiges of causal mask by @ArthurZucker in #29806SuperPoint
] Fix doc example by @amyeroberts in #29816Configuration
📅 Schedule: Branch creation - "before 4am on Monday" (UTC), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired.
This PR has been generated by Renovate Bot.