Releases: illuin-tech/colpali
Releases · illuin-tech/colpali
v0.3.13: ModernVBert
[0.3.13] - 2025-11-15
Added
- Add ModernVBERT to the list of supported models
Fixed
- Fix multi hard negatives training
- Fix multi dataset sampling in order to weight probability of being picked by the size of the dataset
Changed
- Bump transformer, torch and peft support
v0.3.12
[0.3.12] - 2025-07-16
Added
- Video processing for ColQwen-Omni
Fixed
- Fixed loading of PaliGemma and ColPali checkpoints (bug introduced in transformers 4.52)
- Fixed loading of SmolVLM (Idefics3) processors that didn't transmit image_seq_len (bug introduced in transformers 4.52)
v0.3.11
[0.3.11] - 2025-07-04
Added
- Added BiIdefics3 modeling and processor.
- [Breaking] (minor) Remove support for context-augmented queries and images
- Uniform processor docstring
- Update the collator to align with the new function signatures
- Add a
process_textmethod to replace theprocess_queryone. We keep support of the last one for the moment, but we'll deprecate it later - Introduce the ColPaliEngineDataset and Corpus class. This is to delegate all data loading to a standard format before training. The concept is for users to override the dataset class if needed for their specific usecases.
- Added smooth_max option to loss functions
- Added weighted in_batch terms for losses with hard negatives
- Added an option to filter out (presumably) false negatives during online training
- Added a training script in pure torch without the HF trainer
- Added a sampler to train with multiple datasets at once, with each batch coming from the same source. (experimental, might still need testing on multi-GPU)
- Adds score normalization to LI models (diving by token length) for betetr performance with CE loss
- Add experimental PLAID support
Changed
- Stops pooling queries between GPUs and instead pools only documents, enabling training with way bigger batch sizes. We recomment training with accelerate launch now.
- Updated loss functions for better abstractions and coherence between the various loss functions. Small speedups and less memory requirements.
v0.3.10: minor updates & dependency bumps
[0.3.10] - 2025-04-18
Added
- Add
LambdaTokenPoolerto allow for custom token pooling functions. - Added training losses with negatives to InfoNCE type losses
Changed
- Fix similarity map helpers for ColQwen2 and ColQwen2.5.
- [Breaking] (minor) Remove support for Idefics2-based models.
- Disable multithreading in
HierarchicalTokenPoolerifnum_workersis not provided or is 1. - [Breaking] (minor) Make
pool_factoran argument ofpool_embeddingsinstead of aHierarchicalTokenPoolerclass attribute - Bump dependencies for transformers, torch, peft, pillow, accelerate, etc...
v0.3.9
Added
- Allow user to pass custom textual context for passage inference
- Add ColQwen2.5 support and BiQwen2.5 support
- Add support for token pooling with
HierarchicalTokenPooler. - Allow user to specify the maximum number of image tokens in the resized images in
ColQwen2ProcessorandColQwen2_5_Processor.
Changed
- Warn about evaluation being different from Vidore, and do not store results to prevent confusion.
- Remove duplicate resize code in
ColQwen2ProcessorandColQwen2_5_Processor. - Simplify sequence padding for pixel values in
ColQwen2ProcessorandColQwen2_5_Processor. - Remove deprecated evaluation (
CustomRetrievalEvaluator) from trainer - Refactor the collator classes
- Make
processorinput compulsory inColModelTrainingConfig - Make
BaseVisualRetrieverProcessorinherit fromProcessorMixin - Remove unused
tokenizerfield fromColModelTrainingConfig - Bump transformers to
4.50.0and torch to2.6.0to keep up with the latest versions. Note that this leads to errors on mps until transformers 4.50.4 is released.
v0.3.8
v0.3.7
v0.3.6
Description
Loosen default dependencies, but keep stricter dep ranges for the train dependency group.
Features
Added
- Add expected scores in ColPali E2E test
Changed
- Loosen package dependencies
Full Changelog: v0.3.5...v0.3.6
v0.3.5: SmolVM
[0.3.5] - 2024-12-13
Added
- Added support for Idefics3 (and SmolVLM)
Fixed
- Fix typing for
processor.score_multi_vector(allow for both list and tensor inputs). This does not change how the scores are computed. - Fix
tear_down_torchwhen used on a non-MPS machine
v0.3.4
[0.3.4] - 2024-11-07
Added
- General
CorpusQueryCollatorfor BEIR style dataset training or hard negative training. This deprecatesHardNegCollatorbut all changes to the training loop are made for a seemless update.
Changed
- Updates BiPali config files
- Removed query augmentation tokens from BiQwen2Processor
- Modified XQwen2Processor to place
<|endoftext|>token at the end of the document prompt (non-breaking for ColQwen but helps BiQwen). - Removed
add_suffixin the VisualRetrieverCollator and let thesuffixbe added in the individual processors. - Changed the incorrect
<pad>token to<|endoftext|>fo query augmentationColQwen2Processor. Note that previous models were trained with<|endoftext|>so this is simply a non-breaking inference upgrade patch.