-
Compare Siglip1 Siglip2
π53Compare SigLIP1 and SigLIP2 on zero shot classification
-
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
google/siglip2-base-patch16-224
Zero-Shot Image Classification β’ Updated β’ 378k β’ 90 -
google/siglip2-base-patch16-256
Zero-Shot Image Classification β’ 0.4B β’ Updated β’ 64.5k β’ 6
Collections
Discover the best community collections!
Collections including paper arxiv:2502.14786
-
seanghay/khmer_mpwt_speech
Viewer β’ Updated β’ 2.06k β’ 105 β’ 8 -
DeepSeek LLM: Scaling Open-Source Language Models with Longtermism
Paper β’ 2401.02954 β’ Published β’ 53 -
openai/whisper-large-v3-turbo
Automatic Speech Recognition β’ Updated β’ 3.48M β’ β’ 2.84k -
The Ultra-Scale Playbook
π3.71kThe ultimate guide to training LLM on large GPU Clusters
-
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
Scaling Text-Rich Image Understanding via Code-Guided Synthetic Multimodal Data Generation
Paper β’ 2502.14846 β’ Published β’ 14 -
RelaCtrl: Relevance-Guided Efficient Control for Diffusion Transformers
Paper β’ 2502.14377 β’ Published β’ 12
-
LinFusion: 1 GPU, 1 Minute, 16K Image
Paper β’ 2409.02097 β’ Published β’ 34 -
Phidias: A Generative Model for Creating 3D Content from Text, Image, and 3D Conditions with Reference-Augmented Diffusion
Paper β’ 2409.11406 β’ Published β’ 27 -
Diffusion Models Are Real-Time Game Engines
Paper β’ 2408.14837 β’ Published β’ 126 -
Segment Anything with Multiple Modalities
Paper β’ 2408.09085 β’ Published β’ 22
-
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
LongWriter-V: Enabling Ultra-Long and High-Fidelity Generation in Vision-Language Models
Paper β’ 2502.14834 β’ Published β’ 24 -
Qwen2.5-VL Technical Report
Paper β’ 2502.13923 β’ Published β’ 214 -
DICEPTION: A Generalist Diffusion Model for Visual Perceptual Tasks
Paper β’ 2502.17157 β’ Published β’ 52
-
QLIP: Text-Aligned Visual Tokenization Unifies Auto-Regressive Multimodal Understanding and Generation
Paper β’ 2502.05178 β’ Published β’ 10 -
Scaling Text-Rich Image Understanding via Code-Guided Synthetic Multimodal Data Generation
Paper β’ 2502.14846 β’ Published β’ 14 -
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
Efficient LLaMA-3.2-Vision by Trimming Cross-attended Visual Features
Paper β’ 2504.00557 β’ Published β’ 15
-
Compare Siglip1 Siglip2
π53Compare SigLIP1 and SigLIP2 on zero shot classification
-
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
google/siglip2-base-patch16-224
Zero-Shot Image Classification β’ Updated β’ 378k β’ 90 -
google/siglip2-base-patch16-256
Zero-Shot Image Classification β’ 0.4B β’ Updated β’ 64.5k β’ 6
-
LinFusion: 1 GPU, 1 Minute, 16K Image
Paper β’ 2409.02097 β’ Published β’ 34 -
Phidias: A Generative Model for Creating 3D Content from Text, Image, and 3D Conditions with Reference-Augmented Diffusion
Paper β’ 2409.11406 β’ Published β’ 27 -
Diffusion Models Are Real-Time Game Engines
Paper β’ 2408.14837 β’ Published β’ 126 -
Segment Anything with Multiple Modalities
Paper β’ 2408.09085 β’ Published β’ 22
-
seanghay/khmer_mpwt_speech
Viewer β’ Updated β’ 2.06k β’ 105 β’ 8 -
DeepSeek LLM: Scaling Open-Source Language Models with Longtermism
Paper β’ 2401.02954 β’ Published β’ 53 -
openai/whisper-large-v3-turbo
Automatic Speech Recognition β’ Updated β’ 3.48M β’ β’ 2.84k -
The Ultra-Scale Playbook
π3.71kThe ultimate guide to training LLM on large GPU Clusters
-
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
LongWriter-V: Enabling Ultra-Long and High-Fidelity Generation in Vision-Language Models
Paper β’ 2502.14834 β’ Published β’ 24 -
Qwen2.5-VL Technical Report
Paper β’ 2502.13923 β’ Published β’ 214 -
DICEPTION: A Generalist Diffusion Model for Visual Perceptual Tasks
Paper β’ 2502.17157 β’ Published β’ 52
-
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
Scaling Text-Rich Image Understanding via Code-Guided Synthetic Multimodal Data Generation
Paper β’ 2502.14846 β’ Published β’ 14 -
RelaCtrl: Relevance-Guided Efficient Control for Diffusion Transformers
Paper β’ 2502.14377 β’ Published β’ 12
-
QLIP: Text-Aligned Visual Tokenization Unifies Auto-Regressive Multimodal Understanding and Generation
Paper β’ 2502.05178 β’ Published β’ 10 -
Scaling Text-Rich Image Understanding via Code-Guided Synthetic Multimodal Data Generation
Paper β’ 2502.14846 β’ Published β’ 14 -
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Paper β’ 2502.14786 β’ Published β’ 158 -
Efficient LLaMA-3.2-Vision by Trimming Cross-attended Visual Features
Paper β’ 2504.00557 β’ Published β’ 15