Umitcan Sahin PRO

ucsahin

AI & ML interests

Visual Language Models, Large Language Models, Vision Transformers

Recent Activity

liked a model 4 days ago
ByteDance/Sa2VA-4B
liked a model 4 days ago
microsoft/phi-4
liked a model 4 days ago
ByteDance/Sa2VA-8B
View all activity

Organizations

None yet

ucsahin's activity

reacted to m-ric's post with šŸ¤—šŸš€šŸ”„ 6 days ago
view post
Post
4848
Since I published it on GitHub a few days ago,
Hugging Face's new agentic library š˜€š—ŗš—¼š—¹š—®š—“š—²š—»š˜š˜€ has gathered nearly 4k stars šŸ¤Æ

āž”ļø But we are just getting started on agents: so we are hiring an ML Engineer to join me and double down on this effort!

The plan is to build GUI agents: agents that can act on your computer with mouse & keyboard, like Claude Computer Use.

We will make it work better, and fully open. āœØ

Sounds like something you'd like to do? Apply here šŸ‘‰ https://apply.workable.com/huggingface/j/AF1D4E3FEB/
Ā·
reacted to singhsidhukuldeep's post with šŸ”„ 21 days ago
view post
Post
2183
Exciting News in AI: JinaAI Releases JINA-CLIP-v2!

The team at Jina AI has just released a groundbreaking multilingual multimodal embedding model that's pushing the boundaries of text-image understanding. Here's why this is a big deal:

šŸš€ Technical Highlights:
- Dual encoder architecture combining a 561M parameter Jina XLM-RoBERTa text encoder and a 304M parameter EVA02-L14 vision encoder
- Supports 89 languages with 8,192 token context length
- Processes images up to 512Ɨ512 pixels with 14Ɨ14 patch size
- Implements FlashAttention2 for text and xFormers for vision processing
- Uses Matryoshka Representation Learning for efficient vector storage

āš”ļø Under The Hood:
- Multi-stage training process with progressive resolution scaling (224ā†’384ā†’512)
- Contrastive learning using InfoNCE loss in both directions
- Trained on massive multilingual dataset including 400M English and 400M multilingual image-caption pairs
- Incorporates specialized datasets for document understanding, scientific graphs, and infographics
- Uses hard negative mining with 7 negatives per positive sample

šŸ“Š Performance:
- Outperforms previous models on visual document retrieval (52.65% nDCG@5)
- Achieves 89.73% image-to-text and 79.09% text-to-image retrieval on CLIP benchmark
- Strong multilingual performance across 30 languages
- Maintains performance even with 75% dimension reduction (256D vs 1024D)

šŸŽÆ Key Innovation:
The model solves the long-standing challenge of unifying text-only and multi-modal retrieval systems while adding robust multilingual support. Perfect for building cross-lingual visual search systems!

Kudos to the research team at Jina AI for this impressive advancement in multimodal AI!
New activity in ucsahin/TR-VLM-DPO-Dataset about 1 month ago
reacted to merve's post with šŸ”„šŸ‘€šŸ‘ about 2 months ago
view post
Post
2191
The authors of ColPali trained a retrieval model based on SmolVLM šŸ¤  vidore/colsmolvlm-alpha
TLDR;

- ColSmolVLM performs better than ColPali and DSE-Qwen2 on all English tasks

- ColSmolVLM is more memory efficient than ColQwen2 šŸ’—