Crispin Almodovar's picture
5 6

Crispin Almodovar

calmodovar
Β·

AI & ML interests

NLP, log anomaly detection, cyber intelligence

Recent Activity

reacted to openfree's post with πŸš€ 1 day ago
🌟 MoneyRadar - AI-Powered Global News Analysis System πŸ’» Live Demo: https://huggingface.co/spaces/openfree/MoneyRadar 🎯 Core Features 1. πŸ€– 24/7 Automated News Scanning Auto-collection of Top 100 trending news Real-time monitoring across 60 countries Smart filtering of investment-critical news 2. πŸ” Advanced Custom Search Unlimited keyword search capability Country/language-specific search options Real-time trend-based related keywords 3. 🎨 Smart Analysis & Visualization AI-powered sentiment analysis Automated content summarization Investment decision-supporting insights ⚑ Automated Information Collection Key Companies (NVIDIA, APPLE, TESLA, etc.) Earnings/Forecasts Product/Technology announcements Market share changes M&A and major news Financial Markets & Digital Assets Macroeconomic indicators Regulatory changes Market sentiment analysis Major exchange updates πŸ“Š Business Applications Real-time market trend tracking Competitor movement monitoring Early investment opportunity detection Risk early warning system 🌟 Key Differentiators Full Automation Zero manual intervention Real-time data updates Automated result storage/management User-Centric Design Intuitive interface Customizable alerts Mobile optimization Advanced Analytics News cross-checking Historical tracking Trend prediction support Join Community πŸ’¬ "With MoneyRadar, never miss a beat in the global market movements!"
upvoted a collection 2 days ago
DeepSeek-R1
View all activity

Organizations

logfit-project's profile picture test012345's profile picture

calmodovar's activity

reacted to openfree's post with πŸš€ 1 day ago
view post
Post
3883
🌟 MoneyRadar - AI-Powered Global News Analysis System

πŸ’» Live Demo: openfree/MoneyRadar

🎯 Core Features
1. πŸ€– 24/7 Automated News Scanning

Auto-collection of Top 100 trending news
Real-time monitoring across 60 countries
Smart filtering of investment-critical news

2. πŸ” Advanced Custom Search

Unlimited keyword search capability
Country/language-specific search options
Real-time trend-based related keywords

3. 🎨 Smart Analysis & Visualization

AI-powered sentiment analysis
Automated content summarization
Investment decision-supporting insights

⚑ Automated Information Collection
Key Companies (NVIDIA, APPLE, TESLA, etc.)

Earnings/Forecasts
Product/Technology announcements
Market share changes
M&A and major news

Financial Markets & Digital Assets

Macroeconomic indicators
Regulatory changes
Market sentiment analysis
Major exchange updates

πŸ“Š Business Applications

Real-time market trend tracking
Competitor movement monitoring
Early investment opportunity detection
Risk early warning system

🌟 Key Differentiators

Full Automation

Zero manual intervention
Real-time data updates
Automated result storage/management


User-Centric Design

Intuitive interface
Customizable alerts
Mobile optimization


Advanced Analytics

News cross-checking
Historical tracking
Trend prediction support



Join Community πŸ’¬
"With MoneyRadar, never miss a beat in the global market movements!"
reacted to merve's post with πŸ”₯ 1 day ago
view post
Post
2260
Oof, what a week! πŸ₯΅ So many things have happened, let's recap! merve/jan-24-releases-6793d610774073328eac67a9

Multimodal πŸ’¬
- We have released SmolVLM -- tiniest VLMs that come in 256M and 500M, with it's retrieval models ColSmol for multimodal RAG πŸ’—
- UI-TARS are new models by ByteDance to unlock agentic GUI control 🀯 in 2B, 7B and 72B
- Alibaba DAMO lab released VideoLlama3, new video LMs that come in 2B and 7B
- MiniMaxAI released Minimax-VL-01, where decoder is based on MiniMax-Text-01 456B MoE model with long context
- Dataset: Yale released a new benchmark called MMVU
- Dataset: CAIS released Humanity's Last Exam (HLE) a new challenging MM benchmark

LLMs πŸ“–
- DeepSeek-R1 & DeepSeek-R1-Zero: gigantic 660B reasoning models by DeepSeek, and six distilled dense models, on par with o1 with MIT license! 🀯
- Qwen2.5-Math-PRM: new math models by Qwen in 7B and 72B
- NVIDIA released AceMath and AceInstruct, new family of models and their datasets (SFT and reward ones too!)

Audio πŸ—£οΈ
- Llasa is a new speech synthesis model based on Llama that comes in 1B,3B, and 8B
- TangoFlux is a new audio generation model trained from scratch and aligned with CRPO

Image/Video/3D Generation ⏯️
- Flex.1-alpha is a new 8B pre-trained diffusion model by ostris similar to Flux
- tencent released Hunyuan3D-2, new 3D asset generation from images
Β·
reacted to merve's post with πŸ”₯ 25 days ago
view post
Post
4821
supercharge your LLM apps with smolagents πŸ”₯

however cool your LLM is, without being agentic it can only go so far

enter smolagents: a new agent library by Hugging Face to make the LLM write code, do analysis and automate boring stuff!

Here's our blog for you to get started https://huggingface.co/blog/smolagents
reacted to MoritzLaurer's post with πŸ‘ about 1 month ago
view post
Post
2605
Quite excited by the ModernBERT release! 0.15/0.4B small, 2T modern pre-training data and tokenizer with code, 8k context window, great efficient model for embeddings & classification!

This will probably be the basis for many future SOTA encoders! And I can finally stop using DeBERTav3 from 2021 :D

Congrats @answerdotai , @LightOnIO and collaborators like @tomaarsen !

Paper and models here πŸ‘‡https://huggingface.co/collections/answerdotai/modernbert-67627ad707a4acbf33c41deb
Β·
upvoted an article 3 months ago
view article
Article

Visually Multilingual: Introducing mcdse-2b

By marco β€’
β€’ 38
reacted to singhsidhukuldeep's post with πŸ‘€ 3 months ago
view post
Post
2163
While Google's Transformer might have introduced "Attention is all you need," Microsoft and Tsinghua University are here with the DIFF Transformer, stating, "Sparse-Attention is all you need."

The DIFF Transformer outperforms traditional Transformers in scaling properties, requiring only about 65% of the model size or training tokens to achieve comparable performance.

The secret sauce? A differential attention mechanism that amplifies focus on relevant context while canceling out noise, leading to sparser and more effective attention patterns.

How?
- It uses two separate softmax attention maps and subtracts them.
- It employs a learnable scalar Ξ» for balancing the attention maps.
- It implements GroupNorm for each attention head independently.
- It is compatible with FlashAttention for efficient computation.

What do you get?
- Superior long-context modeling (up to 64K tokens).
- Enhanced key information retrieval.
- Reduced hallucination in question-answering and summarization tasks.
- More robust in-context learning, less affected by prompt order.
- Mitigation of activation outliers, opening doors for efficient quantization.

Extensive experiments show DIFF Transformer's advantages across various tasks and model sizes, from 830M to 13.1B parameters.

This innovative architecture could be a game-changer for the next generation of LLMs. What are your thoughts on DIFF Transformer's potential impact?
  • 1 reply
Β·
upvoted an article 6 months ago
view article
Article

Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth

By mlabonne β€’
β€’ 265
upvoted an article 8 months ago
view article
Article

Multimodal Augmentation for Documents: Recovering β€œComprehension” in β€œReading and Comprehension” task

By danaaubakirova β€’
β€’ 17