#Qwen2

2025-11-05

Hướng dẫn tinh chỉnh mô hình Qwen2.5-Coder-1.5B cho phân tích cảm xúc tiếng Trung. Có thể chạy trên Google Colab miễn phí trong 20-30 phút. Độ chính xác tăng từ 91,6% lên 97,8%. #AI #MachineLearning #Qwen2.5 #PhânTíchCảmXúc #GoogleColab #TinhChỉnhMôHình #TríTuệNhânTạo #HọcMáy

i.redd.it/7xx856mftfzf1.png

fosstopia 🇩🇪MichlFranken
2025-10-30
2025-10-15

Ch peque nhá! Tôi vừa chuyển sang dùng Qwen2.5 Code Instruct bản tự-host thành công! M المقابل với Claude đầu tiên (lần nào 1h phải chờ), Qwen2.5 có thể xử lý comuni code, debug, và nhiếp ý nhanh lùi ởстром đường công việc. Ưbrochen ở máy MBook Pro 48GB và PC 2x RTX 5060TI 16GB (không cần quantize). Cài đặt đơn giản, chất lượng tốt cho công việc lẻ lậu.
Tham khảo GitHub: @reliableJARED/qwen_coder
Tags: #AI #Qwen2.5 #CodeAssistant #LocalTech #MáyTínhLâu
#TechTips #OfflineAI #DevelopersCommu

2025-04-25

🧠 #ByteDance ha rilasciato UI-TARS-1.5, un agente multimodale basato su #Qwen2.5-VL-7B che unisce visione e linguaggio con "reasoning". 

👉 I dettagli: linkedin.com/posts/alessiopoma

___ 

✉️ 𝗦𝗲 𝘃𝘂𝗼𝗶 𝗿𝗶𝗺𝗮𝗻𝗲𝗿𝗲 𝗮𝗴𝗴𝗶𝗼𝗿𝗻𝗮𝘁𝗼/𝗮 𝘀𝘂 𝗾𝘂𝗲𝘀𝘁𝗲 𝘁𝗲𝗺𝗮𝘁𝗶𝗰𝗵𝗲, 𝗶𝘀𝗰𝗿𝗶𝘃𝗶𝘁𝗶 𝗮𝗹𝗹𝗮 𝗺𝗶𝗮 𝗻𝗲𝘄𝘀𝗹𝗲𝘁𝘁𝗲𝗿: bit.ly/newsletter-alessiopomar 

#AI #GenAI #GenerativeAI #IntelligenzaArtificiale #LLM 

KINEWS24KiNews
2025-03-31

​Qwen2.5-VL & QVQ-Max: Neue Maßstäbe in der visuellen KI

Fortschrittliche Bild- und Videoanalyse
Präzise Objekterkennung
Verbesserte Dokumentenverarbeitung

.5-VL -Max

Jetzt lesen und folgen!

kinews24.de/qwen2-5-vl-qvq-max/

Alibaba Cloud shakes up the AI scene with **Qwen2.5-Omni-7B!** This cutting-edge multimodal model processes text, images, audio, and video, making it perfect for mobile devices. It's designed for cost-effective AI agents, especially in voice applications for the visually impaired. With a hefty **$53 billion** investment in AI and cloud infrastructure, Alibaba is positioning itself for success in the booming AI market—don’t miss the full story. [Read more](cnbc.com/2025/03/27/alibaba-la) #ArtificialIntelligence #AlibabaCloud #Qwen2 #TechInnovation

N-gated Hacker Newsngate
2025-03-24

Qwen2.5-VL-32B: because nothing says "cutting-edge" like moaning about parameter scales and reinforcement learning 🙄. Apparently, this 32B thing is "smarter" and "lighter" – sounds like a diet ad for AI models. 😂🍩 !
qwenlm.github.io/blog/qwen2.5- .5VL32B

2025-03-16

OLMo 2 32B offers unprecedented transparency in #LLM development:

• 🚀 State-of-the-art results: Outperforms GPT3.5, GPT4o-mini, matches top open-weight models like #Qwen2.5 and approaches #Llama3

2025-03-16

#AI2 releases OLMo 2 32B, trained on 6T tokens with #Tulu3.1 post-training. Matches or exceeds GPT3.5 Turbo while using just 1/3 the compute of #Qwen2.5 32B. Complete open recipe includes data, code, weights and training methodology.

Rod2ik 🇪🇺 🇨🇵 🇪🇸 🇺🇦 🇨🇦 🇩🇰 🇬🇱☮🕊️rod2ik
2025-02-07
2025-01-31

Really going down the #Ollama rabbit hole. #VScode with #qwen2.5-coder, #deepseek-r1 and #localaipilot for fully contained local chat and code completion. Interested in hearing what others are doing for locally hosted ai.

2024-11-29

🎯 #OpenSource Language Model Platform Launch

🔧 Leverages #vLLM technology with custom #GPU scheduler for running various #LLM models
🤖 Supports major models: #Llama3 (405B/70B/8B), #Qwen2 72B, #Mixtral, #Gemma2, #Jamba15, #Phi3

glhf.chat/

2024-11-26

Advanced Reasoning Model: #ai #llm Marco-o1 Pushes Boundaries in Problem-Solving 🧠

🔬 Built on #Qwen2, focusing on open-ended reasoning beyond traditional tasks

💡 Key Innovations:
🤔 #ChainOfThought fine-tuning for structured reasoning
🌳 Monte Carlo Tree Search (#MCTS) for solution space exploration
🔄 Novel reflection mechanisms for self-improvement
🎯 Multiple action granularities for complex problem-solving

📊 Performance Highlights:
📈 +6.17% accuracy on MGSM English dataset
📈 +5.60% accuracy on MGSM Chinese dataset
🌐 Excels in translation tasks, especially with colloquial expressions

🛠️ Technical Features:
• Fine-tuned on 60,266 training samples
• Implements step & mini-step MCTS strategies
• Utilizes confidence scoring for path selection
• Incorporates self-reflection mechanisms

⚡️ Project Status: Research work in progress with continuous optimization
github.com/AIDC-AI/Marco-o1

2024-11-26

Edge-Ready #Vision Language Model Advances Visual #AI Processing 🌟

🧠 #OmniVision (968M params) sets new benchmark as world's smallest #VisionLanguageModel

🔄 Architecture combines #Qwen2 (0.5B) for text & #SigLIP (400M) for vision processing

💡 Key Innovations:
• 9x token reduction (729 → 81) for faster processing
• Enhanced accuracy through #DPO training
• Only 988MB RAM & 948MB storage required
• Outperforms #nanoLLAVA across multiple benchmarks

🎯 Use Cases:
• Image analysis & description
• Visual memory assistance
• Recipe generation from food images
• Technical documentation support

Try it now: huggingface.co/spaces/NexaAIDe
Source: nexa.ai/blogs/omni-vision

2024-11-20

New Cloud Platform for Large Language Model Deployment 🚀

🔧 Run any #opensource #LLM supported by #vLLM on autoscaling #GPU clusters, supporting models up to 640GB VRAM

🤖 Compatible with major models: #Llama3 405B/70B/8B, #Qwen2 72B, #Mixtral 8x22B, #Gemma2 27B, #Phi3, and more

💻 Features include:
- #OpenAI compatible #API
- Custom-built #GPU scheduler
- Support for full-weight and 4-bit AWQ repos
- Multi-tenant architecture for cost efficiency

🆓 Currently free during beta phase, promising competitive pricing post-launch

glhf.chat/landing/home

R (rolandinsh)r@toot.lv
2024-11-14

Pamēģināju jauno #AI #LLM #Qwen2.5 #coder, tas, kurš 32B. Nav slikti, bet izskatās, ka manām / www.MediaBox.lv vajadzībām pieteik arī ar 14B. Smuki izlaboju dažas nepilnības, kas nebija īsti kļūdas, bet labās prakses ievērošana strukturējot #PHP kodu.
Bail domāt par dāvanu sev - jaunu grafisko karti un visu ar to saistīto, lai šajā mestos iekšā vēl dziļāk...
#programmēšana #EsNoLaukiem

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst