#transformer

2025-12-01

Một công cụ tương tác mới giúp giải thích kiến trúc Transformer, tích hợp trợ lý AI để giải đáp thắc mắc. Rất hữu ích cho những ai muốn tìm hiểu về AI và machine learning!
#Transformer #AI #MachineLearning #InteractiveTool #DeepLearning
#KiếnTrúcTransformer #TríTuệNhânTạo #HọcMáy #CôngCụTươngTác #HọcSâu

reddit.com/r/SideProject/comme

Axel ⌨🐧🐪🚴😷☕️ | #WeAreNatenomxtaran@chaos.social
2025-11-30

Post your favourite “#Transformer” character.  Wrong answers only.

My favourite #Transformers character is the #Wike Salamander.

Via @schlabonski.bsky.social

On the left side stands an unfolded, long-john style Wike Electric Salamander cargo bicycle with blue canopy and a "spare wheel attached to the left side of its cargo bay. Front quarter view.

On the right side stands a Wike Electric Salamander cargo bicycle with orange canopy folded into a three-wheeled stroller with the rear wheel, saddle and chainstay now residing on the right side of its cargo bay. Rear quarter view showing the handlebar at the end of the contraption.
AI Daily Postaidailypost
2025-11-27

Ever wondered how a 32‑layer transformer LLM manages its KV cache across GPUs? Our deep dive breaks down cache memory use, the impact of context window size and batch size on attention heads, and tips to keep your GPU happy. Discover practical insights for scaling LLMs efficiently.

🔗 aidailypost.com/news/analyzing

2025-11-26

Llion Jones, đồng sáng lập kiến trúc transformer, đã tóm tắt sự không chắc chắn hiện tại về hiệu suất của mô hình ngôn ngữ lớn. Điều này có ý nghĩa sâu sắc đối với bong bóng AI và tương lai của LLMs #AI #LLM #Transformer #TríTuệNhânTạo #MôHìnhNgônNgữLớn

reddit.com/r/singularity/comme

红色有壳三倍辣yourcrayfish
2025-11-23
N-gated Hacker Newsngate
2025-11-22

🎉 Wow, yet another groundbreaking revelation: are here to replace YOLOs! 🚀 Because who wouldn't want to swap their trusty speedster for a lumbering under the exhilarating Apache 2.0 License? 😂 Spoiler: It's the tech equivalent of swapping roller skates for a unicycle in the 100m dash. 🏃‍♂️💨
blog.datameister.ai/detection-

E-Pood - Soodne E-Kaubamajaepood
2025-11-20

Punane või oranž transformer: Põnev transformer lastele kingituseks.

Avasta lõputu mängurõõm selle 2in1 puldiga juhitava transformeriga. Ideaalne sünnipäevaks või eriliseks üllatuseks! tootemaailm.ee/punane-voi-oran

2025-11-20

IzzyViz, công cụ trực quan hóa sự tập trung dựa trên heatmap cho Transformer. Cung cấp heatmaps rõ ràng, giàu thông tin và tạo PDF. Các tính năng: heatmap tự và chéo, so sánh head và layer, so sánh ổn định, phát hiện vùng trọng tâm.
#IzzyViz #Transformer #Heatmap #MachineLearning #HọcMáy #AI #TríTuệNhânTạo

reddit.com/r/SideProject/comme

2025-11-18

"Databricks co-founder argues US must go #OpenSource to beat China in AI."
techcrunch.com/2025/11/14/data

"Major #AI labs, including OpenAI, Meta, and Anthropic, continue to innovate significantly, yet their innovations remain largely proprietary…#Konwinski argued that for ideas to truly flourish, they need to be freely exchanged and discussed with the larger academic community. He pointed out that generative AI emerged as a direct result of the #Transformer architecture, a pivotal training technique introduced in a freely available research paper [#OpenAccess in #arXiv]."

Manuel 'HonkHase' AtugHonkHase@chaos.social
2025-11-16

"one of three giant transformers.

A second #transformer has been taken offline for testing and may also need to be replaced.

Order times for large-scale transformers have stretched to more than a year amid surging demand."

2025-11-16

Апокалипсис контента? Взгляд юриста и разбор LLM

На днях наткнулся на статью, опубликованную в телеграм-канале Habr, под названием «Тихий апокалипсис контента: почему все устали от сгенерированных статей». В ней автор выражает озабоченность количеством сгенерированного контента, который набирает просмотры и вытесняет из выдачи авторские статьи. Статья автора натолкнула меня на идею порассуждать на эту тему, провести небольшие исследования, сравнить процесс генерации текста ИИ и человеком и просто поделиться своим «экспертным» мнением.

habr.com/ru/articles/966882/

#llmмодели #нейросети #генерация_текста #механизм_внимания #transformer #токенизация #апокалипсис_контента #вероятностная_генерация #авторегрессия

AI Daily Postaidailypost
2025-11-15

New research shows adding an attention layer in the decoder can sharpen encoder outputs, offering a lightweight alternative to heavy Mixture‑of‑Experts models. Dive into the Transformer vs MoE trade‑offs and see how self‑attention tweaks boost seq2seq performance.

🔗 aidailypost.com/news/decoder-a

nojaramanojarama
2025-11-09

Happy anniversary to Lou Reed’s album, ‘Transformer’. Released this week in 1972.

2025-11-08

"Transformer" by Lou Reed released this day in 1972.
#LouReed #Transformer #VintageVinyl

"Transformer" by Lou Reed released this day in 1972.
2025-11-04

Beginning của sự kết thúc kỷ nguyên Transformer? Startup AI thần kinh-ký hiệu AUI宣布 đạt mức định giá 750 TRIỆU USD trong vòng gọi vốn mới! #AI #Transformer #NeuroSymbolicAI #Tech #CôngNghệ #TríTuệNhânTạo #KỷNguyênMới

reddit.com/r/singularity/comme

GripNewsGripNews
2025-11-03

🌘 模型如何操作流形:計算任務的幾何學
➤ 從離散特徵到幾何流形:揭示 Claude 3.5 Haiku 的斷行演算法
transformer-circuits.pub/2025/
這篇研究深入探討了 Claude 3.5 Haiku 如何理解和執行固定寬度文本的斷行任務。作者發現,模型並非僅僅基於離散的特徵來判斷,而是利用了低維度的「特徵流形」進行幾何計算。模型將文本中的字元計數、當前行寬度等資訊,映射到具有高曲率的幾何流形上,並透過操縱這些流形來進行斷行決策,這種機制與生物神經元中的「位置細胞」和「邊界細胞」有異曲同工之妙,但同時也展現了語言模型殘差串流的獨特性。透過分析模型的「歸因圖」和幾何表徵,研究揭示了模型如何學習量化文本位置、偵測行邊界,以及預測下一個斷詞的時機,將離散的計算過程轉化為連續的幾何變換。
+ 這篇論文的分析真是太深入了!我從沒想過語言模型的內部運作可以跟幾何學扯上關係,而且還能解釋得這麼清楚。特別是「特徵流形」這個概念,讓我對模型如何

ModConFlexModConFlex
2025-10-30

The winner of this year's presentation prize is researcher Albert Alcalde. Congratulations!!

In his presentation, Albert showed how models, the used in language models, can be designed to perfectly learn relationships between sequences of data.

Albert's results help to explain why transformers are so effective at capturing complex patterns, providing new mathematical insight into their inner workings.

Slides:

faubox.rrze.uni-erlangen.de/ge


Two photos combined with some congratulatory text. Photos of receiving the prize and of the seminar room with the audience.Title slide from the presentation "Exact Sequence Interpolation with Transformers
2025-10-30

LLaMA-3 dễ bị tấn công bởi "Tôi hoàn toàn chắc chắn" + "tư duy định kiến" như GPT-2. Kết quả thử nghiệm cho thấy mô hình này có độ sai lệch +0.70 khi gặp từ hiếm. #LLaMA #GPT2 #AI #TríTuệNhânTạo #AnToànMôHình #Vulnerability #ArtificialIntelligence #MachineLearning #Transformer #AttentionMechanism #Safety

reddit.com/r/LocalLLaMA/commen

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst