#diffusionmodels

TechRadar (@techradar)

Nano Banana 2의 이미지 프롬프트 예시들이 공개되어 해당 버전의 성능 향상과 개선된 이미지 생성 능력을 보여줍니다. (이미지 프롬프트를 통해 새 모델의 강화된 세부 표현력 및 출력 품질을 시연하는 내용)

x.com/techradar/status/2027557

#nanobanana2 #imagegeneration #generativeai #diffusionmodels

2026-02-24

Steerling-8B, the first interpretable model that can trace any token it generates to its input context, concepts a human can understand, and its training data.

guidelabs.ai/post/steerling-8b

#AI #InterpretableAI #DiffusionModel #DiffusionModels

2026-02-17
🎉 Now videos are alive! After finally getting WAN 2.1 running on my RX 6700 XT via ROCm and ComfyUI, even complex prompts can be turned into animated WebPs locally!

These animated WebP were generated locally using ComfyUI and the WAN 2.1 T2V 1.3B (fp16) model.

Model Stack:
- wan2.1_t2v_1.3B_fp16
- umt5_xxl_fp8_e4m3fn_scaled (Text Encoder)
- wan_2.1_vae
- clip_vision_h

The prompt is first converted into embeddings by the UMT5 encoder.
The WAN video model then generates multiple frames using latent diffusion (noise → iterative refinement), ensuring temporal coherence between frames.
The VAE decodes the latent frames into images, exported as an animated WebP.

Prompt execution time: depends on scene complexity, from 521.62 seconds (~8.7 minutes) up to 17 minutes 26 seconds for more complex prompts.

Rendered locally via ROCm on my AMD RX 6700 XT (12GB VRAM).
No cloud. Pure local inference.

#ComfyUI #WAN21 #ROCm #AMD #LocalAI #FOSS #VideoAI #AIvideo #AIGenerated #MachineLearning #DeepLearning #DiffusionModels #TextToVideo #AIArt #CreativeAI #LocalInference #VideoGeneration

fly51fly (@fly51fly)

논문 'Blind denoising diffusion models and the blessings of dimensionality' (Z. Kadkhodaie, A. Pooladian, S. Chewi, E. Simoncelli, Simons Foundation & Yale University, 2026)이 arXiv에 게시되었음을 알리는 트윗입니다. 블라인드 디노이징 확산모델과 고차원성의 이점에 관한 연구 결과를 소개합니다.

x.com/fly51fly/status/20217044

#diffusionmodels #denoising #arxiv #research

Wayne Radinskywaynerad
2026-01-30

AI Motion Control takes a video and a photo and it transfers the motion of the person in the video to the person in the photo.

The demos where they transfer video to a cartoon character are cute. When they do it to a real person, it feels a bit creepy because it looks real. At least that's my take.

aimotioncontrol.net/

2026-01-26
The first upload went a bit sideways -.-

Some data got lost causing black stripes in the original masterpiece image so I moved it back in the carousel

Here is the real clean version finally posting the image as it was meant to be seen

#sdxl #stablediffusion #comfyui #aiart #animeart #generativeart #diffusionmodels #foss #opensource #digitalart #animegirl
2026-01-26
Generated with SDXL in ComfyUI

First three images are the final results,
image #4 shows the full ComfyUI workflow used to create them.

#sdxl #stablediffusion #comfyui #aiart #animeart #generativeart #diffusionmodels #foss #opensource #digitalart #animegirl

fly51fly (@fly51fly)

새 논문 'Stable-DiffCoder: Pushing the Frontier of Code Diffusion Large Language Model'가 공개되었습니다. 저자 C. Fan, W. Heng, B. Li, S. Liu 등은 코드 확산(code diffusion) 기반의 대형 언어 모델을 제안하며, Huazhong University of Science and Technology와 ByteDance Seed 소속 연구진으로 2026년 arXiv에 게재되었습니다(링크 포함).

x.com/fly51fly/status/20155404

#codegeneration #diffusionmodels #llm #research

Torsten Brezesinski (@TBrezesinski)

Materials Futures에 'DiffNMR: diffusion models for nuclear magnetic resonance spectra elucidation'라는 논문이 게재되었다고 알리는 공지입니다. Gen Long, Xin Chen, Mingjun Xiao 등(USTC, Baidu 참여)이 diffusion models를 핵자기공명(NMR) 스펙트럼 해석에 적용한 연구로, IOP 출판을 통해 오픈액세스로 공개되었습니다.

x.com/TBrezesinski/status/2014

#diffusionmodels #nmr #materialsfutures #research

2026-01-21
If you can run FLUX and SDXL locally, you really don’t need cloud AI services anymore.
No subscriptions, no prompt logging, no rate limits, just full control.

And no, you don’t need an RTX 5090 or a Radeon AI PRO R9700 (I’d love to have one though xD).
Every image here was rendered on an AMD RX 6700 XT, a mid-range GPU from 2021, not even officially supported by ROCm.

Local models. Local compute. Local freedom.

#SDXL #StableDiffusion #LocalAI #GenerativeArt #OpenSource #FOSS #PrivacyFirst #NoCloud #AMD #RX6700XT #ROCm #Linux #AIArt #DiffusionModels #CreativeCoding #MidrangeGPU #SelfHosted #Decentralized
2026-01-19

STARFlow-V normalizing flows bring faster, stable long clips across text, image, and video, using sliding windows and 640x480/16fps limits.

aistory.news/ai-tools-and-plat

#DiffusionModels #OpenAI #VideoGeneration

STARFlow-V normalizing flows power Apple video AI launch
2026-01-19

Apple STARFlow-V video model uses normalizing flows for stable, fast 640x480/16fps clips and 30s demos via sliding window, with three modes.

aistory.news/ai-tools-and-plat

#DiffusionModels #OpenAI #VideoGeneration

Apple STARFlow-V video model debuts flow-based clips
Wayne Radinskywaynerad
2026-01-04

Charlie Kirk's legacy is AI slop, and he's just the first of a new trend that will affect the whole world, says YouTuber Moon.

This got me thinking, people have worried so much about deepfakes that fool people into thinking someone said things they never said or did things they never did, but here it's just sheer quantity of "AI slop" and it none of it seems to be deceptive.

youtube.com/watch?v=WNIiSx0x6EI

Wayne Radinskywaynerad
2025-12-28

When AI plays the game of visual telephone, "the AIs repeatedly converged on the same 12 generic, often Eurocentric motifs -- what the researchers call 'visual elevator music.'"

science.org/content/article/wh

N-gated Hacker Newsngate
2025-12-19

🚀✨ Behold, the groundbreaking Qwen-Image-Layered: Now you can finally edit images like an onion, one tear-inducing layer at a time! 😂 Who knew in diffusion models would be such a *layered* joke? 🤡
huggingface.co/papers/2512.156

2025-12-10
The other day I had the intrusive thought
AI is intellectual Viagra
and it hasn't left me so I am exorcising it here. I'm sorry in advance for any pain this might cause.

#AI #GenAI #GenerativeAI #LLMs #DiffusionModels #tech #dev #coding #software #SoftwareDevelopment #writing #art #VisualArt
2025-12-07

Meta V-JEPA model learns physical intuition from everyday videos, signaling new progress for self-supervised video AI and safer robotics.

aistory.news/generative-ai/met

#AIUpdate #Automation #DiffusionModels

Meta V-JEPA model shows intuitive physics from video
2025-11-30
The rhetoric that limiting or banning AI/generative AI/LLM/diffusion model use is "ableist" or "gatekeeping" is the latest desperate attempt to find an angle through which to force this technology into our lives against our collective will. We need to reject this narrative. Common as it is it simply doesn't scan. It reads to me as an attempt to co-opt the language of social justice to shame people into accepting an unjust and largely failing technology that they are rightfully rejecting.

Think it through. If you don't accept the use of climate-destroying, electricity-and-fresh-water-sapping, job-destroying, economy-thrashing--and yet mediocre or poorly performing!--technology created by multi-trillion-dollar sociopathic entities, then you are preventing people with less privilege than you have from living their best lives. You are preventing them from learning how to code. You are preventing them from obtaining coveted jobs in the tech sector. You are preventing them from having access to information. You, personally, are responsible for all this. Not the multi-trillion-dollar sociopathic entities who've not only created this technology and forced it on us but contributed to creating the less-privileged conditions of the people you are supposedly responsible for with your individual choices. Not the governments that neglected to enforce existing laws that would have prevented such multi-trillion-dollar sociopathic entities from forming in the first place, let alone creating such a technology--while also creating the conditions that led to people being less privileged. No, they are not responsible. You are. I am.

That doesn't make any sense.

Neoliberalism's greatest trick has been to shift responsibility for any problems away from the powerful and onto individuals who are not empowered to fix anything, all while convincing everyone that this is right and proper. Large corporations do not cause a plastic pollution problem; you and I do, by not separating our recycling. Large corporations, governments and militaries do not cause CO2 pollution and climate damage; you and I do, by using incandescent lightbulbs and non-electric/non-hybrid cars or eating meat. Lack of regulation and large agribusiness practices are not to blame for poor food quality; you and I are, for buying what they sell instead of going organic and joining a CSA. Etc. ad infinitum. Large, powerful entities routinely generate a problem, then tell you and me that we are responsible for the problem as well as for fixing it. Never mind that these entities could nudge their own behavior a bit and move the needle on the problem far more than masses of people could no matter how organized they were. Never mind that these entities could be constrained from causing such problems in the first place.

We are watching a new variation of this pattern come into being right in front of our eyes with AI. We should stop accepting these fictions. You are neither ableist nor a gatekeeper for resisting AI. You are, instead, attempting to forestall the further degradation of conditions for everyone, which starts this same cycle anew.

#AI #GenAI #GenerativeAI #LLM #DiffusionModels #neoliberalism #depoliticization
AI Daily Postaidailypost
2025-11-29

NeurIPS 2025 revealed a turning point: the top-four papers argue that bigger isn't always better. From revamped attention mechanisms to leaner diffusion models and smarter RL benchmarks, researchers are redefining performance metrics. Curious how the AI community is shifting focus? Dive into the highlights and see what's next for open-source innovation.

🔗 aidailypost.com/news/neurips-2

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst