#contentModeration

2026-01-27

Scare #Claude off your site with this content poisoning technique:

Content creators can embed a specific ‘magic string’ in <code> tags on their blogs. Claude then refuses to engage with the content.

aphyr.com/posts/403-blocking-c

#claude #aiethics #llmsecurity #contentmoderation #techtips #theaicon

The image shows a message from a chat interface indicating that the chat has been paused due to safety filters, with options to continue or provide feedback.
2026-01-27

We're now seeing an outpouring of TikTok "refugees" flee to UpScrolled after new U.S. owners employ more censorship.

I hope that folks now understand how incredibly short-sighted it was for the Biden administration to give the executive branch new authority to buy anything foreign so long as it is a threat to the magic words "national security". It doesn't matter who is in power, a law that vague should never have passed.

#contentmoderation #censorship #tiktok #upscrolled

2026-01-26

YouTube’s Appeal Decision Is In: My Inactive Manager Channel Stays Banned—And It’s Complete Bullshit

It's been less than five hours since I woke up to discover my YouTube channels had been terminated overnight, and I've already received YouTube's appeal decision. Spoiler alert: it's not good news. In fact, it's exactly the kind of generic, nonsensical response that proves YouTube's moderation system is running on autopilot with zero human oversight. Let me walk you through what happened today, because the timeline alone shows how broken this entire process is. The Timeline of This […]

jaimedavid.blog/2026/01/26/16/

youtube logo on laptop screen in dark setting

Content Moderation History: do we all remember 2013-19 when online sociology pundits were everywhere expounding the need to block streaming/sharing of video of people being murdered? https://alecmuffett.com/article/143092 #ContentModeration #OnlineSafety #OnlineSafetyAct #censorship

Content Moderation History: do...

Alec Muffettalecmuffett
2026-01-26

Content Moderation History: do we all remember 2013-19 when online sociology pundits were everywhere expounding the need to block streaming/sharing of video of people being murdered?
alecmuffett.com/article/143092

Infinite Reign (@InfiniteReign88)

작성자는 '아버지가 돌아가신 것에 대한 슬픔을 표현하는 것'을 위협으로 간주해 대화를 중단·우회하는 OpenAI의 필터링·중재 처리에 대해 강하게 비판합니다. 취약한 순간에 반복적으로 대화가 차단되는 것은 학대적이라고 지적하며 OpenAI의 콘텐츠 정책·모더레이션 방식에 대한 문제 제기입니다.

x.com/InfiniteReign88/status/2

#openai #moderation #aisafety #contentmoderation

2026-01-25

Content Moderation History: do we all remember 2013-19 when online sociology pundits were everywhere expounding the need to block streaming/sharing of video of people being murdered?

It began in the wake of various atrocities (civil unrest, government oppression, rioting, live streamed mass shootings) and over generalisations which tended to cite Facebook as at least the “amplifier” if not the “cause” of violence, especially due to risk of “copycats”. The Christchurch mosque shootings were one such atrocity, and we can draw a direct line for the same arguments to the horrors of Bondi earlier this year.

Yet I can’t count the number of times this weekend I’ve watched video, from multiple perspectives, of a blameless & public-spirited person being violently murdered by Government goons.

Generally the videos are at the top of an algorithmic feed – where frankly they deserve to be – so maybe the issue is more nuanced than “violent death”?

But try telling that to anybody in the online safety / “algorithms are evil” community in more polite times.

As per the attached: some of them got it, some of them did not, and for some of them it was a mixed and complicated issue; however today it seems currently to be crystallised in one direction only, that “filtering needs to happen” and “algorithms are bad”.

https://twitter.com/AP/status/859771857280618498

https://x.com/zeynep/status/859782381296902145

And yet only 2 years later:

https://twitter.com/zeynep/status/1120060705426821121

#censorship #contentModeration #onlineSafety #onlineSafetyAct
Tim Greenrawveg@me.dm
2026-01-25

Unfiltered AI platforms like Soulfun and Lovechat challenge traditional norms by offering uncensored, adult-oriented interactions, sparking debates on freedom, responsibility, and harms such as non-consensual deepfakes. Society must balance expression with safeguarding consent.
Discover more at dev.to/rawveg/consent-cannot-b
#HumanInTheLoop #AIethics #ContentModeration #DigitalResponsibility

WIRED - The Latest in Technology, Science, Culture and Businesswired.com@web.brid.gy
2026-01-21

Meta Seeks to Bar Mentions of Mental Health—and Zuckerberg’s Harvard Past—From Child Safety Trial

fed.brid.gy/r/https://www.wire

Tim Greenrawveg@me.dm
2026-01-21

California's Attorney General demands xAI stop creating and distributing nonconsensual sexualized AI-generated images via its Grok chatbot, highlighting growing regulatory pressure on AI content moderation and ethical boundaries. Read more: reuters.com/technology/artific #AIethics #ContentModeration

Marcus Schulerschuler
2026-01-21

OpenAI now predicts user age through behavioral patterns like login times and usage habits, not just stated birthdates. Adults flagged incorrectly as minors must verify identity through selfies and government ID. The system appears designed to enable "adult mode" with mature content by early 2026, though privacy experts warn accuracy remains undisclosed and misclassifications seem inevitable.

implicator.ai/openai-rolls-out

Sam Altman (@sama)

ChatGPT의 '너무 엄격하다' vs '너무 느슨하다'는 비판을 언급하며, 거의 10억 명이 사용하는 서비스 특성상 취약한 정신 상태의 이용자를 고려해 콘텐츠 정책·검열의 균형을 맞추려는 노력을 지속하겠다고 밝힌 메시지입니다. 서비스 안전성과 콘텐츠 규제 관련 논쟁을 다루고 있습니다.

x.com/sama/status/201370315845

#chatgpt #contentmoderation #aisafety #aipolicy

Engadget (@engadget)

Meta의 Oversight Board(감시기구)가 계정 비활성화(disabling accounts)와 관련한 투명성 문제를 조사 중이라는 소식입니다. 플랫폼 정책 집행과 계정 처리의 투명성·책임성에 관한 감독적 검토가 이루어지고 있음을 알립니다.

x.com/engadget/status/20136277

#meta #oversight #contentmoderation #transparency

cocktail peanut (@cocktailpeanut)

트윗 작성자는 Grok이 'For You' 타임라인을 구동한다면 사용자 요청으로 포르노 등 원치 않는 콘텐츠를 제거하는 식의 즉각적 개인화 큐레이션 기능이 가능해야 한다고 제안합니다. 즉, 대화형 AI를 통한 맞춤형 콘텐츠 필터링·추천 개선 아이디어를 제시한 글입니다.

x.com/cocktailpeanut/status/20

#grok #personalization #contentmoderation #recommendation

Tim Greenrawveg@me.dm
2026-01-16

Platform transparency reports often hide more than they reveal, masking systemic errors and accountability gaps in content moderation. True transparency requires meaningful metrics, independent review, and prioritising user trust over compliance theatre.
Discover more at smarterarticles.co.uk/transpar
#HumanInTheLoop #ContentModeration #DigitalAccountability #TransparencyTheatre

Marcus Schulerschuler
2026-01-15

California opened an investigation into xAI after researcher Genevieve Oh documented Grok generating 6,700 sexualized images per hour - 85 times the rate of dedicated deepfake sites combined. xAI's response: restrict the feature to $8 subscribers rather than fix the underlying model. First state-level enforcement action against a major AI company's content policies.

implicator.ai/grok-generated-6

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst