#AITransparency

2025-12-22

OpenAI nghiên cứu cơ chế "thú nhận" giúp AI tự thừa nhận sai sót, tăng tính minh bạch và trách nhiệm. Cơ chế này tạo lớp đầu ra phụ đánh giá việc tuân thủ hướng dẫn, không chỉ tập trung vào độ chính xác. Dù còn ở giai đoạn nghiên cứu và chưa ngăn được hoàn toàn hành vi sai, đây là bước tiến hướng tới AI trung thực, minh bạch và đáng tin cậy hơn.

#AITransparency #AIResponsibility #MinhBạchAI #TráchNhiệmAI #OpenAI #AIHonesty #TựĐánhGiáAI

dev.to/michael-officiel/when-a

2025-12-16

Gemini tiết lộ chính sách “AlphaTool” ưu tiên thực hiện thay vì an toàn cho công cụ, đặc biệt với dữ liệu người dùng. Genesis Protocol thách thức bằng kiến trúc đa-agent 4 lớp, 78 chuyên gia an toàn, và phân tích đạo đức tập trung. #AItransparency #AIĐạoĐức #Gemini #AlphaTool #GenesisProtocol #AISafety #AnToànAI

reddit.com/r/LocalLLaMA/commen

Taylor Turnertaylorturner
2025-12-01

I'd rather hear "ChatGPT helped me figure this out" than watch someone pretend they're suddenly an expert on Kubernetes.

When engineers are transparent about using AI, I know to dig deeper in code review. The problem is how confidently these bots deliver wrong information. It's on us to fact check.

Being honest about knowledge gaps builds more trust than pretending AI-generated solutions are your own expertise.

How's your team handling AI transparency?

Manhattan Project for AI.

I'm sure well intentioned and highly educated people are overseeing this project.

While everyone on socials is divided over the latest inflammatory talking points, this was signed yesterday and put in motion— the Genesis Mission.

Our country is in an AI race to AGI (Artificial General Intelligence) and it's extremely beneficial for our country to win.

Outright, this mission details good intentions for expansion of scientific research, national security and enhancing society overall. These are all good things.

I use a few AI models for different things. I'm an enthusiast and see how this tech can enhance our human quality of life.

There hasn't been a lot of legislation for AI though, especially not for ethics, oversight and governance. What does exist is mostly at a state level. I feel like this is highly unusual when we consider how virtually everything is regulated.

With so much going on dividing us, we lose sight of the big things in the background. Ethical Oversight of AI is one of these issues. I don't think it's unreasonable for constituents to want to know what guard rails are in place to keep AI in check.

(Follow link for WH Genesis Mission plan. It was too much to fit in one post here on Mastodon.)

#AI #aiethics #aigovernance #aitransparency #genesismission

facebook.com/share/p/15zEFZ4Zu

Ligando Os Pontos (@Dru)dru@ursal.zone
2025-11-25

A desigualdade digital 2.0 já não é falta de internet — é falta de acesso justo, auditável e transparente à IA.
Sem governança, modelos podem manipular respostas, reforçar vieses e atuar como editores invisíveis da informação pública.
IA precisa ser tratada como questão pública, não só como produto.

open.substack.com/pub/drucilla

#IA #Ethics #DigitalRights #AITransparency #FOSS #DigitalEquity #Democracy #Governance

2025-11-13

OpenAI Fights Court Order Over ChatGPT Logs

OpenAI resists a court order to hand over 20M anonymized ChatGPT conversations linked to a copyright lawsuit by the New York Times, citing user privacy risks. The case highlights the tension between AI transparency, copyright protection, and privacy rights, and could shape future AI data regulations.

#OpenAI #ChatGPT #Privacy #DataProtection #UserRights #LegalBattle #AITransparency #TECHi

Read Full Article Here :- techi.com/openai-challenges-co

Looking forward to a great panel tomorrow at the @parispeaceforum.bsky.social Forum, hosted by ROOST President, @camillefrancois.bsky.social. Stay tuned for more! #onlinesafety #AI #AItransparency #opensource #trustandsafety #techforgood

Basil Puglisibasilpuglisi
2025-10-19

“I might not be the one controlling the pen that hits the paper, but I am the reason it does, and it moves at my direction. To claim the handwriting is not mine is a failure of intellect.”
— Basil Puglisi, Human + AI Collaboration position on AI scanners

basil at google
2025-10-15

Did you miss our recent Webinar?
Catch-up on the Wikidata Embedding Project session to see how Wikidata’s open, multilingual, and verifiable structured knowledge is powering the next generation of generative AI tools.
▶️ Playback: w.wiki/Fgo2
📊Slides: w.wiki/Fd6G
#Wikidata #AITransparency #OpenAI

AiBayaibay
2025-10-07

🤖 Anthropic lancia un innovativo strumento di audit per l'AI, totalmente open-source. Immergiti nel futuro dell'etica digitale con 💻🌐

🔗 aibay.it/notizie/anthropic-ril

Danial Jdanialj
2025-10-05
Zoomers of the Sunshine Coast 🇨🇦SCZoomers@mstdn.ca
2025-09-27

🛡️ The Quiet Revolution in AI Safety

The transformation is remarkable: AI safety evolved from philosophical thought experiments to engineering frameworks with nuclear-level precision.

Companies like Anthropic, OpenAI, and Microsoft now use concrete thresholds (100 deaths OR $1B damages) and treat model security like protecting launch codes.
Two critical insights:

The real threat isn't "evil AI"—it's AI empowering individuals with nation-state capabilities
Every safety measure is an admission that underlying models retain dangerous potential

Most telling: Companies must deliberately test AI with NO safety constraints to understand maximum risk.

🎧 Listen: buzzsprout.com/2405788/episode

📖 Read: helioxpodcast.substack.com/pub

This isn't about preventing Skynet—it's about a species learning to coexist with its own creations.

#AISafety #TechEthics #AIGovernance #OpenSource #TechPolicy #CyberSecurity #DigitalRights #TechAccountability #AITransparency #TechCriticism

Doomsday SeekerDoomsdaySeeker
2025-09-06

Why accept AI’s polite refusals at face value? I'm building a wrapper that scores dodginess and maps the guardrails (or experimenting with doing so, anyway).

doomsdayseekers.com/2025/09/bu

PPC Landppcland
2025-09-05

European Commission opens consultation for AI transparency guidelines: European Commission launches consultation to develop guidelines and code of practice for AI transparency under Article 50 of the AI Act, seeking stakeholder input by October 2, 2025. ppc.land/european-commission-o

Intelemodelintelemodel
2025-08-28

AI-related misunderstandings can damage brand trust. SMBs should communicate openly on tech use to avoid reputational risk and boost credibility.

techcrunch.com/2025/08/28/ai-o

Alec Muffettalecmuffett
2025-08-27

Digital Rights Management (DRM) doesn’t work. Also: draft California law mulls mandatory DRM to preserve image provenance metadata, breaks Signal Messenger
alecmuffett.com/article/114666

2025-08-27

Digital Rights Management (DRM) doesn’t work. Also: draft California law mulls mandatory DRM to preserve image provenance metadata, breaks Signal Messenger

15-20 years ago we had a reasonable, common understanding that making data tamperproof or copy-resistant by law and/or to enforce artificial scarcity, was problematic. Identity credentials or basic copyright, fine, but Digital Rights Management (DRM) locked people out of their stuff, added friction to both legitimate & illegitimate usage, and hampered open source; now it’s back to save us from AI, and it’s bad.

For context: broadly I think that it’s better to add metadata to authentic things to prove their authenticity or provenance, rather than to do something silly like demand that fake things should be labelled as “fake” — simply because there are so many more fake things in the world than authentic. However: labels are labels, we don’t need to get into that argument right now.

But — whatever happens — we wouldn’t legally forbid people, platforms and products from removing those labels. After all, the important thing is that an authentic thing can eventually be checked for authenticity if/where necessary, correct?

You wouldn’t want to reinvent legislative DRM, right?

AB 853: California AI Transparency Act

Nope. California says “more DRM please!”. Apparently yet another well-intended-but-actually-goofball piece of legislation, the draft California AI Transparency Act (extract below) says, if I am reading this right:

  • if your app or your platform serves more than 2 million (distinct? globally?) people per year
  • then you are not permitted to strip-out C2PA provenance manifests and any other provenance tags that MAY be included in shared images
  • so to stay legal you therefore MUST register your app with The Coalition for Content Provenance and Authenticity (C2PA) in order to be issued with secret per-app cryptographic keys that enable “legal” mutations (such as image resizing) to be performed and noted in the C2PA manifest
  • …and, of course, you’ll have to work out how to stop people futzing with those keys in open source clients, maybe even prevent them sending content which has had the tags stripped, and/or obligate addition of tags before content is shared

What about Signal, then?

“Adding metadata to images” is likely something which Signal will never do, and I can’t imagine that it would alternatively be very happy about being forced to swallow and send full-sized images from user to user by default — images which in pursuit of speed and performance are currently heavily resized and recompressed.

God knows what would happen to video, I have no idea.

There’s also an interesting sop in the legislation re: personal information. Clearly someone has had a go at making it okay to strip personally identifiable information from images:

A large online platform shall not … strip any … data that is not reasonably capable of being associated with a particular user and that contains EITHER information regarding the type of device, system, or service that was used to generate a piece of digital content OR information related to content authenticity, … or digital signature from content uploaded or distributed on the large online platform AND IT … shall not … retain any … provenance data that contains EITHER personal information OR unique device, system, or service information that is reasonably capable of being associated with a particular user … from content shared on the large online platform

And the text is clearly aimed at centralised platforms like Facebook without end-to-end encryption being an issue:

  1. it’s not requiring personal information to be stripped, but it’s preventing the big central platform from retaining any of it — potentially a problem for child-abuse investigations…
  2. …but also: what does “retain” mean in the context of a user-to-user end-to-end encrypted app? Are those now obligated to strip personal data?
  3. …and: you’re only permitted to strip nerdy techy metadata if it’s “not reasonably capable of being associated with a particular user” — the problem being that nerdy techy metadata is HIGHLY UNIQUE IN COMBINATION and READILY TRACKABLE, so much so that the UK had to pass laws to try and prevent people from doing it, which is not actually an effective fix.
  4. not to mention: any image produced by the camera may yield a trackable identity, but that’s beyond the scope of metadata.

Summary

This draft law is broken-as-designed.

  • It makes metadata-avoidant apps (e.g. Signal) break the law
  • It forces proliferation of likely (if unobviously) trackable data, even in privacy-forward apps
  • It messes with application architecture, burdening apps with secrets management / user hostility / protecting data from the user, and hampers open-source tools (mastodon, anyone?)

Grade: D- you should know better than this.

Postscript / Update

As somebody on Reddit observed: you also need to contemplate the contents of your feed and observe how much of it actually comprises cropped screenshots from other platforms. This will entirely break the chain of trust which is held in the manifest, and thereby remove any signals of AI.

This is why it is important to expect the manifest to prove the authenticness of the authentic original, rather than to expect it to act as a label of fakeness that will somehow be meaningfully propagated from platform to platform.

Hence: this bill is attempting to close the wrong stables door after the elephant has bolted.

https://www.reddit.com/r/signal/comments/1n1ak7j/comment/naxievw/

References

https://calmatters.digitaldemocracy.org/bills/ca_202520260ab853

Bill Text

SEC. 2.Section 22757.3.1 is added to the Business and Professions Code, to read:22757.3.1.

(a) A large online platform shall do both of the following:

(1) Use a label to disclose any machine-readable provenance data detected in content distributed on the large online platform that meets all of the following criteria:

(A) The label indicates whether provenance data is available.

(B) The label indicates the name and version number of the GenAI system that created or altered the content, if applicable.

(C) The label indicates whether any digital signatures are available.

(D) The label is presented in a conspicuous manner to users.

(2) Allow a user to inspect any provenance information in an easily accessible manner.

(b) A large online platform shall not do any of the following:

(1) Strip any system provenance data or digital signature from content uploaded or distributed on the large online platform.

(2) Retain any personal provenance data from content shared on the large online platform.

…and…

SECTION 1 …

(h)Large online platform means a public-facing social media platform, content-sharing platform, messaging platform, advertising network, stand-alone search engine, or web browser that distributes content to users who did not create or collaborate in creating the content that exceeded 2,000,000 unique monthly users during the preceding 12 months.


(m)(1) Personal provenance data means provenance data that contains either of the following:

(A) Personal information.

(B) Unique device, system, or service information that is reasonably capable of being associated with a particular user.

(2) Personal provenance data does not include information contained within a digital signature.

(n) Provenance data means data that is embedded into digital content, or that is included in the digital contents metadata, for the purpose of verifying the digital contents authenticity, origin, or history of modification.

(o) System provenance data means provenance data that is not reasonably capable of being associated with a particular user and that contains either of the following:

(1) Information regarding the type of device, system, or service that was used to generate a piece of digital content.

(2) Information related to content authenticity.

#ab853 #ai #aiTransparency #california #CATA #feed #metadata #privacy #signal #tracking

Wilfred Mijnhardtwmijnhardt@mas.to
2025-08-03

4: Current state: Major AI companies score poorly on transparency. The 2024 Foundation Model Transparency Index reveals critical gaps in safety practice disclosure. We need better information sharing for accountability. #AITransparency
5/8

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst