It's full.
Learn more: https://lydie.cc/data.html
#resist #fascism #datahoarder #digitalpreservation #archive #sciop #torrent #publicdata
It's full.
Learn more: https://lydie.cc/data.html
#resist #fascism #datahoarder #digitalpreservation #archive #sciop #torrent #publicdata
Here's another #FEP for representing torrents on activitypub :)
short, sweet, and with a reference implementation and tests!
towards a federated bittorrent tracker with #sciop !
PR: https://codeberg.org/fediverse/fep/pulls/714
Discussion: https://socialhub.activitypub.rocks/t/fep-d8c8-bittorrent-torrent-objects/8309 (or this thread)
#FEP_d8c8 #BitTorrentOverActivityPub #FederatedP2P #BitTorrent
This is extremely sad y'all. ZERO seeding going on from the SciOp / public data / antifa torrent server. Have a few spare gigs? Even that's something, please help preserve data the fascists are destroying!
#resist #fascism #datahoarder #digitalpreservation #archive #sciop #torrent #publicdata
Finally had some spare time to learn more about the #sciop project recently. I created a #DIY seed box to preserve several #smithsonian datasets at risk of censorship, including over 1.2TB and 94,000 images. I then configured #plex to serve the photos and podcast for my own enjoyment. I even made a little preroll video for my server just for fun. :catjam:
It would be nice if #sciop linked to the actual takedown request for datasets tagged with "takedown_issued"
@nyxmir If the CDC used to have it, it might still be on SciOp.
If you don't know SciOp, it's a community where people seed (through Torrent) data threatened or deleted by Trump and other fascists. The datasets are hefty, but the point is to share and keep them available for everyone:
https://sciop.net/datasets/?query=cdc&sort=-threat
Good luck!
the haters all scoffed when i "embedded a whole set of bitmap fonts for a single use" but i knew i was following the light and the way. #sciop is in its "intelligibility" era, where we will do things like "tell people what is going on" and "comment on stuff" and "chat about moderation decisions" and whatnot.
#Sciop hit a Petabyte (actually a Pebibyte but nobody knows that word) of total proven capacity a week or two ago. That's all the seeders * the size of the things they are seeding. All volunteers, zero dollars in funding, piggybacking off existing resources wherever we can, run on a donated VPS. This is before we even get into federating archives and are still nailing down the basics of the site.
Peer to peer archives are real and they work, period. 216TiB of threatened cultural, climate, queer, and historical information held in common. That's a people powered archive, and you're welcome in it - to take from, to add to, and help sustain if you can.
Edit: if this is the first you're hearing of sciop, it's at https://sciop.net
First post on the #sciop blog - on adding webseeds from the web interface and why this is cool for bridging archives and bringing bigger systems into the swarm. If you can't run a bittorrent client but still want to seed, this post is for you!
Im making a blog for #sciop because we keep doing dope shit and not writing it down. Putting a call out for guest artists who want to contribute fake GeoCities era banner ads
Worried about #Smithsonian data and collections? We are too. Our friends over at #SafeguardingResearchAndCulture have been hard at work helping with #DataRescue and adding Smithsonian information to #SciOp.
Check out their available datasets and please spread the word: sciop.net/datasets/
Datasets - SciOp
if you are seeding anything on #sciop (or anywhere else too) using qbittorrent (and probably other clients too), you should increase your max torrent size to something like 2GB - that's what's causing the recurring problem that many people have flagged to us where their torrents seem to disappear from their client after restarting:
https://github.com/arvidn/libtorrent/issues/8012
tools > options > advanced, set both torrent file size limit and bdecode token limit very high
v2 torrents are very very good for archives, but they are more rarely used in piracy, so there is comparatively less optimization pressure for them. so this explains why our seed stats are so spiky, because we encourage hybrid/v2, and by default any v2 torrent larger than a few dozen GB will just go poof on restart.
edit: this was actually fixed in qbt 5.1.2, so you can also just update
Last week trump announced plans to "review" 8 Smithsonian museums. Today he doubled down, very explicit about the intent to revise history to reflect the ethno-nationalist fantasy of US history.
You can do something about that! We are backing up the digital archives of those museums on sciop: https://sciop.net/tags/smithsonian
You can take direct action to preserve the historical artifacts the right wants to destroy:
1) you can download a copy and seed it, every seeder counts. Subscribe to the smithsonian RSS feed to auto-download torrents as they are scraped.
2) we have also written a crawler connected to sciop that distributes the scraping work, and automatically creates and uploads a validated torrent that piggybacks off the s3 bucket as a webseed source while it lasts (instructions in reply).
The data from the 8 threatened museums is on the order of ~10 TB, and we have split it up by jpg/tif so people without much spare storage can join in on the jpg's at least. The full contents of the public smithsonian bucket is ~700TB, so if we want to have a full independent copy we'll need lots more seeders.
All this code is being written flat out, on the run, as it's needed by volunteers with exactly zero resources, so it's not polished or well documented, and if you're interested in helping damp the flames of the book burning by contributing to any of the code or docs, we'd love to have you.
The slides of my talk at #WHY2025 "Safeguarding Research & Culture: Save public data from the digital bookburnings!" are now online:
https://hu.berlin/SRC-WHY2025
Recording here (27min):
https://media.ccc.de/v/why2025-238-safeguarding-research-culture-save-public-data-from-the-digital-bookburnings
(Wow, awesome work by @c3voc 💜)
More context: https://program.why2025.org/why2025/talk/B8DANE/
Some of you may have seen the news re National #Climate Assessment Reports
https://www.cnn.com/2025/08/07/climate/wright-national-climate-assessments-updating
A friendly reminder:
They are all accessible here in this archive from November https://globalchange.govarchive.us/
As well as on #SciOp (from April)
https://sciop.net/datasets/globalchange-gov-webrip/pdf
Digital archival projects are crucial in the fight against fascism. I wrote about the why and the how.
And if you're reading this, that means you have a computer, so you too can contribute!
https://carefullmusings.bearblog.dev/the-urgency-of-digital-archiving/
#ArchiveTeam #SciOp #fascism #archive #resistance #DigitalPreservation
On SciOp, looking for data to protect from digital book burning, I find that these USA gov Coronavirus sites now point to a single site, promoting the lab leak conspiracy theory, attacking Faulci, China, science, common sense and being one of the most cringeworthy pieces of disinformation I've seen on the net.
I'll link to SciOp, not to the bullshit site:
https://sciop.net/datasets/coronavirus-gov
#politics #science #SciOp #trump #coronavirus #COVID19 #torrent #p2p #fascism #usa
man i just had a series of extremely good ideas* that are very simple and very implementable for #sciop that i think will cause an absolutely disgusting amount of (good, intrinsically deduplicating, actually decrease server load by creating a supporting swarm of peers) public data scraping to happen and basically lower the barrier to scouting endangered datasets to zero
*if you received the message flood of me having them you are not allowed to tell people if they are actually bad