Fixing Noisy Logs with #OpenTelemetry Log #Deduplication
https://www.dash0.com/guides/opentelemetry-log-deduplication-processor
Fixing Noisy Logs with #OpenTelemetry Log #Deduplication
https://www.dash0.com/guides/opentelemetry-log-deduplication-processor
Hab mein PyHardLinkBackup komplett neu geschrieben. Ursprรผnglich 2015 gestartet und bis 2020 genutzt, schlief es jetzt fast 6 Jahre...
Aber als ich รผber alte damit erstellte backups gestolpert bin, hab ich mir gedacht, das Konzept ist doch ganz nรผtzlich.
Also kompletter rewrite: https://github.com/jedie/PyHardLinkBackup
TIL: #XFS kann #Snapshots aber keine #Compression, aber dafรผr #deduplication, wenn auch noch experimental
Wer ein #snapshot artiges Backup fรผr #Linux sucht, kรถnnte sich #kopia ansehen.
รber Regeln sehr fein granuliert einstellbar.
Es hat mich allerdings jetzt fast eine Woche gekostet, es so zum laufen zu bringen, wie ich es mir gewรผnscht habe. Aber mit viel #scriptโen hat alles geklappt.
#deduplication und #kompression, schnell und easy.
Sehr zu empfehlen.
And once in a while I cleanup the external libraries with #Czkawka
This is an amazing software for #deduplication of image folders.
@stratosphere it's always interesting what's constantly trying to hack...
Also one little feedback: It would be cool to see some "#deduplication" or rather reduction of entries by using #CIDR notation which should save a lot of lines = table entries compared to the single IPs.
Sick: Indexed deduplicated binary storage for JSON-like data structures
#HackerNews #Sick #Indexed #Binary #Storage #JSON #Deduplication #DataStructures
The machine forgets. The Ghost does not.
https://deadswitch.tomsitcafe.com/2025/10/borg-backup-intro.html
Part 1 : #PySpark Data Pre-processing Essentials #filtering || #Deduplication || Data Cleansing.
Learn PySpark data pre-processing with our tutorial! Learn the art of filtering and deduplication, essential techniques for cleaning ... source
Ich hab mal versucht, die Speicheroptimierung durch #Deduplication beim #Backup mit #restic zu quantifizieren. Dies nach einer Laufzeit von knapp 2 Jahren.
Herausgekommen ist: 22,4%
# restic stats latest
repository d989459c opened successfully, password is correct
scanning...
Stats in restore-size mode:
Snapshots processed: 1
Total File Count: 438037
Total Size: 23.271 GiB
# restic stats latest --mode raw-data
repository d989459c opened successfully, password is correct
scanning...
Stats in raw-data mode:
Snapshots processed: 1
Total Blob Count: 265960
Total Size: 18.409 GiBHoffe, das richtig interpretiert zu haben.
Some design tests for ๐ฅ ๐ฉ diff'ing files hosted on Hugging Face.
With the new storage backend (Xet), similar files share many data "chunks".
As HF provides an API to get the list of these chunks, you can now compute the diff and display a chart in the browser without downloading the files. Very fast.
Test here: https://observablehq.com/@severo/parquet-content-defined-chunking-blog-post-reprodu
Details about the three charts in the responses.
๐ ๐๐ฒ๐ป๐ป๐ฒ๐ป ๐ฆ๐ถ๐ฒ ๐๐ฐ๐ต๐ผ๐ป ๐ฑ๐ถ๐ฒ ๐๐ป๐๐ฒ๐ด๐ฟ๐ฎ๐๐ถ๐ผ๐ป ๐๐ผ๐ป ๐ฃ๐ฟ๐ผ๐ ๐บ๐ผ๐ ๐๐ฎ๐ฐ๐ธ๐๐ฝ ๐ฆ๐ฒ๐ฟ๐๐ฒ๐ฟ (๐ฃ๐๐ฆ) ๐บ๐ถ๐ ๐ฃ๐ฟ๐ผ๐ ๐บ๐ผ๐ ๐ฉ๐ถ๐ฟ๐๐๐ฎ๐น ๐๐ป๐๐ถ๐ฟ๐ผ๐ป๐บ๐ฒ๐ป๐?
Wer ๐ฃ๐ฟ๐ผ๐
๐บ๐ผ๐
๐ฉ๐ nutzt, sollte ๐ฃ๐๐ฆ einsetzen:
๐งฉ๐ก๐ฎ๐ต๐๐น๐ผ๐๐ฒ ๐๐ป๐๐ฒ๐ด๐ฟ๐ฎ๐๐ถ๐ผ๐ป
โ๏ธ๐๐ฎ๐ฐ๐ธ๐๐ฝ-๐ ๐ผ๐ฑ๐ถ ๐ณรผ๐ฟ ๐บ๐ฎ๐
๐ถ๐บ๐ฎ๐น๐ฒ ๐๐ผ๐ป๐๐ฟ๐ผ๐น๐น๐ฒ
๐พ๐๐ฒ๐ฑ๐๐ฝ๐น๐ถ๐ฐ๐ฎ๐๐ถ๐ผ๐ป & ๐ฃ๐ฒ๐ฟ๐ณ๐ผ๐ฟ๐บ๐ฎ๐ป๐ฐ๐ฒ ๐บ๐ถ๐ ๐ฃ๐๐ฆ
๐๐๐น๐ฒ๐ฒ๐ฐ๐ถ๐ป๐ด ๐ณรผ๐ฟ ๐ฉ๐ -๐๐ฎ๐ฐ๐ธ๐๐ฝ๐
๐ฆ๐๐ผ๐ป๐๐ฎ๐ถ๐ป๐ฒ๐ฟ-๐๐ฎ๐ฐ๐ธ๐๐ฝ๐ ๐บ๐ถ๐ ๐๐ต๐ฎ๐ป๐ด๐ฒ ๐๐ฒ๐๐ฒ๐ฐ๐๐ถ๐ผ๐ป
๐๐๐ผ๐บ๐ฝ๐ฟ๐ฒ๐๐๐ถ๐ผ๐ป & ๐ฅ๐ฒ๐๐ฒ๐ป๐๐ถ๐ผ๐ป
๐ ๐ฆ๐ถ๐ฒ ๐บรถ๐ฐ๐ต๐๐ฒ๐ป ๐๐ต๐ฟ๐ฒ ๐๐ฎ๐ฐ๐ธ๐๐ฝ-๐ฆ๐๐ฟ๐ฎ๐๐ฒ๐ด๐ถ๐ฒ ๐บ๐ผ๐ฑ๐ฒ๐ฟ๐ป๐ถ๐๐ถ๐ฒ๐ฟ๐ฒ๐ป?
Sprechen Sie mit uns ๐ 0441-309197-69
#ProxmoxVE #PBS #VMBackup #ContainerBackup #Deduplication #SnapshotBackup
Borg - Deduplicating Archiver with Compression and Encryption
#HackerNews #Borg #Deduplication #Compression #Encryption #Backup
J'ai lu le trรจs bon billet d' @athoune sur #Kloset, moteur de stockage de backup de #Plakar
Testing some more concurrent operations on the new #Synology #DS925+ as part of the upcoming #review
Dual #ActiveBackupForBusiness remote #windows11 machines (testing the #deduplication in the process), along a simple file copy onto an external #USB #SSD #OWC enclosure.
(yes 3rd party external drives work just fine!)
The NAS is nice and calm with no performance issues at all.
Happening now! In this month's #CommunityCall, #OpenAIREGraph Training & Engagement Officer Stefania Amodeo takes us on a journey through the Graph #deduplication process that uses a combination of advanced algorithms & human expertise. Missed out? All past #CommunityCall materials can be found on the Graph portal, today's to be uploaded in the coming days.
Past Calls https://graph.openaire.eu/community-calls
anyone know of an implementation of the CPM-SW deduplication algorithm of this paper?
[edited to change the url to the abstract page rather than direct to the pdf]
The Problem Isn't Email, It's Microsoft Exchange -- it turns out my 2011-vintage rant still rings true, now also available trackerless: https://nxdomain.no/~peter/the_problem_isnt_email_its_microsoft_exchange.html #inefficiency #timewasted #email #archiving #microsoft #exchange #compliance #deduplication #unsolvedproblems #smtp #mail #annoyances
DB2 Query Deduplication: Optimizing Large Datasets with ROW_NUMBER()
Learn efficient DB2 Query Deduplication using ROW_NUMBER() for large datasets. Optimize your queries with CTEs & indexing for smoother performance. #DB2Query #Deduplication #ROW_NUMBER #SQLoptimization #largedatasets
https://tech-champion.com/database/db2-query-deduplication-optimizing-large-datasets-with-row_number/
Learn how to efficiently deduplicate large DB2 datasets using ROW_NUMBER() and optimize query performance. ...
Testing #Synology #ActiveProtectManager and #DP320 appliance by making the ISO bare metal recovery media.
Let's see how fast this little machine can recover a 50GB setup.
The actual #backup and #deduplication are working really well.