#Duplicates

PrecisionSQLPrecisionSQL
2025-12-26

UNION ALL Preserves Duplicates?!

UNION ALL keeps ALL rows! UNION removes duplicates. This one keyword difference changes the entire result set. This will break your data!

youtube.com/watch?v=_Sa8Kgc-nuU

2025-10-15

A random #gaming #console post:

I am a recent convert to a modded #SNES (thank you fellow #veteran and retro gaming community member) passed along to me, and been playing the hell out of the games that he passed along for "offsite #backup /#archival storage". I started looking for other ROMs for other games that I have, found some ROM packs, imported into #Hakchi2. Now I have a ton of ROMs but soooo many duplicates. I looked though the options/capabilities within Hakchi2 but I'm not seeing anything for detecting & purging #duplicates detect which ones do or do not have art, and other characteristics to help filter duplicates.

Am I missing something, or is this a manual process for cultivating your ROMs? Seems my limited understanding & experience with ROM sets created the mess I am trying to get my head wrapped around. As a casual gamer, I was looking to explore the different games that I never got to experience in the heyday; not sure what pre-made sets would be of benefit TBH.

So OMG down the #rathole I went - found the dat-o-matic and am getting the DAT file imported into #ROMcenter... I found a couple of decent vids on youtube that has me on a hopefully good path on getting these sorted, along with using universal ROM Cleaner as well as #ROMcenter. But I ended up spending all my time last night working on this rather than applying for jobs. 🤦‍♂️

Also that 1TB SSD in the laptop's data drive is now almost full with #ROMs & #ROM packs, so I have to get things filtered and deleted out.

Anyone else wrangled this challenge to success?
#retroGaming #casualGamer #retro
#WallOfRetro
#retroTech #retroTechnology
#nerdsOfVintage #happyNerding
#computer #tech #retroTech #retroTechnology
#nerdsOfVintage #happyNerding
#makeShitMonday #showmewhatyougot

Kevin Karhan :verified:kkarhan@infosec.space
2025-08-22

@pancake Relative.

  • I think in that case there's rather a systemic problem of underresourced #FLOSS projects: More funding and personnel would greatly help in said effort.

The only thing #StaleBot generates are #duplicates due to new tickets being opened with the exact same problem

I don't deny the issue at hand; I just think that such a massive pile needs to be combed through better by using tags and prioritization. Even if that means one has split the backlog into half a dozen different baskets that read "feedback", "suggestion", "bug", "security", "ui/ux", "building", "l18n/i18n", "accessibility", "reproduceability", etc...

N-gated Hacker Newsngate
2025-07-14

🚀 Breaking news: someone reinvented the wheel by creating yet ANOTHER photo organizer that detects . 🎉 Of course, they value your privacy... as they track and analyze every click you make to "improve" your experience. 🙄
organizer.flipfocus.nl/

2025-04-19

How Can I Resolve Duplicate Invoice Row Issues in SQL?
Tackling Duplicate Invoice Row Issues in SQL is crucial for data integrity. Employ methods like DISTINCT, GROUP BY, and JOINs to ensure unique datasets. Analyzing relationships and joins helps in identifying and managing duplicates effectively.
tech-champion.com/programming/
...

elshara Silverheartelshara
2025-04-11

Wow is I make a post and it 20 times across everywhere.

2025-03-29

How Can I Resolve Duplicate Invoice Row Issues in SQL?
Tackling Duplicate Invoice Row Issues in SQL is crucial for data integrity. Employ methods like DISTINCT, GROUP BY, and JOINs to ensure unique datasets. Analyzing relationships and joins helps in identifying and managing duplicates effectively.
tech-champion.com/programming/
...

2025-02-20

Handling Duplicate Keys in DB2: Efficiently Inserting Data While Ignoring Duplicates
Learn efficient techniques for DB2 Insert Ignore Duplicates. Explore methods like MERGE statements & exception handling to maximize performance & minimize disruptions.
tech-champion.com/database/db2

AtoZ Libraryatozlibrary
2024-04-08

Discover 4 quick ways to find duplicates in Google Sheets and maintain clean data.
Visit: atozlibrary.in/4-ways-to-find-
Learn how to use custom formulas in conditional formatting, highlight duplicates using Countif Function, leverage pivot tables for duplicate analysis, and consider whitespace issues. Streamline your workflows and improve data accuracy with these methods!

2024-04-05

When you're on Ancestry, do you ever:
* Swear that someone else *must* have changed your tree (even though you know that's not possible)?

* Think that people are missing from your tree or that they "somehow" got disconnected from other people in your tree?

* Accept Hints from other people's trees, or maybe you used to do so, early on?

There's a good chance you have duplicates in your tree, and you might not even realize it.

A lot of Genealogists have been having problems with a glitch in Ancestry, where it creates duplicates of people already in their trees. Here's how to fix the problem now, in the easiest way possible, and how to avoid it in the future.

Do NOT delete the duplicates! This can mess up relationships in your tree, and cause you to lose sources & hint matches. Instead, try this…

#AncestryCom #FamilyTree #Research #Duplicates #Genealogist #Ancestry #Genealogy @geneadons #Geneadon #Geneadons #FamilyTrees

diggingintoyourfamilytree.word

2023-10-21

Saturday night craziness: implementing things in shell that are already available in C.

**Tool for detecting duplicate folders**

github.com/ilario/finddirdupes

Beware: it is much slower than `rmlint -D`.

#dupes #linux #shell #bash #duplicates #lint #rmlint

2023-10-10

I finished getting ~4TB of my data from Amazon Photos last night(it's just one part of the whole though). The transfer has been going on since last Friday and finally finished during the night last night. So, I am going over the lot with #dupeGuru to weed out #duplicates and get those ~4TB to a lesser number.
The first scan isn't finished, but damn that's a lot of duplicates already!

2023-09-08

I have a large amount of #files and #folders spread across 3 external hard #drives. Some of the files are duplicates but I don't know which ones and where. Can anyone recommend some software that can easily find #duplicates across external hard drives?

2023-09-03

I repurposed one of his old laptops (that overheats, sadly) as frontend, making room for an Ubuntu partition.
I installed the trusty #recoll to index text documents and emails.
For images, I ended up using #KPhotoAlbum, that has very flexible tagging possibilities, making it useful on more than photo albums. Chronological view and image stacking are also great.
I used #Geeqie, then more recently #Czkawka for finding image #duplicates, that I then stacked manually in KPhotoAlbum.

2023-07-24

anyone know anything about #data #deduplication. i have 1 hard drive where I've gathered everything all my #backups, music, docs, emails etc etc etc. but i have a few copies of everything. want a program that simply scans the #drive, deletes all #duplicates, and keeps a copy of only 1 of everything. does the damn thing exist as i cant find anything simple enough? #techHelp #ComputerHelp #software #tech #techie

2023-07-18

Recommended #opensource #file #duplicates detection and deletion: #rmlint

Why? - Extremely fast · #CLI · Candidate file filtering by #name, #size, #modification #time · Configurable criteria for determining the original file · Paranoia mode offered (byte-by-byte comparison) · Flexible output #formats, including #bash deletion script, #json, #CSV · Excellent #documentation and #tutorials

github.com/sahib/rmlint

More recommendations: tuxwise.net/recommended-softwa

DeaDSouL :fedora:DeaDSouL@fosstodon.org
2023-05-27

I needed to remove all duplicates from ~/.bash_history, and since it contains the date and time for each command, I ended up with lots of dates and times with no commands (since they were duplicates and got removed).

And I didn't want to check 34k+ of lines manually. So, I wrote a Python script to solve that.

in case you're interested, here is the code: gitlab.com/ds.python/useful-sc

#BASH #history #bashhistory #python #duplicates #duplicate #dedup #code #python3 #linux #gnulinux

@skaficianado

At least it is #unique .

No need for stupid numbers at the end to avoid #duplicates

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst