New study shows smart chatbots can figure out who you really are from just a few posts… and it only costs a couple of dollars.
Nothing of substance here. Stylometric analysis was already a thing. Easy enough to defeat and with correct opsec you can avoid it. Burn accounts regularly. Use accounts for specific topics or share accounts with others. Dont post personal details online , on the internet everyone is a cat.
what kind of opsec can defeat it?
I certainly am!
I don’t like having to be vague about my age, nationality, job etc, because I’d rather be honest and relate to others online, but sadly it’s a necessity in the modern landscape
That’s why you have multiple accounts. Some that are for above table things which can have your personal details and others for eating the rich
One account that can be correlated to place/city, willing to discuss local news and issues.
One account that can be correlated to family status, willing to mention details about relationships.
One account that can be correlated to career, willing to mention details about educational background, industry news, the job market, the workplace, etc.
One account that can be correlated to each distinct hobby or interest. Some interests can correlate among themselves (like an all sports account that discusses multiple sports) and are safe to discuss on a single account. Like my current account that is tech oriented, including some stuff about games or Linux or networking or even the tech industry. But keep the different interests on separate accounts.
Then different accounts for topics that you consider controversial or private.
And, preferably, spread all those accounts across multiple instances so that instance admins can’t link accounts from metadata (client, OS, IP address, email verification), use completely unique usernames, and avoid unique markers like esoteric phrases, unique autocorrect errors, etc.
Even if an adversary can link two accounts, they probably can’t link all of them.
Step 2: Search the whole internet: It quietly checks LinkedIn, Google, other Reddit accounts, etc., to find possible real people who match those clues.
Oh. Whew.
Yeah, call this article what it is, clickbait fear mongering.
To be honest, internet search got so shitty, soon it will be a really impressive skill to search the internet efficiently
I thought Lemmy comments might be indexed anyway, but neither Kagi nor DDG turned up anything for my username. Wonder if it’s different for other instances?
Edit: Kagi actually does show some threads I’ve participated in by enabling the “Fediverse Forums” filter
Not surprising considering those challenges against AI scrapers likely also effect Search Engine crawlers. Stuff can get through if it’s federated on other servers that don’t have such measures but if you don’t participate in communities on those instances it’s less likely.
Searched “lemmy floquant” on DDG and one of your comments is the second result.
First result was the user profile of the same name but different instance.
and if they are wrong they still get the couple of dollars so win win. I can unmask anyone you want online from just a few posts and a name randomizer.
Looks like the LLM can be used to cross reference data from your pseudo private account to your public account. What a surprise
It’s a good thing that I work for Dick’s Fish & Chips located on the main street of a bustling city in Antarctica. I wouldn’t want the LLM to get it wrong.
The same Dick’s Fish & Chips where in 1998, The Undertaker threw Mankind off Hell In A Cell, and plummeted 16 ft through an announcer’s table?
That Dick’s Fish & Chips?
Yup. The only difference between this and what any individual could already do is just time and scale.
Data brokers and government surveillance organizations have already had specialized tools to do this sort of thing for a while now, it’s just that LLMs reduce the complexity and specialization needed to actually make an implementation that works well as an individual person.
upto
If the ‘journalist’ can’t use a spell-check, I don’t trust his opinions on automation.
AI can’t even get facial identification correct. It might claim to be able to identify people by online presence, but I wouldn’t doubt if it guesses incorrectly most of the time.
On real Hacker News users, the AI correctly linked the secret username to the real person 67% of the time, and when it made a guess, it was right 90%. The paper also states that matching the same person’s Reddit posts from different years or groups met with 68% success.
Based on the research they had a 60 something % accuracy. But the test data was for HackerNews accounts which linked to LinkedIn. I would guess that anyone linking their anonymous account to their LinkedIn profile isn’t really trying to hide themselves.
Oh no, who’s making them use all those terrible services that whore them out like that?






