So much of internet traffic is ran by bots/AI, it's only getting better at mimicking real humans. I'm not sure if privacy would be possible if people were to take a stance against AI driven traffic

  • Your post has absolutely nothing to do with your question. The premise makes zero sense. Why would the proliferation or amount of traffic made by bots have anything to do with your personal privacy?

    When bot activity outnumbers human activity by a large margin, I assume people would have to find ways to create a space for only humans to communicate, and verifying humans to whitelist them would require giving up some level of identification

    I give up social media long before a requirement to tie aliases to a real person is the standard.

    that will only be if you want to prove its you but your never gonna be able to unless the do fingerprint or something. mostly everything else will probably stay the same. your gonna have bots generating a ton of false info and false id's its probably already to late for them then when asi gets here that could identify things probably. i just hope sites start dropping age verification en mass they should never have followed it to begin with

    Well, one way would be;

    1. You make a comment and a bot analyzes it.

    2. A bot writes a reply to see how you respond to certain prompts.

    3. Then you reply and it gets analyzed again.

    4. Your writing patterns then get analyzed and matched with comments on multiple services, like Reddit, Instagram, Facebook etc. A shared email address across services makes it even easier.

    5. Now a bot knows exactly how you write and interact with people online. A few years of this, and any comment made by anyone on any site can instantly be linked to a person, no matter what account you're using.

    But this is all tin foil hat. I don't necessarily believe this but this is absolutely possible. There are already people working as forensic linguists, this would just be a robotic mass implementation of that.

    TL;DR: Comment section scrapers and bot commenters can be used to fingerprint each person, fully non-digital (IE. No browser, IP, or account fingerprint needed. Just you, the author.)

    This assumes intelligence.

    “ai” is a marketing term, not intelligence.

    LLMs are based on basically taking an enormous amount of data (say Reddit posts or Twitter posts), converting the words into numbers (that’s the LLM part) then feeding the results into a lossy compression algorithm that is designed to predict what the next number in a sequence would be (a so called token). To make this computationally feasible (not petabytes) the least probable tokens are pruned. Thus only the most common sequences are kept, converting petabytes to gigabytes. Then any input you enter is fed through the pruned tree to produce a “random” output which is converted back into words by running the LLM backwards. By closed loop feeding new inputs into the “learning” offline it can incorporate those Or the inputs can be used immediately to bias/prime the filter to tune the responses to be inline with the entire input history.

    That’s right, lossy data compression is literally being accused of “intelligence”.

    We can play around with the tuning if thus but ultimately this is as good as it gets. It’s not going to get “better” or “more intelligent” other than we can create ever larger pruned data sets or we can tune for longer inout predictions or tinker with the overall design. But simply put even a trained monkey is more intelligent. There isn’t actual intelligence so no innovation, non”understanding”. Anyone that spends even seconds looking at the output, never mind interacting with it, will realize it’s just a program.

    lossy data compression is literally being accused of “intelligence”

    This is a funny sentence. Do you think the brain is lossless? I am not disagreeing with the rest of what you said, but this part got to me.

    No? My comment assumed simple text-in-analysis-text-out, totally doable with simple LLM. The scenario I described can be easily done even today. Massive servers needed to fingerprint everyone, but that's never stopped anyone before.

    You just described the current monetary model of all social media companies except you replaced it with "bot"

    Now a bot knows exactly how you write and interact with people online. A few years of this, and any comment made by anyone on any site can instantly be linked to a person, no matter what account you're using.

    How does this take multiple tones into the equation, and in a data breach, would it be safe to say another AI could use that info to mimic real humans and write horrible things?

    You might find this interesting (or terrifying lol)

    From what I understand most people have identifying traits in the way they write even if they consciously try not to. I don't know very much about it but I watch too much true crime stuff and it comes up pretty often.

    "Modern stylometry uses computers for statistical analysis, and artificial intelligence and access to the growing corpus of texts available via the Internet." -Wikipedia

    Also says this under the adversarial stylometry section:

    "The privacy risk is expected to grow as machine learning techniques and text corpora develop."

    Could be used for some pretty good things but with the way things have been going I bet it won't be.

    in a data breach, would it be safe to say another AI could use that info to mimic real humans and write horrible things?

    A gym teacher already tried to use AI voice generation to get his boss fired a while back, this shit is going to be a pretty bad problem soon I think.

    Sorry if this is a jumbled mess or if I misunderstood anything I'm tired as hell and should probably get some sleep.

    Definitely, yes. If an AI scraped every single comment a person made over a few years, it could mimic them perfectly.

    Cool, but he placed the conditional on taking a stand against AI

  • Maybe, but you can avoid putting information out there.

  • I don’t see how your two ideas are connected

    In an age (in the future) where most comments and posts from social media will be made by bots/AI, the only way for humans to verify themselves is by revealing identifying info in order to access a human only internet

  • A temporary solution to verify that both ends are humans could be to start writing backwards on social networks. LLMs generate their paragraphs one word at a time, predicting which is the most probable word that shall go next to the previously generated word. If you ask AIs to generate a paragraph in reverse (starting from last word and ending with first word), they struggle... but humans could easily get used to this. Eventually AI will pick this up, but for the time being, it could be a limitation that could be creatively exploited.

  • The great and sad irony here is that a bot has posted this…..