Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm reminded of Neal Stephenson's book Anathem [1], which some people didn't like (compared to say Snow Crash or Cryptonomicon) but I really enjoyed. Apart from the mystical elements it raised some interesting points.

At one point it was talking about the Reticulum (Internet), botnet ecologies and how some of these would sublty modify information such that you couldn't really trust anything you read. I find this an intriguing and somewhat alarming idea.

Facebook was good early on because bots weren't sophisticated enough to create real-looking Facebook profiles so if you had a Facebook profile you had a person. Now they're better but still a human can pick a bot pretty easily.

But bots will only get better.

One idea that's been raised is the notion of trust. Usage patterns are analyzed to identify bot or bot-like behaviour. You see this in identifying sock or meat puppets on reddit, HN and other social news sites.

People have talked about the hordes of Twitter accounts used to make content appear more relevant than it is. This sort of thing is already happening.

But how do you pick out the bots when they're the majority and thus there is little to no statistical norm to compare them to real people?

At what point will bots start modifying Wikipedia articles and have other bots or, more likely, some stooges, approve those changes? What about establishing fake sites with wrong information and having them appear valid to search engines?

This will I believe be a big problem.

One particularly problematic aspect to this is that the site owners themselves, within reason, have little incentive to expose bots (or inactive accounts for that matter) because so much importance is placed on metrics like "# of active accounts".

This only becomes problematic if the user experience suffers beyond a certain threshold.

It's going to be interesting to see just how much of a problem this becomes and what we do to solve or at least mitigate it.

[1]: http://www.amazon.com/Anathem-Neal-Stephenson/dp/0061474096



so much importance is placed on metrics like "# of active accounts".

Also known as asshole metrics (which are, are at most, 40% true). They only exist to manipulate you into thinking the company is more important than reality. Most companies will tell you "total registered users" but few will give "active users." If they do give you active users, then they'll count "active" as something silly like "we saw their cookie during the month" meaning they were just delivered some external asset passively.

Stop lying and make something valuable (except lies can keep companies in business much longer than most people realize).


Hellbanning [1] is a solution - let the bot think it's connecting to other users, but in reality it's completely invisible - even better when you address a small net of these - they're banned to the same hell so they can see each other and interact, but are all invisible to the greater populace.

[1] http://en.wikipedia.org/wiki/Hellbanning


That works until you unintentionally hellban a legitimate user.


Ultimately it is just the same problem as with digesting information created by humans. Humans also misinform, out of malice or because they don't know better. The solution is rationality, I suppose. You have to verify the claims that are being made. The bots of the future will just have to get better at that.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: