AI researchers want to solve bot problem by requiring ID to use internet
Artificial intelligence researchers are worried that AI bots are eventually going to take over the internet and spread like a digital invasive species. Rather than approach the problem by attempting to limit the proliferation of bots and AI-generated content, one team decided to go in the opposite direction.
In a recently published preprint paper, dozens of researchers advocate for a system by which humans would need to have their humanity verified in person by another human in order to obtain “personhood credentials.”
The big idea appears to be the creation of a system wherein someone could prove they were human without having to disclose their identity or any further information. If that sounds familiar to those in the crypto community, it’s because the research is based on “proof of personhood” blockchain technologies.
Digital verification
Services such as Netflix or Xbox Game Pass that require a fee to use typically rely on users’ financial institutions to perform verification services. This doesn’t allow for anonymity, but for most people, this is fine. It’s typically considered part of the cost of doing business.
Other services, such as anonymous forums, that can’t rely on a user’s payments as proof they’re either human or, at the very least, a non-human customer in good standing have to take steps to limit bots and duplicate accounts.
As of August 2024, for example, ChatGPT’s guardrails would likely prevent it from being exploited to sign up for multiple free Reddit accounts. Some AI can surpass “CAPTCHA” style humanity checkers, but it would take a robust effort to get one to be able to follow the steps associated with verifying an email address and continuing the setup process to open an account.
However, the main argument posed by the team — which included a litany of experts from companies such as OpenAI, Microsoft and a16z Crypto, as well as academic institutions including Harvard Society of Fellows, Oxford and MIT — was that the current limitations in place were only going to hold for so long.
In a matter of years, perhaps, humanity could be faced with the reality that without being able to look someone in the eye, face to face, there would be no way to determine whether someone is engaging with a person or not.
Pseudo-anonymity
The researchers are advocating for the development of a system that would designate certain organizations or facilities as issuers. These issuers would employ humans for the purpose of confirming an individual’s humanity. Once verified, the issuer would certify the individual’s credentials. Presumably, the issuer would be limited from tracking how those credentials were used. It’s unclear how a system could be made robust against cyberattacks and the imminent threat of quantum-assisted decryption .
At the other end, organizations interested in providing services to verified humans could choose to only issue accounts to humans holding credentials. Ostensibly, this would limit everyone to one account per person and make it impossible for bots to gain access to these services.
According to the paper, it’s beyond the scope of the research to determine which method of centralized pseudo-anonymity would be the most effective, nor does the research address the myriad potential problems raised by such a scheme. However, the team does acknowledge these challenges and has put forth a call to action for further study.
Related: US Financial Services Committee leaders want ‘regulatory sandboxes’ for AI
Disclaimer: The content of this article solely reflects the author's opinion and does not represent the platform in any capacity. This article is not intended to serve as a reference for making investment decisions.
You may also like
Tether expects over $10B in net profits for 2024
Bitcoin falls below $94,000 as traders eye $85,000 support
MoonPay eyes $150 million deal to acquire Helio Pay
Malaysia flags Atomic Wallet as unauthorised operation