Artificial Intelligence Can Now Take Full Control of Your PC
Imagine giving a chatbot the power to control and manage your computer. This exciting, but potentially dangerous idea is making people nervous.
Anthropic, the company behind Claude 3, has already captured attention this year with its cutting-edge chatbot, which launched across Europe in May. Claude 3 set a high bar with its ability to summarize entire dissertations in seconds, perform complex calculations, and craft well-written content effortlessly.
Now, the company has unveiled Claude 3.5 Sonnet, which extends even further, granting the bot full control over a computer's operations. While this is undeniably a major achievement, it also brings up serious questions about security.
Claude is no longer just a conversational AI; it has become an active digital assistant capable of operating the keyboard and mouse, as well as running applications directly on a computer. Unlike other chatbots, it can fully interact with installed programs and perform intricate tasks on your device.
Anthropic's Chief Scientist, Jared Kaplan, stated in an interview with Wired, "We're entering an era where AI models can use all the tools humans rely on to perform tasks." During trials, Claude 3.5 demonstrated its versatility by planning travel itineraries through web browsers, managing calendar events, and even building websites from scratch using Visual Studio Code.
While Claude 3.5's abilities are impressive, it isn't perfect. Testing shows it still struggles with certain tasks, like coding, and occasionally misses essential details when organizing tasks. TechCrunch reported that, in a scenario requiring the bot to book and alter flight reservations, Claude completed less than half of the tasks assigned, showing that it's far from infallible.
But should we feel comfortable granting AI such extensive control? Though it isn't all powerful yet, it raises valid concerns about the risks of giving a bot unrestricted access to our systems.
Anthropic's approach is cautious. According to the company, "Allowing these current, more restricted and relatively safe models to access computers is a proactive step. It lets us identify and address any emerging issues, all while progressively improving security protocols." However, this strategy implies that users must assume a degree of risk themselves. Are they fully informed of the potential dangers? And if something goes wrong, who will bear the responsibility, Anthropic, the user, or both?