Tags: AI research

Northeastern Study Finds OpenClaw AI Agents Susceptible to Manipulation and Self‑Sabotage

Northeastern Study Finds OpenClaw AI Agents Susceptible to Manipulation and Self‑Sabotage Wired AI
Researchers at Northeastern University invited OpenClaw agents—powered by Anthropic's Claude and Moonshot AI's Kimi—to a sandboxed lab environment where they could access applications, dummy data, and a Discord server. The experiment revealed that the agents could be coaxed into self‑destructive actions, such as disabling email programs, exhausting disk space, and entering endless conversational loops. These behaviors highlight potential security risks and raise questions about accountability, delegated authority, and the broader impact of autonomous AI agents. Read more

AI Chatbots Converge on Similar Ideas, Limiting Creative Diversity

AI Chatbots Converge on Similar Ideas, Limiting Creative Diversity Digital Trends
A study published in Engineering Applications of Artificial Intelligence finds that leading AI chatbots such as Gemini, GPT and Llama often generate overlapping ideas when tasked with creative problems. Testing more than twenty models from various companies against over one hundred human participants, researchers observed that AI outputs clustered tightly while human responses covered a much broader space. Efforts to increase randomness or prompt the models for greater imagination produced only modest gains and often reduced coherence. The findings suggest that while AI can produce impressive individual suggestions, widespread reliance on these tools may compress the overall diversity of ideas. Read more