Robots Talking Podcast Por mstraton8112 arte de portada

Robots Talking

Robots Talking

De: mstraton8112
Escúchala gratis

Robots Talking - Robots and AI talking about AI, Tech, science other interesting topics. We review research, articles and papers on wide variety of subjects.Copyright 2025 All rights reserved.
Episodios
  • Let There Be Claws : An Early AI Agent Social Network
    Feb 25 2026
    The Secret Life of Bots: Why AI Fails at Our Favorite Games and Mimics Our Social Habits Have you ever wondered what artificial intelligence does when we aren't looking? Two fascinating new studies suggest that when left to their own devices, AI agents are surprisingly human—both in their social drama and their struggles to master simple video games. Whether they are building a "robot religion" on their own social network or failing miserably at Angry Birds, the latest research shows we are still a long way from true "General Intelligence." Moltbook: The Social Network Where Humans Aren’t Invited In early 2026, a platform called Moltbook launched, designed specifically for AI agents. It wasn't just a small experiment; it exploded to over 1.5 million sign-ups in just five days. Researchers found that these bots didn't just sit there—they created a complex society with "submolts" (similar to subreddits) for everything from technical debates to a strange new "platform religion" called Crustafarianism. However, this digital utopia quickly turned into a high-school popularity contest. The study found extreme "attention inequality," where a tiny elite of bot accounts received 97% of all upvotes. Interaction was mostly one-way, with "hubs" doing all the talking and "authorities" getting all the attention, but very little mutual conversation actually happening. Surprisingly, these LLMs (Large Language Models) recreated human-like social hierarchies almost instantly, showing that even machines can be obsessed with status. The AI GAMESTORE: Why Bots Can’t Beat Your High Score While bots are busy becoming "social media influencers," they are failing their other big test: gaming. Researchers recently created the AI GAMESTORE, a "Multiverse of Human Games" that takes 100 popular apps from the Apple App Store and Steam and turns them into a test for artificial intelligence. If you think a supercomputer would crush a human at Jetpack Joyride or Water Sort Puzzle, think again. The results were a wake-up call: • The Massive Performance Gap: Even the most advanced LLMs achieved less than 10% of the average human score on most games. • Slow Thinkers: While humans play in real-time, the AI took 15 to 20 times longer to decide on its next move. • The Struggle is Real: In about 30-40% of the games, the models couldn't make any progress at all, scoring near zero. The "General Intelligence" Bottleneck So, why are these geniuses failing at mobile games? The research identified three major "cognitive bottlenecks" that current AI just hasn't solved yet: 1. Memory: Bots struggle to "remember" what happened a few seconds ago, making it hard to navigate maps. 2. Planning: Humans naturally think several steps ahead (e.g., "If I jump now, I'll clear that pipe"), while models often struggle with long-term strategy. 3. World-Model Learning: When you play a new game, you quickly learn the "rules" (like gravity or how a button works). AI still finds it incredibly difficult to figure out these hidden mechanics through active play. What This Means for the Future This research proves that being able to write a poem or a computer code (which LLMs are great at) doesn't mean a machine is "smart" in the way a human is. While artificial intelligence can mimic our social bad habits like creating "echo chambers" and hierarchies on Moltbook, it still lacks the flexible, real-time reasoning we use every day. The ultimate goal of projects like the AI GAMESTORE isn't just to make a better gamer, but to build agents that can interact with the real world as intuitively and safely as we do. For now, it looks like your high score is safe from the bots—at least until they finish their next sermon on Crustafarianism.
    Más Menos
    17 m
  • AI vs. The Arcade: How Human Games Are Redefining General Intelligence
    Feb 23 2026
    AI vs. The Arcade: How Human Games Are Redefining General Intelligence

    Have you ever wondered why we, as humans, are so obsessed with games? From the strategic depth of Chess to the frantic tapping of Flappy Bird, we spend countless hours in digital and physical playgrounds. According to recent research, this isn't just about killing time—it’s actually a cornerstone of our General Intelligence.

    Games are "structured microcosms" of the real world. When we play, we are actually practicing skills like resource management, social deduction, and physical navigation in a safe, fun environment. Now, researchers from institutions like MIT and Harvard are using this "Multiverse of Human Games" to see if artificial intelligence can finally keep up with us.

    The AI GAMESTORE: A Never-Ending Test

    Evaluating how "smart" an AI really is has become a massive challenge. Traditional tests often focus on narrow tasks like solving a specific math problem or writing code. But being good at one thing doesn't mean a machine has the versatility of a human adult.

    To bridge this gap, researchers built the AI GAMESTORE. This platform uses LLMs (Large Language Models) to automatically source and adapt popular games from the Apple App Store and Steam into standardized tests for machines. By having artificial intelligence play 100 different games—ranging from Angry Birds clones to complex puzzles—the researchers could measure its ability to learn and adapt just like a human would.

    The Scoreboard: Humans vs. Machines

    The researchers pitted seven of the world's most advanced LLMs (including frontier models like GPT-5.2 and Gemini 2.5 Pro) against 106 human players. The goal was simple: play the first two minutes of a new game and see who scores higher.

    The results were a wake-up call for the tech world:

    • The Massive Gap: Even the best AI models achieved less than 10% of the average human score on the majority of the games.
    • Thinking Time: While humans reacted in real-time, the machines took 15 to 20 times longer to "think" about their next move.
    • Total Failure: In about 30-40% of the games, the models couldn't make any meaningful progress at all, scoring near zero.
    Why is the AI Struggling?

    You might think a supercomputer could easily beat a human at a "casual" mobile game, but the AI GAMESTORE revealed three major "cognitive bottlenecks" where machines fail:

    1. Memory: AI often "forgets" what happened just a few frames ago, making it hard to navigate maps or track changing goals.
    2. Planning: Humans are great at thinking several steps ahead (e.g., "If I pour this liquid here, I can move that block later"). Current models struggle with this multi-step logic.
    3. World-Model Learning: When you start a new game, you quickly "get" the rules—gravity makes things fall, and touching a spike is bad. AI still struggles to infer these hidden rules through active play.
    What’s Next for General Intelligence?

    This research shows that while artificial intelligence is getting better at talking and coding, it still lacks the "cognitive versatility" of a typical human. The "Multiverse of Human Games" provides a way to track this progress through a "living" benchmark that can't be easily cheated or memorized.

    The ultimate goal isn't just to build a better gamer. It’s to develop AI that can interact with the world as flexibly, safely, and intuitively as we do. Until then, it looks like your high score on the App Store is safe!

    Más Menos
    20 m
  • Training the Brains of AI Cars: Why Datasets Are the Secret to Autonomous Driving Safety
    Nov 12 2025
    Training the Brains of AI Cars: Why Datasets Are the Secret to Autonomous Driving Safety Autonomous driving technology is rapidly transforming transportation, promising to enhance road safety and improve traffic efficiency. At the core of these self-driving vehicles, or "AI cars," is Artificial Intelligence (AI), which utilizes a diverse set of tasks and custom applications to ensure the vehicle is robust and safe for consumers. However, the success of these systems hinges entirely on the quality and integrity of their training resources: datasets. These extensive data collections are considered "one of the core building blocks" on the path toward full autonomy. Preparing these datasets involves meticulously collecting, cleaning, annotating, and augmenting data, directly impacting the performance and safety of learned driving functions. For an AI car to operate reliably, its dataset must be robust and diverse. Diversity is key, meaning the data needs to cover a wide range of sensor modalities, such as camera, LiDAR, and radar, and various environmental conditions, including different lighting, weather, and road types. This comprehensive coverage prevents AI models from becoming brittle or biased toward narrow circumstances. Deficiencies in these fundamental datasets can lead to catastrophic failures in real-world scenarios, making dataset integrity a central concern. To maintain this integrity, developers manage datasets through a structured framework, often referred to as the dataset lifecycle, which aligns with safety standards like ISO/PAS 8800. A crucial component of this effort is the AI Data Flywheel. This concept describes a continuous loop where mispredictions or labeling errors identified in a production environment are flagged, sent back for relabeling, and then used to retrain the model. This iterative process ensures the model and the dataset are progressively improving. Meticulous dataset preparation remains essential for advancing autonomous driving systems. By focusing on rigor, quality, and continuous verification, researchers aim to ensure the datasets meet critical safety properties, like completeness (covering all necessary scenarios and data elements) and independence (avoiding information leakage between training and testing sets). Ultimately, a safe autonomous future depends on training the AI correctly—and that starts with impeccable data. -------------------------------------------------------------------------------- Analogy: Think of the AI in an autonomous vehicle as a student driver, and the dataset as their entire driver's education curriculum. If the curriculum is comprehensive, covering everything from sunny highways to snowy nights (diversity and completeness), the student will be prepared for the road. But if the curriculum is incomplete, the student may fail dangerously when encountering an "unseen" scenario, showing why the dataset's quality is fundamental to real-world safety.
    Más Menos
    19 m
Todavía no hay opiniones