UX Roundup: Preschool Robots | Multi-Metaphor AI | Meta Releases New AI | Cows & AI | Task Horizons | VC All-In on AI
- Jakob Nielsen
- 20 minutes ago
- 16 min read
Summary: Preschool robots build agency by reducing recovery time after failure | Multiple AI roles improve learning without flattening voice | Meta’s Muse Spark suggests the frontier race is shifting toward interface orchestration as well as raw model scale | Halter turns cattle management into a closed-loop AI system | GPT task horizon lower than Claude’s | Venture capital investments now even more dominated by AI

UX Roundup for April 13, 2026 (Nano Banana 2)
Instructional Robot Helps Preschool Children Develop Agency
Most preschool educational technology is a usability disaster. Conventional instruction relies on passive, teacher-centered delivery or flat multimedia that fails to account for young children’s need for active engagement. Faced with a passive interface, children quickly lose motivation, abandon the task, and fail to learn.
A recent study tested a superior interaction model: an AI-robot-supported task-based learning (TBL) approach. By evaluating how 42 five- and six-year-old users interacted with a physical humanoid robot named “Lele,” Jia-Hua Zhao and colleagues from Universidad Rey Juan Carlos in Madrid, Spain, uncovered insights into how we can design educational systems that don’t just broadcast information, but actually foster user agency and mastery motivation.
The User Test
The study deployed a classic comparison test in a preschool health education context. Both groups were tasked with learning about the five senses using tablet-based interactive animations, and the same human teacher taught both groups of kids.
The Control Group used a traditional multimedia approach guided only by a human teacher.
The Experimental Group was guided by the AI robot, which acted as a learning companion, introduced the tasks, and provided real-time, personalized feedback using natural language processing, facial expressions, and physical gestures.

The study isolates the variable that matters: same teacher, same school, same tablet lesson, but one group also had a physical robot. The point was not merely whether children learned the five senses. It was whether the interface increased agency.. (NotebookLM)
The Behavioral Data: Friction vs. Flow
The behavioral sequence data revealed a stark contrast in user experience. Children in the control group frequently hit usability roadblocks. When they failed a task, they had to wait for help from a busy teacher or peers. This friction led to severe drop-off rates; these children were far more likely to exhibit negative behaviors, give up on challenges entirely, and engage in off-topic chatting.

In the condition without a physical robot to help the kids, any challenges would break the learning momentum, since the children didn’t have enough agency to overcome the problem. (NotebookLM)
Conversely, the robot group demonstrated significantly higher levels of task persistence, better problem-solving abilities, and more positive emotional responses. The robot’s interface essentially kept the children in a state of productive flow.
Findings
Both the teachers and the children themselves rated the kids in the robot condition as having substantially higher mastery motivation (a critical component of agency) after completing the project. (There were no baseline differences before the study.) Ratings on a 1–5 scale (5 being best) were:
Teacher’s rating: 4.6 for the robot group vs. 3.8 for the control group.
Children’s self-ratings: 4.5 for the robot group vs. 3.3 for the control group.
Both differences were highly significant (p<0.001)

Since the children in this study had not yet learned to read, their responses were given on a scale of 5 faces rather than the usual Likert-type verbal scale. (NotebookLM)
More detailed data analysis in the paper showed that the robot-supported learning approach was better than the traditional multimedia task-based learning approach in terms of promoting children’s persistence, emotional response, and problem-solving abilities.
Designing for Agency: The Take-Away Lessons
Agency is the most important skill for the AI age, since AI will possess full competency to complete any task on its own but will still need to be directed by humans (see also my video on how to develop human agency.) Agency is the ability of a user to take independent action, navigate challenges, and direct their own experience. In child psychology, this is closely tied to mastery motivation. If we want to build technology that gives children agency, we must look at exactly what the AI robot did right.
Lesson 1: Micro-Interactions Must Reframe Failure
When a child in the experimental group failed a task, the system did not simply log an error. The AI robot triggered an incentive feedback mechanism, using comforting speech (e.g., “Don't worry, Lele will be with you”) paired with a physical hug gesture. Behavioral logs proved this specific interaction pathway successfully transitioned children out of frustration and directly back into trying the task again. To build agency, you must design interfaces that gracefully handle user failure. Timely, process-oriented feedback prevents abandonment and empowers the user to try again.

The physical robot encouraged the kids to try again, teaching them the agency to overcome challenges. Success or failure is not something that’s imposed on you by outside forces outside your control. You have an internal locus of control, which means you influence the outcome through your actions. (NotebookLM)
Lesson 2: Embodied Presence Lowers Emotional Friction
A flat avatar on a screen lacks physical presence. Physical robots provide embodied interaction, which feels significantly more natural and immersive to a human user. Because the robot was humanoid and utilized anthropomorphic gestures, the children treated it as a peer-like partner. This companionship reduced the anxiety and fear associated with making mistakes. Agency requires risk-taking. A physically present, non-judgmental interface creates a zone of emotional safety where children feel confident enough to take the initiative.
Lesson 3: Task-Based Architecture Forces Active Choice
The system was not designed as a talking encyclopedia. It utilized a Task-Based Learning (TBL) framework. The robot assigned specific, contextual missions, like dragging animated items to the correct sense organ, and then stepped back. The user was in control of the tablet. The robot only stepped in to provide scaffolding when the child actively sought help. You cannot design for agency if the user is a passive consumer. You must provide interactive, goal-oriented tasks where the child is the primary actor and the technology is merely the facilitator.
Lesson 4: Remove Bottlenecks to Problem-Solving
In traditional classrooms, a single teacher creates a bottleneck for help-seeking. The AI robot removed this bottleneck. Equipped with natural language processing, it could instantly identify keywords in a child's question and pull up relevant multimodal learning materials. Instant, personalized support allows users to maintain momentum. When children can reliably access the tools they need to solve their own problems in real-time, their independent competence skyrockets.
The deeper lesson is not that every classroom needs a cute robot. It is that agency depends on recovery latency: the time between “I’m stuck” and “I know what to do next.” In the control condition, that latency was governed by the teacher’s bandwidth. In the robot condition, it collapsed. Adults abandon software for the same reason children abandon lessons: the path back from confusion is too slow.
The Bottom Line
AI robots are not meant to replace the classroom management provided by human teachers for emotional support, value orientation, and classroom culture-building. However, as an interface for learning, they are vastly superior to flat multimedia. By combining tangible, embodied interaction with immediate, personalized feedback, robots prevent the usability failures that cause children to give up. Ultimately, designing for learner agency means building systems that demand active participation, destigmatize failure, and give children the immediate tools they need to succeed on their own.

Benefits of using a physical robot as an instructional aide, according to the Spanish study. (NotebookLM)
Multiple AI Roles Improve Learning
In my article 4 Metaphors for Working with AI, I explained that to maximize productivity and ensure a good user experience, we must deliberately shift our mental models. Depending on the task, we should treat the AI as an Intern (to delegate drudgery requiring supervision), a Coworker (to complete bigger projects in collaboration with humans), a Coach (for collaborative brainstorming), or a Tutor (to guide our learning).

My 4 metaphors for working with AI: intern, coworker, teacher, coach. (NotebookLM)
But what if choosing just one metaphor at a time is a fundamental UX failure?
Up to now, the tech industry has trapped us in a strictly monogamous relationship with our AI. You open a chat window, and it is just you and a single machine. This solitary paradigm is artificially constraining. Human learning and creativity have never been solitary; they are inherently social, shaped by debates, diverse perspectives, and the friction of observing others’ mistakes.

Human learning is often social, so relying on a solitary AI may not be the best way to provide individualized instruction. The multi-AI approach, explored in this paper, seemed to work better. (NotebookLM)
A new University of Toronto paper, “Beyond the AI Tutor: Social Learning with LLM Agents,” suggests that the next UX leap may come from combining AI roles rather than choosing one. Across two controlled studies (one on SAT-level math and one on essay writing), the researchers found gains that a single-agent setup could not fully match.

Two new studies (presented in a single paper) show the benefits of combining two different AI metaphors in educational applications. (NotebookLM)
Study 1: Math and the Power of Flawed Peers
In the first experiment, 315 participants tackled complex, SAT-level math problems. The researchers divided them into four conditions:
Control: No AI.
AI Peers Only: Users interacted with two AI “students” attempting to solve the problem alongside them.
AI Tutor Only: Users interacted with a single authoritative AI guide.
Tutor + Peers: A multi-agent setup where the user interacted with both the AI guide and the AI peers.
Here is the brilliant UX twist: the AI “peers” were deliberately designed to be flawed. One made arithmetic mistakes; the other made conceptual errors.
If you believe the goal of AI is simply to spit out correct answers, this sounds like terrible design. Why introduce an incorrect AI? Because of how the human mind works. When learners see an AI peer make a mistake, it triggers schema conflict. Users must actively evaluate the reasoning rather than passively accepting an authoritative answer. (All 3 AIs were the same underlying model, GPT-5.2, but with different prompts, causing them to act differently toward the students. Thus, the difference was not in the level of AI intelligence, but in their instructional role.)

Convergent thinking can be too straight and narrow, especially for creative tasks, such as those used in Study 2. The different AI models used in Study 1 triggered a schema conflict in the students, which pushed them into independent, divergent thinking. (NotebookLM)
The results speak for themselves. On the final, unassisted test, accuracy increased monotonically:
Control: 42%
Peers Only: 48%
Tutor Only: 59%
Tutor + Peers: 65%
Combining the Tutor metaphor with the Peer metaphor (which acts much like an error-prone Intern) creates a learning environment that neither can achieve alone. The Tutor provides essential guardrails, while the flawed Peers force the user into active critical thinking. Furthermore, users in the “Peers Only” condition reported higher confidence than their performance actually warranted. This proves that the Peer metaphor is incredibly powerful for boosting user self-efficacy and motivation.
Study 2: Breaking “AI Sludge” in Writing
The second experiment tackled a divergent task: writing creative and argumentative essays. The 247 participants were assigned to write either without AI, with a single AI (acting as a standard Assistant), or with a Duo of two AIs operating simultaneously.
In the Duo condition, the two AIs acted as specialized Sparring Partners. (What I called “Coaches” in my model of the 4 metaphors for using AI.) For example, in creative writing, one AI focused strictly on imagination and voice, while the other focused on narrative structure and pacing.
We already know that single-AI assistance improves average writing quality, but it comes with a massive downside: homogenization. Leaning too heavily on a single AI leads to a reversion to the mean, which some call “AI sludge,” where everybody’s ideas start to sound identical. The study confirmed this. Using a single AI increased the similarity of ideas across users significantly compared to the control group.

Simple use of AI to help people write does improve writing quality but makes everybody sound the same, erasing their unique voice. The use of multiple AIs in this study overcame this negative effect. (NotebookLM)
But the Duo condition neutralized this homogenization. Users writing with two specialized AI Sparring Partners produced essays that were just as high-quality as the single-AI group, but with the same idea-level diversity as unassisted human writers.
When we rely on a single AI, we are funneled into its preferred representational space. By employing multiple specialized AI Sparring Partners, we break the consensus. The user regains their intellectual authorship because they must synthesize competing feedback rather than mindlessly accepting a single AI’s rewrite.
Takeaway Lessons: The Multi-Metaphor Future
For UX designers, the lesson is clear: stop treating AI as a single chatbot and start designing orchestrated multi-agent ecosystems. The most promising systems will combine metaphors on purpose: for example, an AI Intern to prepare the material, an AI Coach to challenge the strategy, and an AI Tutor to catch methodological errors.
Here are the core interaction design takeaways:
1. Productive Friction is a Feature
In learning and creative ideation, a completely frictionless experience leads to passive consumption. Introducing multiple, sometimes flawed or conflicting AI perspectives adds the desirable difficulty necessary for deep cognitive engagement. AI that makes legible mistakes builds human confidence.
This represents a radical reversal of the last 25 years of UX design. Since my good friend Steve Krug’s famous “Don’t Make Me Think” paradigm, the holy grail has been to eliminate friction entirely. But when the machine does all the thinking, a frictionless UI breeds human complacency. Designing for AI means we must intentionally inject some difficulty back into the interface to keep the human in the loop and cognitively engaged.

A small amount of friction can be good because it cultivates cognitive engagement. Too much friction, and users become discouraged and abandon the project. (NotebookLM)
2. Role Specialization Beats General Intelligence
A single, general-purpose AI tends to regress to the mean. Splitting the AI into distinct personas with narrow, specialized roles prevents homogenization and preserves human authorship.
3. The Usability Tax: Cognitive Load
The study revealed one critical warning for usability: cognitive load. In both experiments, participants interacting with multiple agents reported feeling substantially more overwhelmed and confused. Staring at a chat interface where multiple different AIs are talking to each other and to you is mentally taxing.

Intermixing multiple AI models and metaphors within a single long, scrolling text field was too much of a usability burden. We need a new UX for multi-metaphor AI. (NotebookLM)
This is our next great usability challenge. We cannot simply dump three chatbots into a scrolling text window and expect a good user experience. We must design interaction architectures that make multi-source AI feedback legible and manageable. (The project in my previous news item that located some AI agency with a physically distinct robot may be overkill, but it does provide inspiration.)
The likely design answer is staged collaboration, not simultaneous chatter. Let one agent propose, a second critique, and a third synthesize only when needed. The benefit of multiple agents comes from contrast; the usability will depend on when and how that contrast is revealed.
The era of the one-on-one AI chat is ending. The future belongs to rich, multi-agent collaborations. It’s time for our interfaces to catch up to the social reality of how humans actually learn and work.

A variety of specialized AI tools embodying different metaphors for interacting with AI are likely better than a single generic AI for many purposes. Determining these purposes will require more research. (NotebookLM)
Meta Releases Frontier AI Model
Meta (Facebook) has released a new AI model called Muse Spark. Its scores on a variety of benchmarks are strong, placing it in the ranks of frontier AI models, even if it doesn’t quite equal the most capable models like GPT 5.4 Pro and Gemini 3.1 Pro.
I put Muse to two tests: summarizing the research paper on the use of multiple AI metaphors in education (which I covered in the previous news item), and providing feedback on my article “Redesigning Workflows for AI.” In both cases, Muse provided useful answers, but not as strong as those I received from Gemini. Its writing suffers from the standard AI tendency to use a “it’s not this, it’s that” schema, but to its credit, Muse uses fewer em-dashes than the other AI models — though maybe it goes too far in banning this useful grammar element.
The harder task was providing insightful suggestions for my article. I incorporated one minor idea, but rejected most of Muse’s proposed revisions. It mainly offered redundant elaborations on points I had already covered elsewhere in the draft.
Meta’s announcement of the new model contains two interesting pieces of information: they confirmed that the AI scaling laws continue to hold for all three forms of AI scaling: pre-training, reinforcement learning, and test-time reasoning. For all three, results improved every time they added more compute. (I had expected as much, but always nice to see the AI scaling laws confirmed by an independent lab.) As a result, Meta plans to use even more compute for its next-generation model.

The AI scaling laws work. AI gets more intelligent by the year. However, as discussed in the previous news item, the user interfaces to AI haven’t kept up. Meta is at least beginning to move beyond pure single-box chat. Muse Spark ships inside Meta AI with mode switching, multimodal perception, and parallel subagents. That is still not a mature new UX, but it is more than a 2022 blank text box. The real question is whether Meta can turn those capabilities into a workflow users can actually understand. (NotebookLM)
However, in addition to confirming scaling, Meta also confirmed the concept of algorithmic efficiency: More compute helps, but using the same compute more efficiently also helps, and as we gain more experience implementing high-end AI models, we learn how to do so better. In Meta’s case, they reached the capability level of its previous model, Llama 4 Maverick, with over an order of magnitude less compute than when they trained that model a year ago. This is very good news, given the compute famine all the big AI labs (even Google) are suffering these days, as users’ demand for AI grows faster than the labs can build new data centers. (And definitely faster than enough electricity generation capacity can be built outside China.)
For UX professionals, algorithmic efficiency is the key to the next hardware paradigm. As models require an order of magnitude less compute to achieve frontier-level reasoning, we move closer to running high-end AI locally on edge devices. This bypasses cloud latency entirely, unlocking fluid, real-time AI user experiences, such as the responsive physical preschool robots discussed earlier.
Meta has been lying low for a year, since the disappointing launch of Llama 4 in April 2025. But never count out Mark Zuckerberg! A year later, he and his merry band of AI superintelligence researchers are back with a strong model and aggressive plans to scale compute to get even better soon. I am very happy to see renewed competition in the AI space, since all the other contenders have their various downsides. xAI’s Grok is also scaling very fast, so we can hope to go from 3 to 5 non-Chinese AI frontier models later this year. (Plus whatever we get from China, though they have been better in video and images than with general-purpose AI. It’ll probably take a few more years before Huawei’s AI hardware equals or surpasses NVIDIA, and until then, the very best AI models will likely remain American.)

Meta releases a new high-end AI model, getting back into the game. (Nano Banana 2)
AI Helps Farmers Manage Cows
Halter is a New Zealand agri‑tech company that makes solar‑powered, GPS‑enabled smart collars for cattle. Its “Cowgorithm” system uses machine learning on the collar sensor streams to model individual animal behavior. This AI system has been trained on extremely large corpora of livestock behavior data (on the order of billions of hours) to learn patterns in cows’ movement, grazing, and responses to cues.

Halter’s signature feature is “virtual fencing,” where the farmer draws a boundary on the accompanying mobile app, and the cows will stay within that area. Halter claims to have implemented 568,000 km (353,000 miles) of virtual fencing, which would have cost about $7 billion to build as physical fences. (Nano Banana Pro)
By learning how each cow responds, the Cowgorithm adapts cue timing and intensity, and helps optimize grazing patterns (e.g., residual grass height, rotation timing) to increase pasture utilization and weight gain. The system also functions as a health and fertility monitor, using movement and activity signatures to detect illness or heat events earlier than would typically be visible to farmers.
Halter just raised a large Series E funding round of $220M at a $2B valuation to scale the Cowgorithm platform globally.
From a UX perspective, Halter is a masterclass in “Zero UI.” The ultimate end user (the cow) has no screen, no buttons, and no graphical interface. Instead, the UI consists entirely of ambient sensory cues (sound and vibration) that gently shape behavior. It is a powerful reminder that as AI grows more sophisticated, the most effective interfaces might completely bypass the screen.
There you have it: AI for cows = $2 billion.

Who says that farming is a low-tech business? AI helps cows. (Nano Banana Pro)
GPT 5.4 Has a Lower Time Horizon than Claude Opus 4.6
METR has released its latest measures of AI time horizons. This measures the largest tasks an AI model can complete with 50% accuracy, expressed as the number of hours it would take a human expert to perform them. (This is not the same as the time the AI uses: it’s almost always much faster than the human.)
GPT 5.4 (running at xhigh reasoning) scored a time horizon of 5 hours and 42 minutes. This is up from 3 hours and 23 minutes for GPT 5.0, and is a disappointingly low improvement. GPT 5.4 was released in March 2026, and by comparison, Claude Opus 4.6, which was released a month earlier, was measured with a time horizon of 11 hours and 59 minutes, or 110% better.

Claude utterly dominates GPT on the time horizon metric. (Nano Banana 2)
Let’s hope the rumored “Spud” model OpenAI is supposedly releasing soon does better.

Will “Spud” achieve liftoff and go further than Claude? Silicon Valley rumors are heavy that we shall see soon. (Nano Banana 2)
Venture Capital Investments Up Substantially
A16Z, which is the leading venture capital firm, has released statistics on VC funding worldwide. In the first quarter of 2026, VC investments reached $300 billion, a new record. For the last three years, all the quarters have hovered around $100 billion, so the increase is substantial.
80% of VC investments now go to AI companies, up from 55% a year ago. The opportunity is palpable.

AI now accounts for 80% of rapidly growing venture capital investments. (Nano Banana 2)
Another interesting statistic from the same article: Data center capacity is growing, but only at about 22% per year, which seems insufficient to keep pace with advancing AI capabilities. The compute famine will only get worse.
Right now, the world’s data center capacity is split roughly in thirds between AI training, AI inference, and pre-AI tasks (e.g., YouTube streaming). All three are expected to grow, but AI inference will grow the fastest, at 35% per year, according to a McKinsey prediction.

The compute famine will hit hard. (Nano Banana 2)
Conclusion: Escaping the Text Box
Looking across this week’s news items, a powerful unifying theme emerges: AI capabilities are scaling exponentially, but our user interfaces are stuck in the past.
As Meta’s newest release proves, the raw “brains” of frontier models are growing vastly more powerful and algorithmically efficient. Yet the tech industry continues to cram these super-intelligences into the exact same single-stream text-box UI we used in 2022. The intelligence of the machine is no longer our primary bottleneck; the UX is.
To unlock the true value of AI, we must escape the text box. The most effective interfaces of the future will abandon the solitary screen in favor of contextual, systemic designs. For preschoolers, the ideal interface isn't a screen at all, but a physically embodied robot that offers a hug when you fail. For complex knowledge work, the ideal interface isn’t an omniscient oracle, but an ecosystem of specialized, sometimes-flawed agents that foster productive friction. And for agricultural management, the ideal interface is entirely invisible, operating ambiently in the background as the cows focus on grazing, not the UI.
The era of conversational ping-pong with a single chatbot is ending. The future of AI UX is embodied, multi-layered, and environmental.

There are signs that AI may soon escape its confining text-box UI. (Nano Banana Pro)
