Blog

Robots & Pencils Launches “Rewired: The New AI Architecture of Higher Education” 

As the world’s top education innovators gather at ASU’s Agentic AI Summit and EDUCAUSE, Robots & Pencils unveils a bold blueprint for the intelligent university. 

Robots & Pencils, an Applied AI Engineering Partner that helps universities and enterprises modernize applications and increase the speed of productivity, today announced the launch of Rewired: The New Architecture of Higher Education. This three-part thought leadership series challenges universities to reinvent how they define, deliver, and prove learning in the age of AI. 

As AI reshapes every dimension of learning from admissions to advising, research to retention, Robots & Pencils offers a vision for what intelligent universities can become. 

Start reading Rewired: The New AI Architecture of Higher Education.  

Arriving as higher education leaders converge for the Agentic AI and the Student Experience Summit at Arizona State University and the EDUCAUSE Annual Conference, Rewired explores how institutions can move from digital transformation to institutional intelligence, building systems that learn, adapt, and evolve alongside their students. 

“The next era of higher education will be defined by who learns fastest,” said Kristina Gralak, Client Strategy Analyst at Robots & Pencils and author of the series. “Agentic AI is transforming what it means to be student-centered. The universities that win will rewire their infrastructure for intelligence, creating systems that personalize experiences, validate skills, and connect learning to lifelong opportunity.” 

The three essays within Rewired trace higher education’s most urgent frontiers: 

“Kristina’s series captures the intersection of vision and engineering,” said Jeff Kirk, Executive Vice President of Applied AI at Robots & Pencils. “Every institution seeks to enhance the student experience, yet few realize that progress begins with the invisible systems: the data, cloud, and AI engines that make intelligence possible. Rewired shows what it takes to connect strategy with reality.” 

From intelligent tutoring systems to AI-powered credential networks, Rewired outlines how forward-thinking universities can turn experimentation into institutional evolution. It is a call to action for higher education leaders to design for the lifelong learners of tomorrow and to embrace an AI-driven future where universities think, adapt, and evolve as intelligently as the students they serve.  

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing 

The Invisible Infrastructure That Determines Higher Education Success 

Part 3 of our series Rewired: The New AI Architecture of Higher Education

Part 1: The New AI Architecture of Higher Education | Part 2: How Higher Education Proves Value in the Skills Economy

You can have the perfect enrollment strategy. You can deliver credentials that employers both trust and understand. But none of it matters if your systems frustrate students at every turn. 

The State of Higher Education 2025 highlights how AI is already transforming administrative operations. Institutions are cutting admissions decision times from weeks to days. That efficiency gain matters, but it’s pointing at something bigger. The most transformative applications of AI in higher education will happen in the invisible systems that touch students every day and determine whether institutions can actually deliver on their promises of personalized pathways, skills verification, and career outcomes. 

The Invisible Systems that Determine Everything 

Think about what student-facing infrastructure should look like: registration that anticipates scheduling conflicts before they derail a semester, financial aid that explains packages in plain language and flags missing steps in real time, advising that surfaces degree progress at midnight without requiring an appointment, and career services that connect learning to opportunity throughout the educational journey rather than just senior year. 

Now consider what most students actually experience. Most universities operate on infrastructure built before students expected real-time information, before mobile-first design, and before APIs enabled systems to communicate seamlessly. Advising platforms can’t access degree audit tools. Financial aid offices require documentation already submitted during admissions because systems don’t share data. Registration workflows assume students know course prerequisites that aren’t clearly mapped anywhere accessible. 

This friction is the difference between serving traditional students adequately and serving diverse learners well. A 19-year-old living on campus might tolerate process-heavy systems because they have time to navigate them. A 35-year-old parent working full-time while taking evening classes cannot. 

When Systems Don’t Talk  

Here’s what disconnected systems look like in practice: A student registers for next semester’s courses. The registration system confirms enrollment, but the degree audit tool doesn’t update for 48 hours. The student panics, thinking they’ve registered wrong, and emails their advisor, who also can’t see the registration because their advising platform pulls data overnight. By the time systems sync, the student has already spent hours searching for answers that should have been instantly available. 

Or consider the transfer student navigating data silos. Transcript evaluation sits in one system. The academic advisor works in another. The degree audit reflects only current-institution courses. Financial aid can’t see transfer credits until manually entered elsewhere. Each office operates with partial information, and the student becomes the integration layer, having to shuttle information between departments, resubmit documentation, and try to piece together what no system can provide. 

These challenges define daily operations for institutions managing disconnected systems, and they’re a key reason students choose to leave. Academic quality and affordability still matter, but experience now defines whether education feels achievable or exhausting.  

Building Systems that Create Advantage 

Better experiences lead to stronger retention, which enables sustained enrollment, which funds continued improvement, which attracts students who see a responsive institution. This cycle creates compounding advantages. 

As the State of Higher Education 2025 report notes, students want “an integrated and seamless experience on campus like they have with Amazon 1-Click, Netflix preferences, and Instagram likes.” The goal is not consumerization, but rather alignment with the baseline expectations of how digital systems should function in 2025. 

The institutions that invest in operational intelligence now will differentiate themselves in ways competitors can’t quickly replicate. Competitors can replicate program offerings, but integrated systems that learn from student behavior and adapt over time create advantages that take years to build. 

From Disconnected Systems to Institutional Data Intelligence  

The challenge institutions face goes beyond isolated student-facing systems. It’s a fundamental question about how data flows across the entire institution and whether that data can inform better decision-making at every level. 

The EDUCAUSE 2025 Horizon Report: Data and Analytics Edition identifies the shift “toward unified data models and integrated data ecosystems” as critical for institutional effectiveness. The report notes significant barriers remain: “slow adoption of common data standards, lack of in-house expertise, tight budgets, and concerns about privacy and security when connecting different data sources.” 

But institutions that overcome these barriers will build systems that “respond more quickly, spot and support at-risk students earlier, and evaluate programs more effectively as a whole.” This is what infrastructure modernization actually means: not just upgrading individual systems, but creating the connective tissue that enables institutional learning. 

Imagine infrastructure that functions like a learning organism. Student outcomes from last semester inform course scheduling for next semester. Advising patterns surface which interventions work for specific populations. Registration data reveals course conflicts before hundreds encounter them. Each cycle generates insights that make the next more effective. 

The EDUCAUSE report warns that “rapid AI adoption is introducing new risks” but is equally clear about the path forward: institutions must “develop clear policies and build cross-functional governance structures that include voices from IT, academic affairs, compliance, and student services.” This is the work of infrastructure modernization: integrating intelligence across systems while maintaining human oversight, transparency, and accountability. 

The Infrastructure Challenge for Lifelong Learners  

Traditional systems assume continuous enrollment, students who enter as freshmen and graduate four years later. These assumptions are embedded in everything from registration workflows to student information systems to advising models. 

Serving lifelong learners requires fundamentally different infrastructure. Systems need to remember students across years of non-enrollment. Credential systems must stack learning experiences accumulated across time and institutions. Registration workflows need to accommodate students taking one course while working full-time. 

The platform approach outlined in the first article in this series now defines the path forward for institutions ready to scale lifelong learning. Without unified infrastructure, institutions will continue to relegate adult learners to separate systems that feel like second-class experiences. The institutions that build infrastructure for lifelong learning will turn the enrollment cliff and broader demographic changes into drivers of innovation and competitive advantage.  

The Infrastructure Behind Skills-Based Credentials 

The second article of our series outlined the opportunity in skills-based credentials. But credential transformation depends entirely on infrastructure most institutions don’t yet have. Making educational outcomes relevant to employers requires systems that track competency development across courses and verify skill demonstration through assessed work. These systems must translate learning outcomes into employer language and enable dynamic credential pathways as employment demands evolve. 

Right now, course outcomes exist in syllabi. Assessment data sits in learning management systems. Career outcomes are tracked separately. None of these systems talk to each other, and none can generate the comprehensive, verifiable credentials students need. Building this infrastructure requires more than technical expertise. It depends on registrars, academic affairs, career services, IT, and institutional research working from unified data models. 

Where to Start  

Transformation gains traction through precise, coordinated initiatives that evolve into integrated systems over time. 

Start with a data integration pilot in one high-friction area, such as transfer credit evaluation, financial aid processing, or advising workflows. Build the connections that eliminate manual handoffs. Use that pilot to establish governance patterns and technical standards that can scale. 

Map the student journey to identify friction points. Follow students through registration, financial aid, advising, and enrollment. Document every place they encounter disconnected information or redundant data entry. These pain points become your integration roadmap. 

Most importantly, build with student-facing impact in mind. Every integration should make something tangibly better, such as faster information, clearer guidance, reduced manual work, or more responsive service. Infrastructure projects that deliver only backend efficiencies will struggle to sustain commitment. Projects that demonstrably improve student experiences will build momentum for continued transformation. 

The Infrastructure Imperative 

This series has outlined a clear progression: who to serve (lifelong learners at all career stages), how to prove value (skills-based credentials and AI-powered career connection), and what makes it possible (operational infrastructure that executes strategy at scale). 

The institutions that lead will approach transformation as an interconnected system. Success with diverse learners comes from modern infrastructure, and lasting credential innovation emerges from systems built to verify skills throughout learners’ lives. 

Infrastructure serves as a core differentiator, converting strategic vision into operational strength. It’s the difference between institutions that adapt to demographic change and those that watch enrollment decline while running on systems built for students who no longer represent their future. 

The work is demanding. It requires sustained commitment, cross-functional collaboration, and investment in capabilities that many institutions have historically under-resourced. Continuing to operate on disconnected systems while competitors advance with unified platforms limits growth and long-term resilience. 

Transformation begins with the essential work of modernizing systems, integrating data, and building platforms that serve lifelong learners. That’s where real differentiation happens, and that’s what determines institutional success in the decade ahead. 

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing. 


Key Takeaways 


FAQs 

Why does infrastructure modernization matter for student success? 
Modern systems remove friction in core experiences such as registration, advising, and financial aid. When data flows seamlessly, students receive faster responses, clearer guidance, and more personalized support. 

What does operational intelligence mean for higher education? 
Operational intelligence describes systems that automate processes and learn from them. When institutions integrate data across departments, they gain the ability to anticipate student needs, identify risks earlier, and continuously improve operations. 

How does infrastructure connect to skills-based credentials? 
Skills-based learning depends on interoperable data. Institutions need infrastructure that connects course outcomes, assessments, and verified competencies, creating credentials that employers understand and trust. 

Where should institutions start with modernization? 
Start with a pilot that addresses a visible student challenge such as transfer credit evaluation or financial aid delays. Use that project to establish governance patterns, integration standards, and measurable improvements that demonstrate value across the institution. 

What defines a future-ready institution? 
A future-ready institution treats infrastructure as a living system that learns and adapts. It measures success by student outcomes, institutional agility, and the ability to serve learners continuously throughout their careers.  

How Higher Education Proves Value in the Skills Economy 

Part 2 of our series Rewired: The New AI Architecture of Higher Education

Part 1: The New AI Architecture of Higher Education | Part 3: The Invisible Infrastructure That Determines Higher Education Success

Higher education faces a trust problem. College-going rates have dropped from 70% to 62% since 2016. When you ask students why, two themes dominate: affordability concerns and uncertainty about return on investment. Universities have responded by defending the value of degrees with more vigor and better marketing, but this strategy misunderstands what’s shifting. Students still want to learn, but they also want to know whether what they are learning matters to employers and how it connects to real employment opportunities. Degrees used to provide that assurance implicitly. Employers valued degrees, so students trusted their worth. But as employers shift toward skills-based hiring, that implicit value is eroding. Students now need explicit proof that their education translates into capabilities employers actually want. 

Meanwhile, employers are adopting skills-based hiring at accelerating rates. They care less about where you went to school and more about what you can do. This creates an opportunity for institutions willing to reimagine credentials entirely and use AI to connect learning to career outcomes in real time. 

The Credential Revolution  

The degree is evolving to become modular, transparent, and aligned to real-world capabilities. Today’s students demand degree programs where industry-aligned certifications are embedded throughout, not tacked on at the end. They want digital credentials that verify specific competencies in formats employers can instantly understand. They need evidence of skills activated, not just courses completed. 

This requires solving a problem most institutions are only beginning to articulate: making educational outcomes relevant and legible to employers. Right now, a degree signals institutional affiliation and field of study, but nothing more. Hiring managers need a clear view into whether a graduate can analyze datasets, lead cross-functional teams, or communicate complex ideas to non-technical audiences. 

Institutions know these things. Course learning outcomes exist. Assessment data sits in learning management systems. Capstone projects demonstrate applied competencies. But this evidence is trapped in internal systems, inaccessible to anyone outside the institution. Students leave with a diploma that says what they studied, not what they can do. 

Consider what this looks like from a student’s perspective. A sociology major graduates knowing they can conduct qualitative research, analyze social patterns, manage community-based projects, and synthesize complex information for diverse audiences. But their diploma says “Bachelor of Arts in Sociology.” Their transcript lists course titles and grades. They spend months after graduation trying to articulate their actual capabilities in resumes and interviews because their institution never made those skills visible or verifiable to employers. 

Institutions that build interoperable credential systems with digital credentials that verify specific competencies, stackable certifications embedded throughout degree programs, and verified skill demonstrations will define a new model for learning. They will become the trusted translators between education and employment. They will award degrees and validate capabilities that matter, serving students throughout their careers as they return for new credentials and competencies. 

Some institutions are already moving in this direction. Computer science programs embed AWS or Google Cloud certifications alongside degree requirements. Business schools offer IBM badges and Six Sigma certifications as integrated components of coursework. Universities partner with platforms like Credly and Canvas Credentials to issue competency-based digital badges that students can share directly with employers. 

Arizona State University is taking this even further with its Trusted Learner Network (TLN), building infrastructure for distributed ledger-based, verifiable credentials that can follow students throughout their lifelong learning journey—not just credentials from ASU, but a vision of interoperable credential exchange across institutions, employers, and learning providers. This is what credential infrastructure looks like when institutions think beyond single transactions to lifelong relationships. 

But most institutions are still treating credentials as isolated experiments rather than core infrastructure. A certificate program here, a digital badge pilot there, maybe some industry partnerships in high-demand fields. What’s missing is the institutional commitment to make skills verification foundational to how students progress through their education and how alumni demonstrate their capabilities throughout their careers. 

This transforms the institutional relationship from a four-year transaction to a lifelong partnership. Alumni leave with more than a degree, they maintain a credential relationship with the institution, returning for micro-credentials, professional certifications, and competency validations as their careers evolve. This is the infrastructure that makes lifelong learning operationally viable, a unified system where a 22-year-old recent graduate and a 45-year-old mid-career professional engage with the same credential ecosystem. 

Where AI Readiness Becomes Competitive Advantage 

Recent research surfaces a critical gap. Students are already using AI tools extensively in their academic work for research, writing, and problem-solving. Meanwhile, fewer than 20% of faculty feel confident teaching with or about AI. Most institutions are treating this as a training problem: a few workshops on prompt engineering, some guidance on academic integrity, maybe a pilot program or two. 

That response entirely misses the opportunity. The institutions that will differentiate themselves are doing more than training faculty on AI tools. They’re integrating AI into how students learn, how advisors guide, and how the institution operates. The difference is between treating AI as a tool to learn about versus treating it as the intelligence layer that makes every system more responsive. 

Consider what this looks like operationally. Right now, when a student struggles in a course, they might get flagged for early intervention. For example, they may receive an automated email suggesting the tutoring center, or maybe an advisor reaches out to recommend better study habits or office hours. That’s reactive and generic. 

An AI-informed institution operates differently. The system recognizes the struggle in real-time and surfaces personalized tutoring resources at the moment intervention is needed. These are not generic study tips, but alternative approaches to the material aligned with how that student learns best. When the student registers for next semester, the system adjusts course recommendations to sequence their learning more effectively while still maintaining progress toward their degree. The advisor still has the conversation, but now they’re working with intelligence about what approaches are actually effective for this student. 

The difference is more than better outcomes. It’s operational efficiency at scale. An advisor managing 400 students can’t manually track how each student learns best, which interventions are working, and what course sequences will set them up for success. But an AI-informed system can surface exactly which students need proactive outreach, what specific guidance would be most relevant, and how to sequence their learning path most effectively. The advisor’s time shifts from administrative triage to high-value relationship building. 

The challenge is organizational. It requires integrating intelligence across disconnected systems like advising platforms, learning management systems, career services tools, and student information systems. It requires training staff to use AI-informed insights without replacing their professional judgment. And it necessitates building workflows where AI augments human interaction rather than creating another dashboard no one checks. 

I’ve watched institutions pilot AI capabilities that never scale beyond the pilot. A chatbot answers basic questions but cannot access student records. An early alert system generates so many flags that advisors cannot possibly respond to them all, leading them to ignore the alerts entirely. An AI-powered degree planning tool recommends optimal course sequences but operates in a separate system, disconnected from the advising and registration workflows students actually use. 

The competitive advantage comes from embedding AI into how every system serves students. That requires treating AI integration as an operational transformation, not a technology deployment. And it requires infrastructure built to make intelligence actionable, not just theoretical. 

Proving Value Through Skills and Intelligence 

The institutions that solve the ROI crisis will be the ones that make learning outcomes transparent and connected to employment. They’ll build credential systems that translate education into employer-legible skills and use AI to connect students with career pathways from day one, not just senior year. Industry certifications will be embedded throughout their degree programs rather than treating them as add-ons. 

This transformation requires institutions to fundamentally rethink how they measure success, from degrees awarded to skills activated, from course completion to demonstrated capability, and from graduation metrics to career readiness at every stage. It requires building credential systems that prove competency, not just attendance, and treating career preparation as foundational to education, not a separate service bolted on at the end. 

The institutions leading this work will be the ones that understand proving value is no longer a marketing problem, but an infrastructure problem. You can’t demonstrate skills if you don’t have systems to verify and credential them. You can’t connect learning to careers if your academic systems don’t talk to your career services platforms. You can’t serve students throughout their lifelong learning journey if your infrastructure is designed exclusively for traditional four-year degree seekers. 

The next article in this series examines the operational infrastructure that makes all of this possible. The invisible systems that determine whether students persist or leave, whether institutions can deliver on these promises at scale, and whether the transformation from traditional education to intelligent learning ecosystems actually works in practice. 

Read part 3 of our Rewired series, The Invisible Infrastructure That Determines Higher Education Success.  If you missed our first article in this series, check out The New AI Architecture of Higher Education.  

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing. 


Key Takeaways 


FAQs 

Why do credentials need to change when degrees still matter to employers? 

Employers increasingly hire based on demonstrated skills rather than degree prestige. They need to understand what a graduate can actually do, not just where they studied. Verifiable digital credentials that translate coursework into specific competencies help employers make better decisions and help graduates prove their capabilities clearly. 

What makes AI fluency different from AI adoption in higher education? 

AI adoption means using tools like ChatGPT or administrative automation. AI fluency means weaving intelligent systems into how students learn, advisors guide, career services operate, and institutions run. It’s the difference between adding technology and reimagining how education works when intelligence can personalize, predict, and adapt at scale. 

How do institutions make educational data legible to employers? 

Through interoperable credential systems that translate courses into demonstrated competencies. Instead of transcripts showing only course titles and grades, modern credentials verify specific skills like data analysis, cross-functional leadership, or technical communication. Digital badges and stackable certifications create a common language between education and employment. 

What does AI-powered career services look like in practice? 

AI-powered career services track labor market trends in real time, connect coursework to emerging job opportunities, help students build competency portfolios throughout their education, surface relevant alumni mentors based on career interests, and personalize guidance based on individual strengths and market demand. The technology enables career planning from freshman year instead of senior year scrambling. 

The New AI Architecture of Higher Education 

Part 1 of our series Rewired: The New AI Architecture of Higher Education 

Part 2: How Higher Education Proves Value in the Skills Economy | Part 3: The Invisible Infrastructure That Determines Higher Education Success

The State of Higher Education 2025 report confirms what institutions have been tracking for years: the enrollment cliff is here. Peak high school enrollment arrived with the Class of 2025, and from now through 2041, the number of graduates will decline by 13%

Institutions knew this was coming. The story they aren’t ready to hear is what it requires: not better retention strategies or more aggressive recruiting, but fundamental reinvention of who they serve and how they serve them. Most institutions see the enrollment cliff as a crisis to be managed. I see it as the catalyst for higher education’s most exciting transformation in decades. 

The report captures a sector at an inflection point. Demographic shifts, AI advancement, and evolving student expectations are converging to create the conditions for fundamental reinvention. The barrier isn’t awareness or willingness, it’s execution. Institutions move slowly. Their systems are disconnected. Their infrastructure is rigid, designed for a traditional student population that no longer represents their future. 

The transformation requires work most institutions have barely started: reimagining who their students are, modernizing how systems serve them, and redefining what counts as proof of learning. 

The Student You’re Not Designing For 

I’ve sat in countless conversations with enrollment and student success teams. The pattern is always the same: everyone is focused on meeting this term’s targets, fixing immediate friction points, optimizing for the students already enrolled. There’s barely time to think about next month, let alone reimagine who you could serve five years from now. 

When leaders do push for serving non-traditional populations, such as adult learners, part-time students, and those with significant transfer credits, the instinct is often to squeeze these students into existing systems. Use the same registration workflows. Same advising model. Same assumptions about what ‘student success’ means. The result? You’ve diversified your enrollment numbers but not your infrastructure. 

This is the trap that keeps institutions focused on a shrinking market. As the traditional undergraduate population declines, a massive population of learners remains underserved: 

These learners represent the future majority of higher education, and they bring fundamentally different expectations. They need to learn while working full-time, while managing families, while living far from campus. They require flexibility as a condition of participation. And they expect university systems to work like every other digital experience in their lives: responsive, intelligent, and adaptive. 

Online-only enrollment has already surpassed 5 million students, and online master’s degrees now exceed in-person programs. The pandemic validated what these learners already knew: flexible learning is the only viable path for students juggling multiple commitments. What institutions treated as emergency response in 2020 has become permanent expectation in 2025. 

Being “student-centric” requires building systems with institutional memory, platforms that recognize a returning student, pre-populate forms with known information, and give advisors visibility into a student’s full academic journey. The technology to do this exists in every other sector. Higher education’s challenge is the complexity of dismantling deeply embedded silos while keeping operations running. 

The institutions that will thrive aren’t the ones fighting to preserve systems designed for traditional learners. They’re the ones willing to do the hard work of building platforms that serve a 19-year-old college freshman and a 45-year-old professional returning for a certification with equal intelligence, systems that recognize both learners, understand their different needs, and adapt accordingly. 

The Platform Play Higher Ed Hasn’t Made 

Online education has proven its viability. The next frontier is integration. Online and on-campus work best as different modes within a unified learning platform that follows students wherever they are in life. 

Right now, most universities treat online programs as separate business units with distinct registration systems, student services, and cultures. I’ve seen this friction play out in painful ways. A junior takes a summer internship out of state and wants to stay on track by taking one online course. Suddenly they’re navigating a completely different registration portal, calling a separate help desk, and dealing with advisors who can’t see their on-campus transcript.  

Or consider the undergraduate alum applying to an online master’s program at the same institution. They’re re-entering all the information the university already has, speaking with advisors who have no visibility into their four years of history. Same institution, but the student experiences it as if starting from zero. 

The friction is real, and it’s expensive. Every moment of confusion, every duplicated form, every advisor who doesn’t have complete context is a moment where the student considers whether continuing is worth the hassle. 

The opportunity sits in building modular, always-on learning environments where micro-credentials, degrees, and continuous upskilling integrate seamlessly. Picture this: A student completes a graduate certificate in data analytics. Three years later, they return for an MBA. The certificate credits automatically apply, their prior work is visible to new faculty, and the advising team can build on previous conversations rather than starting fresh. The student doesn’t have to re-explain themselves. They’re simply continuing a relationship the institution remembers. 

This isn’t hypothetical. Some institutions are building this now, and it’s becoming their competitive advantage. 

This vision requires treating education as a lifelong relationship rather than a four-year transaction. It means building systems that remember students, adapt to their changing needs, and make re-entry feel seamless rather than starting from scratch. The institutions that crack this will turn alumni into lifelong learners and turn education into something that compounds in value over time. 

This fundamentally shifts how institutions think about their role. Instead of a four-year engagement, you’re building relationships that span careers. Alumni who return for stackable credentials every few years represent the best kind of growth: learners you’ve already served well, who understand how your programs work, and who are advocating for your institution with their employers. This is how institutions build enrollment resilience in a shifting demographic landscape. 

What This Looks Like in Practice 

Transformation at this scale relies on strategic planning and attention to detail. It happens when your data architecture can track a learner across programs, modalities, and decades. When your student information system doesn’t silo traditional and non-traditional students into separate workflows and data structures. When your advising model scales to support someone taking one course just as effectively as someone enrolled full-time. 

The institutions getting this right are treating it as a technology transformation, not just a strategy refresh. They’re building unified data layers, modernizing APIs, and creating seamless user experiences. They’re measuring success by how little friction a learner experiences, not just by enrollment and retention numbers. 

Building the Foundation for What’s Next 

The universities that thrive over the next decade will be the ones that expand their definition of students to include learners at every career stage. They’ll create unified platforms where online and on-campus blend seamlessly, building experiences that serve diverse populations with equal care. 

Transformation happens in the essential work of modernizing systems, integrating data, and building platforms for lifelong learning. It happens when institutions shift their focus from what they’ve always done to designing for who they could serve. 

The institutions leading this work will be the ones that respond to the enrollment cliff by expanding who they serve. The ones that understand serving lifelong learners requires purpose-built infrastructure. The ones ready to measure success by skills activated rather than degrees awarded. 

The opportunity is clear: institutions that expand their definition of ‘student’ and build unified platforms for lifelong learning will own the next decade. But expanding who you serve only matters if learners believe your programs are worth their investment. In the next article, we’ll explore how institutions prove value in a skills economy—how they make learning outcomes transparent, credentials employer-legible, and career pathways visible from day one. 

Read part 2 of our Rewired series, How Higher Education Proves Value in the Skills Economy.

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice. We’d love to be a partner in that journey. Request an AI briefing.  


Key Takeaways 


FAQs 

How can universities grow enrollment during the demographic cliff? 

Growth comes from expanding who you define as a student. First-time adult learners, students with transfer credits, professionals seeking micro-credentials, and alumni returning for reskilling represent massive underserved populations. Institutions that build systems serving these learners as well as traditional undergraduates will find new revenue streams throughout the demographic transition. 

How do institutions serve traditional students and lifelong learners simultaneously? 

By building unified platforms where different learner types access personalized experiences through the same underlying systems. An 18-year-old residential student and a 40-year-old professional seeking a certificate have different needs, but both benefit from intelligent advising, clear pathways, and responsive operations. The technology should adapt to the learner, not force the learner to adapt to rigid categories. 

What does a unified learning platform actually include? 

A unified platform integrates registration, advising, credential tracking, and student services across all learning modes. It remembers student history regardless of how long they’ve been away, allows seamless transitions between degree programs and micro-credentials, and personalizes communication and support based on individual circumstances. The goal is making re-entry as natural as initial enrollment. 

Why is lifelong learning more valuable than traditional four-year models? 

Lifelong learning creates recurring revenue streams and deeper alumni relationships. Students who return multiple times throughout their careers generate sustained tuition revenue while building stronger institutional loyalty. Education becomes a compounding relationship rather than a single transaction, increasing lifetime value per student. 

How I Learned to Stop Worrying and Love AI Code: A Designer’s Journey 

How Designers Are Using AI Code Tools: From Figma to Functional Prototypes 


The team Zoom call felt like an intervention. “Just try it,” they said. “Everyone open Claude Code.” My palms were sweating. Twenty years of visual design, and I’d only clumsily played around with code. I was the furthest thing from a developer you could imagine. 

Two hours later, I couldn’t stop. I’d built three working prototypes. My ideas, the ones that lived and died in Figma for years, were suddenly real. Interactive. Alive. 

This is the story of how I went from code-phobic to code-addicted in a single afternoon. And why every designer reading this is about to follow the same path. 

The Designer–Developer Divide 

For decades, we’ve accepted a fundamental lie: Designers design, developers develop. The gap between these worlds felt like a chasm requiring years of computer science education to cross. HTML, CSS, JavaScript were foreign languages spoken in basement servers and terminal windows. 

I believed this myth completely. My job was making things beautiful. Someone else’s job was making them work. This division of labor felt natural, inevitable, and even efficient. Why would I learn to code when developers already did it so well? 

That myth cost me every idea I couldn’t prototype myself, every interaction I couldn’t test, and every vision that got lost in translation. Twenty years of creative constipation, waiting for someone else to birth my ideas. 

Five Minutes to AI-Powered Prototyping 

“Open your terminal,” they said. Haha. I’d only ever really seen it used in The Matrix. The black window appeared. The cursor blinked in judgment. Type ‘claude’ and tell it what you want to build. 

My first prompt was embarrassingly simple: “Make me a color palette generator.” I expected nothing. Error messages, maybe. Definitely not working code. 

But there it was. A functioning app. My app. Built with my words. 

The next prompt came faster: “Add a feature that saves palettes locally.” Done. “Make the colors animate when they change.” Done. Each success made me braver. Each response made me hungrier. 

By the end of that call, I wasn’t just using AI to code. I was thinking in code. The barrier I’d spent two decades accepting had evaporated in minutes. 

The New Addiction: Vibe Coding 

They call it “vibe coding,” this conversational dance with AI. You describe what you want. The AI builds it. You refine. It rebuilds. No syntax to memorize. No documentation to parse. Just pure creative expression flowing directly into functional reality. 

I became obsessed. That first night, I built seven prototypes. Not because anyone asked. Not because I needed them. Because I could. Every design idea I’d shelved, every interaction I’d dreamed about was suddenly possible. 

The feeling was intoxicating. After years of creating static mockups, watching my designs move and respond felt like gaining a superpower. Click this button, trigger that animation. Hover here, reveal that detail. My aesthetic decisions instantly became experiential. 

When Designers Start Coding 

Something profound happens when the person with design taste controls implementation. The endless back-and-forth disappears. The “that’s not quite what I meant” conversations vanish. The design is the product is the code. 

UXPin’s research shows designers can now “generate fully functional components with just a few inputs.” But that clinical description misses the emotional reality. It’s not about generating components. It’s about giving creative vision direct access to digital reality. 

I started noticing details I’d never considered before. The precise timing of transitions. The subtle response to user actions. The difference between functional and delightful. When you control every aspect of implementation, you start designing differently. You start designing more ambitiously, more precisely, and with more courage.  

AI Code Tools That Make It Possible 

The technology enabling this transformation is staggering. Visual Copilot converts Figma designs directly to React code. Codia processes designs 100x faster than manual coding. These aren’t incremental improvements. They’re paradigm shifts disguised as product features. 

But the tools are just enablers. The real revolution happens in your mind. That moment when you realize the prison was self-imposed. The guards were imaginary. The key was always in your pocket. 

Natural language is the new programming language. If you can describe what you want, you can build it. If you can envision it, you can ship it. The only barrier left is imagination. 

The Future of Designer-Coders 

Organizations clinging to traditional designer-developer divisions are about to face a reckoning. While they coordinate handoffs and manage miscommunications, designers who code are shipping. Iterating. Learning. Building. 

This shift amplifies designers. Developers can focus on complex systems and architecture. Designers can implement their vision directly. Everyone works at a higher level of abstraction and impact. 

The competitive advantage is obvious. Teams with designer-coders ship better products faster. Not because they’re more efficient, but because they’re more effective. Vision and execution unified in a single mind. 

Your First Steps with AI Coding 

I know what you’re thinking. “But I’m not technical.” Neither was I. “But I don’t understand programming.” You don’t need to. “But I’m just a designer.” That’s exactly why you’re perfect for this. 

The same skills that make you a great designer, understanding users, crafting experiences, and obsessing over details, make you a natural at AI-powered development. You already think in systems and interactions. Now you can build them. 

Start small. Open a terminal and type a prompt. Build something stupid. Then build something slightly less stupid. Within hours, you’ll be building things that matter. Within days, you’ll wonder how you ever worked without this power. 

The Designer You’ll Become with AI 

Six months later, I barely recognize my old workflow. Static mockups feel like cave paintings. Design documentation seems like elaborate fiction. The idea of handing off my vision for someone else to interpret? Unthinkable. 

My role hasn’t changed. I’m still a visual designer. But my capability has transformed. I create experiences versus just imagining them. I propose ideas and prove them. I don’t just design products and ship them. 

The code anxiety is gone. Every limitation that once constrained me now seems artificial. The only question left is what to build next. 

Your journey starts with a single prompt. What will yours be? 

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing. 


Key Takeaways 


FAQs 

Q: What are AI code tools? 
AI code tools (like Claude Code, GitHub Copilot, or Visual Copilot) let you describe what you want in natural language, then generate working code automatically. 

Q: How can designers use AI code tools? 
Designers can turn Figma mockups or written prompts into functional prototypes, animations, and interactions—without learning traditional programming. 

Q: Does this replace developers? 
No. Developers focus on complex architecture, scaling, and systems. AI coding empowers designers to own interaction and experience details, speeding collaboration. 

Q: Why does this matter for organizations? 
Teams that adopt AI prototyping iterate faster, align design and development more tightly, and ship higher-quality products with fewer miscommunications. 

Q: What skills do designers need to start? 
Curiosity and creativity. If you can describe an idea clearly, you can build it with AI code tools. 

The Death of the Design Handoff: How AI Turns Tastemakers into Makers 

Every designer knows the ritual. You pour weeks into pixel-perfect mockups. You document every interaction, annotate every state, and build out comprehensive design systems. Then you hand it all to development and pray. 

Three sprints later, what comes back looks… different. Not wrong exactly, but not right either. The spacing feels off. The animations lack finesse. That subtle gradient you agonized over? Gone. The developer followed your specs perfectly, yet somehow the soul got lost in translation. 

Designers have always accepted this degradation as the cost of building digital products. We tried creating processes to minimize it, like design tokens, component libraries, and endless documentation, but we never stopped to question the handoff itself. 

Until now. AI just made the entire ritual obsolete. 

AI Ends the Design Handoff 

The design-to-development pipeline has always been messy, more like a game of telephone in a storm than a straight line. A designer’s vision turns into static mockups, those mockups get turned into specs, and then the specs are coded by someone who wasn’t there when the creative calls were made. 

Every step adds noise. Every handoff blurs the details. By the time the design reaches a user, the intent has been watered down through too many layers of translation. To manage the loss, we added layers. Product managers translate between teams, QA engineers catch mistakes, and design systems impose order. But taste cannot be standardized. 

AI design-to-code tools eliminate this process entirely. When a designer can move directly from Figma to functional code, the telephone line disappears. One vision, one implementation, and zero interpretation. 

Developers Spend Half Their Time on UI 

Here’s a truth we rarely say out loud. Developers spend 30–50% of their time on UI implementation. They’re not solving tough algorithms or designing big system architectures. They’re taking what’s already laid out in Figma and turning it into code. It takes skill and attention, but it’s work that repeats more than it invents. 

I’m not criticizing developers. I’m criticizing this process. We’ve asked our most technical team members to spend a third of their time as human transpilers, converting one formal language (design) into another (code). The real tragedy? They’re good at it. So good that we never stopped to ask if they should be doing it at all. 

When Airbnb started generating production code from hand-drawn sketches, they weren’t just saving time. They were liberating their engineers to focus on problems that actually require engineering. 

The Rise of the Tastemaker-Maker 

Something big shifts when designers can bring their own vision to life. The feedback loop shrinks from weeks to minutes. When something doesn’t look right, you can fix it immediately. If inspiration strikes, you can send it to staging and get real reactions in hours instead of weeks. What used to take whole sprints now fits inside a single coffee break. 

It’s tempting to frame this as designers turning into developers, but that misses the point. What’s really happening is that taste itself can now be put into action. The person who knows why a button feels right at 48 pixels, or why an animation needs a certain ease, or why an error state demands a particular shade of red, can actually make those choices real. 

That shift is giving rise to a new kind of role: the tastemaker-maker. They’re not confined to design or development but move fluidly between both. They hold the vision and the skills to bring it to life. They think in experiences and build in code. 

What Happens When Handoffs Disappear 

The implications ripple outward. When handoffs disappear, so do the roles built around managing them. The product manager who translates between design and development. The QA engineer who catches implementation mismatches. The technical lead who estimates UI development time. 

Teams start reorganizing around vision rather than function. Instead of design teams and development teams, you get product teams led by tastemaker-makers who can move from concept to code without translation. Supporting them are engineers focused on what AI can’t do: solving novel technical challenges, building robust architectures, optimizing performance at scale. 

This is job elevation. Developers stop being expensive markup translators and become true engineers. Designers stop being documentation machines and become product builders. Everyone moves up the value chain. 

AI Design to Code Speeds Shipping 

Companies using AI design-to-code tools report shipping features 3x faster with pixel-perfect accuracy. That’s a step function change in capability. While your team is still playing telephone, your competitors are shipping experiences that feel inevitable because they were never compromised by translation. 

The gap compounds daily. Each handoff you eliminate saves time on that project and builds institutional knowledge about what becomes possible when vision and execution converge. Your competitors are shipping faster and learning faster. 

How to Reorganize Without Handoffs 

Adopting AI design-to-code tools is the easy part. The hard part is reimagining your organization without handoffs. Start here: 

Identify your tastemaker-makers. They already exist in your organization. These are the designers who code on the side with strong aesthetic sense. Give them AI tools and watch them soar. 

Reorganize around products, not functions. Small teams with end-to-end ownership beat large teams with perfect handoffs every time. 

Measure differently. Stop counting tickets closed and start counting experiences shipped. Quality and velocity aren’t tradeoffs when the same person owns both. 

The End of the Design Handoff Era 

The design handoff was a bug in digital product development. A workaround for the technological limitation that the person who could envision the experience couldn’t build it. That limitation just died, and with it, an entire way of working that we tolerated for so long we forgot it was broken. 

The future belongs to those who can both dream and deliver. The handoff is dead. Long live the makers. 

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing. 


Key Takeaways 


FAQs 

What is a design handoff? 
The process where designers deliver mockups and specifications to developers, who then translate them into code. 

Why is the handoff inefficient? Each translation from design to documentation to implementation introduces information loss, slowing delivery and compromising quality. 

How do AI design-to-code tools change the process? 
They allow direct conversion from design tools like Figma into functional code, eliminating the translation step. 

What is a tastemaker-maker? 
A hybrid role that combines a designer’s vision with the ability to implement in code, collapsing feedback loops and accelerating iteration. 

Does this replace developers? 
No. It elevates developers to focus on complex engineering challenges, while routine UI translation is handled by AI. 

What’s the business impact? 
Companies using these tools report shipping 3x faster with higher fidelity—creating both a speed and learning advantage. 

Your Team’s AI Productivity Is About to Break Your Management Structure 

The Hidden Bottleneck 

It starts small. A manager logs in to find ten versions of the same deck, each slightly more polished than the last. By afternoon, three more have landed, polished by AI and dropped into the same crowded folder. Multiply that across a dozen employees, and what looked like momentum now feels like quicksand. 

AI has made people more prolific than ever. The crisis is no longer creation. It’s curation. 

Work used to move at a human pace: weekly deliverables, quarterly reviews, annual plans. AI just changed the speed limit, and that old cadence can’t keep up. MIT reports that 90% of employees now use unsanctioned AI tools for productivity gains. One person using AI can now outpace a full team. But as output rises, managers are left sorting through a flood of material. One study found teams spend 2.4 hours per day—nearly 30% of the week—just searching for the right information. When everyone is creating but few are curating, alignment and quality slip through the cracks.  

The Middle Manager’s Breaking Point 

This tension shows up most sharply in the middle. Managers used to be the connective tissue of organizations; now they’re underwater. Their traditional role of reviewing work, ensuring coherence, and maintaining quality no longer scales. It’s a common refrain from today’s leaders: teams are so prolific, managers simply can’t keep up. And when first-line managers can’t keep up, directors and executives above them lose sight of what’s really happening. The pyramid bends under the weight of its own output. 

The Productivity Paradox 

On paper, AI promises $4.4 trillion in productivity potential. In practice, many companies see dips before they see gains. The technology works. The real challenge lies in the structure around it. 

Creation has never been cheaper. We still have limits, especially around attention and judgment, the things that give creation its value. So, teams start cutting corners by spot-checking work, letting AI police itself, or just letting things through without a real review. 

The deeper problem? We’re using yesterday’s management playbook to navigate today’s nonstop output. Widening the highway without adding exits moves traffic faster… for a while. But the jams just reappear farther downstream in even bigger knots. Capability is no longer the constraint. Management capacity is. 

New Levers for Leaders 

What can organizations actually do about it? 

The first shift is in how progress gets measured. Volume no longer tells the story; what matters is whether the work actually moves strategic goals forward. Counting docs and drafts misses the point. 

The second shift is treating curation as real work. Tagging and organizing might not be glamorous, but they’re what keep AI-generated abundance usable instead of overwhelming. 

The third shift is elevating judgment. The real value comes not from creating yet another draft, but from deciding which draft matters and why. 

Finally, quality has to be a shared responsibility. Peer review and team-owned standards often beat the old model, where every piece of work climbs a slow chain of approvals before it ships. AI can point to anomalies, but people still define what “good” looks like. 

This isn’t just a productivity challenge; it’s a purpose problem. When roles shrink to prompting and passing along outputs, people lose connection to the work. Middle managers, once anchors of coordination and context, risk becoming bottlenecks. The real value lies in interpretation: guiding teams to make sense of abundance and channel it toward impact. 

The Path Forward 

The organizations that succeed are not the ones producing the most AI content. They are the ones curating with clarity, aligning work to strategy, and building structures strong enough to absorb exponential output without breaking. 

In an age of infinite creation, we’re no longer short on drafts or ideas. What’s scarce now is attention, judgment, and trust. 

AI has made abundance the easy part. The real leadership test is building systems that can turn that abundance into progress. 

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing. 


Key Takeaways 


FAQs 

Why is AI productivity breaking traditional management structures? 
AI enables individuals to produce the volume of ten, overwhelming management systems built for human-speed workflows. Structures designed for weekly deliverables and quarterly reviews cannot scale to AI’s pace. 

What is the real bottleneck: creation or curation? 
The challenge is curation. AI makes content generation cheap and fast, but deciding what matters, aligning it to strategy, and maintaining quality now consume more time and energy than creation. 

Why are middle managers most affected? 
Middle managers traditionally ensure coherence, review work, and maintain quality. With AI-driven output multiplying, this role no longer scales, leaving managers swamped and executives disconnected from what is happening on the ground. 

What is the productivity paradox of AI? 
AI has the potential to unlock trillions in value, yet many companies initially see dips in productivity. More output does not automatically mean more progress. Without curation, abundance creates confusion and slows decision-making. 

How can leaders adapt management for the AI era? 
By shifting from counting deliverables to measuring outcomes, investing in structured curation, redesigning roles around judgment, and embracing peer-driven models of quality. These approaches align AI productivity with organizational purpose. 

Robots & Pencils to Sponsor and Exhibit at Agentic AI and the Student Experience Conference 

AI-first consultancy joins higher ed leaders to explore how agentic AI is reshaping the student journey 

Robots & Pencils, an AI-first, global digital innovation firm specializing in cloud-native web, mobile, and app modernization, today announced its sponsorship and participation in Agentic AI and the Student Experience, hosted by Arizona State University (ASU). As a Silver Sponsor, exhibitor, and active participant, Robots & Pencils will engage with education leaders from around the world October 22–24, 2025, at the Omni Tempe Hotel at ASU. 

See how Robots & Pencils blends AI, cloud, and design to shape the future of education. 

The three-day event convenes higher education professionals and technology innovators to explore how agentic AI, systems that not only respond but proactively decide and solve problems, is revolutionizing the student experience. 

“Higher education is at a turning point, and agentic AI represents a breakthrough opportunity to enhance every stage of the student journey, from admissions to graduation and beyond,” said Leonard Pagon, CEO of Robots & Pencils. “We’re proud to join ASU, AWS, and other higher-education leaders to showcase what’s possible when cloud-native design, intelligent systems, and human-centered experiences come together. This is about accelerating AI readiness and charting the future of the student experience.” 

Robots & Pencils brings deep expertise to the higher education sector, having partnered with ASU on a multi-year transformation to unify academic data, streamline credential management, and expand student engagement through secure, scalable platforms. As an AWS Partner, the firm builds AI-ready, cloud-native systems that deliver speed, security, and scale across higher- ed institutions. 

“Education stands at the edge of a new frontier with agentic AI, where AI systems are proactive, adaptive and deeply personalized to enhance the student experience,” said Lev Gonick, Chief Information Officer at ASU and executive sponsor for the event. “What began as a call to convene has grown into a global gathering of more than 500 education and industry leaders who will chart the next chapter of AI in education,” Gonick continued.   

Robots & Pencils will host conversations at its exhibit table in the conference lobby, where attendees can explore use cases, see demonstrations, and connect with experts on campus modernization and AI readiness. Higher education leaders attending the event are encouraged to reach out in advance to request one-on-one meetings at robotsandpencils.com/asu2025

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing 

AI Agents Are Users Too: Rethinking Research for Multi-Actor Systems 

The first time an AI assistant rescheduled a meeting without human input, it felt like a novelty. Now it happens daily. Agents draft documents, route tickets, manage workflows, and interact on our behalf. They are no longer hidden in the background. They have stepped into the front lines, shaping experiences as actively as the people they serve. 

For research leaders, that changes the question. We have always studied humans. But when an agent performs half the task, who is the user? 

Agents on the Front Lines of Experience 

AI agents reveal truths that interviews cannot. Their activity logs expose where systems succeed and where they stumble. A rerouted request highlights a friction point. A repeated error marks a design flaw. Escalations and overrides surface moments where human judgment still needs to intervene. These are not anecdotes filtered through memory. They are live records of system behavior. 

And that’s why we need to treat agents as participants in their own right. 

A New Kind of Participant 

Treating agents as research participants reframes what discovery looks like. Interaction data becomes a continuous feed, showing failure rates, repeated queries, and usage patterns at scale. Humans remain the primary source of insight: the frustrations, the context, and the emotional weight. Agent activity adds another layer, highlighting recurring points of friction within the workflow and offering evidence that supports and extends what people share. Together, they create a more complete picture than either could alone. 

Methodology That Respects the Signal 

Of course, agent data is not self-explanatory. Logs are noisy. Bias can creep in if models were trained on narrow datasets. Privacy concerns must be addressed with care. The job of the researcher remains critical: separating signal from noise, validating patterns, and weaving human context into machine traces. Instead of replacing human perspective, agent data can enrich and ground it, adding evidence that makes qualitative insight even stronger. This reframing doesn’t just affect research practice, it also changes how we think about design. 

Designing for Multi-Actor Systems 

Products are no longer built for humans alone. They must work for the people who use them and the agents that increasingly mediate their experience. A customer may never touch a form field if their AI assistant fills it in. An employee may never interact directly with a dashboard if their agent retrieves the results. Design must account for both participants. 

Organizations that learn to research this new ecosystem will see problems sooner, adapt faster, and scale more effectively. Those that continue to study humans alone risk optimizing for only half the journey. 

The New Research Frontier 

Research has always been about listening closely. Today, listening means more than interviews and surveys. It means learning from the digital actors working beside us, the agents carrying out tasks, flagging failures, and amplifying our actions. 

The user is no longer singular. It is human and machine together. Understanding both is the only way to design systems that reflect the reality of work today. 

This piece expands the very definition of the user. For the other shifts redefining research, see our earlier explorations on format, how to move beyond static deliverables, and scope, how AI dissolves the depth vs. breadth tradeoff. 

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing.   


Key Takeaways


FAQs

Why consider AI agents as research participants?
AI agents actively shape workflows and user experiences. Their activity logs reveal friction points, errors, and escalations that human feedback alone may miss. Including them as research participants offers a more complete picture of how systems actually perform.

Do AI agents replace human participants in research?
No. Humans remain the primary source of context, emotion, and motivation. Agent data adds a complementary layer of evidence, enriching and grounding what people already share.

What types of insight can AI agents provide?
Agents surface recurring points of friction, repeated errors, and escalation patterns. These signals highlight where workflows break down, offering evidence to support and extend human feedback.

What role do researchers play when analyzing agent data?
Researchers remain critical. They filter noise, validate patterns, address bias, and ensure agent activity is interpreted with proper human context. The shift broadens qualitative practice rather than replacing it.

What is a multi-actor system in research?
A multi-actor system is one where both humans and AI agents interact to complete tasks. Designing for these systems means studying the interplay between people and machines, ensuring both participants are accounted for.

How does including agents in research improve design?
By listening to both humans and agents, organizations can spot problems sooner, adapt faster, and create systems that reflect the true complexity of modern workflows.

How AI Ends the Depth vs. Breadth Research Tradeoff 

The transcripts pile up fast. Ten conversations yield sticky notes that cover the wall, each quote circled, each theme debated. By twenty, the clusters blur. At thirty, the team is saturated, sifting through repetition in search of clarity. The insights are still valuable, but the effort to make sense of them begins to outweigh the return. 

This has always been the tradeoff, go deep with a few voices or broaden the scope and risk losing nuance. Leaders accepted that limitation as the cost of qualitative research. 

That ceiling is gone. 

The Ceiling Was Human Labor 

Generative research has always promised what numbers cannot capture, the story beneath the metric. But human synthesis is slow. Each new transcript multiplies complexity, until the process itself becomes the limiter. Teams stopped at 20 or 30 conversations not because curiosity ended, but because the hours to make sense of them did. Nuance gave way to saturation. 

Executives signed off on smaller studies and called it pragmatism. In truth, it was constraint. 

AI Opens the Door to Scale 

Large language models change the equation. Instead of weeks of sticky notes and clustering, AI can surface themes in hours. It highlights recurring ideas, connects outliers, and organizes insights without exhausting the team. The researcher’s role remains. Judgment still matters, but the ceiling imposed by human-only synthesis disappears. 

Instead of losing clarity as the number grows, each additional conversation now sharpens the signal, strengthening patterns, surfacing weak signals earlier, and giving leaders the confidence to act with richer evidence. 

Discovery Becomes Active 

The real breakthrough is not only scale, but also timing. With AI-enabled synthesis, insights emerge as the study unfolds. After the first dozen conversations, early themes are visible. Gaps in demographics or use cases show up while there is still time to adjust. By week two, the research is already feeding product decisions. 

Instead of waiting for a final report, teams get a living stream of discovery. Research shifts from retrospective artifact to active driver of strategy. 

Nuance at Speed 

For organizations, this ends the false binary. Depth and breadth no longer compete. A bank exploring new digital features can capture voices across demographics in weeks, not months. A health-tech team can fold dozens of patient experiences into the design cycle in real time. A software platform can test adoption signals across continents without sacrificing cultural nuance. 

The payoff is more than efficiency. It is confidence. When executives see both scale and nuance in the evidence, they act faster and with greater conviction. 

The New Standard 

The era of choosing between depth or breadth is behind us. AI frees research leaders from the constraints of small samples or limited perspectives. With AI as a synthesis partner, the standard shifts: hundreds of voices, interpreted with clarity, delivered at speed. 

For teams still focused on fixing the format problem, our previous piece, The $150K PDF That Nobody Reads, explores how static reports constrain research. Our next article examines an even bigger shift: what happens when your users are no longer only people.

The pace of AI change can feel relentless with tools, processes, and practices evolving almost weekly. We help organizations navigate this landscape with clarity, balancing experimentation with governance, and turning AI’s potential into practical, measurable outcomes. If you’re looking to explore how AI can work inside your organization—not just in theory, but in practice—we’d love to be a partner in that journey. Request an AI briefing.   


Key Takeaways


FAQS

What is the depth vs. breadth tradeoff in qualitative research?
The depth vs. breadth tradeoff refers to the long-standing belief that teams must choose between conducting a small number of interviews with rich nuance (depth) or a larger sample with less detail (breadth). Human synthesis struggles to handle both simultaneously, forcing this choice.

How does AI change the depth vs. breadth tradeoff?
AI dissolves the tradeoff by enabling researchers to process hundreds of conversations quickly while still preserving nuance. Instead of diluting insight, scale strengthens pattern recognition and surfaces weak signals earlier.

Why has qualitative research been constrained to small sample sizes?
Human synthesis is time-consuming. After 20–30 interviews, transcripts become overwhelming, and important signals get lost in the noise. This labor bottleneck led leaders to view small samples as “pragmatic,” even though it was really a constraint of capacity.

Does AI replace the role of the researcher?
No. AI accelerates synthesis, but the researcher remains critical for judgment, interpretation, and ensuring context and nuance are applied correctly. AI acts as a partner that expands capacity rather than a replacement.

What is the impact of AI-enabled synthesis on decision-making?
With faster synthesis and preserved nuance, research insights emerge in real time rather than only in final reports. Leaders gain richer evidence earlier, which supports faster, more confident decisions.

What does this mean for the future of qualitative research?
The old tradeoff between depth and breadth is over. AI makes it possible to achieve both simultaneously, shifting the standard for research to hundreds of voices interpreted with clarity and delivered at speed.