futures thinking

Three Horizons Framework & Futures Wheel Explained

There's a reason some organizations consistently seem to see disruption coming — and it's usually not because they're smarter or better funded. It's because they've built structured habits of thinking about change in multiple time horizons simultaneously, and they've learned how to trace the cascading consequences of a single shift before it becomes a crisis.

Two of the oldest and most reliable tools for doing exactly that are the Three Horizons Framework and the Futures Wheel. In this episode of Modem Futura, hosts Sean Leahy and Andrew Maynard break both down in accessible, conversational detail — and show what becomes possible when you use them together.

The Three Horizons Framework

Originally developed by Bill Sharpe and widely used in professional foresight and strategic planning, divides the landscape of change into three overlapping zones. Horizon 1 represents the dominant present — the systems, structures, and assumptions that govern how the world works today. Horizon 3 is the emergent fringe: weak signals, nascent ideas, and early-stage shifts that are observable but not yet mainstream. And Horizon 2 is the transitional space between them — turbulent, hard to define, and full of both opportunity and risk.

The model doesn't tell you what the future will bring. What it offers is a way of *positioning* trends, signals, and innovations in relation to change — helping individuals and organizations understand what to watch, what to act on, and what to prepare for.

The Futures Wheel

Developed by Jerome Glenn in 1971, works differently but complementarily. Starting from a specific change or trend, it maps outward through first, second, and third-order consequences — building a rich, networked picture of how a single shift might ripple through a system over time. It's a brainstorming and sense-making tool, not a prediction engine, and it's at its most powerful when used with diverse groups who bring different perspectives to the same question.

Used individually, each tool offers genuine insight. Used together, they offer something more: a way of understanding not just *what* a signal might do, but *when* and *through which pathways* it might do it.

Whether you're a founder trying to figure out which wave to ride, a strategist scanning for disruption, or simply someone trying to make better decisions in an uncertain world, these tools are worth adding to your thinking practice.

🎧 Listen to the full episode wherever you get your podcasts, or watch on YouTube.


Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.


🎧 Apple Podcast: https://apple.co/4sosMdQ

🎧 Spotify: https://open.spotify.com/episode/58Fdc2SrWodBTbwfxK8Pwm?si=leiCnhRsQxuv-_hnxEeNjQ

📺 YouTube: https://youtu.be/eVk6L_VfAkY

🌐 Website: https://www.modemfutura.com/   

Fluid Futures: Navigating an AI-Mediated World

What Happens When AI Stops Being a Tool and Starts Being the World?

There's a useful distinction that keeps getting lost in conversations about artificial intelligence: the difference between augmentation and mediation.

Augmentation is familiar. It's the calculator model — AI helps you work faster, smarter, better. You remain the agent. The tool amplifies your capacity.

Mediation is something else. When AI mediates your world, it's not just helping you do things — it's shaping the system you're doing them inside of. What information surfaces. What options appear. What feels like the obvious next move. You're not using the environment anymore. You're inside one that AI has constructed, and it's shifting around you in real time.

This distinction is at the heart of Exploring the Futures of Technology 2.0, the new report from the Copenhagen Institute of Future Studies — and it's the central thread of the latest episode of Modem Futura.

On this episode, my co-host Andrew Maynard, fresh from attending the report's launch in Copenhagen, joined me to work through ten signals the report identifies as defining the near future: the shift from static to liquid content, the rise of agentic organizations, neurotechnology and cognitive integration, synthetic simulations replacing real-world research populations, physical AI entering embodied space, the geopolitics of technological access, AI-mediated cybersecurity threats, the sustainability challenges of AI infrastructure, and quantum computing as the wildcard at the edge of everything.

What holds these signals together isn't a single prediction. It's a pattern: the world is becoming fluid, and the frameworks we built for a more static environment — static reports, static institutions, static skillsets — are increasingly inadequate for navigating it.

One of the episode's sharpest observations is about the cost of cognitive offloading. As we hand more of our decision-making and information retrieval to AI systems, we risk losing the capacity to recognize when something's wrong. Not because AI is malicious, but because we've stopped practicing the skills that would let us notice. Like losing the ability to read a map. Except the stakes are considerably higher.

The conversation doesn't resolve these tensions — and that's exactly the point. Futures thinking, at its best, isn't about prediction. It's about staying awake to what's changing, naming the tensions, and refusing to optimize for a world that no longer exists.

If you want the full report, the Copenhagen Institute has made it freely available. And if you want the conversation around it — the episode is a good place to start.

Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.

🎧 Apple Podcast: https://apple.co/4bSGsZP

🎧 Spotify: https://open.spotify.com/episode/4sdx83QUD6pIs9IXb9G0VY?si=JdbwVHUKRg2mFO0Gsi_EFw

📺 YouTube: https://youtu.be/-2enUvPYmHo

🌐 Website: https://www.modemfutura.com/   

The Futures Cone: A Framework for Exploring What Could Be

How one deceptively simple tool can transform the way you think about uncertainty, possibility, and the choices that shape tomorrow.

There's a habit most of us share when it comes to thinking about the future: we treat it as a destination. A singular, somewhat “predictable” place that today's trends are quietly marching toward. It's a useful shorthand — but as a mental model, it's quietly limiting.

The Futures Cone, a foundational tool in the field of futures studies, offers a different way of seeing. Rather than imagining the future as a point, it asks you to imagine it as a cone — wide open, expanding outward from the present moment, filled with layers of possibility that range from the likely to the genuinely unthinkable.

How the Cone Works

The narrowest point is now. As the cone extends outward through time, it widens to reveal different regions of possible futures, each defined by how much disruption or change would be required to bring them about:

Projected futures — the baseline; what happens if nothing changes

Probable futures — where current trends are pointing

Plausible futures — what could happen given known forces and trajectories

Possible futures — speculative, requiring future knowledge we don't yet have

Preposterous futures — the outer edge; scenarios that challenge our deepest assumptions about what is physically or socially feasible

Threaded through all of these is the Preferable future — not a separate ring, but a cross-section that asks: given everything in this cone, what do we actually want? Where do our values point?

The Dator-Clarke Line

One of the most provocative ideas associated with the cone is what's referred to as the Dator-Clarke Line — drawn from futurist James Dator's claim that any genuinely useful idea about the future should, at first glance, appear ridiculous. Paired with Arthur C. Clarke's observation that the only way to find the limits of the possible is to push into the impossible, it suggests that the most valuable futures work happens precisely in the uncomfortable space at the edge of the cone.

The practical implication is significant: if every idea your team generates sounds reasonable, you probably haven't stretched far enough. The preposterous isn't a failure of imagination — it's a boundary worth exploring.

Why This Tool Matters Now

In a period defined by technological acceleration, geopolitical uncertainty, and rapid social change, the instinct to "project forward" can feel reassuring — but it's also where strategic blind spots form. The Futures Cone doesn't resolve that uncertainty. Instead, it gives individuals, teams, and organizations a shared language for navigating it: a structured way to ask not just "what will happen?" but "what could happen, what might we prefer, and what are we willing to do about it?"

This is the subject of Episode 74 of Modem Futura, in which we walk through the cone layer by layer — and then demonstrate it live with a thought experiment that starts with frogs and ends somewhere near the moons of Jupiter.

Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.

🎧 Apple Podcast: https://apple.co/4bz1tIC

🎧 Spotify: https://open.spotify.com/episode/20Hz36eLfZ90M6EifUrRuu?si=swNkzHWZSLCelVSKsAyj7A

📺 YouTube: https://youtu.be/wc_e3dsY-vw

🌐 Website: https://www.modemfutura.com/   

What Old iPods and Tiny Cameras Teach Us About Technology, Ownership, and Being Human

There's a moment in this episode of Modem Futura where two grown adults are hunched over a miniature Polaroid camera, watching a blurry selfie slowly develop — and laughing about it. It's objectively a terrible photograph. But it captures something that most modern technology has quietly optimized away: surprise, imperfection, and the distinctly human joy of not knowing exactly what you're going to get.

This episode began with a box of old iPods — tangled cables, dead batteries, and all — and evolved into a wide-ranging conversation about what we trade away every time we upgrade to something faster, thinner, and more connected. The themes are ones that touch anyone who has ever felt a pang of something unnamed while scrolling through an infinite library of music and being unable to choose a single song.

Ownership in the age of access. The iPods in the conversation are air-gapped — no internet connection, no cloud sync, no subscription. The music on them belongs to their owner in a way that a Spotify library simply does not. This distinction matters more than it might seem, especially when you consider that digital books, photos, and music can disappear when a service shuts down or an account holder passes away. The question of digital legacy — who inherits your cloud — is one most people haven't thought through yet.

Craft, care, and the "fast food" of technology. Sean raises a pointed observation about a recently released video game that shipped with fewer features than its predecessor from a decade ago. It's a pattern that extends well beyond gaming: the pressure to release fast increasingly overrides the commitment to release well. When did "good enough" become the standard?

The paradox of abundance. One of the episode's most compelling threads is the tension between scarcity and surplus. Limited storage on an old iPod forced intentional curation — playlists that became personal time capsules. Unlimited streaming offers everything and, paradoxically, can deliver less meaning. Andrew's students, however, offer a counterpoint: raised in abundance, they've developed their own sophisticated habits of curation and care. Perhaps the pendulum is already swinging.

Imperfection as a feature. The tiny Kodak keychain camera. The Polaroid with its gloriously blurry output. The analog photograph whose chemistry introduces an element of chance. These aren't failures of technology — they're reminders that the most human experiences are often the least predictable ones.

This episode doesn't offer prescriptions. It offers an invitation: to notice, to question, and to be intentional about the role technology plays in your life before someone else makes that choice for you.

🎧 Listen to Episode 73 of Modem Futura — available on Apple Podcasts, Spotify, and wherever you listen.

Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.

🎧 Apple Podcast: https://apple.co/4b3P8L8

🎧 Spotify: https://open.spotify.com/episode/2UNsDaZox2jdEb1QYN1m44?si=FUkqjQ0gSEecnYyrjKfoVA

📺 YouTube: https://youtu.be/UKC7UHkGNJQ

🌐 Website: https://www.modemfutura.com/   

Thriving with AI: Two Futures Thinking Tools for Navigating Uncertainty

Illustration of Sean and Andrew presenting their workshop title slide

The question is no longer whether AI will reshape education. It already has. The more interesting question — and the harder one — is how educators, leaders, and institutions can navigate that transformation with clarity, purpose, and agency.

In this episode of Modem Futura, hosts Sean Leahy and Andrew Maynard walk listeners through a workshop they developed for ASU's 2026 Folk Fest titled "Thriving with AI: Ethical, Transparent, and Human-Centered Learning." Rather than demonstrating AI platforms or advocating for a particular stance, the session offers two practical thinking tools designed to help individuals make sense of complexity and make intentional decisions — regardless of where they fall on the AI adoption spectrum.

Foresight Methodologies

The Futures Triangle, originally developed by futurist Sohail Inayatullah, is a foresight method that maps three forces shaping any change landscape: the pull of the future (emerging visions and possibilities), the push of the present (trends, pressures, and mandates driving change), and the weight of history (the traditions, values, and institutional structures that resist or ground that change). By making these forces visible, individuals and teams can better orient themselves within the dynamics of change rather than simply reacting to them.

The Intent Map, drawn from Jefferey Abbott and Andrew Maynard's book AI and the Art of Being Human, complements the triangle by shifting from orientation to action. A simple two-by-two matrix, it asks users to identify four elements: their core values (what they won't compromise), their desired outcomes (what success looks like), their guardrails (the hard boundaries they won't cross), and their metrics (how they'll know if it's working). Critically, the framework recognizes that metrics don't have to be numerical — sometimes the most meaningful indicators of success are qualitative, like a student who can't stop thinking about what they learned.

What makes these tools particularly valuable is their accessibility. Both can be sketched on a scrap of paper. Both work for individuals and teams. And both are domain-agnostic — while the episode frames them in the context of education, they apply equally well to organizational strategy, technology adoption, and personal decision-making.

The episode is anchored by two provocative 2035 headlines: one in which AI tutors outperform human teachers and faculty roles come under review, and another in which human-AI partnership produces the most critically thinking generation in history. The question the workshop poses isn't which headline is more likely. It's which one you want — and what intentional choices you need to make to move toward it.

Thriving with AI, as the hosts frame it, isn't about mastering the latest platform. It's about staying awake to what matters.


Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.

🎧 Apple Podcast: https://apple.co/3ZXgT2P

🎧 Spotify: https://open.spotify.com/episode/1b1Q0W7YVSGZA2ELYj6g6C?si=wL1sXb-DQsSluBkLYCu9tg

📺 YouTube: https://youtu.be/zi_zvXCt9sY

🌐 Website: https://www.modemfutura.com/   


Asimov's "The Fun They Had" and the Real Cost of AI-Driven Education

Illustration of Asimov's Fun They Had boy reading by mechanical teacher

The History of our Future

More than seventy years ago, Isaac Asimov imagined a future where children learn in isolation, guided by personalized mechanical tutors, and books are relics of a forgotten age. His 1951 short story, "The Fun They Had," is set in 2155, but its questions feel startlingly current.

In the story, a young girl named Margie discovers a paper book and learns about a time when children went to school together—sat in classrooms, were taught by human teachers, and shared the experience of learning with their peers. Her own education is efficient, personalized, and lonely. Her mechanical teacher can diagnose her struggles and recalibrate its approach, but it cannot inspire her, connect with her, or make her feel like she belongs to something larger than a lesson plan.

Asimov didn’t predict AI as we know it. But he predicted the question that matters most: in our rush to optimize education, are we designing out the very things that make learning meaningful?

This is precisely the tension at the heart of today's conversation about AI in education. The promise of AI-powered tutors is real and, in many cases, genuinely valuable: adaptive pacing, instant feedback, content tailored to individual needs. But when personalization becomes the dominant paradigm—when every learner is on a separate track, in a separate space, at a separate time—the communal dimensions of education begin to disappear.

Natural Human Impulses for Learning (not schooling)

John Dewey argued more than a century ago that learning is driven by four natural impulses: inquiry, communication, construction, and expression. Most of these are inherently social. They depend on friction, dialogue, surprise, and the presence of other people. No amount of algorithmic sophistication can fully replicate the moment a teacher's unexpected enthusiasm shifts a student's entire trajectory, or the experience of working through difficulty alongside peers who share the same struggle.

Asimov's story also raises a subtler question about what endures. The book Margie discovers has survived two centuries. The static words on the page—unchanging, tactile, physical—carry a kind of permanence that digital media cannot easily match. This resonates with the growing cultural appetite for analog experiences: vinyl records, film photography, even old iPods. These are not acts of technological rejection. They are expressions of a deeper need for embodied engagement, deliberate choice, and the kind of friction that gives experience its texture.

Where do we go next?

None of this means AI has no place in education. It does, and increasingly will. But Asimov's story is a quiet reminder that the most important things about learning—curiosity, connection, belonging, the joy of shared discovery—are not problems to be optimized. They are human experiences to be protected.

The question is not whether AI can teach us. It's whether, in building systems that teach us more efficiently, we are designing out the very things that made learning worth having in the first place.

*Episode 71 of Modem Futura explores these themes through Asimov's story and a wider conversation about technology, nostalgia, and what it means to learn as a human being.*

Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.

🎧 Apple Podcast: https://apple.co/4s1lDk1

🎧 Spotify: https://open.spotify.com/episode/20I5j2DliUnZAbWDiVw7y8?si=WoEW_Zb2SPiynHYb4d8XHA

📺 YouTube: https://youtu.be/TDQc15Muwto

🌐 Website: https://www.modemfutura.com/   

Vibe Coding and the Return of Personal Software

Vintage styled computer terminal with the text "What will you build today?" displayed on screen

The Echo of Early Personal Computing

There was a brief, electric moment in the history of computing—roughly the late 1970s through the mid-1980s—when ordinary people could sit down at a keyboard and make a machine do something it hadn't done before. The Commodore 64, the BBC Micro, the Apple II: these were limited, clunky, and profoundly empowering. For a generation, they opened the door to a kind of creative agency that felt almost magical.

That door closed, gradually, as software became professionalized. The gap between what you could imagine and what you could build widened into a canyon. If you wanted a tool that didn't exist, you needed a developer—or you went without.

Vibe coding is reopening that door.

The term refers to the practice of describing what you want in natural language and letting a generative AI—tools like Claude, ChatGPT, or Copilot—write the code for you. No syntax to memorize. No debugging by hand. You describe your intent, and working software comes back in seconds.

In this episode of Modem Futura, we explore what this shift means—not just technically, but humanly. I demonstrates tools he built from single prompts (also referred to as a a ‘one-shot’): a horizon-scanning app for futures research and a two-by-two uncertainty matrix used in strategic foresight. Both were functional on the first attempt. Both took less time to create than it takes to describe them.

The Inherited Power Problem

But the episode resists the temptation to treat this as a simple good-news story. The hosts dig into the real tensions: AI-generated code that no one fully understands, security vulnerabilities baked into apps that reach market before anyone reviews them, the new threat landscape of prompt injection, and the philosophical question of wielding power you haven't earned the literacy to evaluate—what the hosts call "inherited power."

There are also rich implications for education. Rather than relying on off-the-shelf apps that never quite fit, instructors and students alike can now build purpose-specific tools—and in doing so, develop a more grounded understanding of what these AI systems can and cannot do.

The deeper question the episode surfaces is less about code and more about agency. For decades, software was something done to us—platforms we adapted to, interfaces we learned, ecosystems we bought into. Vibe coding hints at a possible reversal: software shaped by the individual, for the individual, in the moment they need it.

Whether that future is liberating or reckless—or both—depends on the kind of literacy, caution, and imagination we bring to it.

Listen to the full conversation on Modem Futura.

Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.

🎧 Apple Podcast: https://apple.co/4rbOr9r

🎧 Spotify: https://open.spotify.com/episode/28DMXJsM2kEBA2QDxuDmtJ?si=AJpR7zCpRgS2KCCfWwjjWg

📺 YouTube: https://youtu.be/lQGYaiThuBk?si=nRbHVEQk9dwL3gXr

🌐 Website: https://www.modemfutura.com/   

Pluribus and the Philosophy of the Happy Apocalypse: What Apple TV's New Sci-Fi Asks About Individuality, Consent, and Being Human

What if happiness is the threat?

Most apocalypse stories share a common grammar: society collapses, resources become scarce, and survival demands violence. We've internalized this template so thoroughly that it shapes how we imagine catastrophe itself.

Apple TV's Pluribus, created by Vince Gilligan (Breaking Bad, Better Call Saul), disrupts that grammar entirely. Its apocalypse isn't marked by destruction or suffering. It's marked by peace. By synchronization. By happiness—at a planetary scale.

An alien signal arrives carrying an RNA sequence. Humanity, being humanity, synthesizes it. Within days, most of the global population transforms into a unified hive mind. Not zombies. Not drones. Just billions of people sharing consciousness, moving together, experiencing what appears to be genuine contentment.

About a dozen people remain unconverted. And the series follows one of them—Carol Sterka, played by Rhea Seehorn—as she grapples with being the most unhappy person on earth.

On a recent episode of the Modem Futura podcast, we explored what Pluribus surfaces about individuality, consent, collective identity, and the stories we tell ourselves about what makes a human life worth living. What follows are some of the tensions that emerged.

What is in a Name: Many Without the One

The title "Pluribus" comes from the Latin phrase E Pluribus Unum—"out of many, one"—which appears on American currency as a motto of national unity.

But the show drops both the "E" (out of) and the "Unum" (one). What remains is simply "Pluribus": the many. It's a subtle signal that this isn't a story about diversity coming together into unity. It's a story about what happens when "the many" becomes literal—when individual minds merge into a single, collective consciousness.

That linguistic choice frames everything that follows.

Who Becomes the Monster?

One of the most productive lenses for understanding Pluribus is Richard Matheson's 1954 novel I Am Legend. Not the Will Smith film adaptation, but the original text, which ends with a devastating realization: the protagonist, who has spent the story hunting the "monsters" who have replaced humanity, comes to understand that from their perspective, he is the monster. The one who kills in the night. The one who refuses to accept the new order.

Carol Sterka occupies similar territory. She's convinced she needs to "set things right"—to restore humanity to its pre-hive state. But the show keeps surfacing an uncomfortable question: right for whom? The hive mind has eliminated war, poverty, and suffering. Billions of people who lived in misery are now at peace.

If Carol succeeds in reversing the transformation, she's not saving people. She's condemning them to return to lives many of them would never have chosen.

The Consent Paradox

The hive mind in Pluribus operates under an interesting constraint: it cannot lie, and it will not assimilate anyone without their explicit permission.

This sounds like respect for autonomy. And in some sense, it is. But the hive mind also desperately wants everyone to join (even explaining that it’s a ‘biological’ imperative). So what emerges is a kind of relentless, patient persuasion—always honest, always gentle, and always oriented toward a predetermined outcome.

There's something uncomfortably familiar in this dynamic. We navigate versions of it constantly: platforms that "personalize" our experience toward their engagement metrics, systems that "recommend" content optimized for their retention goals, interfaces designed to make one choice frictionless and alternatives invisible.

The hive mind's honesty doesn't make its agenda less persistent. It just makes the agenda transparent.

The Sustainability Problem

Midway through the season, Pluribus introduces a complication: the hive mind will only consume things that have already died naturally. No killing. No harvesting. Just waiting for life to end on its own terms.

Which means, at planetary scale, they're slowly starving.

This creates a strange inversion. Carol, the last holdout, has skills and knowledge that could help solve the problem. But she's too consumed by her mission to "fix" things to collaborate with the very beings who need her help.

There's something painfully recognizable in that dynamic—the way ideological certainty can prevent us from engaging productively with people whose worldview differs from our own, even when collaboration would benefit everyone.

Is the Individual Still in There?

One of the more haunting threads in Pluribus involves the question of whether individual identities persist within the hive mind.

Carol's "chaperone"—a member of the hive who presents as an individual named Zosia—occasionally exhibits moments that feel less like collective consciousness and more like... a person surfacing. A memory that seems too specific. A reaction that seems too singular. (The Mango ice cream scene is a particular interesting one where for a moment - the real Zosia seems to surface).

Another character (Manousos) experiments with radio frequencies, attempting to extract individuals back out of the collective, seemingly trying to hack the near field electromagnetic connections the “others” have with one another.

The show doesn't resolve this, but rather leaves it as a season 1 cliffhanger as it seems some progress is made. But it raises the question: if you could pull someone out of a state of collective happiness and return them to individual consciousness, would that be rescue or harm? Liberation or trauma?

There's no easy answer. And Pluribus is wise enough not to pretend there is.

The AI Parallel (That Isn't Really About AI)

Vince Gilligan has stated that Pluribus isn't intended as an AI allegory. The original concept predates the current wave of generative AI by years.

And yet.

The show's exploration of collective intelligence, of optimization toward contentment, of systems that genuinely want to help but whose help involves transformation into something other than what you were—all of it resonates with questions we're already asking about artificial intelligence and its role in human flourishing.

The hive mind's impulse to "fix" things, to smooth over friction, to optimize for happiness—that's not so different from Silicon Valley's persistent faith that the right algorithm can solve human problems. The show doesn't moralize about this. It simply shows what it might feel like to be on the receiving end of that faith.

The hive mind might be the best thing that ever happened to humanity. Or it might be the end of everything that made humanity worth preserving. The show suggests both readings are available, and neither is obviously wrong. In the end, this is my favorite part of the show - it catalyzes great conversations… it pushes us to examine very human elements by forcing us to entertain scenarios in which we question what it means to be human. Now we just have to wait a seemingly excruciating long time until Season 2 will be ready – until then, stay curious!

Subscribe and Connect!

Subscribe to Modem Futura wherever you get your podcasts and connect with us on LinkedIn. Drop a comment, pose a question, or challenge an idea—because the future isn’t something we watch happen, it’s something we build together. The medium may still be the massage, but we all have a hand in shaping how it touches tomorrow.

🎧 Apple Podcast: https://apple.co/4k0l1bo

🎧 Spotify: https://open.spotify.com/episode/5ymC2VZJUz7iLTvYj89CXa?si=52mn5UiBRH-gbkpSEnV4Tw

📺 YouTube: https://youtu.be/xsxJWN5FO-U

🌐 Website: https://www.modemfutura.com/   

Related Reading

  • I Am Legend by Richard Matheson

  • Solaris by Stanisław Lem

  • The Borg episodes of Star Trek: The Next Generation

Techno-Humans and the Energy Futures We’re Designing

Techno-Humans and the Energy Futures We’re Designing

What if the clean energy transition isn’t just a technology problem—but a techno-human design challenge that determines who benefits, who’s left out, and whether our cities can thrive?

Modem Futura Year in Review: What 2025 Taught Us About Being Human

As we step toward 2026, we recorded a “Year in Review” episode of Modem Futura to pause the treadmill, look back, and ask a bigger question: what did this year reveal about the future of being human?

This wasn’t a victory lap. It was a reflection on what resonated, what surprised us, and what it means to build a future-focused show while the future keeps moving.

metrics matter… and they don’t

Yes, growth matters — it helps ideas travel. But podcast analytics are often incomplete and inconsistent, and they rarely capture what impact actually looks like. The most meaningful signals are still human: messages, emails, thoughtful disagreement, and reviews that help someone new discover the show.

If you want to support the show: subscribing, sharing, and leaving a rating/review are still the most helpful actions.
— Modem Futura

The themes that defined our year:

AI, beyond the hype: We kept returning to the same tension — generative tools are everywhere, but “AI” isn’t just a feature set. It’s a cultural force that shapes identity, agency, creativity, and values. We try hard to avoid both the hype machine and the doom loop, and instead stay in the messy middle where the most useful questions live.

Education and learning: We lean into what learning actually is (not just schooling), including John Dewey’s idea that humans are wired for inquiry, communication, construction, and expression. When AI arrives in every document and device, what does it do to those impulses — especially for kids?

Technology in the physical world: From autonomous‑vehicle safety systems that quietly drift out of calibration, to EVs and the persistent “flying car” dream, we explore what happens when shiny promises meet real‑world constraints.

Big questions, no apologies: Yes, we go there — simulation hypotheses, black holes, de‑extinction, space travel, and the edges of what science can (and can’t) explain. These episodes aren’t about “being right.” They’re about expanding the space of possible futures we can imagine.

If there’s one takeaway, it’s this: the future isn’t something that happens to us — it’s something we build together.That’s why we keep showing up each week: to create a shared space for curiosity, skepticism, wonder, and responsible imagination.

If you’ve been listening, thank you. If you’re new here, welcome. And if an episode sparked a thought you can’t shake — share it with a colleague, a student, a friend, or your community. As we step into 2026, we’re excited to keep exploring the possible, probable, and preferable futures — with you.