Toggle light / dark theme

Schellman AI Summit 2025 · Luma

Join Adam Perella and I at the Schellman AI Summit on November 18th, 2025 at Schellman HQ in Tampa Florida.

Your AI doesn’t just use data; it consumes it like a hungry teenager at a buffet.

This creates a problem when the same AI system operating across multiple regulatory jurisdictions is subject to conflicting legal requirements. Imagine your organization trains your AI in California, deploys it in Dublin, and serves users globally.

This means that you operate in multiple jurisdictions, each demanding different regulatory requirements from your organization.

Welcome to the fragmentation of cross-border AI governance, where over 1,000 state AI bills introduced in 2025 meet the EU’s comprehensive regulatory framework, creating headaches for businesses operating internationally.

As compliance and attestation leaders, we’re well-positioned to offer advice on how to face this challenge as you establish your AI governance roadmap.

Cross-border AI accountability isn’t going away; it’s only accelerating. The companies that thrive will be those that treat regulatory complexity as a competitive advantage, not a compliance burden.

How do you trust a robot you’ve never met?

Many of the environments where human-facing universal robots can provide benefits — homes, hospitals, schools — are sensitive and personal. A tutoring robot helping your kids with math should have a track record of safe and productive sessions. An elder-care assistant needs a verifiable history of respectful, competent service. A delivery robot approaching your front door should be as predictable and trustworthy as your favorite mail carrier. Without trust, adoption will never take place, or quickly stall.

Trust is built gradually and also reflects common understanding. We design our systems to be explainable: multiple AI modules talk to each other in plain language, and we log their thinking so humans can audit decisions. If a robot makes a mistake — drops the tomato instead of placing it on the counter — you should be able to ask why and get an answer you can understand.

Over time, as more robots connect and share skills, trust will depend on the network too. We learn from peers, and machines will learn from us and from other machines. That’s powerful but just like parents are concerned about what their kids learn on the web, we need good ways to audit and align skill exchange for robots… Governance for human–machine societies isn’t optional; it’s fundamental infrastructure.

Human Flourishing In The Age Of AI And Robots — The Futurists X Summit 2025

See my Comment below for a link to David Orban’s 20 minute talk.


In this keynote, delivered at The Futurists X Summit, on September 22 in Dubai, David Orban maps how AI and humanoid robotics shift us from steady exponential progress to an acceleration of acceleration—what he calls the Jolting Technologies Hypothesis. He argues we’re not in a zero-sum economy; as capability compounds and doubling times shrink, we unlock new degrees of freedom for individuals, firms, and society. The challenge is to steer that power with clear narratives, robust safety, and deliberate design of work, value, and purpose.

You’ll hear:
• Why narratives (optimism vs. doom) shape which futures become real.
• How shortening doubling times in AI capabilities pull forward timelines once thought 20–30 years out.
• Why trust in AI is task-relative: if +5% isn’t enough, aim for 10× reliability.
• The coming phase transformation as intelligence becomes infrastructure (homes, mobility, industry).
• Concrete social questions (e.g., organ donation post–road-death decline) that demand AI-assisted governance.
• Why the nature of work will change: from jobs as status to human aspiration as value.

Key ideas:
• Humanoid robots at scale: rapid iteration, non-fragile recovery, and human-complementary performance.
• Designing agency: go from idea → action with near-instant execution; experiment, learn, and iterate fast.
• From zombies to luminaries: use newfound freedom to architect lives worth living.

Resources & Links:

ASI Risks: Similar premises, opposite conclusions | Eliezer Yudkowsky vs Mark Miller

A debate/discussion on ASI (artificial superintelligence) between Foresight Senior Fellow Mark S. Miller and MIRI founder Eliezer Yudkowsky. Sharing similar long-term goals, they nevertheless reach opposite conclusions on best strategy.

“What are the best strategies for addressing risks from artificial superintelligence? In this 4-hour conversation, Eliezer Yudkowsky and Mark Miller discuss their cruxes for disagreement. While Eliezer advocates an international treaty that bans anyone from building it, Mark argues that such a pause would make an ASI singleton more likely – which he sees as the greatest danger.”


What are the best strategies for addressing extreme risks from artificial superintelligence? In this 4-hour conversation, decision theorist Eliezer Yudkowsky and computer scientist Mark Miller discuss their cruxes for disagreement.

They examine the future of AI, existential risk, and whether alignment is even possible. Topics include AI risk scenarios, coalition dynamics, secure systems like seL4, hardware exploits like Rowhammer, molecular engineering with AlphaFold, and historical analogies like nuclear arms control. They explore superintelligence governance, multipolar vs singleton futures, and the philosophical challenges of trust, verification, and control in a post-AGI world.

Moderated by Christine Peterson, the discussion seeks the least risky strategy for reaching a preferred state amid superintelligent AI risks. Yudkowsky warns of catastrophic outcomes if AGI is not controlled, while Miller advocates decentralizing power and preserving human institutions as AI evolves.

Shadow AI Discovery: A Critical Part of Enterprise AI Governance

MITs State of AI in Business report revealed that while 40% of organizations have purchased enterprise LLM subscriptions, over 90% of employees are actively using AI tools in their daily work. Similarly, research from Harmonic Security found that 45.4% of sensitive AI interactions are coming from personal email accounts, where employees are bypassing corporate controls entirely.

This has, understandably, led to plenty of concerns around a growing “Shadow AI Economy”. But what does that mean and how can security and AI governance teams overcome these challenges?

Contact Harmonic Security to learn more about Shadow AI discovery and enforcing your AI usage policy.

OpenAI’s GPT-5 Flop, AI’s Unlimited Market, China’s Big Advantage, Rise in Socialism, Housing Crisis

Questions to inspire discussion.

📊 Q: How did GPT-5 perform compared to GPT-4? A: GPT-5 was narrowly ahead of GPT-4 in artificial analysis, but GPT-4 was significantly better in “humanity’s last exam” and RKGI2, which measures tasks relatively easy for humans but hard for AIs.

🌐 Q: What is the key architectural improvement in GPT-5? A: GPT-5 has a multimodal architecture that can self-select the underlying model for a task, providing a simple, clean interface without users needing to understand technical details.

AI industry growth and economic impact.

💰 Q: How much is being invested in the AI industry annually? A: The AI industry is experiencing astronomical growth, with hundreds of billions of dollars being deployed annually, and a projected trillion dollars in the next 5 years on data centers and AI infrastructure.

📈 Q: Are there already economic returns on AI investments? A: Economic returns on AI investments are already evident, with companies like Meta and Microsoft reporting significant revenue growth and productivity gains.

Cryonics in Space, Cryostasis Repair Science & Revival Ethics

Cryonics in space, cryostasis repair science, and revival ethics and planning are converging in 2025 to shape a bold new vision for life extension and post-biological freedom.

Join us Thursday, July 31 at 6 PM EST for a virtual service featuring two of cryonics’ leading voices:

Rudy Hoffman – Immortality Through Innovation.
Rudy opens with the visionary idea of cryonics in space and shares how today’s planning tools—annuity structures, revival trusts, and insurance-backed systems—support long-term access to biostasis. He ends with a powerful call to preserve freedom in the era of revival governance.

Alex Crouch – The Bridges to Reanimation.
Founder of Revival Research Group, Alex outlines the six bridges of cryostasis repair science, covering nanotech repair, AI orchestration, simulation, and bioprinting. His roadmap aims to make revival a transparent, collaborative goal.

Opening remarks by Neal Vanderee, officiator of the Church of Perpetual Life, connecting science, spirit, and future readiness.

Schedule:

Elon Just Made Tesla Unstoppable

Questions to inspire discussion.

🚀 Q: How might Elon Musk’s diverse projects contribute to Tesla’s value? A: Musk’s involvement in AI, energy, transportation, and communication through projects like Tesla, SpaceX, and Neuralink demonstrates his capacity to make progress on multiple fronts, potentially creating significant value for Tesla.

Political Involvement and Economic Strategy.

🏛️ Q: Why is Elon Musk getting involved in politics? A: Musk’s political involvement aims to create a better political system on Earth, addressing the unsustainability of US government spending and debt to avoid a fiscal doom loop.

📊 Q: What is Musk’s strategy to improve the US economy? A: Musk plans to accelerate GDP growth through AI-driven growth, humanoid bots, and reducing government spending and waste, potentially breaking free from the constant 7% growth line of the US stock market.

💰 Q: How could reducing government spending benefit the economy? A: By cutting wasteful spending and implementing a balanced budget requirement, the US could potentially grow its economy faster than its spending, reducing interest costs and freeing up money for other investments.

What Happens After Superintelligence? (with Anders Sandberg)

Anders Sandberg joins me to discuss superintelligence and its profound implications for human psychology, markets, and governance. We talk about physical bottlenecks, tensions between the technosphere and the biosphere, and the long-term cultural and physical forces shaping civilization. We conclude with Sandberg explaining the difficulties of designing reliable AI systems amidst rapid change and coordination risks.

Learn more about Anders’s work here: https://mimircenter.org/anders-sandberg.

Timestamps:
00:00:00 Preview and intro.
00:04:20 2030 superintelligence scenario.
00:11:55 Status, post-scarcity, and reshaping human psychology.
00:16:00 Physical limits: energy, datacenter, and waste-heat bottlenecks.
00:23:48 Technosphere vs biosphere.
00:28:42 Culture and physics as long-run drivers of civilization.
00:40:38 How superintelligence could upend markets and governments.
00:50:01 State inertia: why governments lag behind companies.
00:59:06 Value lock-in, censorship, and model alignment.
01:08:32 Emergent AI ecosystems and coordination-failure risks.
01:19:34 Predictability vs reliability: designing safe systems.
01:30:32 Crossing the reliability threshold.
01:38:25 Personal reflections on accelerating change.

Dr. Thomas Ehmer, Ph.D. — Merck KGaA Darmstadt, Germany — Quantum Computing Innovation In Pharma

Quantum Computing Innovation In Pharma — Dr. Thomas Ehmer, Ph.D. — Merck KGaA, Darmstadt, Germany


Dr. Thomas Ehmer, Ph.D. (https://www.linkedin.com/in/tehmer/) is a seasoned technology strategist with over two decades of experience in IT innovation, business development, and R&D within the pharmaceutical industry, and co-founder of the Quantum Interest Group, at Merck KGaA Darmstadt, Germany (https://www.emdgroup.com/en).

Dr. Ehmer currently is in the Sector Data Office — AI Governance and Innovation Incubator at Merck KGaA Darmstadt, Germany, where he scouts emerging and disruptive technologies, demonstrating their potential value for R&D applications, with a focus on quantum technologies.

Throughout his career at Merck KGaA Darmstadt, Germany, Dr. Ehmer has played a pivotal role in shaping IT strategy, business process optimization, and digital transformation across the entire pharmaceutical value chain, currently focusing on transparent AI and how and where emerging technology can help patients live a better life. His expertise spans technology scouting, business analysis, and IT program leadership, having successfully driven major global projects.

Beyond his corporate career, Dr. Ehmer is an active private seed investor and has contributed to quantum computing research and applications in drug discovery, authoring publications on the potential of quantum computing and machine learning in pharmaceutical R&D (https://onlinelibrary.wiley.com/doi/10.1002/9783527840748.ch26).

/* */