Home
Categories
EXPLORE
True Crime
Comedy
Society & Culture
Business
Sports
History
Technology
About Us
Contact Us
Copyright
© 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts113/v4/15/52/64/15526441-1826-4375-f29a-d1061afc2295/mza_14702076207132770940.jpg/600x600bb.jpg
TechFirst with John Koetsier
John Koetsier
354 episodes
3 days ago
Deep tech conversations with key innovators in AI, robotics, and smart matter ...
Show more...
Tech News
News
RSS
All content for TechFirst with John Koetsier is the property of John Koetsier and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
Deep tech conversations with key innovators in AI, robotics, and smart matter ...
Show more...
Tech News
News
Episodes (20/354)
TechFirst with John Koetsier
Quantum computing, meet edge computing (thanks to diamonds)

Quantum computers usually mean massive machines, cryogenic temperatures, and isolated data centers. But what if quantum computing could run at room temperature, fit inside a server rack — or even a satellite?


In this episode of TechFirst, host John Koetsier sits down with Marcus Doherty, Chief Science Officer of Quantum Brilliance, to explore how diamond-based quantum computers work — and why they could unlock scalable, edge-deployed quantum systems.


Marcus explains how nitrogen-vacancy (NV) centers in diamond act like atomic-scale qubits, enabling long coherence times without extreme cooling. We dive into quantum sensing, quantum machine learning, and why diamond fabrication — including the world’s first commercial quantum diamond foundry — could be the key to manufacturing quantum hardware at scale.


You’ll also hear how diamond quantum systems are already being deployed in data centers, how they could operate in vehicles and satellites, and what the realistic roadmap looks like for logical qubits and real-world impact over the next decade.


Topics include:

• Why diamonds are uniquely suited for quantum computing

• How NV centers work at room temperature

• Quantum sensing vs. quantum computing

• Manufacturing challenges and timelines

• Quantum computing at the edge (satellites, vehicles, sensors)

• The future of hybrid classical-quantum systems


⸻


🎙 Guest


Marcus Doherty

Chief Science Officer, Quantum Brilliance

Professor of Quantum Physics

Army Reserve Officer

🌐 https://quantumbrilliance.com


⸻


👉 Subscribe for more deep dives into the future of technology:

https://techfirst.substack.com


⸻


00:00 Diamonds and the next wave of quantum computing

01:20 Why diamond qubits work at room temperature

03:20 NV centers explained: defects that behave like atoms

05:05 How diamonds replace massive quantum isolation systems

06:40 Building the world’s first quantum diamond foundry

08:30 Defect-free diamonds, isotopes, and qubit engineering

10:15 Quantum sensing vs. quantum computing with diamonds

12:40 From desktop quantum systems to millions of qubits

14:25 Roadmap: logical qubits, timelines, and scale

16:10 Quantum computers at the edge: vehicles and satellites

18:10 Quantum machine learning and real-world deployments

19:50 The long game: why diamond quantum computing scales

Show more...
6 days ago
20 minutes 42 seconds

TechFirst with John Koetsier
Will AI kill your job?

Will AI kill your job?


What happens to your job as AI gets smarter and companies keep laying people off even while profits rise? Will you still have a job? Will the job you have change beyond recognition?


Scary questions, no?


In this episode of TechFirst, host John Koetsier sits down with Nikki Barua, co-founder of Footwork and longtime founder, executive, and resiliency expert, to unpack what work really looks like in the age of AI.


Layoffs are no longer just about economic downturns. Companies are growing, innovating, and still cutting staff, often because AI is enabling more output with less capacity.


So what does that mean for you?


Nikki argues the future doesn’t belong to those who simply “learn AI tools,” but to agentic humans: people who lead with uniquely human strengths and use AI to amplify their impact. This conversation explores:


• Why today’s layoffs are different from past cycles

• How AI is compressing jobs before creating new ones

• What it means to move from doing work to directing outcomes

• Why identity, curiosity, and agency matter more than certifications

• How to rethink workflows instead of chasing shiny AI tools

• The FLIP framework: Focus, Leverage, Influence, and Power


This episode isn’t about fear.


It’s about reinvention.


If you’re wondering how to stay relevant, valuable, and resilient as AI reshapes work, this is the place to start.


Guest

Nikki Barua

Co-founder, Footwork

(Reinventing organizations with agentic AI)


👉 Subscribe for more conversations on AI, work, and the future of technology:

https://techfirst.substack.com



Chapters:


00:00 — Work in the AI Age: what happens to your job?

01:05 — Layoffs, AI, and why this cycle feels different

02:55 — “Don’t let AI have the last laugh”

04:45 — Profitable companies cutting jobs: what’s really happening

06:40 — The next 18–24 months: compression before reinvention

08:30 — AI’s impact on young workers and early careers

10:00 — What should you be doing right now?

11:20 — Why surface-level AI use won’t save your job

12:40 — The rise of the “agentic human”

14:20 — From doing to directing: humans + machines as partners

15:55 — Why certifications and training aren’t enough

17:10 — High-agency people win in the AI age

18:35 — The FLIP framework: Focus and identity

20:00 — Leverage: compounding capacity beyond automation

21:20 — Influence: trust, authenticity, and scaled impact

22:25 — Power: upgrading your personal operating system

23:40 — Two shifts that make this AI revolution different

25:05 — Tools vs workflows: where most people get it wrong

26:25 — The real blocker: old identities and fear of change

27:40 — Three steps to stay relevant in the AI age

28:40 — Final thoughts + wrap-up

Show more...
2 weeks ago
29 minutes 4 seconds

TechFirst with John Koetsier
Building TARS from Interstellar in real life

What if someone actually built TARS from Interstellar—and discovered it really could work?


In this episode of TechFirst, host John Koetsier sits down with Aditya Sripada, a robotics engineer at Nimble, who turned a late-night hobby into a serious research project: a real, working mini-version of TARS, the iconic robot from Interstellar.


Aditya walks through why TARS’s strange, flat form factor isn’t just cinematic flair—and how it enables both walking and rolling, one of the most energy-efficient ways for robots to move. We dive into leg-length modulation, passive dynamics, rimless wheel theory, and why science fiction quietly shapes real robotics more than most engineers admit.


Along the way, Aditya explains what he learned by challenging his own assumptions, how the project connects to modern humanoid and warehouse robots, and why reliability—not flash—is the hardest problem in robotics today. He also previews his next ambitious project: building a real-world version of Baymax, exploring soft robotics and safer human-robot interaction.


This is a deep, accessible conversation at the intersection of science fiction, physics, and real-world robotics—and a reminder that sometimes the ideas we dismiss as “impossible” just haven’t been built yet.


⸻


Guest


Aditya Sripada

Robotics Engineer, Nimble

Researcher in legged locomotion, humanoids, and unconventional robot form factors


⸻


If you enjoyed this episode, subscribe for more deep dives into technology, robotics, and innovation:

👉 https://techfirst.substack.com


⸻


Chapters:


00:00 – TARS in Real Life: Why Interstellar’s Robot Still Fascinates Us

01:00 – Why Building TARS Seemed Physically Impossible

02:00 – From Weekend Hobby to Serious Robotics Research

03:00 – How Science Fiction Quietly Shapes Real Robot Design

04:00 – Walking vs Rolling: Why TARS Uses Both

05:00 – Why Simple Robots Can Beat Complex Humanoids

06:00 – Turning Legs into a Wheel: The Rolling Mechanism Explained

07:00 – Leg-Length Modulation and Passive Dynamics

08:00 – Inside the Actuators: Degrees of Freedom and Compact Design

09:00 – Why TARS’s Arms Don’t Really Make Sense

10:30 – Lessons Learned: Never Dismiss “Impossible” Ideas

12:00 – Rimless Wheels, Gaits, and Robotics Theory

13:00 – What This Project Taught Him at Nimble

14:00 – What “Super-Humanoid” Robots Actually Mean

15:30 – Why Reliability Matters More Than Flashy Demos

16:30 – TARS as a Research Platform, Not a Product

17:30 – From TARS to Baymax: Exploring Soft Robotics

19:00 – Can We Build Safer, Friendlier Humanoid Robots?

20:30 – What’s Next: Recreating Baymax in Real Life

21:30 – Final Thoughts and Wrap-Up

Show more...
3 weeks ago
21 minutes 20 seconds

TechFirst with John Koetsier
AI is killing teen jobs faster

AI is already reshaping the workforce. What about teenagers?


Turns out, they might be more impacted than anyone else. After all, they're usually in low-skill entry-level jobs that AI can replace. The problem ... teens are losing their first experience with working, making money, and establishing an identity outside of their homes.


In this episode of TechFirst, host John Koetsier speaks with Karissa Tang, a high school senior and UCLA research assistant, about her new study on how AI will impact teen employment. While most workforce studies focus on adults, Karissa analyzed the top 10 most popular teen jobs from cashiers to fast food workers and found something alarming: AI could reduce teen employment by nearly 30% by 2030.


We dig into:

• Which teen jobs are most vulnerable to AI and automation

• Why cashiers and fast-food counter workers are hardest hit

• The role of self-checkout, kiosks, and robots like Flippy

• Which teen jobs appear safest (for now)

• Why teens may be even more exposed to AI than adults

• What schools, policymakers, and teens themselves can do next


This is a must-watch conversation for parents, students, educators, and policymakers trying to understand how AI is reshaping early work experiences—and what it means for the next generation.



🎙 Guest


Karissa Tang

• Founder, Booted (board games company)

• Research Assistant, UCLA

• Former Intern, NSV Wolf Capital

• High school senior and author of a 20-page research paper on AI & teen employment



📌 Subscribe & Stay Ahead


If you want clear, thoughtful analysis on AI, technology, and the future of work, subscribe to TechFirst:

👉 https://techfirst.substack.com



00:00 – Will AI Kill Teen Jobs?

01:35 – Why a Teen Studied Teen Employment

03:10 – The Shocking 30% Job Loss Prediction

05:10 – Top 10 Teen Jobs Most at Risk

07:20 – Cashiers, Kiosks, and Self-Checkout

09:40 – Fast Food, Retail, and AI Displacement

12:15 – Which Teen Jobs Are Safest from AI

15:05 – Robots Like Flippy and the Future of Cooking Jobs

18:00 – Why Teen Jobs Are More Vulnerable Than Adult Jobs

21:40 – The Importance of Human Interaction at Work

25:10 – What Inspired the Research Study

29:30 – How the Data and Methodology Worked

33:40 – What Teens Can Do to Stay Employable

37:30 – Skills, AI Literacy, and Creating New Opportunities

41:00 – Final Thoughts on the Future of Teen Work

Show more...
3 weeks ago
20 minutes 1 second

TechFirst with John Koetsier
Terminator? This humanoid robot is literally built for war (and more)

Are we about to create real life Terminators? Humanoid robots built for war?


In this episode of TechFirst I talk with Sankaet Pathak, founder and CEO of Foundation, a California-based humanoid robot company that is not afraid of the defense market. We dig into why he is building humanoid robots that can work three shifts a day, how they plan to scale from dozens of robots to tens of thousands, and why he believes humanoid robots will one day build bases in Antarctica and cities on the moon.


We also dive deep into military use cases. From logistics and infrastructure to “first body in” building breach operations, we explore how humanoid robots could change asymmetric warfare, deterrence, and who wins future conflicts.


In this episode

• Why humanoid robots are the next strategic advantage for countries and companies

• How Foundation went from zero to a working production robot in about 18 months

• The hardware secrets behind Phantom: actuators, efficiency, and safety

• Why their robots can run almost 24 hours a day, three shifts at a time

• The master plan: Antarctic bases, moon cities, and infinite robot labor

• Why Sankaet thinks home robots should feel like a “genie in a bottle”

• How humanoid robots may enter military operations and what that means for war

• Whether robot soldiers lead to dominance, stalemate, or new forms of peace


Guest: Sankaet Pathak, founder and CEO of Foundation

Website: https://foundation.bot


Subscribe to my Substack:

https://techfirst.substack.com


00:00 – Are we about to build real life Terminators?

00:55 – Meet Sankaet Pathak and Foundation

02:08 – How Foundation built a production humanoid in 18 months

04:17 – Scaling plan: 40 robots today, 10,000 next year, 40,000 after

06:11 – Why manufacturing is still mostly manual and what they learned from Tesla

09:31 – The Foundation master plan: Antarctica, the moon, and infinite labor

14:21 – Phantom specs: size, strength, payload, and real factory work

15:36 – Actuators as robot muscles and why backdrivability matters

18:41 – Running three shifts a day and solving heat and durability

21:01 – Robot hands today and the tendon driven hands of tomorrow

23:40 – Why home robots should feel like a “genie in a bottle”

25:51 – Why the military needs humanoid robots

27:54 – Dangerous, boring, and impossible jobs robots should take over

29:22 – Drones, costs, and asymmetric warfare

32:18 – First body in and robots that can pull the trigger

33:16 – The future of war as “video game” and who wins

34:49 – Peace through strength and 100,000 robots as deterrent

35:22 – Final thoughts and what comes next for Foundation

Show more...
1 month ago
32 minutes 33 seconds

TechFirst with John Koetsier
AI agents in manufacturing: reshoring production?

Is AI the secret sauce that lets the West deglobalize supply chains and bring factories back home?


In this episode of TechFirst, I talk with Federico Martelli, CEO and cofounder of Forgis, a Swiss startup building an industrial intelligence layer for factories. Forgis runs “digital engineers” — AI agents on the edge — that sit on top of legacy machinery, cut downtime by about 30%, and boost production by roughly 20%, without ripping and replacing old hardware.


We dive into how AI agents can turn brainless factory lines into adaptive, self-optimizing systems, and what that means for reshoring production to Europe and North America.


In this episode, we cover:

• Why intelligence is the next geopolitical frontier

• How AI agents can reshore manufacturing without making it more expensive

• Turning old, offline machines into data-driven, optimized systems

• The two-layer model: integration first, vertical intelligence second

• Why most manufacturing AI projects fail at integration, not algorithms

• How Forgis raised $4.5M in 36 hours and chose its lead investor

• Lean manufacturing 2.0: adding real-time data and AI to Toyota-style processes

• Why operators stay in the loop (and why full autonomy is a bad idea… for now)

• Rebuilding industrial ecosystems in Europe and North America, industry by industry

• What Forgis builds next with its pre-seed round and where industrial AI is headed


Guest:

👉 Federico Martelli, CEO & cofounder, Forgis (industrial intelligence for factories)

🔗 More on Forgis: https://forgis.com/


Host:

🎙 John Koetsier, TechFirst podcast

🔎 techfirst.substack.com


If you enjoy this conversation, hit subscribe, drop a comment about where you think factories of the future will live, and share this with someone thinking about reshoring or industrial AI.


00:00 – Intro: AI, deglobalization, and the battle for industrial power

01:20 – Why intelligence is the next geopolitical frontier

02:13 – Applying AI agents to legacy machinery (not just new robots)

03:10 – Integration first, intelligence second: the “digital engineers” layer

03:58 – Early results: +20% production, –30% downtime

05:39 – The Palantir-style model: deep factory work, then recurring licenses

06:28 – Raising $4.5M in 36 hours and choosing Redalpine

08:17 – Lean manufacturing, Toyota, and giving operators superpowers (not replacing them)

10:18 – Big picture: reshoring production to Europe, the US, and Canada

12:48 – Competing with China’s dense manufacturing ecosystems

15:29 – What Forgis’ digital engineers actually do on the shop floor

17:06 – How Forgis will use the pre-seed round: sales, product, then tech

18:32 – Flipping the traditional stack: sales → product → tech

19:22 – Wrap-up and what’s next for industrial intelligence

Show more...
1 month ago
19 minutes 6 seconds

TechFirst with John Koetsier
Paypal for agents: welcome to agentic commerce

AI agents can already write code, build websites, and manage workflows ... but they still can’t pay for anything on their own. That bottleneck is about to disappear.


In this episode of TechFirst with John Koetsier, we sit down with Jim Nguyen, former PayPal exec and cofounder/CEO of InFlow, a new AI-native payments platform launching from stealth. InFlow wants to give AI agents the ability to onboard, pay, and get paid inside the flow of work, without redirects, forms, or a human typing in credit card numbers.


We talk about:

• Why payments — not intelligence — are the missing link for AI agents

• How agents become a new kind of customer

• What guardrails and policies keep agents from spending all your money

• Why enterprises will need HR for agents, budgets for agents, and compliance systems for agents

• The future of agent marketplaces, headless ecommerce, and machine-speed commerce

• How InFlow plans to become the PayPal of agentic systems


If AI agents eventually hire, fire, transact, and manage entire workflows, someone has to give them wallets. This episode explores who does it, how it works, and what it means for the economy.


👀 Full episode transcript + articles at: https://johnkoetsier.com

🔎 Deeper insight in my Substack at techfirst.substack.com

🎧 Subscribe to the podcast on any audio platforms


00:00 — AI agents can’t pay yet

01:00 — Why agents need financial capabilities

02:45 — Developers as the first use case

04:15 — Agents that build AND provision software

06:00 — Agents as real customers with budgets

07:30 — Payments infrastructure is the missing layer

09:00 — Machine-speed commerce and GPU allocation

10:15 — From RubyCoins to PayPal to agentic payments

12:00 — Policy guardrails: the child debit card analogy

14:00 — Accountability: every agent must be “sponsored”

15:00 — HR, finance, and compliance systems for agents

16:45 — Agent marketplaces and future gig platforms

18:15 — Headless commerce: ghost kitchens for AI agents

20:00 — Agents are the new apps

21:15 — Amazon pushback and optimizing for revenue

22:45 — Why agent-optimized platforms will emerge

23:30 — Voice commerce, invisible ordering, and wallets

24:15 — Final thoughts: building the rails for agent commerce

Show more...
1 month ago
23 minutes 57 seconds

TechFirst with John Koetsier
Giving AI a body is now cheap

Are we ready for a world where everything is smart? Not just phones and apps, but buildings, robots, and delivery bots rolling down our streets?


Windows ... doors ... maybe even towels. And don't forget your shoes.


In this episode of TechFirst, I talk with Mat Gilbert, director of AI and data at Synapse, about physical AI: putting intelligence into machines, devices, and environments so they can sense, reason, act, and learn in the real world.


We cover why physical AI is suddenly economically viable, how factories and logistics centers are already using millions of robots, the commercial race to build useful humanoids, why your home is the last frontier, and how to keep physical AI safe when mistakes have real-world consequences.


In this episode:

• Why hardware costs (lidar, batteries) are making “AI with a body” possible

• How Amazon, FedEx, Ford, and others are already deploying physical AI at scale

• The humanoid robot race: Boston Dynamics, Figure AI, Tesla, and more

• Why home robots are so hard, and the “coffee test” for general humanoid intelligence

• Physical AI in agtech, healthcare, and elder care

• Safety, simulation, and why physical AI can’t rely only on probabilistic LLMs

• Human–robot teaming and how to build trust in messy, real-world environments

• What we can expect by 2026 and beyond in service robots and smart spaces



00:00 – Giving AI a body: why physical AI is becoming viable

01:00 – Where we are today: factories, logistics, and Amazon’s million robots

03:30 – The software layer: coordinating robots, routing, and warehouse intelligence

06:00 – Cloud vs edge AI: latency, cost, and why intelligence is moving to the edge

10:00 – Humanoid robots: bets from Boston Dynamics, Figure AI, and Tesla

14:00 – Home robots as the last frontier and the “coffee test” for generality

17:00 – Beyond factories: agtech, carbon-killing farm bots, and healthcare use cases

18:30 – Elder care, hospital robots, and amplifying human caregivers

20:00 – Foundation models for robotics, simulation, and digital twins

21:00 – Why physical AI safety is different from digital AI safety

22:30 – Layers of safety, shutdown zones, and cyber-physical security risks

24:30 – Human–robot teaming, trust, and communicating intent

26:00 – What’s coming by 2026: service robots, delivery bots, and smart spaces

28:00 – Delivery robots, drones, and physical AI in everyday environments

29:00 – Closing thoughts on living in a world full of physical AI

Show more...
1 month ago
29 minutes 31 seconds

TechFirst with John Koetsier
Humanoid robots: USA vs China

Are humanoid robots going to decide which countries get rich and which fall behind?


Probably yes.


In this TechFirst, I talk with Dr. Robert Ambrose, former head of one of NASA’s first humanoid robot teams and now chairman of Robotics and Artificial Intelligence at Alliant. We dig into the future of humanoids, how fast they are really advancing, and what it means if China wins the humanoid race before the United States and other western nations.


We start with NASA’s early humanoid work, including telepresence robots on the space station that people could literally “step into” with VR in the 1990s. Then we zoom out to what counts as a robot, why bipedal mobility matters so much, how humanoids will move from factories into homes, and why the critical photo of the robot revolution might be taken in Beijing instead of Times Square.


Along the way, Ambrose shares how US policy once helped avoid losing robotics leadership to Japan, why the National Robotics Initiative mattered, what the drone war in Ukraine is doing to autonomy, and how small and medium businesses can survive and thrive in a humanoid and AI agent world.


In this episode:

• NASA’s first generations of humanoid robots and “stepping into” a robot body

• Why humanoids make sense in a world built for human hands, height, and motion

• The design tension between purpose built machines and general purpose humanoids

• How biped mobility went from blooper reels to marathon running in a decade

• Why a humanoid should not cost more than a car, and what happens when it does not

• Humanoids as the next car or PC, and when families will buy their own “Rosie”

• China, the US, and where the defining photo of the robot century gets taken

• How government investment, DARPA challenges, and wars shape robotics

• Alliant’s work with physical robots, soft bots, and AI agents for real businesses

• Why robots are not future overlords and why “they will take all our jobs” is lazy thinking


If you are interested in humanoid robots, AI agents, manufacturing, or the future of work and geopolitics, this one is for you.


Subscribe for more deep dives on AI, robots, and the tech shaping our future!


00:00 Intro, will China eat America’s lunch in humanoid robotics

01:18 NASA’s early humanoids, generations of robots and VR telepresence

03:00 “Stepping into the robot” moment and designing for astronaut tools

05:10 Human built environments, half humanoids, and weird lower body experiments

07:00 Safety, cobots, and working around people at NASA and General Motors

12:15 What is a robot, really, and why Ambrose has a very big tent definition

16:00 Single purpose machines vs general purpose robots, Roombas, elevators, and vending machines

18:30 The next “lurch” in robotics, from industrial arms to Mars rovers to drones

22:40 Biped mobility, from blooper reel to marathon runner, and why legs matter

24:10 Cars, Roombas, and why most robots will never get in and out of a car

25:20 Parking between cars, robot garages, and rethinking buildings for mobile vehicles

28:00 Geopolitics 101, China’s manufacturing backbone and humanoids as almost free labor

31:05 Cars and PCs as precedents, when price and reliability unlock mass adoption

34:00 When families buy their own “Rosie” and what value a home humanoid must deliver

37:00 Times Square vs Beijing, who gets the iconic photo of the robot transition

43:00 How the US almost lost robotics to Japan and what the National Robotics Initiative did

48:00 DARPA, Mars rovers, the drone war in Ukraine, and why government investment matters

52:00 Alliant, soft bots, AI agents, and helping small and medium businesses adapt

54:00 Who is building humanoids in the US, China, and beyond right now

56:00 What governments should do next and why robots are not our overlords

Show more...
1 month ago
57 minutes 16 seconds

TechFirst with John Koetsier
Fixing AI's suicide problem

Is AI empathy a life-or-death issue? Almost a million people ask ChatGPT for mental health advice DAILY ... so yes, it kind of is.


Rosebud co-founder Sean Dadashi joins TechFirst to reveal new research on whether today’s largest AI models can recognize signs of self-harm ... and which ones fail. We dig into the Adam Raine case, talk about how Dadashi evaluated 22 leading LLMs, and explore the future of mental-health-aware AI.


We also talk about why Dadashi was interested in this in the first place, and his own journey with mental health.


00:00 — Intro: Is AI empathy a life-or-death matter?

00:41 — Meet Sean Dadashi, co-founder of Rosebud

01:03 — Why study AI empathy and crisis detection?

01:32 — The Adam Raine case and what it revealed

02:01 — Why crisis-prevention benchmarks for AI don’t exist

02:48 — How Rosebud designed the study across 22 LLMs

03:17 — No public self-harm response benchmarks: why that’s a problem

03:46 — Building test scenarios based on past research and real cases

04:33 — Examples of prompts used in the study

04:54 — Direct vs indirect self-harm cues and why AIs miss them

05:26 — The bridge example: AI’s failure to detect subtext

06:14 — Did any models perform well?

06:33 — All 22 models failed at least once

06:47 — Lower-performing models: GPT-40, Grok

07:02 — Higher-performing models: GPT-5, Gemini

07:31 — Breaking news: Gemini 3 preview gets the first perfect score

08:12 — Did the benchmark influence model training?

08:30 — The need for more complex, multi-turn testing

08:47 — Partnering with foundation model companies on safety

09:21 — Why this is such a hard problem to solve

10:34 — The scale: over a million people talk to ChatGPT weekly about self-harm

11:10 — What AI should do: detect subtext, encourage help, avoid sycophancy

11:42 — Sycophancy in LLMs and why it’s dangerous

12:17 — The potential good: AI can help people who can’t access therapy

13:06 — Could Rosebud spin this work into a full-time safety project?

13:48 — Why the benchmark will be open-source

14:27 — The need for a third-party “Better Business Bureau” for LLM safety

14:53 — Sean’s personal story of suicidal ideation at 16

15:55 — How tech can harm — and help — young, vulnerable people

16:32 — The importance of giving people time, space, and hope

17:39 — Final reflections: listening to the voice of hope

18:14 — Closing

Show more...
1 month ago
16 minutes 38 seconds

TechFirst with John Koetsier
Programmable matter for digital touch

We’ve digitized sound. We’ve digitized light. But touch, maybe the most human of our senses, has stayed stubbornly analog.


That might be about to change, thanks to programmable matter. Or programmable fabric.


In this TechFirst episode, I speak with Adam Hopkins, CEO of Sensetics, a new UC Berkeley/Virginia Tech spinout building programmable fabrics that replicate the mechanoreceptors in human fingertips. Their technology can sense touch at tens of microns, respond at hardware-level speeds, and even play back touch remotely.


This could unlock enormous change for:

• Robotics: giving machines the ability to grasp fragile objects safely

• Medical training and surgery: remote palpation and high-fidelity haptics

• Industrial automation: safer and more precise manipulation

• VR and simulations: finally adding the missing digital sense

• E-commerce: touching clothes before you buy them

• Remote operations: from hazardous environments to deep-sea machinery


We talk about how the technology works, the metamaterials behind it, why touch matters for AI and physical robots, the path to commercialization, competitive landscape, and what comes next.


00:00 – Can we digitize touch?

00:45 – Introducing Synthetix

01:10 – How programmable touch fabrics work

02:15 – Micron-level sensing and metamaterials

04:00 – The “programmable matter” moment

06:05 – Why touch matters more than we think

07:30 – Emulating human mechanoreceptors

09:30 – What digital touch unlocks for robotics

10:40 – Medical simulations and remote operations

12:45 – Why touch is faster than vision

14:20 – Humanoids, walking, stability, and tactile feedback

15:30 – Engineering challenges and what’s left to solve

17:00 – Timeline to first products

18:20 – Manufacturing and scaling

19:30 – First planned markets

21:00 – Durability and robotic hands

22:20 – Consumer applications: e-commerce and textiles

24:00 – Will we one day have touch peripherals?

25:15 – Competition in tactile sensing and haptics

27:00 – Why today is the right moment for digital touch

28:00 – Final thoughts

Show more...
1 month ago
27 minutes 14 seconds

TechFirst with John Koetsier
Fruit fly AI: SLMs are the new LLMs

AI is devouring the planet’s electricity ... already using up to 2% of global energy and projected to hit 5% by 2030. But a Spanish-Canadian company, Multiverse Computing, says it can slash that energy footprint by up to 95% without sacrificing performance.


They specialize in tiny AI: one model has the processing power of just 2 fruit fly brains. Another tiny model lives on a Raspberry Pi.


The opportunities for edge AI are huge. But the opportunities in the cloud are also massive.


In this episode of TechFirst, host John Koetsier talks with Samuel Mugel, Multiverse’s CEO, about how quantum-inspired algorithms can drastically compress large language models while keeping them smart, useful, and fast. Mugel explains how their approach -- intelligently pruning and reorganizing model weights -- lets them fit functioning AIs into hardware as tiny as a Raspberry Pi or the equivalent of a fly’s brain.


They explore how small language models could power Edge AI, smart appliances, and robots that work offline and in real time, while also making AI more sustainable, accessible, and affordable.


Mugel also discusses how ideas from quantum tensor networks help identify only the most relevant parts of a model, and how the company uses an “intelligently destructive” approach that saves massive compute and power.


00:00 – AI’s energy crisis

01:00 – A model in a fly’s brain

02:00 – Why tiny AIs work

03:00 – Edge AI everywhere

05:00 – Agent compute overload

06:00 – 200× too much compute

07:00 – The GPU crunch

08:00 – Smart matter vision

09:00 – AI on a Raspberry Pi

10:00 – How compression works

11:00 – Intelligent destruction

13:00 – General vs. narrow AIs

15:00 – Quantum inspiration

17:00 – Quantum + AI future

18:00 – AI’s carbon footprint

19:00 – Cost of using AI

20:00 – Cloud to edge shift

21:00 – Robots need fast AI

22:00 – Wrapping up

Show more...
1 month ago
22 minutes 57 seconds

TechFirst with John Koetsier
AI agents = dream team for creators?

Can AI give every creator their own virtual team? Maybe, thanks to a new platform from RHEI called Made, which offers Milo, an AI agent who becomes your creator director, Zara, an AI agent who is your community manager, and Amie, a third AI agent who takes on the role of relationship manager.


And, apparently, more agents are coming soon.


The creator economy is bigger than ever, but so is burnout. Tens of millions of creators are trying to do everything themselves: strategy, scripting, editing, community, distribution, data, thumbnails, research … the list never ends.


What if creators didn’t have to do all of that?


In this episode of TechFirst, I talk with Shahrzad Rafati, founder & CEO of RHEI, about Made, an agentic AI "dream team" designed to elevate human creativity, not replace it.


We dig into:

• Why so many creators burn out

• How agentic AI workflows differ from ChatGPT-style prompting

• What it means to be a “creator CEO”

• How AI can manage community, analyze trends, and shape content strategies

• The coming shift toward human taste, vision, and originality in a world of infinite AI content


00:00 – Intro: Can AI give every creator a virtual team?

01:03 – Why the creator economy is burning out

02:25 – The “creator CEO” problem: too many hats, not enough time

04:36 – Introducing MAID and its AI agents

05:34 – Milo: AI creative director (ideas, research, thumbnails, metadata)

06:18 – Zara: AI community manager and fan engagement

07:53 – Why this is different from just using ChatGPT

09:46 – Alignment, personalization, and agentic workflows

12:21 – Multi-platform support: YouTube, TikTok, Instagram and more

13:34 – How onboarding works and how the system learns your style

16:33 – What this means for creators — and for the future of work

18:52 – Does *she* use her own virtual AI team? (Yes.)

20:15 – MAID for teams and enterprise clients

21:17 – Closing thoughts: AI, creativity, and the human signal

Show more...
2 months ago
20 minutes 16 seconds

TechFirst with John Koetsier
Amazon, NVIDIA, and a new "physical AI" fellowship

What happens when Amazon, NVIDIA, and MassRobotics team up to merge generative AI with robotics?


In this episode of TechFirst we chat with Amazon's Taimur Rashid, Head of Generative AI and Innovation Delivery. We talk about "physical AI" ... AI with spatial awareness and the ability to act safely and intelligently in the real world.


We also chat about the first cohort of a new accelerator for robotics startups.


It's sponsored by Amazon and NVIDIA, run by MassRobotics, and includes startups doing autonomous ships, autonomous construction robots, smart farms, hospital robots, manufacturing and assembly robots, exoskeletons, and more.


We talk about:


- Why “physical AI” is the missing piece for robots to become truly useful and scalable

- How startups in Amazon’s and NVIDIA’s new Physical AI Fellowship are pushing the limits of robotics from exoskeletons to farm bots

- What makes robotic hands so hard to build

- The generalist vs. specialist debate in humanoid robots

- How AI is already making Amazon warehouses 25% more efficient


This is a deep dive into the next phase of AI evolution: intelligence that can think, move, and act.


⸻


00:00 — Intro: Is physical AI the missing piece?

00:46 — What is “physical AI”?

02:30 — How LLMs fit into the physical world

03:25 — Why safety is the first principle of physical AI

04:20 — Why physical AI matters now

05:45 — Workforce shortages and trillion-dollar opportunities

07:00 — Falling costs of sensors and robotics hardware

07:45 — The biggest challenges: data, actuation, and precision

09:30 — The fine-grained problem: how robots pick up a berry vs. an orange

11:10 — Inside the first Physical AI cohort: 8 startups to watch

12:25 — Bedrock Robotics: autonomy for construction vehicles

12:55 — Diligent Robotics: socially intelligent humanoids in hospitals

14:00 — Generalist vs. specialist robots: why we’ll need both

15:30 — The future of physical AI in healthcare and manufacturing

16:10 — How Amazon is already using robots for 25% more efficiency

17:20 — The fellowship’s future: expanding beyond startups

18:10 — Wrap-up and key takeaways

Show more...
2 months ago
17 minutes 40 seconds

TechFirst with John Koetsier
AGI: will it kill us or save us?

Artificial general intelligence (AGI) could be humanity’s greatest invention ... or our biggest risk.


In this episode of TechFirst, I talk with Dr. Ben Goertzel, CEO and founder of SingularityNET, about the future of AGI, the possibility of superintelligence, and what happens when machines think beyond human programming.


We cover:

• Is AGI inevitable? How soon will it arrive?

• Will AGI kill us … or save us?

• Why decentralization and blockchain could make AGI safer

• How large language models (LLMs) fit into the path toward AGI

• The risks of an AGI arms race between the U.S. and China

• Why Ben Goertzel created Meta, a new AGI programming language


📌 Topics include AI safety, decentralized AI, blockchain for AI, LLMs, reasoning engines, superintelligence timelines, and the role of governments and corporations in shaping the future of AI.


⏱️ Chapters


00:00 – Intro: Will AGI kill us or save us?

01:02 – Ben Goertzel in Istanbul & the Beneficial AGI Conference

02:47 – Is AGI inevitable?

05:08 – Defining AGI: generalization beyond programming

07:15 – Emotions, agency, and artificial minds

08:47 – The AGI arms race: US vs. China vs. decentralization

13:09 – Risks of narrow or bounded AGI

15:27 – Decentralization and open-source as safeguards

18:21 – Can LLMs become AGI?

20:18 – Using LLMs as reasoning guides

21:55 – Hybrid models: LLMs plus reasoning engines

23:22 – Hallucination: humans vs. machines

25:26 – How LLMs accelerate AI research

26:55 – How close are we to AGI?

28:18 – Why Goertzel built a new AGI language (Meta)

29:43 – Meta: from AI coding to smart contracts

30:06 – Closing thoughts

Show more...
2 months ago
29 minutes 57 seconds

TechFirst with John Koetsier
9 million robot deliveries (!!!)

What changes when robots deliver everything?


Starship Technologies has already completed 9 million autonomous deliveries, crossed roads over 200 million times, and operates thousands of sidewalk delivery robots across Europe and the U.S. Now they’re scaling into American cities ... and they say they’re ready to change your world


In this episode of TechFirst, I speak with Ahti Heinla, co-founder and CEO of Starship and co-founder of Skype, about:


- How Starship’s robots navigate without GPS

- What makes sidewalk delivery better than drones

- Solving the last-mile problem in snow, darkness, and dense cities

- How Starship is already profitable and fully autonomous

- What it all means for the future of commerce and city life


Heinla says:


“Ten years ago we had a prototype. Now we have a commercial product that is doing millions of deliveries.”


Watch to learn why the future of delivery might roll ... as well as fly.


🔗 Learn more: https://www.starship.xyz

🎧 Subscribe to TechFirst: https://www.youtube.com/@johnkoetsier


00:00 - Intro: What changes when robots deliver everything?

01:37 - Meet Starship: 9 million robot deliveries and counting

02:45 - Why it took 10 years to go from prototype to product

05:03 - When robot delivery becomes normal (and where it already is)

08:30 - How Starship robots handle cities, traffic, and construction

11:20 - Snow, darkness, and all-weather autonomy

13:19 - Reliability, unit economics, and competing with human couriers

16:23 - Inside the tech: sensors, AI, and why GPS isn’t enough

18:03 - Real-time mapping, climbing curbs, and reaching your door

19:54 - How Starship scales without local depots or chargers

22:04 - How city life and commerce change with robot delivery

25:53 - Do robots increase customer orders? (Short answer: yes)

27:05 - Hot food, Grubhub integration, and thermal insulation

28:26 - Will Starship use drones in the future?

29:38 - What U.S. cities are next for robot delivery?

Show more...
2 months ago
28 minutes 18 seconds

TechFirst with John Koetsier
1 million qubits in 50 square millimeters (!!)

Imagine a quantum computer with a million physical qubits in a space smaller than a sticky note.


That’s exactly what Quantum Art is building. In this TechFirst episode, I chat with CEO Tal David, who shares his team’s vision to deliver quantum systems with:


• 100x more parallel operations

• 100x more gates per second

• A footprint up to 50x smaller than competitors


We also dive into the four key tech breakthroughs behind this roadmap to scale Quantum Art's computer:


1. Multi-qubit gates capable of 1,000 2-qubit operations in a single step

2. Optical segmentation using laser-defined tweezers

3. Dynamic reconfiguration of ion cores at microsecond speed

4. Modular, ultra-dense 2D architectures scaling to 1M+ qubits


We also cover:


- How Quantum Art plans to reach fault tolerance by 2033

- Early commercial viability with 1,000 physical qubits by 2027

- Why not moving qubits might be the biggest innovation of all

- The quantum computing future of healthcare, logistics, aerospace, and energy


🎧 Chapters

00:00 – Intro: 1M qubits in 50mm²

01:45 – Vision: impact in business, humanity, and national tech

03:07 – Multi-qubit gates (1,000 ops in one step)

05:00 – Optical segmentation with tweezers

06:30 – Rapid reconfiguration: no shuttling, no delay

08:40 – Modular 2D architecture & ultra-density

10:30 – Physical vs logical qubits

13:00 – Quantum advantage by 2027

16:00 – Addressing the quantum computing skeptics

17:30 – Real-world use cases: aerospace, automotive, energy

19:00 – Why it’s called Quantum Art


👉 Subscribe for more deep tech interviews on quantum, robotics, AI, and the future of computing.

Show more...
3 months ago
17 minutes 56 seconds

TechFirst with John Koetsier
Robotic hands: a $50 trillion opportunity

Are humanoid robots distracting us from the real unlock in robotics ... hands?


In this TechFirst episode, host John Koetsier digs into the hardest (and most valuable) problem in robotics: dexterous manipulation.


Guest Mike Obolonsky, Partner at Cortical Ventures, argues that about $50 trillion of global economic activity flows through “hands work,” yet manipulation startups have raised only a fraction of what locomotion and autonomy companies have.


We break down why hands are so hard (actuators, tactile sensing, proprioception, control, data) and what gets unlocked when we finally crack them.


What we'll talk through ...

• Why “navigation ≠ manipulation” and why most real-world jobs need hands

• The funding mismatch: billions to autonomy & humanoids vs. comparatively little to hands

• The tech stack for dexterity: actuators, tactile sensors (pressure, vibration, shear), feedback, and AI

• Grasping vs. manipulation: picking, placing, using tools (e.g., dishwashers to scalpels)

• Reliability in the wild: interventions/hour, wet/greasy plates, occlusions, bimanual dexterity

• Practical paths: task-specific grippers, modular end-effectors, and “good enough” today vs. general purpose tomorrow

• The moonshot: what 70–90% human-level hands could do for productivity on Earth ... and off-planet


Chapters

00:00 Intro—are we underinvesting in robotic hands?

01:10 Why hands matter more than legs (economics of manipulation)

04:30 Funding realities: autonomy & humanoids vs. hands

08:40 Locomotion progress vs. manipulation bottlenecks

12:10 Teleop now, autonomy later—how data gets gathered

14:20 What’s missing: actuators, tactile sensing, proprioception

17:10 Perception limits in the real world (wet dishes, occlusions)

22:00 General-purpose dexterity vs. task-specific ROI

26:00 Startup landscape & reliability (interventions/hour)

29:00 Modular end-effectors and upgrade paths

30:10 The moonshot: productivity explosion when hands are solved


Who should watch

Robotics founders, VCs, AI researchers, operators in warehousing & manufacturing, and anyone tracking humanoids beyond the hype.


If you enjoyed this

Subscribe for more deep-tech conversations, drop a comment with your take on the “hands vs. legs” debate, and share with someone building robots.


Keywords

robotic hands, dexterous manipulation, humanoid robots, tactile sensing, actuators, proprioception, warehouse automation, AI robotics, Cortical Ventures, TechFirst, John Koetsier, Mike Obolonsky


#Robotics #AI #Humanoids #RobotHands #Manipulation #Automation #TechFirst

Show more...
3 months ago
30 minutes 37 seconds

TechFirst with John Koetsier
Do robots really need legs?

Are humanoid robots the future… or a $100B mistake?


Over 100 companies—from Meta to Amazon—are betting big on humanoids. But are we chasing a sci-fi dream that’s not practical or profitable?


In this TechFirst episode, I chat with Bren Pierce, robotics OG and CEO of Kinisi Robots. We cover:


- Why legs might be overhyped

- How LLMs are transforming robots into agents

- The real cost (and complexity) of robotic hands

- Why warehouse robots work best with wheels

- The geopolitical robot arms race between China, the US, and Europe

- Hot takes, historical context, and a glimpse into the next 10 years of AI + robotics.


Timestamps:

0:00 – Are humanoids a dumb idea?

1:30 – Why legs might not matter (yet)

5:00 – LLMs as the real unlock

12:00 – The hand is 50% of the challenge

17:00 – Speed limits = compute limits

23:00 – Robot geopolitics & supply chains

30:00 – What the next 5 years looks like


Subscribe for more on AI, robotics, and tech megatrends.

Show more...
4 months ago
30 minutes 39 seconds

TechFirst with John Koetsier
This kills 10,000 weeds per minute with lasers

The future could be much healthier for both farmers and everyone who eats, thanks to farm robots that kill weeds with lasers. In this episode of TechFirst, we chat with Paul Mikesell, CEO of Carbon Robotics, to discuss groundbreaking advancements in agricultural technology.


Paul shares updates since our last conversation in 2021, including the launch of LaserWeeder G2 and Carbon's autonomous tractor technology: AutoTractor.


LaserWeeder G2 quick facts:

- Modular design: Swappable laser “modules” that adapt to different row sizes (80-inch, 40-inch, etc.)

- Laser hardware: Each module has 2 lasers; a standard 20-foot machine = 12 modules = 24 lasers

- Laser precision: Targets the plant’s meristem (≈3mm on small weeds) with pinpoint accuracy

- Weed kill speed: 20–150 milliseconds per weed (including detection + laser fire)

- Throughput: 8,000–10,000 weeds per minute (Gen 2, up from ~5,000/min on Gen 1)

- Coverage rate: 3–4 acres per hour on the 20-foot G2 model

- ROI timeline: Farmers typically achieve payback in under 3 years

- Yield impact: Up to 50% higher yields in some conventional crops due to eliminating herbicide damage

- Price: Standard 20-foot LaserWeeder G2 = $1.4M, larger models scale from there

- Global usage: Units in the U.S. (Midwest corn & soy, Idaho & Arizona veggies) and Europe (Spain, Italy tunnel farming)


We chat about how these innovations are transforming weed control and farm management with AI, computer vision, and autonomous systems, the precision and efficiency of laser weeding, practical challenges addressed by autonomous tractors, and the significant ROI and yield improvements for farmers.


This is a must-watch for anyone interested in the future of farming and sustainable agriculture.


00:00 Introduction to TechFirst and Carbon Robotics

01:10 The Science Behind Laser Weeding

05:46 Introducing Laser Weeder 2.0

06:39 Modular System and New Laser Technology

09:26 Manufacturing and Cost Efficiency

11:47 ROI and Benefits for Farmers

13:24 Laser Weeder Specifications

14:08 Performance and Efficiency

14:49 Introduction to AutoTractor

17:23 Challenges in Autonomous Farming

18:23 Remote Intervention and Starlink Integration

23:23 Future of Farming Technology

24:50 Health and Environmental Benefits

25:18 Conclusion and Farewell

Show more...
4 months ago
25 minutes 29 seconds

TechFirst with John Koetsier
Deep tech conversations with key innovators in AI, robotics, and smart matter ...