MemCast

AI Interfaces Of The Future | Design Review

Exploring cutting-edge AI interfaces that move beyond chat UIs, featuring voice agents, autonomous AI workflows, adaptive interfaces, and AI-generated video production.

36m·Guest Raphael Shad·Host Aaron·

Verbs Over Nouns

1 / 17

The shift from static UI elements (nouns) to dynamic workflows (verbs) as the foundation of AI interfaces, requiring new design paradigms to visualize and control autonomous actions.

Software today is mostly nouns; AI requires verbs for workflows and actions.
  • Static interfaces use nouns like text forms, buttons
  • AI introduces verbs like auto-complete, auto-suggest, gather info
  • Current tools lack ways to visually represent verbs on screen
software of today or kind of like up until this point was mostly kind of like just clear things you can point out on the screen um that are you know kind of nouns like text forms drop downs buttons Etc and with AI what really changes is I think so much of the design of what AI does is kind of more verbs um it's more the workflows Auto Complete Auto suggest um go out and gather some information for me Etc and we don't really have the tooling yet to kind of draw verbs on the screen Rafael Shad
these are all verbs we're creating videos we have agents going out executing tasks and so much of it is how do you keep the user in the loop and in control while AI does its magic Host
AI interfaces are at a touch-device 2010 moment, requiring full reimagining of software components.
  • Similar to how touch devices forced UI redesigns
  • Every component is being rethought for AI-native interactions
  • Current static interfaces are being replaced by dynamic, action-oriented designs
it almost feels like back in like 2010 or so when touch um devices really kind of came on the market and everything had to reinvented kind of Touch first and we're at one of those moments again where like all of software all the components that we kind of took for granted um they are really being reimagined and reshaped by the builders and startups and designers out there right now Host
when we start first kind of started to get this llm technology everything was sort of like a chat box and people just kind of like prompting it and now within just a few short like like a few short months or or one two years we see this explosion of AI interface and AI components that really kind of are built AI natively um totally different modalities how to interact with this new teolog with the llms um and really just endless opportunity for uh iteration and uh building a new world of software Rafael Shad
Current tools lack visualization for AI workflows as verbs.
  • No existing design patterns for drawing verbs on screen
  • Traditional UI elements (buttons, forms) are nouns, not actions
  • Requires new metaphors for autonomous agent workflows
we don't really have the tooling yet to kind of draw verbs on the screen and so that's what's really fascinating how you know this software is now emerging in this new AO World Rafael Shad
from a high level what are the differences between kind of the say static web-based uh 2D interfaces that we're used to today with where things are going in the future Host
Opening question about interface evolution

From Nouns to Verbs: The Shift in AI Interface Design

2 / 17

Traditional interfaces focused on static elements (nouns) like buttons and forms, while AI interfaces emphasize actions (verbs) like auto-complete and autonomous workflows. This shift requires new design paradigms to visualize dynamic processes.

AI interfaces focus on workflows and actions rather than static UI elements
  • Traditional software UI consists of static elements like buttons, forms, and dropdowns (nouns)
  • AI interfaces emphasize dynamic processes like auto-complete, auto-suggest, and information gathering (verbs)
  • Current design tooling isn't optimized for visualizing these dynamic workflows
  • The challenge is representing actions visually in interfaces where the AI handles complex processes
software of today or kind of like up until this point was mostly kind of like just clear things you can point out on the screen um that are you know kind of nouns like text forms drop downs buttons Etc and with AI what really changes is I think so much of the design of what AI does is kind of more verbs um it's more the workflows Auto Complete Auto suggest um go out and gather some information for me Etc Raphael Shad
we don't really have the tooling yet to kind of draw verbs on the screen and so that's what's really fascinating how you know this software is now emerging in this new AO World Raphael Shad
Latency becomes a critical UI element in conversational interfaces
  • In voice interfaces, response latency directly impacts perceived naturalness
  • Sub-200ms responses feel human-like while longer delays reveal the robotic nature
  • Some interfaces now expose latency metrics to help developers understand thresholds
  • Visual feedback during voice interactions helps maintain user confidence
the latency is the interface in some ways and that how fast it responds to you the longer it takes the less it feels like a natural conversation and the more it feels like you're talking to a robot Aaron
they always rendered um kind of like a little label that shows you instantly for each each answer the milliseconds of the delay um really kind of building you an intuition you know how many milliseconds feels natural ver it kind of feels like oh I'm talking to a robot Raphael Shad

Nouns vs Verbs in AI Interfaces

3 / 17

Traditional software interfaces are built around static 'nouns' like buttons and forms, but AI introduces dynamic 'verbs' that execute workflows autonomously. This shift requires entirely new design paradigms to visualize and control AI-driven actions, marking a fundamental reset for software interfaces similar to the touch revolution of 2010.

AI interfaces shift from static 'nouns' to dynamic 'verbs'

Traditional software interfaces are built around static elements like text fields, buttons, and dropdowns—nouns that users interact with. AI introduces workflows where software autonomously performs actions like gathering information, auto-completing tasks, or executing processes. This shift requires new design paradigms to visualize and manage these verb-based actions, as current UI tools aren't built to represent dynamic workflows. The core challenge is translating abstract AI behaviors into tangible user controls.

software of today or kind of like up until this point was mostly kind of like just clear things you can point out on the screen um that are you know kind of nouns like text forms drop downs buttons Etc and with AI what really changes is I think so much of the design of what AI does is kind of more verbs um it's more the workflows Auto Complete Auto suggest um go out and gather some information for me Etc and we don't really have the tooling yet to kind of draw verbs on the screen and so that's what's really fascinating how you know this software is now emerging in this new AO World Rafael
these are all verbs we're creating videos we have agents going out executing tasks and so much of it is how do you keep the user in the loop and in control while AI does its magic and we've seen some pretty amazing interfaces to get that level of control and and make sure it's doing the right thing that leads to incredible output that would have taken days years it almost feels it almost feels like back in like 2010 or so when touch um devices really kind of came on the market and everything had to reinvented kind of Touch first and we're at one of those moments again where like all of software all the components that we kind of took for granted um they are really being reimagined and reshaped by the builders and startups and designers out there right now future is going to be incredible Host
Current UI tools lack verbs, requiring new design paradigms

Existing design tools are optimized for static elements, but AI-driven workflows require dynamic, context-aware interactions that don't fit traditional UI patterns. Designers must invent new ways to represent processes like 'go gather information' or 'auto-complete this task', which aren't just clickable buttons but ongoing actions. This is a fundamental shift requiring rethinking how users interact with software beyond static screens.

we don't really have the tooling yet to kind of draw verbs on the screen and so that's what's really fascinating how you know this software is now emerging in this new AO World Rafael
it almost feels it almost feels like back in like 2010 or so when touch um devices really kind of came on the market and everything had to reinvented kind of Touch first and we're at one of those moments again where like all of software all the components that we kind of took for granted um they are really being reimagined and reshaped by the builders and startups and designers out there right now Host
AI interfaces require reimagining software components from scratch

Just as touch interfaces in 2010 forced a complete redesign of software (e.g., no more right-click menus), AI is now forcing a similar reset. Components like buttons, forms, and navigation menus are being replaced by dynamic, context-aware interactions that adapt to user needs. This isn't incremental improvement but a fundamental rethinking of how users interact with software across all domains.

back in like 2010 or so when touch um devices really kind of came on the market and everything had to reinvented kind of Touch first and we're at one of those moments again where like all of software all the components that we kind of took for granted um they are really being reimagined and reshaped by the builders and startups and designers Host
within just a few short like like a few short months or or one two years we see this explosion of AI interface and AI components that really kind of are built AI natively um totally different modalities how to interact with this new teolog with the llms Rafael

Voice Interface Nuances

4 / 17

Effective voice interfaces require attention to latency, multimodal feedback, and interruption handling to maintain natural conversation flow and user trust.

Latency is the interface — longer response times break the illusion of a human conversation.
  • Delays make interactions feel robotic
  • Real-time responsiveness is critical for natural feel
  • Developers should expose latency metrics for debugging
latency is is an issue huh that's what kind of like breaks the illusion of this being a real person yeah Host
the latency is the interface in some ways and that how fast it responds to you the longer it takes the less it feels like a natural conversation and the more it feels like you're talking to a robot the whole point is to make it seem like you're talking to a human Rafael Shad
Multimodal cues are essential for voice interfaces to indicate active listening.
  • Visual feedback for microphone status (e.g., recording indicator)
  • No visual cues during voice input/output leads to confusion
  • Screen-based feedback complements audio for clarity
when I was speaking um it wasn't there was no visual feedback um uh making it clear that my voice is actually recognized by the microphone um and then similarly when the uh voice was answering um there was no sort of like visual indication um that that's what's happening Host
important I guess to kind of pair multimodal cues um so not just rely on voice um in these type of scenarios where you do have a screen uh on the phone that would be a different scenario Rafael Shad
Latency metrics in dev mode build intuition for developers.
  • Showing milliseconds of delay helps developers understand performance
  • Metrics provide transparency into system responsiveness
  • Dev mode features aid in debugging and optimization
they always rendered um kind of like a little label that shows you instantly for each each answer the milliseconds of the delay um really kind of building you an intuition you know how many milliseconds feels natural ver it kind of feels like oh I'm talking to a robot Host
the latency is the interface in some ways and that how fast it responds to you the longer it takes the less it feels like a natural conversation and the more it feels like you're talking to a robot the whole point is to make it seem like you're talking to a human Rafael Shad

Voice Interface Latency & Multimodal Feedback

5 / 17

Latency and visual feedback are critical for voice interfaces to feel natural. Delays break immersion, while multimodal cues (like visual indicators) ensure users understand system state. Effective interruption handling and immediate feedback are essential for human-like interactions.

Latency is the core of voice interface quality

In voice interactions, response time directly affects perceived naturalness. Delays as short as hundreds of milliseconds make the system feel robotic, while near-instant responses (under 200ms) create the illusion of human conversation. This latency is not just a technical metric but a critical design element that shapes user trust and engagement—longer delays break immersion and force users to question whether the system is working.

latency is the interface in some ways and that how fast it responds to you the longer it takes the less it feels like a natural conversation and the more it feels like you're talking to a robot Rafael
latency is is an issue huh that's what kind of like breaks the illusion of this being a real person Host
Multimodal feedback is essential for voice interactions

Voice interfaces must provide visual cues alongside audio to confirm input/output states. Without visual indicators (e.g., microphone active, processing status), users can't tell if the system is listening or responding, leading to confusion. This is especially critical in screen-based environments where users expect visual feedback for all actions, unlike phone-only voice interactions where audio alone suffices.

when I was speaking um it wasn't there was no visual feedback um uh making it clear that my voice is actually recognized by the microphone um and then similarly when the uh voice was answering um there was no sort of like visual indication um that that's what's happening so for example if our laptop was a mute uh we were not sure whether demo is broken or what's going on so important I guess to kind of pair multimodal cues um so not just rely on voice um in these type of scenarios where you do have a screen uh on the phone that would be a different scenario Host
important I guess to kind of pair multimodal cues um so not just rely on voice um in these type of scenarios where you do have a screen uh on the phone that would be a different scenario Rafael
Interrupt handling requires real-time processing

Current voice agents often fail to handle interruptions gracefully, continuing to speak even when the user tries to cut in. This disrupts natural conversation flow and highlights the need for systems that can pause, reprocess inputs, and dynamically adjust responses. Effective interruption handling is a key differentiator between robotic and human-like voice interfaces.

it didn't pause uh when you were interrupting and then two um it entirely missed um your your question when uh when it actually got done with with its own sort of agenda Host
it entirely missed um your your question when uh when it actually got done with with its own sort of agenda yeah Rafael

Voice Interfaces: The New Frontier

6 / 17

Voice AI interfaces are achieving human-like interaction quality, enabling natural conversations with software. However, challenges remain around latency, interruption handling, and multimodal feedback.

Voice interfaces require multimodal feedback to maintain user confidence
  • Pure voice interfaces without visual feedback create uncertainty
  • Users can't tell if the system is listening or responding without visual cues
  • Combining voice with visual indicators creates more robust interactions
  • The modality should match the device context (phone vs screen)
when I was speaking um it wasn't there was no visual feedback um uh making it clear that my voice is actually recognized by the microphone um and then similarly when the uh voice was answering um there was no sort of like visual indication um that that's what's happening Raphael Shad
important I guess to kind of pair multimodal cues um so not just rely on voice um in these type of scenarios where you do have a screen uh on the phone that would be a different scenario Raphael Shad
Natural conversation requires handling interruptions gracefully
  • Human conversations involve frequent interruptions and overlaps
  • Current voice AI struggles with mid-speech interruptions
  • Systems either ignore interruptions or lose context
  • Future interfaces need to manage conversational flow more dynamically
when you're talking to a human um the latency is really important and also interruptions and it felt pretty fast and pretty natural when we were conversing I wonder what would happen if we tried to interrupt it would it be able to handle it Aaron
two things happened one um it didn't pause uh when you were interrupting and then two um it entirely missed um your your question when uh when it actually got done with with its own sort of agenda Raphael Shad

Visual Workflow Modeling

7 / 17

Canvases and flowcharts provide intuitive ways to design, monitor, and control complex AI agent workflows through visual representation of branching logic and multi-dimensional processes.

Canvases are a new document type ideal for modeling AI agent workflows.
  • Visual pan/zoom interfaces allow complex process mapping
  • Color-coded blocks distinguish input, actions, outputs
  • Enables non-linear, multi-dimensional process design
Canabis has really emerged as a really interesting kind of almost new document type um that seems to lend itself pretty well to not just kind of for design tools or or kind of brainstorming tools but lends itself really well for these sort of modeling these kind of like AI processes yeah Host
it's great because it gives us the user a visual overflow of exactly what steps the agent is going to take and we can control what it should do at each of these steps Rafael Shad
Branching logic is the key power of AI workflow modeling.
  • Linear flows are insufficient for complex agent decisions
  • Multi-dimensional branching handles real-world unpredictability
  • Visual tools must support non-linear, conditional paths
the power is in sort of like the multi-dimensionality in the branching um and so for as a starter template to kind of like explain the power of this tool to mod these processes I think one that is multi-dimensional would really showcase the power of this Rafael Shad
it's always historically been static and it seems like what's new is actually making it interactive Host
Flowcharts resurface in AI era as interactive tools.
  • Legacy flowchart techniques from chip design are being reused
  • Modern interactivity adds real-time control and feedback
  • Combines historical paradigms with new AI capabilities
it's interesting to kind of like see this Paradigm kind of getting resurfaced in the AI era so you know we didn't inent invent this today but we're building on a lot of Legacy um and on the Giant on the shoulder of giants here yeah and it's always historically been static and it seems like what's new is actually making it interactive Rafael Shad
flowcharts Etc probably like chip designers like 50 years ago they're like oh yeah we used to you know kind of model our things like that and so it's interesting to kind of like see this Paradigm kind of getting resurfaced in the AI era Host

Adaptive Contextual UIs

8 / 17

Interfaces that dynamically adjust based on content context reduce cognitive load by showing only relevant controls. Consistent keyboard shortcuts maintain usability despite changing UI elements, but clear focus states prevent unintended actions when typing.

UIs that adapt to content context reduce cognitive load

Traditional interfaces show all possible options regardless of context, overwhelming users. Adaptive UIs dynamically surface only relevant actions based on current content—like email-specific response buttons or document-specific formatting tools. This reduces clutter and streamlines workflows, but requires precise context understanding to avoid unpredictability.

Microsoft Word right where like the thing that everybody is so familiar with is a billion buttons on the top row because they're never sure which one you might need because they don't know the context of how you're editing and with AI now we don't need to show all the buttons we can just show you the buttons that are relevant Host
the interface then dynamically changes which typically isn't you know static software typically wasn't the case and so here it's kind of like the input is the actual content and then the output of the AI llm is then the UI to interact back with that content Rafael
Keyboard shortcuts maintain consistency in adaptive UIs

Even as UI elements change based on context, consistent keyboard shortcuts (e.g., pressing 'Y' to confirm) allow users to interact without relearning new controls. This preserves muscle memory while enabling dynamic behavior—critical for high-efficiency workflows like email processing where speed matters.

the buttons and the responses are technically changing for every single email but the the keys that you're pressing do not and so you can kind of keep your hand right there and and know what to expect each time Host
being able to access all these adaptive kind of like uh uh options by just keyboard shortcut with a single letter um is uh is is really on point Rafael
Input focus ambiguity causes unintended actions

Adaptive UIs risk accidental actions when keyboard input is ambiguous—e.g., pressing 'Y' to type a letter in a text field versus confirming a button. Clear visual indicators of focus state are essential to prevent unintended commands, especially in high-speed workflows where users expect immediate feedback.

what if I think that my cursor is focused inserting text and I want to kind of reply yes then basically my first y keystroke like submits a button right and so there's always this challenge of really being very clear when an input element is focused and you're typing versus now typing on the keyboard will just do stuff in your UI Host
really being very clear when an input element is focused and you're typing versus now typing on the keyboard will just do stuff in your UI Rafael

Visualizing AI Workflows

9 / 17

As AI agents perform complex, autonomous tasks, new interface paradigms like canvas-based flowcharts emerge to help users understand and control these processes.

Canvas interfaces are ideal for modeling complex AI decision trees
  • Traditional linear documentation can't capture branching AI workflows
  • Canvas interfaces allow spatial organization of multi-step processes
  • Color coding helps distinguish different types of actions/nodes
  • The paradigm resembles chip design flowcharts from decades past
Canabis has really emerged as a really interesting kind of almost new document type um that seems to lend itself pretty well to not just kind of for design tools or or kind of brainstorming tools but lends itself really well for these sort of modeling these kind of like AI processes Raphael Shad
the canvas and modeling these kind of like AI ancient decision trees gets really really powerful when it isn't something you could just kind of like linearly write in a document like a recipe first do this then do this then do this but really the power is in sort of like the multi-dimensionality in the branching Raphael Shad
Zoom levels should adapt to show relevant workflow detail
  • At high zoom levels, detailed text becomes unreadable
  • Interfaces should collapse nodes to colored blocks when zoomed out
  • Different zoom levels should show different fidelity of information
  • This maintains overview while preserving ability to drill into details
because it is uh canis um kind of having different Zoom levels showing different Fidelity so right now we're so zoomed out I can't read any of the small text why not just kind of hide it and make the note almost collapse it into just you know in this case a brown Block in this case a Yellow Block to kind of give different Zoom levels different fidelities Raphael Shad

Data Extraction & Trust

10 / 17

Structured data outputs from AI agents require clear source attribution and transparency to build user trust, especially when handling sensitive or critical information.

Every spreadsheet cell can have its own AI agent for data extraction.
  • AI processes data at cell-level granularity
  • Eliminates need for predefined columns; dynamic column creation
  • Enables on-demand data gathering from multiple sources
it's like a spreadsheet on steroids Host
it's almost like every cell of the spreadsheet gets its own AI agent to get the data that we want which is pretty incredible Rafael Shad
Source attribution is critical for validating AI-generated data.
  • Inline references allow users to verify information
  • Reduces hallucination risks by showing provenance
  • Builds trust through transparency in data sourcing
by having a source closely attached that you can just you know Click on each of these right here you can see immediately where the sources came from it helps us to be able to validate and Trust the data that the AI agent is bringing back Host
it's also interesting you know you mentioned before about how um you know we always had flowcharts and these are like modern flowcharts with the canvases and it's interesting too that you know a lot of the citing sources in the footnotes is not a new thing that's been around since the beginning of books but now it's actually being used in a new way to actually validate and verify information in real time that an agent brings back which is really cool Rafael Shad
Dynamic column creation enables human-guided data extraction.
  • Users define new data points on the fly
  • Agents fetch and populate columns automatically
  • Eliminates rigid pre-defined schemas
we can do is add columns and have sort of the agent go out again not on a sort of like static uh you know set of columns that were predefined but our columns like things we want to know kind of putting the human back into the loop Host
it's like a spreadsheet on steroids Host

Visual Workflow Modeling for AI Agents

11 / 17

Canvas-based interfaces enable complex AI agent decision trees through spatial layouts, zoom levels, and color coding. Legacy flowchart paradigms are resurfacing in AI for dynamic, interactive workflows that replace static diagrams with executable processes.

Canvases enable complex agent decision trees

Traditional linear workflows fail to represent the branching, multi-dimensional logic of AI agents. Visual canvases allow designers to model complex decision paths, conditional branches, and parallel tasks in a spatial layout. This makes it easier to understand, debug, and iterate on agent behavior—especially for non-technical users who need to see the entire process flow at once.

it's great because it gives us the user a visual overflow of exactly what steps the agent is going to take and we can control what it should do at each of these steps Rafael
the power is in sort of like the multi-dimensionality in the branching um and so for as a starter template to kind of like explain the power of this tool to mod these processes I think one that is multi-dimensional would really showcase the power of this Host
Zoom levels and color coding improve workflow clarity

Effective visual workflows use zoom levels to show high-level overviews or detailed steps, and color-coded elements to distinguish input, action, and output nodes. Without these, complex diagrams become unreadable. A legend or consistent visual language is critical for users to quickly interpret the workflow structure without getting lost in details.

using colar um to show different type of notes um kind of like input actions um output Etc I almost feel like I would want like a legend like which color is what Host
because it is uh canis um kind of having different Zoom levels showing different Fidelity so right now we're so zoomed out I can't read any of the small text why not just kind of hide it and make the note almost collapse it into just you know in this case a brown Block in this case a Yellow Block to kind of give different Zoom levels different fidelities Rafael
Legacy flowchart paradigms are resurfacing in AI

While the concept of flowcharts isn't new (used by chip designers decades ago), AI agents are reviving this approach for dynamic, interactive workflows. Modern tools now allow real-time editing and execution of these diagrams, transforming static diagrams into living, executable processes that guide AI behavior in real time.

it's interesting to kind of like see this Paradigm kind of getting resurfaced in the AI era so you know we didn't inent invent this today but we're building on a lot of Legacy um and on the Giant on the shoulder of giants here Rafael
it's always historically been static and it seems like what's new is actually making it interactive Host

Adaptive Interfaces: UI That Responds to Content

12 / 17

AI enables interfaces that dynamically adapt based on content context, moving beyond static layouts to personalized interaction flows.

Email interfaces can adapt response options based on message content
  • Traditional email clients offer static reply options
  • AI can analyze email content to suggest context-specific responses
  • Response buttons adapt to each email's needs (e.g., scheduling options)
  • Keyboard shortcuts maintain consistency despite changing button labels
it's pulling up the user's email and it's suggesting specific responses to that email based on the content of that email it's it's almost changing what the reaction buttons are exactly Aaron
the buttons and the responses are technically changing for every single email but the the keys that you're pressing do not and so you can kind of keep your hand right there and and know what to expect each time Aaron
Prompt builders should offer structured input options
  • Freeform text prompts are powerful but intimidating
  • Many users don't know specialized terminology (e.g., 'glass morphic')
  • Interface could offer visual building blocks alongside text input
  • This lowers the barrier while maintaining flexibility
having maybe a a interface here um that gives me sort of like selection and ideas kind of almost like pills maybe these are design terms that can almost like you know like Lego bricks can like drag in versus just I need to know these terms and just type them or learn them from the examples Raphael Shad

Trust & Transparency in AI Outputs

13 / 17

Inline source citations and footnotes validate AI-generated data, transforming passive references into active verification tools. Per-cell AI agents in spreadsheets dynamically fetch specific data points, while academic-style footnotes ensure accountability for real-time information.

Inline sources validate AI-generated data

AI systems that cite sources directly within outputs (e.g., footnotes in spreadsheets) build trust by allowing users to verify information instantly. This is especially critical for factual data where hallucinations are common—users can quickly check the origin of each data point without leaving the interface, reducing uncertainty about accuracy.

this is another common pattern that we see where if AI is going out and doing a thing how do you know you can trust the results that it brings back you know sometimes it hallucinates sometimes it gets the wrong thing and so by having a source closely attached that you can just you know Click on each of these right here you can see immediately where the sources came from it helps us to be able to validate and Trust the data that the AI agent is bringing back Host
it's also interesting you know you mentioned before about how um you know we always had flowcharts and these are like modern flowcharts with the canvases and it's interesting too that you know a lot of the citing sources in the footnotes is not a new thing that's been around since the beginning of books but now it's actually being used in a new way to actually validate and verify information in real time that an agent brings back which is really cool Rafael
Footnotes as real-time validation in software

Just as academic papers use footnotes to cite sources, modern AI interfaces integrate inline references that link directly to the data origin. This transforms passive citations into active verification tools—users can click to see the source page, ensuring transparency and accountability for AI-generated content in real-time workflows.

when you Googled kind of in the past um you just had a list of websites a list of basically the references or the links and they were your destination but now that you ask kind of a chat chat box um and you get the answer back you kind of want to have the links um the references kind of like inlined and I believe it was maybe perplexity kind of to to do that pattern first where you had like these little round uh numbered dots right in line with the answer Host
this is a really nice pattern that um is sort of like used here um and it could even be used in other context or even be inlined here um I guess in a spreadsheet it works kind of to pull it out into in into its own popover um when you are kind of more uh looking for space and to condense it then sort like the pattern of having the footnotes almost directly in the answer is a really really successful pattern Rafael
Per-cell AI agents in spreadsheets enhance precision

By treating each spreadsheet cell as an independent AI agent, systems can fetch specific data points on demand without predefined columns. This allows users to dynamically add columns (e.g., 'funding raised'), with each cell's AI agent sourcing the correct information—turning spreadsheets into intelligent, self-updating data tables that adapt to user needs.

it's almost like every cell of the spreadsheet gets its own AI agent to get the data that we want which is pretty incredible it's like a spreadsheet on steroids Host
we took a prompt as an input um and we got the spreadsheet structured data as an output and in the background it went to these websites scraped it assembled this this spreadsheet and now we can do is add columns and have sort of the agent go out again not on a sort of like static uh you know set of columns that were predefined but our columns like things we want to know kind of putting the human back into the loop Rafael

AI Video Generation: Trading Fidelity for Iteration Speed

14 / 17

AI video production tools use clever UX patterns to enable rapid iteration despite the computational demands of high-quality output generation.

Blurred previews enable fast iteration before full rendering
  • High-quality AI video generation takes significant time (minutes)
  • Showing blurred previews with audio allows immediate feedback
  • Users can iterate on script and timing before committing to render
  • This maintains creative flow despite technical constraints
they're trading off um basically uh Fidelity for immediacy and basically putting the human kind back in the loop because if it was just a generate button right we would wait for 12 minutes uh figure out that something is not quite right and then kind of like you know give the machine a new prompt and wait until it comes back Raphael Shad
blurring the video is is a is a is a great uh design uh approach to do that Raphael Shad
AI video generation enables personalized content at scale
  • Deepfake technology can clone voices and appearances with minimal samples
  • Videos can be dynamically personalized (e.g., changing names)
  • This enables mass customization previously impossible with traditional production
  • Ethical considerations around consent and misuse become critical
they just need a few minutes of uh video of me or whoever talking and then they can basically process it automatically in their models to create their deep fi and you were saying something completely different than this year that's incredible Aaron

Adaptive UIs

15 / 17

Context-aware interfaces dynamically change based on content, showing only relevant actions and reducing cognitive load by eliminating unnecessary UI elements.

Adaptive UIs show only relevant actions based on content context.
  • Eliminates static button clutter (e.g., Microsoft Word's toolbar)
  • UI changes dynamically based on current task or document
  • Reduces cognitive load by focusing on what's needed now
Microsoft Word right where like the thing that everybody is so familiar with is a billion buttons on the top row because they're never sure which one you might need because they don't know the context of how you're editing and with AI now we don't need to show all the buttons we can just show you the buttons that are relevant Host
some of the uh adap interface that we see emerge um based on the content of for example an email or document the interface then dynamically changes which typically isn't you know static software typically wasn't the case and so here it's kind of like the input is the actual content and then the output of the AI llm is then the UI to interact back with that content Rafael Shad
Hot keys remain consistent even as UI elements change.
  • Keyboard shortcuts stay fixed despite dynamic button changes
  • Allows users to maintain muscle memory for common actions
  • Balances adaptability with predictable interaction patterns
the buttons and the responses are technically changing for every single email but the the keys that you're pressing do not and so you can kind of keep your hand right there and and know what to expect each time Host
being able to access all these adaptive kind of like uh uh options by just keyboard shortcut with a single letter um is uh is is really on point Rafael Shad
Adaptive UIs balance dynamic changes with consistent interaction patterns.
  • Buttons change per context but keyboard shortcuts remain fixed
  • Users value consistency in interaction patterns
  • Prevents confusion while allowing context-aware actions
the buttons and the responses are technically changing for every single email but the the keys that you're pressing do not and so you can kind of keep your hand right there and and know what to expect each time Host
the challenge of course are kind of predictability people love to have like their you know billion buttons and like the exact place and so Host

Iterative Human-AI Collaboration

16 / 17

Progressive fidelity (blurred previews) accelerates iteration by letting users confirm direction before full generation. Incremental updates preserve existing work, while prompt feedback highlights respected vs ignored elements to refine instructions and improve AI understanding.

Progressive fidelity improves iteration speed

AI systems that show low-fidelity previews (e.g., blurry video) while generating high-fidelity outputs let users iterate quickly without waiting for full renders. This balances immediacy with quality—users can confirm the direction early, then refine before final generation, avoiding wasted time on incorrect outputs.

trading off Fidelity for immediacy and basically putting the human kind back in the loop because if it was just a generate button right we would wait for 12 minutes uh figure out that something is not quite right and then kind of like you know give the machine a new prompt and wait until it comes back so this is a really uh clever trick uh to really kind of create this iterative human machine collaboration interface Rafael
the easier or the faster part in generating this is actually in creating the voice and the hard part is it takes many minutes to actually process and generate the video with the right lip movement to match the text that you've entered and so rather than showing you something you know lips moving that is off from what you what you've uh put in they first show you just kind of a blurry version with the audio so you can get a sense of like what it's going to be like then you click generate here and then that you know it says 12 minutes right here is how long it's GNA take Host
Incremental updates preserve existing work

When modifying AI-generated outputs (e.g., changing a color in a design), systems that only update the changed elements—rather than regenerating everything—save time and maintain consistency. This delta-based approach is crucial for iterative design, where small tweaks shouldn't require full reprocessing of complex assets.

what if I just took this and what if I said okay explain your changes okay so so so make this make the sidebar blue mhm run revision okay okay now we're waiting again hopefully this will be faster hopefully it's like an incremental change right where we can sort of only submit to Delta and not like do sort of like a single shot do the whole thing over again um not just for kind of to weight but also for you know resources um preserve kind of the the existing um design that we did like and did want to turn blue um so let's see how I can s like deal with diffs and and for consistency too because you know especially when you're prompting to create Graphics one of the challenges is if you want to change one element at a hat on a person it's hard to keep the rest of it consistent that's a common challenge now and so if they're able to do this then I I think that that speaks pretty highly Host
if they're able to do this then I I think that that speaks pretty highly yeah so so so any interface designer or technical team kind of figure out the challenge of how to kind of add sub prompts or how to only change kind of iteratively um that that that that's kind of like really the the frontier Rafael
Prompt feedback highlights respected vs ignored elements

When AI generates outputs from prompts, showing which parts of the prompt were successfully executed versus ignored (e.g., through visual highlights) helps users refine their instructions. This feedback loop allows humans to learn how to communicate effectively with AI systems, improving future prompts and reducing trial-and-error.

what did it index on from The Prompt and execute on and what did it sort like maybe your fail if there could be sort of like that feedback loop then it can help the human to refine their prompt and kind of learn how to you know interact with the machine and help the AI to figure out what it did well and should keep doing and what it didn't do well and should get more data to improve Host
the output generated what are the things the machine actually respected from your prompt and where did it sort of like ignore or struggle with kind of giving that feedback back to the human and the prompt maybe with like little squiggly lines or kind of things or with color kind of showing you know what did it index on from The Prompt and execute on and what did it sort like maybe your fail Rafael

Video Generation Workflow

17 / 17

Balancing real-time feedback with full-generation latency requires clever UI design, such as blurred previews, to enable iterative refinement before final output.

Blurring video previews trades fidelity for immediacy during generation.
  • Low-fidelity previews allow quick iteration on scripts
  • Full generation happens later, reducing wait time
  • Maintains user engagement during long processing times
the easier or the faster part in generating this is actually in creating the voice and the hard part is it takes many minutes to actually process and generate the video with the right lip movement to match the text that you've entered and so rather than showing you something you know lips moving that is off from what you what you've uh put in they first show you just kind of a blurry version with the audio so you can get a sense of like what it's going to be like then you click generate here and then that you know it says 12 minutes right here is how long it's GNA take Host
trading off um basically uh Fidelity for immediacy and basically putting the human kind back in the loop because if it was just a generate button right we would wait for 12 minutes uh figure out that something is not quite right and then kind of like you know give the machine a new prompt and wait until it comes back so this is a really uh clever trick uh to really kind of create this iterative human machine collaboration interface um Rafael Shad
Iterative human-machine collaboration is key in video generation.
  • Blurred previews allow quick script adjustments
  • Users can refine content before full rendering
  • Reduces wasted time on incorrect outputs
clever trick uh to really kind of create this iterative human machine collaboration interface um Rafael Shad
blurring the video is is a is a is a great uh design uh approach to do that Host
Text-to-video systems need better script-to-action mapping.
  • Manual selection of body language gestures is tedious
  • Future systems should auto-detect gestures from script
  • Reduces friction in creating expressive video content
you have to select it manually yeah so I selected it manually you could imagine in the future they would autodetect it right right that would be really interesting Host
you can almost imagine how you could highlight certain parts of the script and then from a drop- down there kind of choose suggest but then also kind of standard um uh part of this this Library kind of like you know just try point to myself I'm I'm I'm what is it I'm crushing it um or I'm I'm crushed um and so so there's like a lot of kind of like interplay with the text interface to the left Host
⚙ Agent-readable JSON index — click to expand
{
  "memcast_version": "0.1",
  "episode":  {
    "id": "DBhSfROq3wU",
    "title": "AI Interfaces Of The Future | Design Review",
    "podcast": "Y Combinator",
    "guest": "Raphael Shad",
    "host": "Aaron",
    "source_url": "https://www.youtube.com/watch?v=DBhSfROq3wU",
    "duration_minutes": 37
  },
  "concepts":  [
    {
      "id": "verbs-over-nouns",
      "title": "Verbs Over Nouns",
      "tags":  [
        "3d-ai"
      ]
    },
    {
      "id": "from-nouns-to-verbs-the-shift-in-ai-interface-design",
      "title": "From Nouns to Verbs: The Shift in AI Interface Design",
      "tags":  [
        "ai-alignment",
        "design-paradigms"
      ]
    },
    {
      "id": "nouns-vs-verbs-in-ai-interfaces",
      "title": "Nouns vs Verbs in AI Interfaces",
      "tags":  [
        "ai-alignment",
        "user-experience",
        "design-paradigms"
      ]
    },
    {
      "id": "voice-interface-nuances",
      "title": "Voice Interface Nuances",
      "tags":  [
        "ai-adoption",
        "adaptive-ui"
      ]
    },
    {
      "id": "voice-interface-latency-multimodal-feedback",
      "title": "Voice Interface Latency & Multimodal Feedback",
      "tags":  [
        "latency",
        "multimodal-design"
      ]
    },
    {
      "id": "voice-interfaces-the-new-frontier",
      "title": "Voice Interfaces: The New Frontier",
      "tags":  [
        "ai-adoption",
        "latency",
        "adaptive-ui"
      ]
    },
    {
      "id": "visual-workflow-modeling",
      "title": "Visual Workflow Modeling",
      "tags":  [
        "ai-adoption",
        "workflow-design"
      ]
    },
    {
      "id": "adaptive-contextual-uis",
      "title": "Adaptive Contextual UIs",
      "tags":  [
        "adaptive-ui",
        "cost-efficiency",
        "context-aware"
      ]
    },
    {
      "id": "visualizing-ai-workflows",
      "title": "Visualizing AI Workflows",
      "tags":  [
        "ai-adoption",
        "workflow-design"
      ]
    },
    {
      "id": "data-extraction-trust",
      "title": "Data Extraction & Trust",
      "tags":  [
        "3d-ai",
        "confidence"
      ]
    },
    {
      "id": "visual-workflow-modeling-for-ai-agents",
      "title": "Visual Workflow Modeling for AI Agents",
      "tags":  [
        "ai-adoption",
        "workflow-design",
        "multimodal-design"
      ]
    },
    {
      "id": "adaptive-interfaces-ui-that-responds-to-content",
      "title": "Adaptive Interfaces: UI That Responds to Content",
      "tags":  [
        "adaptive-ui",
        "context-aware"
      ]
    },
    {
      "id": "trust-transparency-in-ai-outputs",
      "title": "Trust & Transparency in AI Outputs",
      "tags":  [
        "ai-adoption",
        "data-quality"
      ]
    },
    {
      "id": "ai-video-generation-trading-fidelity-for-iteration-speed",
      "title": "AI Video Generation: Trading Fidelity for Iteration Speed",
      "tags":  [
        "3d-ai",
        "latency",
        "iterative-design"
      ]
    },
    {
      "id": "adaptive-uis",
      "title": "Adaptive UIs",
      "tags":  [
        "adaptive-ui",
        "context-aware"
      ]
    },
    {
      "id": "iterative-human-ai-collaboration",
      "title": "Iterative Human-AI Collaboration",
      "tags":  [
        "ai-adoption",
        "iterative-design"
      ]
    },
    {
      "id": "video-generation-workflow",
      "title": "Video Generation Workflow",
      "tags":  [
        "3d-ai",
        "latency"
      ]
    }
  ]
}