
"Today's models experience reality in a single thread. Until the user finishes typing or speaking, the model waits with no perception of what the user is doing or how the user is doing it. Until the model finishes generating, its perception freezes, receiving no new information until it finishes or is interrupted. This creates a narrow channel for human-AI collaboration that limits how much of a person's knowledge, intent, and judgement can reach the model, and how much of the model's work can be understood."
"Picture trying to resolve a crucial disagreement over email rather than in person."
"At Thinking Machines, we believe we can solve this bandwidth bottleneck by making AI interactive in real time across any modality. This enables AI interfaces to meet humans where they are, rather than forcing humans to contort themselves to AI interfaces."
Current models perceive reality in a single thread, waiting for the user to finish typing or speaking before processing. While the model generates, its perception freezes and it receives no new information until completion or interruption. This creates a narrow channel for human-AI collaboration, restricting how much of a person’s knowledge, intent, and judgment reaches the model and how much of the model’s work the person can understand. Real-time, multi-modality interaction can address this bandwidth bottleneck by allowing AI to respond continuously as new information arrives. Such interfaces can adapt to humans instead of forcing humans to reshape their communication to fit rigid AI input and output timing.
#ai-interaction #real-time-multimodal-systems #human-ai-collaboration #bandwidth-bottleneck #model-perception
Read at The Verge
Unable to calculate read time
Collection
[
|
...
]