Techniques and Trends in AI-Powered Search by Faye Zhang at QCon SF
Briefly

Faye Zhang highlighted the rapid development in AI as a search tool, noting it grew from 1% of the population using it in early 2024 to 8% by October. This growth underscores the significant role AI will play in search by 2027 when it's expected to exceed 60%. This trend reflects a broader shift in technology as users increasingly turn to AI for more efficient and personalized search experiences.
Zhang delved into the concept of multi-modal interaction, stating that AI search goes beyond text with models now accepting images, video, and speech. She showcased Meta’s Chameleon model and explained that the strategy often includes mapping all input modalities to a unified embedding space, as seen with Meta's ImageBind model. This shift opens new avenues for how users interact with AI, fundamentally changing traditional search paradigms.
To tackle the challenge of real-time user refinement in searches, Zhang explained a proposed architecture incorporating a vision transformer and a T5 language model. This dual approach allows understanding of image features and language simultaneously, creating a more intuitive search process. She emphasized that T5's ability to efficiently manage embeddings and text interaction could greatly enhance user experience during searches, making it easier to iterate based on specific needs.
Throughout her talk, Zhang emphasized the need for AI systems that effectively manage user interactions in natural, intuitive ways. For example, when searching for sunglasses, a user could quickly specify different criteria—like price or shape—after receiving initial search results. Addressing these iterative needs with advanced AI techniques shows promise for future search experiences, blending user intent with intelligent system capabilities.
Read at InfoQ
[
|
]