Tech industry
fromTNW | Artificial-Intelligence
23 hours agoMeta's Muse Spark is here - and it's closed source
Meta has launched Muse Spark, a multimodal AI model developed by Meta Superintelligence Labs to compete with leading AI companies.
Qwen3.5 is available via Hugging Face and is released under an open-source license. With this, Alibaba is explicitly targeting developers and research institutions that want to work with the model themselves. The system can process very long prompts, up to 260,000 tokens, and can be scaled further with additional optimizations. This makes it suitable for complex applications such as extensive document analysis and code generation.
In this tutorial, you'll learn how to build an interactive Streamlit Python-based UI that connects seamlessly with your vLLM-powered multimodal backend. You'll write a simple yet flexible frontend that lets users upload images, enter text prompts, and receive smart, vision-aware responses from the LLaVA model - served via vLLM's OpenAI-compatible interface. By the end, you'll have a clean multimodal chat interface that can be deployed locally or in the cloud - ready to power real-world apps in healthcare, education, document understanding, and beyond.
Gemini has been integrated across nearly all of Google's offerings -- and now it's time for Google Maps' AI facelift. On Wednesday, the company launched four upgrades to Google Maps that make it easier for users to get where they want to go, including new multimodal features, such as conversational natural language prompts to find a stop en route or Lens to identify new places at your destination.