Alibaba Tongyi Lab have released MAI-UI—a family of foundation GUI agents. It natively integrates MCP tool use, agent user interaction, device–cloud collaboration, and online... In this tutorial, we ...
Shobha is a data analyst with a proven track record of developing innovative machine-learning solutions that drive business value.
The year 2025 marks a turning point for Voice AI Agents, with technology reaching levels of naturalness, context-awareness, and commercial adoption that were unimaginable a decade ago. Powered by ...
Alibaba’s Tongyi Lab has open-sourced Tongyi-DeepResearch-30B-A3B, an agent-specialized large language model built for long-horizon, deep information-seeking with web tools. The model uses a ...
An open-source framework that couples LLM-driven program mutations with evolutionary search to automate algorithm discovery and optimization. Code and report are public. 2) How does it achieve higher ...
MIT Researchers Enhanced Artificial Intelligence (AI) 64x Better at Planning, Achieving 94% Accuracy
Across domains, the research team showcase up to 66% absolute improvement over untuned baselines. Detailed validator feedback outperforms binary signals, and longer feedback budgets further help. PDDL ...
ACE positions “context engineering” as a first-class alternative to parameter updates. Instead of compressing instructions into short prompts, ACE accumulates and organizes domain-specific tactics ...
What is an AI Agent? An AI Agent is an autonomous software system that can perceive its environment, interpret data, reason, and execute actions to achieve specific goals without explicit human ...
Canary-1b-v2: Multilingual ASR + Translation (En ↔ 24 Languages) Canary-1b-v2 is a billion-parameter Encoder-Decoder model trained on Granary, delivering high-quality transcription and translation ...
xAI introduced Grok-4-Fast, a cost-optimized successor to Grok-4 that merges “reasoning” and “non-reasoning” behaviors into a single set of weights controllable via system prompts. The model targets ...
Zhipu AI has officially released and open-sourced GLM-4.5V, a next-generation vision-language model (VLM) that significantly advances the state of open multimodal AI. Based on Zhipu’s 106-billion ...
Results that may be inaccessible to you are currently showing.
Hide inaccessible results