Alibaba’s Qwen team has released the Qwen3.5 Small Model Series, a collection of Large Language Models (LLMs) ranging from 0.8B […]
Category: Open Source
Alibaba Team Open-Sources CoPaw: A High-Performance Personal Agent Workstation for Developers to Scale Multi-Channel AI Workflows and Memory
As the industry moves from simple Large Language Model (LLM) inference toward autonomous agentic systems, the challenge for devs have […]
A Coding Implementation to Build a Hierarchical Planner AI Agent Using Open-Source LLMs with Tool Execution and Structured Multi-Agent Reasoning
In this tutorial, we build a hierarchical planner agent using an open-source instruct model. We design a structured multi-agent architecture […]
Perplexity Just Released pplx-embed: New SOTA Qwen3 Bidirectional Embedding Models for Web-Scale Retrieval Tasks
Perplexity has released pplx-embed, a collection of multilingual embedding models optimized for large-scale retrieval tasks. These models are designed to […]
Nous Research Releases ‘Hermes Agent’ to Fix AI Forgetfulness with Multi-Level Memory and Dedicated Remote Terminal Access Support
In the current AI landscape, we’ve become accustomed to the ‘ephemeral agent’—a brilliant but forgetful assistant that restarts its cognitive […]
Open source vulnerabilities double with AI code creation
Open source has become effectively universal, appearing in 98 percent of codebases, meaning almost every application now inherits third-party risk. […]
Liquid AI’s New LFM2-24B-A2B Hybrid Architecture Blends Attention with Convolutions to Solve the Scaling Bottlenecks of Modern LLMs
The generative AI race has long been a game of ‘bigger is better.’ But as the industry hits the limits […]
Meta AI Open Sources GCM for Better GPU Cluster Monitoring to Ensure High Performance AI Training and Hardware Reliability
While the tech folks obsesses over the latest Llama checkpoints, a much grittier battle is being fought in the basements […]
Alibaba Qwen Team Releases Qwen 3.5 Medium Model Series: A Production Powerhouse Proving that Smaller AI Models are Smarter
The development of large language models (LLMs) has been defined by the pursuit of raw scale. While increasing parameter counts […]
Composio Open Sources Agent Orchestrator to Help AI Developers Build Scalable Multi-Agent Workflows Beyond the Traditional ReAct Loops
For the past year, AI devs have relied on the ReAct (Reasoning + Acting) pattern—a simple loop where an LLM […]
