Voice AI is becoming one of the most important frontiers in multimodal AI. From intelligent assistants to interactive agents, the […]
Category: Large Language Model
Google AI Releases VaultGemma: The Largest and Most Capable Open Model (1B-parameters) Trained from Scratch with Differential Privacy
Google AI Research and DeepMind have released VaultGemma 1B, the largest open-weight large language model trained entirely with differential privacy […]
BentoML Released llm-optimizer: An Open-Source AI Tool for Benchmarking and Optimizing LLM Inference
BentoML has recently released llm-optimizer, an open-source framework designed to streamline the benchmarking and performance tuning of self-hosted large language […]
Meet mmBERT: An Encoder-only Language Model Pretrained on 3T Tokens of Multilingual Text in over 1800 Languages and 2–4× Faster than Previous Models
Table of contents Why was a new multilingual encoder needed? Understanding the architecture of mmBERT What training data and phases […]
NVIDIA AI Releases Universal Deep Research (UDR): A Prototype Framework for Scalable and Auditable Deep Research Agents
Why do existing deep research tools fall short? Deep Research Tools (DRTs) like Gemini Deep Research, Perplexity, OpenAI’s Deep Research, […]
Baidu Releases ERNIE-4.5-21B-A3B-Thinking: A Compact MoE Model for Deep Reasoning
Baidu AI Research team has just released ERNIE-4.5-21B-A3B-Thinking, a new reasoning-focused large language model designed around efficiency, long-context reasoning, and […]
Building a Speech Enhancement and Automatic Speech Recognition (ASR) Pipeline in Python Using SpeechBrain
In this tutorial, we walk through an advanced yet practical workflow using SpeechBrain. We start by generating our own clean […]
MBZUAI Researchers Release K2 Think: A 32B Open-Source System for Advanced AI Reasoning and Outperforms 20x Larger Reasoning Models
A team of researchers from MBZUAI’s Institute of Foundation Models and G42 released K2 Think, is a 32B-parameter open reasoning […]
Meta Superintelligence Labs Introduces REFRAG: Scaling RAG with 16× Longer Contexts and 31× Faster Decoding
Table of contents Why is long context such a bottleneck for LLMs? How does REFRAG compress and shorten context? How […]
Tilde AI Releases TildeOpen LLM: An Open-Source Large Language Model with Over 30 Billion Parameters and Support Most European Languages
Latvian language-tech firm Tilde has released TildeOpen LLM, an open-source foundational large language model (LLM) purpose-built for European languages, with […]
