Large Language Models (LLMs) and Vision-Language Models (VLMs) transform natural language understanding, multimodal integration, and complex reasoning tasks. Yet, one […]
Category: Large Language Model
MinMo: A Multimodal Large Language Model with Approximately 8B Parameters for Seamless Voice Interaction
Advances in large language and multimodal speech-text models have laid a foundation for seamless, real-time, natural, and human-like voice interactions. […]
New LLM-powered engine helps secure complex cloud environments
While moving systems to the cloud delivers many benefits, it also leads to complex dynamic environments that can be a […]
Alibaba Qwen Team just Released ‘Lessons of Developing Process Reward Models in Mathematical Reasoning’ along with a State-of-the-Art 7B and 72B PRMs
Mathematical reasoning has long been a significant challenge for Large Language Models (LLMs). Errors in intermediate reasoning steps can undermine […]
Enhancing Language Model Performance and Diversity Through Multiagent Fine-Tuning
LLMs, such as GPT-3.5 and GPT-4, have shown exceptional capabilities in language generation, comprehension, and translation tasks. Despite these advancements, […]
OpenBMB Just Released MiniCPM-o 2.6: A New 8B Parameters, Any-to-Any Multimodal Model that can Understand Vision, Speech, and Language and Runs on Edge Devices
Artificial intelligence has made significant strides in recent years, but challenges remAIn in balancing computational efficiency and versatility. State-of-the-art multimodal […]
Outcome-Refining Process Supervision: Advancing Code Generation with Structured Reasoning and Execution Feedback
LLMs excel in code generation but struggle with complex programming tasks requiring deep algorithmic reasoning and intricate logic. Traditional outcome […]
UC Berkeley Researchers Released Sky-T1-32B-Preview: An Open-Source Reasoning LLM Trained for Under $450 Surpasses OpenAI-o1 on Benchmarks like Math500, AIME, and Livebench
The rapid advancements in artificial intelligence have opened new possibilities, but the associated costs often limit who can benefit from […]
Apple Researchers Introduce Instruction-Following Pruning (IFPruning): A Dynamic AI Approach to Efficient and Scalable LLM Optimization
Large language models (LLMs) have become crucial tools for applications in natural language processing, computational mathematics, and programming. Such models […]
Meet Search-o1: An AI Framework that Integrates the Agentic Search Workflow into the o1-like Reasoning Process of LRM for Achieving Autonomous Knowledge Supplementation
Large reasoning models are developed to solve difficult problems by breaking them down into smaller, manageable steps and solving each […]
