Conversational artificial intelligence is centered on enabling large language models (LLMs) to engage in dynamic interactions where user needs are […]
Category: Large Language Model
xAI says an “unauthorized” prompt change caused Grok to focus on “white genocide”
When analyzing social media posts made by others, Grok is given the somewhat contradictory instructions to “provide truthful and based […]
GenAI vulnerable to prompt injection attacks
New research shows that one in 10 prompt injection atempts against GenAI systems manage to bypass basic guardrails. Their non-deterministic […]
Researchers from Tsinghua and ModelBest Release Ultra-FineWeb: A Trillion-Token Dataset Enhancing LLM Accuracy Across Benchmarks
The data quality used in pretraining LLMs has become increasingly critical to their success. To build information-rich corpora, researchers have […]
xAI’s Grok suddenly can’t stop bringing up “white genocide” in South Africa
Where could Grok have gotten these ideas? The treatment of white farmers in South Africa has been a hobbyhorse of […]
This AI Paper Investigates Test-Time Scaling of English-Centric RLMs for Enhanced Multilingual Reasoning and Domain Generalization
Reasoning language models, or RLMs, are increasingly used to simulate step-by-step problem-solving by generating long, structured reasoning chains. These models […]
Reinforcement Learning, Not Fine-Tuning: Nemotron-Tool-N1 Trains LLMs to Use Tools with Minimal Supervision and Maximum Generalization
Equipping LLMs with external tools or functions has become popular, showing great performance across diverse domains. Existing research depends on […]
OpenAI Releases HealthBench: An Open-Source Benchmark for Measuring the Performance and Safety of Large Language Models in Healthcare
OpenAI has released HealthBench, an open-source evaluation framework designed to measure the performance and safety of large language models (LLMs) […]
PrimeIntellect Releases INTELLECT-2: A 32B Reasoning Model Trained via Distributed Asynchronous Reinforcement Learning
As language models scale in parameter count and reasoning complexity, traditional centralized training pipelines face increasing constraints. High-performance model training […]
This AI Paper Introduces Effective State-Size (ESS): A Metric to Quantify Memory Utilization in Sequence Models for Performance Optimization
In machine learning, sequence models are designed to process data with temporal structure, such as language, time series, or signals. […]
