Deep Learning

485 articles found

Google's TurboQuant Slashes LLM Memory by 5x and Boosts Speed 8x With No Accuracy Loss

Google's TurboQuant Slashes LLM Memory by 5x and Boosts Speed 8x With No Accuracy Loss

Mar 25, 2026
MarkTechPost

Google's TurboQuant is revolutionizing AI efficiency, slashing large language model memory usage by over 5x and boosting speed up to 8x with zero accuracy loss, using a data-oblivious quantization algorithm requiring no dataset-specific tuning — maintaining perfect retrieval accuracy across 104,000 tokens in benchmark tests.

Base LLMs Show Strong Semantic Confidence Accuracy, But Fine-Tuning and Chain-of-Thought Reasoning Destroy It

Base LLMs Show Strong Semantic Confidence Accuracy, But Fine-Tuning and Chain-of-Thought Reasoning Destroy It

Mar 25, 2026
Apple Machine Learning Research

New research reveals that base large language models possess strong semantic confidence accuracy, but popular techniques like fine-tuning and chain-of-thought reasoning actively destroy this calibration, raising urgent questions about the reliability of widely deployed AI systems.

Previous
Page 2 of 49
Next
Showing 11 - 20 of 485 articles