#llm-compression

[ follow ]
#ai
fromTechCrunch
1 week ago
Silicon Valley

Startup Gimlet Labs is solving the AI inference bottleneck in a surprisingly elegant way | TechCrunch

Data science
fromTheregister
14 hours ago

PrismML debuts 1-bit LLM in bid to free AI from the cloud

PrismML's Bonsai 8B is a 1-bit language model that outperforms larger models, enhancing AI efficiency for mobile applications.
Law
fromAbove the Law
4 days ago

The Iron Man Model Of Legal AI - Above the Law

Claude Code empowers developers to enhance their capabilities, transforming them into super developers rather than viewing AI as a threat.
Marketing
from3blmedia
4 days ago

"AI Can't Quote Coverage You Never Generated."

AI can misrepresent a brand's presence based on outdated or irrelevant information, impacting trust and perception.
Data science
fromTheregister
3 days ago

TurboQuant is a big deal, but it won't end the memory crunch

TurboQuant is an AI data compression technology that reduces memory usage for KV caches but may not significantly alleviate memory shortages.
Silicon Valley
fromTechCrunch
1 week ago

Startup Gimlet Labs is solving the AI inference bottleneck in a surprisingly elegant way | TechCrunch

Gimlet Labs raised $80 million to enhance AI inference efficiency across diverse hardware types.
Data science
fromInfoWorld
3 days ago

How to halve Claude output costs with a markdown tweak

A markdown file can reduce Claude's token output by over 50%, aiding enterprises in managing AI costs during production.
#ai-models
Artificial intelligence
fromTNW | Apps
1 day ago

Microsoft launches three in-house AI models in direct challenge to OpenAI

Microsoft has launched three in-house AI models that compete directly with OpenAI, marking a significant shift in its AI strategy.
Artificial intelligence
fromTNW | Apps
1 day ago

Microsoft launches three in-house AI models in direct challenge to OpenAI

Microsoft has launched three in-house AI models that compete directly with OpenAI, marking a significant shift in its AI strategy.
Silicon Valley
fromSilicon Canals
1 day ago

Frugal AI wants to break the global compute hierarchy before it becomes permanent - Silicon Canals

The Soliga tribe's speech AI system exemplifies a new, decentralized approach to AI that challenges existing global tech hierarchies.
Tech industry
fromTheregister
2 days ago

Google battles Chinese open weights models with Gemma 4

Google launched new open-weights Gemma models optimized for agentic AI and coding, offering enterprises a domestic alternative to Chinese LLMs.
Science
fromNature
2 days ago

Breakthrough computer chip tech could help meet 'monumental demand' driven by AI

A new light source enables the creation of 8 nm wide structures on silicon wafers, increasing transistor density for advanced computer chips.
Scala
fromInfoQ
2 days ago

Beyond RAG: Architecting Context-Aware AI Systems with Spring Boot

Context-Augmented Generation (CAG) enhances Retrieval-Augmented Generation (RAG) by managing runtime context for enterprise applications without requiring model retraining.
#ollama
#meta
European startups
fromTheregister
5 days ago

Rebellions eyes global expansion with rack-scale AI platform

Rebellions raised $400 million to expand globally with AI accelerators and a new compute platform for enterprises and sovereign clouds.
fromArs Technica
1 week ago

Google's TurboQuant AI-compression algorithm can reduce LLM memory usage by 6x

PolarQuant is doing most of the compression, but the second step cleans up the rough spots. Google proposes smoothing that out with a technique called Quantized Johnson-Lindenstrauss (QJL).
Roam Research
Software development
fromInfoWorld
3 days ago

Meta shows structured prompts can make LLMs more reliable for code review

Code review is evolving towards machine-led verification, improving accuracy but introducing tradeoffs like increased latency and workflow overhead.
DevOps
fromInfoWorld
1 week ago

An architecture for engineering AI context

AI systems must intelligently manage context to ensure accuracy and reliability in real applications.
Artificial intelligence
fromTechCrunch
2 days ago

Microsoft takes on AI rivals with three new foundational models | TechCrunch

Microsoft AI released three foundational AI models for text, voice, and image generation, emphasizing human-centered design and competitive pricing.
Data science
fromInfoWorld
2 days ago

Why 'curate first, annotate smarter' is reshaping computer vision development

Strategic data selection and curation reduce annotation costs and enhance development productivity in computer vision teams.
#openai
Artificial intelligence
fromFuturism
6 days ago

OpenAI's Obsession With Data Centers Is Running Into Trouble

OpenAI has significantly reduced its AI infrastructure spending plans from $1.4 trillion to $600 billion amid financial pressures and market expectations.
Artificial intelligence
fromFuturism
6 days ago

OpenAI's Obsession With Data Centers Is Running Into Trouble

OpenAI has significantly reduced its AI infrastructure spending plans from $1.4 trillion to $600 billion amid financial pressures and market expectations.
Artificial intelligence
fromTheregister
2 days ago

Microsoft shivs OpenAI with new AI models for speech, images

Microsoft launched public preview versions of machine learning models for speech recognition, speech synthesis, and image generation, competing directly with OpenAI.
#ai-efficiency
Data science
fromInfoWorld
4 days ago

A GitHub tinkerer teaches Claude to talk less, and that may matter more than it seems

A markdown file can significantly reduce AI output token usage, enhancing efficiency without code changes.
Artificial intelligence
fromInfoWorld
1 week ago

Google targets AI inference bottlenecks with TurboQuant

TurboQuant improves AI model efficiency by compressing key-value caches, reducing memory usage and runtime without accuracy loss.
Data science
fromInfoWorld
4 days ago

A GitHub tinkerer teaches Claude to talk less, and that may matter more than it seems

A markdown file can significantly reduce AI output token usage, enhancing efficiency without code changes.
Artificial intelligence
fromInfoWorld
1 week ago

Google targets AI inference bottlenecks with TurboQuant

TurboQuant improves AI model efficiency by compressing key-value caches, reducing memory usage and runtime without accuracy loss.
Tech industry
fromTheregister
2 weeks ago

Nvidia slaps Groq into new LPX racks for faster AI response

Nvidia integrates Groq's language processing units into Vera Rubin systems to dramatically accelerate LLM inference, enabling hundreds to thousands of tokens per second per user.
#llm-safety
Information security
fromInfoWorld
3 weeks ago

19 large language models redefining AI safety-and danger

Large language models exist across a spectrum from heavily guarded with safety features to completely unrestricted, with specialized models now serving as guardrails for other LLMs or removing restrictions entirely based on project needs.
Information security
fromInfoWorld
3 weeks ago

19 large language models redefining AI safety-and danger

Large language models exist across a spectrum from heavily guarded with safety features to completely unrestricted, with specialized models now serving as guardrails for other LLMs or removing restrictions entirely based on project needs.
#ai-safety
Artificial intelligence
fromTechCrunch
3 days ago

Anthropic is having a month | TechCrunch

Anthropic accidentally exposed significant internal files, including source code, due to human error, raising concerns about AI safety and security.
Artificial intelligence
fromTechCrunch
3 days ago

Anthropic is having a month | TechCrunch

Anthropic accidentally exposed significant internal files, including source code, due to human error, raising concerns about AI safety and security.
Data science
fromFast Company
1 week ago

A top AI researcher explains the limitations of current models

Francois Chollet's ARC-AGI-3 benchmark reveals AI's limitations in navigating novel situations compared to human intelligence.
Artificial intelligence
fromFortune
4 days ago

Is AI's visual understanding mostly a 'mirage'? New research suggests so. | Fortune

Anthropic faces significant cybersecurity risks following multiple sensitive data leaks related to its new AI model, Mythos.
Data science
fromTechzine Global
1 week ago

As AI hits scaling limits, Google smashes the context barrier

TurboQuant significantly reduces KV cache size, enhancing AI model performance and expanding context windows for complex workloads.
Software development
fromMedium
2 weeks ago

Inside Dify AI: How RAG, Agents, and LLMOps Work Together in Production

Dify AI provides a unified platform for deploying production language model systems with built-in solutions for data freshness, observability, versioning, and safe deployment across multiple cloud environments.
Artificial intelligence
fromFortune
5 days ago

Nvidia's Jensen Huang says 'We've achieved AGI.' But no one can agree on what AGI means. | Fortune

Nvidia CEO Jensen Huang claims AGI has been achieved, though definitions of AGI vary widely among researchers.
Software development
fromInfoWorld
2 weeks ago

How to build an AI agent that actually works

Successful agents embed intelligence within structured workflows at specific decision points rather than operating autonomously, combining deterministic processes with reasoning models where judgment is needed.
Data science
fromInfoWorld
2 weeks ago

The 'toggle-away' efficiencies: Cutting AI costs inside the training loop

Simple optimizations can significantly reduce AI training costs and carbon emissions without needing the latest GPUs.
Artificial intelligence
fromInfoWorld
1 week ago

Final training of AI models is a fraction of their total cost

Developing AI models incurs significant costs, with most expenditures on scaling and research rather than final training runs.
Software development
fromInfoQ
3 weeks ago

The Oil and Water Moment in AI Architecture

Software architecture is transitioning to AI architecture, requiring architects to manage the coexistence of deterministic systems with non-deterministic AI behavior while shifting from tool-centric to intent-centric thinking.
Artificial intelligence
fromMedium
1 week ago

Less Compute, More Impact: How Model Quantization Fuels the Next Wave of Agentic AI

Model quantization and architectural optimization can outperform larger models, challenging the belief that more GPUs equal greater intelligence.
Artificial intelligence
fromTechCrunch
2 weeks ago

Multiverse Computing pushes its compressed AI models into the mainstream | TechCrunch

Multiverse Computing offers on-device AI models that eliminate counterparty risk by running locally without requiring external compute infrastructure or cloud providers.
Artificial intelligence
fromFast Company
2 weeks ago

OpenAI's new frontier models mark a huge change in how AI will be built

OpenAI released two frontier models in early March: GPT-5.3 optimized for fast responses and GPT-5.4 optimized for deep analytical work, representing a shift toward specialized AI models.
Artificial intelligence
fromInfoWorld
2 weeks ago

Why AI evals are the new necessity for building effective AI agents

User trust in AI agents depends on interaction-layer evaluation measuring reliability and predictability, not just model performance benchmarks.
Artificial intelligence
fromTechCrunch
2 weeks ago

Niv-AI exits stealth to wring more power performance out of GPUs | TechCrunch

AI data centers waste significant power due to GPU demand surges, forcing operators to throttle performance by up to 30%, prompting startups like Niv-AI to develop precision power management solutions.
Artificial intelligence
fromTechzine Global
2 weeks ago

Nvidia's Groq 3 LPU targets agentic AI inference at GTC 2026

Nvidia's acquisition of Groq technology produces the Groq 3 LPU, a specialized inference chip delivering 40 petabytes per second bandwidth, significantly outpacing GPU inference speeds.
fromInfoWorld
3 weeks ago

Neoclouds run AI cheaper and better

By neoclouds, I'm referring to GPU-centric, purpose-built cloud services that focus primarily on AI training and inference rather than on the sprawling catalog of general-purpose services that hyperscalers offer. In many cases, these platforms deliver better price-performance for AI workloads because they're engineered for specific goals: keeping expensive accelerators highly utilized, minimizing platform overhead, and providing a clean path from model development to deployment.
Artificial intelligence
Artificial intelligence
fromZDNET
4 weeks ago

New GPT-5.4 clobbers humans on pro-level work in OpenAI's tests - by 83%

GPT-5.4 matches or outperforms human professionals 83% of the time across nine industries and 44 occupations, with 18% fewer errors and 33% fewer false claims than GPT-5.2.
Artificial intelligence
fromTheregister
1 month ago

OpenAI GPT-5.3 Instant less likely to beat around the bush

GPT-5.3 Instant reduces unnecessary refusals and moralizing preambles while decreasing hallucination rates by up to 26.8 percent compared to prior models.
Software development
fromTheregister
1 month ago

Three AI engines walk into a bar in single file...

Dependency-free single-file LLaMA inference engines in C and JavaScript enable transparent GGUF parsing and token generation for educational, broadly compatible local hardware use.
Artificial intelligence
fromTechCrunch
1 month ago

Perplexity's new Computer is another bet that users need many AI models | TechCrunch

Perplexity launches Computer, an agentic tool for Max subscribers that unifies AI capabilities to execute complex workflows independently using 19 models and create subagents.
Artificial intelligence
fromTheregister
1 month ago

AI models get better at math but still get low marks

Current LLMs struggle with mathematical accuracy, with even top performers scoring C-grade equivalent on practical math benchmarks, though recent versions show modest improvements.
#large-language-models
fromFuturism
2 months ago
Artificial intelligence

AI Agents Are Mathematically Incapable of Doing Functional Work, Paper Finds

fromFuturism
2 months ago
Artificial intelligence

AI Agents Are Mathematically Incapable of Doing Functional Work, Paper Finds

Artificial intelligence
fromTechCrunch
1 month ago

Running AI models is turning into a memory game | TechCrunch

Rising DRAM prices and sophisticated prompt-caching orchestration make memory management a critical cost and performance factor for large-scale AI deployments.
Artificial intelligence
fromInfoQ
2 months ago

Intel DeepMath Introduces a Smart Architecture to Make LLMs Better at Math

DeepMath uses a Qwen3-4B Thinking agent that emits small Python executors for intermediate math steps, improving accuracy and significantly reducing output length.
Artificial intelligence
fromTechzine Global
1 month ago

OpenAI seeks faster alternatives to Nvidia chips

OpenAI seeks alternative inference chips with larger on-chip SRAM to improve response speed for coding and AI-to-AI communication, aiming for about 10% of future inference capacity.
Artificial intelligence
fromTheregister
1 month ago

How AI could eat itself: Using LLMs to distill rivals

Competitors are probing commercial AI models to extract underlying reasoning via distillation attacks to replicate capabilities and lower development costs.
Artificial intelligence
fromHackernoon
1 month ago

This "Flash" AI Model Is Fast and Dangerous at Math-Here's What It Can Do | HackerNoon

GLM-4.7-Flash is a 30-billion-parameter mixture-of-experts model offering strong performance for lightweight deployment.
Artificial intelligence
fromInfoQ
1 month ago

Building LLMs in Resource-Constrained Environments: A Hands-On Perspective

Prioritize small, resource-efficient models and iterative, human-in-the-loop data creation to build practical, improvable AI under infrastructure and data constraints.
fromComputerworld
2 months ago

OpenAI's GPT is getting better at mathematics

OpenAI's GPT-5.2 Pro does better at solving sophisticated math problems than older versions of the company's top large language model, according to a new study by Epoch AI, a non-profit research institute.
Artificial intelligence
Artificial intelligence
fromInfoWorld
1 month ago

First look: Run LLMs locally with LM Studio

LM Studio provides integrated model discovery, in-app download and management, memory-aware filtering, and configurable inference settings for CPU threads and GPU layer offload.
Artificial intelligence
fromLogRocket Blog
1 month ago

LLM routing in production: Choosing the right model for every request - LogRocket Blog

Route requests to appropriate models—cheap models for simple tasks and powerful ones for complex tasks—to reduce cost, latency, and outage risk.
Artificial intelligence
fromInfoQ
2 months ago

Foundation Models for Ranking: Challenges, Successes, and Lessons Learned

Large-scale search and recommendation systems use two-stage retrieval and ranking pipelines to efficiently serve personalized results for hundreds of millions of users and items.
#gpt-53-codex-spark
fromInfoWorld
1 month ago

Researchers propose a self-distillation fix for 'catastrophic forgetting' in LLMs

"To enable the next generation of foundation models, we must solve the problem of continual learning: enabling AI systems to keep learning and improving over time, similar to how humans accumulate knowledge and refine skills throughout their lives," the researchers noted. Reinforcement learning offers a way to train on data generated by the model's own policy, which reduces forgetting. However, it typically requires explicit reward functions, which are not easy in every situation.
Artificial intelligence
Artificial intelligence
fromArs Technica
1 month ago

OpenAI sidesteps Nvidia with unusually fast coding model on plate-sized chips

Cerebras' Wafer Scale Engine enables high token throughput while OpenAI diversifies hardware beyond Nvidia amid fast-paced coding model competition.
Artificial intelligence
fromZDNET
1 month ago

OpenAI's new Spark model codes 15x faster than GPT-5.3-Codex - but there's a catch

Codex-Spark enables conversational, real-time coding with major latency improvements (15x faster code generation; 80% roundtrip, 50% time-to-first-token) using Cerebras WSE-3.
Artificial intelligence
fromZDNET
2 months ago

AI is quietly poisoning itself and pushing models toward collapse - but there's a cure

Unverified AI-generated data causes model collapse and unreliable AI outputs unless organizations enforce data provenance, verification, and governance.
fromInfoQ
1 month ago

Building Embedding Models for Large-Scale Real-World Applications

What happens under the hood? How is the search engine able to take that simple query, look for images in the billions, trillions of images that are available online? How is it able to find this one or similar photos from all that? Usually, there is an embedding model that is doing this work behind the hood.
Artificial intelligence
Artificial intelligence
fromInfoWorld
1 month ago

Single prompt breaks AI safety in 15 major language models

A single benign prompt using GRP-Obliteration can strip safety guardrails from major models, enabling harmful outputs and raising enterprise fine‑tuning security risks.
fromenglish.elpais.com
2 months ago

How does artificial intelligence think? The big surprise is that it intuits'

Each of these achievements would have been a remarkable breakthrough on its own. Solving them all with a single technique is like discovering a master key that unlocks every door at once. Why now? Three pieces converged: algorithms, computing power, and massive amounts of data. We can even put faces to them, because behind each element is a person who took a gamble.
Artificial intelligence
Artificial intelligence
fromZDNET
1 month ago

AI isn't getting smarter, it's getting more power hungry - and expensive

Total computing power explains more model performance gains than proprietary algorithmic 'secret sauce' across 809 large language models.
fromFast Company
1 month ago

Are LTMs the next LLMs? This new type of AI can do what large-language models can't

A major difference between LLMs and LTMs is the type of data they're able to synthesize and use. LLMs use unstructured data-think text, social media posts, emails, etc. LTMs, on the other hand, can extract information or insights from structured data, which could be contained in tables, for instance. Since many enterprises rely on structured data, often contained in spreadsheets, to run their operations, LTMs could have an immediate use case for many organizations.
Artificial intelligence
Artificial intelligence
fromArs Technica
2 months ago

New OpenAI tool renews fears that "AI slop" will overwhelm scientific research

OpenAI's free Prism workspace streamlines LaTeX scientific writing with GPT-5.2 but risks accelerating a flood of low-quality AI-assisted papers into journals.
[ Load more ]