Close Menu
  • Home
  • AI
  • Art & Style
  • Economy
  • Entertainment
  • International
  • Market
  • Opinion
  • Politics
  • Sports
  • Trump
  • US
  • World
What's Hot

European Darts Grand Prix: Gerwyn Price wins record 10th European title with 8-6 win over Ross Smith | Darts News

April 19, 2026

Private credit rift threatens to deepen private equity crisis

April 19, 2026

RBC Heritage: Matt Fitzpatrick holds off Scottie Scheffler in playoff to earn second win of the season on the PGA Tour | Golf News

April 19, 2026
Facebook X (Twitter) Instagram
Smart Breaking News on AI, Business, Politics & Global Trends | WhistleBuzz
Facebook X (Twitter) Instagram
  • Home
  • AI
  • Art & Style
  • Economy
  • Entertainment
  • International
  • Market
  • Opinion
  • Politics
  • Sports
  • Trump
  • US
  • World
Smart Breaking News on AI, Business, Politics & Global Trends | WhistleBuzz
Home » Tensormesh raises $4.5 million to squeeze more inference out of AI server load
AI

Tensormesh raises $4.5 million to squeeze more inference out of AI server load

Editor-In-ChiefBy Editor-In-ChiefOctober 25, 2025No Comments3 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Telegram Email Copy Link
Follow Us
Google News Flipboard
Share
Facebook Twitter LinkedIn Pinterest Email


As the push for AI infrastructure reaches incredible scale, the pressure to squeeze as much inference out of GPUs as possible is greater than ever. And for researchers with expertise in a particular technology, now is a great time to raise funding.

That’s part of the driving force behind Tensormesh, which emerged from stealth this week with $4.5 million in seed funding. The investment was led by Laude Ventures, with additional angel funding provided by database pioneer Michael Franklin.

Tensormesh is using its funding to build a commercial version of its open source LMCache utility, launched and maintained by Tensormesh co-founder Yihua Cheng. When used successfully, LMCache can reduce inference costs by up to 10x. This ability has made LMCache a staple in open source deployments, drawing integrations from powerhouses like Google and Nvidia. Now Tensormesh plans to leverage its academic reputation into a viable business.

The core of this product is the key-value cache (or KV cache). This is a memory system used to process complex input more efficiently by condensing it into key values. In traditional architectures, the KV cache is discarded at the end of each query, which Tensormesh co-founder and CEO Junchen Jiang argues is a major source of inefficiency.

“It’s like a very smart analyst reading all the data, but forgetting what he learned after each question,” Jiang says.

Instead of discarding that cache, Tensormesh’s system preserves it and allows you to redeploy it when your model performs a similar process with another query. Since GPU memory is at a premium, this means distributing the data across multiple different storage layers, but the benefit is significantly more inference power for the same server load.

This change is especially powerful for chat interfaces, as the model must continually reference a chat log that grows as the conversation progresses. The agent system has a similar problem, with a growing log of actions and goals.

In theory, these changes could be made by AI companies on their own, but the technical complexity makes this a difficult task. As the Tensormesh team studies the process and considers the complexity of the details itself, the company believes there will be a lot of demand for a ready-to-use product.

“Keeping the KV cache on a secondary storage system and reusing it efficiently without slowing down the overall system is a very challenging problem,” Jiang says. “We’ve seen people hire 20 engineers and spend three to four months building a system like that. Or they can use our product to build it very efficiently.”



Source link

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Editor-In-Chief
  • Website

Related Posts

OpenAI’s existential questions | Tech Crunch

April 19, 2026

12 Month Window | Tech Crunch

April 19, 2026

TechCrunch Mobility: Uber enters asset maximization era

April 19, 2026
Add A Comment
Leave A Reply Cancel Reply

News

Canada’s Prime Minister says close economic relationship with the US has become a ‘weak link’ | International Trade News

By Editor-In-ChiefApril 19, 2026

Mark Carney said Canada needs to build economic relationships with other countries as its relationship…

President Trump warns of ‘threats’ as Iran reasserts control of the Strait of Hormuz | US and Israel’s war against Iran News

April 18, 2026

Mexico, Spain, Brazil seek protection of Cuba’s sovereignty | Donald Trump News

April 18, 2026
Top Trending

OpenAI’s existential questions | Tech Crunch

By Editor-In-ChiefApril 19, 2026

OpenAI has been all over the news lately, whether it’s about acquisitions,…

12 Month Window | Tech Crunch

By Editor-In-ChiefApril 19, 2026

On a recent episode of the excellent podcast “No Priors,” co-hosted by…

TechCrunch Mobility: Uber enters asset maximization era

By Editor-In-ChiefApril 19, 2026

Welcome to TechCrunch Mobility. A hub for learning about the future of…

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Welcome to WhistleBuzz.com (“we,” “our,” or “us”). Your privacy is important to us. This Privacy Policy explains how we collect, use, disclose, and safeguard your information when you visit our website https://whistlebuzz.com/ (the “Site”). Please read this policy carefully to understand our views and practices regarding your personal data and how we will treat it.

Facebook X (Twitter) Instagram Pinterest YouTube

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Facebook X (Twitter) Instagram Pinterest
  • Home
  • Advertise With Us
  • Contact US
  • DMCA Policy
  • Privacy Policy
  • Terms & Conditions
  • About US
© 2026 whistlebuzz. Designed by whistlebuzz.

Type above and press Enter to search. Press Esc to cancel.