Nvidia's new texture compression tech slashes VRAM usage by up to 95%

Alfonso Maruccia

Posts: 1,579   +453
Staff
Forward-looking: Nvidia has been working on a new method to compress textures and save GPU memory for a few years now. Although the technology remains in beta, a newly released demo showcases how AI-based solutions could help address the increasingly controversial VRAM limitations of modern GPUs.

Nvidia's Neural Texture Compression can provide gigantic savings in the amount of VRAM required to render complex 3D graphics, even though no one is using it (yet). While still in beta, the technology was tested by YouTube channel Compusemble, which ran the official demo on a modern gaming system to provide an early benchmark of its potential impact and what developers could achieve with it in the not-so-distant future.

As explained by Compusemble in the video below, RTX Neural Texture Compression uses a specialized neural network to compress and decompress material textures dynamically. Nvidia's demo includes three rendering modes: Reference Material, NTC Transcoded to BCn, and Inference on Sample.

  • Reference Material: This mode does not use NTC, meaning textures remain in their original state, leading to high disk and VRAM usage.
  • NTC Transcoded to BCn (block-compressed formats): Here, textures are transcoded upon loading, significantly reducing disk footprint but offering only moderate VRAM savings.
  • Inference on Sample: This approach decompresses texture elements only when needed during rendering, achieving the greatest savings in both disk space and VRAM.

Compusemble tested the demo at 1440p and 4K resolutions, alternating between DLSS and TAA. The results suggest that while NTC can dramatically reduce VRAM and disk space usage, it may also impact frame rates. At 1440p with DLSS, Nvidia's NTC transcoded to BCn mode reduced texture memory usage by 64% (from 272MB to 98MB), while NTC inference on sample drastically cut it to 11.37MB, a 95.8% reduction compared to non-neural compression.

Also read: Why Are Modern PC Games Using So Much VRAM?

The demo ran on a GeForce RTX 4090 GPU, where DLSS and higher resolutions placed additional load on the Tensor Cores, affecting performance to some extent depending on the setting and resolution. However, newer GPUs may deliver higher frame rates and make the difference negligible when properly optimized. After all, Nvidia is heavily invested in AI-powered rendering techniques like NTC and other RTX applications.

The demo also shows the importance of cooperative vectors in modern rendering pipelines. As Microsoft recently explained, cooperative vectors accelerate AI workloads for real-time rendering by optimizing vector operations. These computations play a crucial role in AI model training and fine-tuning and can also be leveraged to enhance game rendering efficiency.

Permalink to story:

 
Can't wait for the RTX 6090 for $5120 with 512MB of memory.

On a more serious note, sounds like another technology that likely worsens the image and has a heavier toll on older cards making them obsolete. So probably for consumers not very interesting but NVIDIA gets to save money on VRAM and ensures consumer grade cards stay limited in AI workloads keeping the data center margins high.
So uhmm win win win for NVIDIA, well played.
 
Last edited:
Can't wait for the RTX 6090 for $5120 with 512MB of memory.

On a more serious note, sounds like another technology that likely worsens the image and has a heavier toll on older cards making them obsolete. So probably for consumers not very interesting but NVIDIA gets to save money on VRAM and ensures consumer grade cards stay limited in AI workloads keeping the data center meetings high.
So uhmm win win win for NVIDIA, well played.

TBF, VRAM is probably the most expensive component on modern GPUs, so in theory (assuming decent market pressure) this would reduce the need for more VRAM leading to lower prices.
 
This BETA tech is probably a necessary step, but not sure applicable to every game. Really needs to go hand in hand with more VRAM - and more modest, less latency settings.
Nvidia is happy to throw your wattage at a problem, but not VRAM - ie they want to give you less for more. All cards except 5090 only had slight bumps in things the really mattered to Frankenstein on their fix up fake frames
 
TBF, VRAM is probably the most expensive component on modern GPUs, so in theory (assuming decent market pressure) this would reduce the need for more VRAM leading to lower prices.
This estimate puts 5090 GPU and memory at about $350usd each.
 
The blame lies with the game devs. They’re the ones who insist on making games so VRAM-dependent.


And they will start writing games with even more intense graphics, which will then require the video card makers to increase the amount of vram, and the cycle will continue.
 
Video and image compression has been worked on for decades by some brilliant people and now Nvidia has achieved 95% compression ratio for video game assets? This tech is ground breaking for all walks of computing and it’s only for games?
 
nVidia's approach to selling GPU power is now to insist on having more of it because you're no longer getting any local storage so have to do everything on the fly via more horsepower. My approach to buying GPU power is to get it from AMD who at least aren't so blatant in their anti-consumer practices. I would rather pay more for a huge pile of VRAM up front and never again, than pay continuously via my power company to keep my CPU and GPU thrashing.
 
It's NVidias' software engineers at least as much as their hardware people that have got them where they are. Even if AMD do produce a good card this time around there is always that nagging doubt about drivers, FSR etc. Fingers crossed the 90 series from AMD is aggressively priced (and works). God knows, we need something to stop the awful gouging being done by NVidia now.
 
I think they are working on utilizing the AI transistors on their GPUs that traditional graphics don't use, to leverage them to create value for gamer owners. Both DLSS and now this. I think it's good to have, even if I don't use it.
 
Video and image compression has been worked on for decades by some brilliant people and now Nvidia has achieved 95% compression ratio for video game assets? This tech is ground breaking for all walks of computing and it’s only for games?
did you know that the porn industry is one of the first industries to use groundbreaking tech? it'll trickle down into other industries
 
Back