Learning Library

← Back to Library

Moore's Law Dead, Computational Storage Solution

Key Points

  • Moore’s Law has effectively ended: after five decades of regular, cheaper, low‑power CPU upgrades, new processors now arrive less frequently, cost more, and consume 200‑400 W, challenging sustainability goals.
  • As raw transistor scaling stalls, manufacturers are packing more functionality (compression, encryption, database, ML/DL workloads) into CPUs, but this adds complexity and power draw without solving performance bottlenecks.
  • The prevailing architecture still forces massive data movement—entire datasets must be fetched from storage, streamed over the network, and scanned by the CPU even when only a tiny fraction is relevant, wasting bandwidth, memory, and CPU cycles.
  • The “computational storage” paradigm proposes moving analytics and filtering logic into the storage layer so that irrelevant records are discarded before they reach the server, dramatically cutting I/O and processing overhead.
  • By offloading tasks like search, compression, and encryption to storage devices, organizations can mitigate the limits imposed by the end of Moore’s Law and achieve more efficient, sustainable compute workloads.

Full Transcript

# Moore's Law Dead, Computational Storage Solution **Source:** [https://www.youtube.com/watch?v=lCuJnhZaTv8](https://www.youtube.com/watch?v=lCuJnhZaTv8) **Duration:** 00:09:57 ## Summary - Moore’s Law has effectively ended: after five decades of regular, cheaper, low‑power CPU upgrades, new processors now arrive less frequently, cost more, and consume 200‑400 W, challenging sustainability goals. - As raw transistor scaling stalls, manufacturers are packing more functionality (compression, encryption, database, ML/DL workloads) into CPUs, but this adds complexity and power draw without solving performance bottlenecks. - The prevailing architecture still forces massive data movement—entire datasets must be fetched from storage, streamed over the network, and scanned by the CPU even when only a tiny fraction is relevant, wasting bandwidth, memory, and CPU cycles. - The “computational storage” paradigm proposes moving analytics and filtering logic into the storage layer so that irrelevant records are discarded before they reach the server, dramatically cutting I/O and processing overhead. - By offloading tasks like search, compression, and encryption to storage devices, organizations can mitigate the limits imposed by the end of Moore’s Law and achieve more efficient, sustainable compute workloads. ## Sections - [00:00:00](https://www.youtube.com/watch?v=lCuJnhZaTv8&t=0s) **Moore's Law Declared Dead** - The presenter explains that although CPUs still evolve, escalating costs and power demands have ended the historic 18‑to‑24‑month performance gains that defined Moore's Law. - [00:03:05](https://www.youtube.com/watch?v=lCuJnhZaTv8&t=185s) **Computational Storage Offloads CPU Work** - The speaker describes how moving tasks like compression, encryption, RAID, and erasure coding from the main processor to smart NICs/DPUs (computational storage) frees memory and CPU cycles, reducing bandwidth and latency overhead. - [00:06:11](https://www.youtube.com/watch?v=lCuJnhZaTv8&t=371s) **SSD Compute Offloads Data Processing** - The speaker explains that modern SSDs contain powerful compute resources (ARM cores, ASICs, FPGAs) that can use excess internal bandwidth to compress, filter, scan, and analyze data—offloading tasks like ransomware detection and targeted queries from the CPU. - [00:09:20](https://www.youtube.com/watch?v=lCuJnhZaTv8&t=560s) **Optimism Beyond Slowing Moore's Law** - The speaker acknowledges Moore's Law is decelerating yet asserts that ongoing optimization and innovation will sustain a brilliant future of technical advancement, ending with a typical video call‑to‑action. ## Full Transcript
0:00Moore's Law is dead. 0:03Now, I know some of you are thinking, well, is it really dead? 0:09We've had a tremendous run in the last 50 years. 0:14And so are we really saying goodbye to Moore's Law? 0:19Well, we have had a tremendous run. 0:22Let's just draw a compute system over here with memory. 0:29With the CPU with storage down here, and the CPU is connected to the storage device over a network. 0:39And I'm very sorry, network people. 0:41I have reduced you to a single line. 0:44And although the network is a critical part of the system, it sometimes seems like it's too narrow for us. 0:54And even with all of the growth in the network and the storage, we still are at a point where things are slowing down. 1:04Let's look at the CPU, for example. 1:07For 50 years, we could count on a new release every 18 to 24 months, 1:12and we would get more function, we would get more cores, we would get more offload engines inside the CPU. 1:20And not only that, we would pay less for it. 1:24And the power, the wattage, would be about the same. 1:28So every two years we could have more functionality and pay about the same or less for the same server. 1:38That is what is dead. 1:41We may still get releases of new CPU's, but 1:45what's happening now is that they tend to cost more each generation to get that extra power. 1:52And powerful is starting to take on new meaning as CPUs are starting to take 200, 300, 400 or more watts. 2:03It is making it difficult in a world where sustainability is becoming critical. 2:10And so as this becomes more constrained, what we have also done is to put more and more function inside. 2:20So now we have the compression, we have encryption, we do all of the database work. 2:27We have machine learning and deep learning. 2:29So inside this storage box, all the data has to come up here. 2:36It has to be processed by the CPU, it has to be analyzed, 2:41and only the data that is needed is, is what the CPU is going to work on. But everything has to be brought up. 2:50Take, for example, 2:52that you want to look in your your huge amount of data for all the customers who bought a handbag in 2022 3:00so that you can put ads for handbags as they do their Christmas shopping. 3:06Well, to do that, you're going to have to bring my record in. 3:09And I have never bought a handbag in all my life. 3:12And so I'm sitting in memory taking up space along with all of the other people who have never bought handbags. 3:18And you're having to look at each one of these saying, Nope, not him, not him, not him aah yes, him or her.. 3:25and that then gets processed. 3:27But all of that data taking up bandwidth, speed, taking up memory and taking up precious CPU cycles. 3:35That is the problem that we are facing, that although we have beautiful processing systems, 3:41we are spending a lot of time doing things which really could be done other places, 3:48and that is the nature of computational storage. 3:53Let's see where computational storage can help. 3:56Inside the CPU system here - this server, are things that we call HBAs. 4:03Here is a NIC [network interface card] for Ethernet, and here is fiber channel or some other protocol. 4:11If we expand that NIC and 4:17itself has ARM processing, memory and ASIC [application-specific integrated circuits]. 4:24And in that ASIC, there could be offload like compression or encryption or other things. 4:29And so now think about this. 4:31We could take functionality that we're running here, 4:35like a storage stack, like the raid or the erasure coding, 4:40and we could move that into this, and it is in line here. 4:45So data coming from the storage can be processed 4:49by the sNIC, they're today called DPUs as well, Data Processing Units. 4:55So this is an example of where computational storage offload can exist. 5:02But let's go further. 5:03Let's look at a storage box. 5:07A external storage box can have hard disk drives in it and it can have tape. 5:13Tape has made a beautiful ride over the past 50 years. 5:17But in order to really process this quickly in these days, it's SSDs. 5:26And you can have dozens of them in one of these storage arrays. 5:30There's also controllers. 5:32And these controllers are full blown servers in their own right with Intel or AMD processors in them. 5:41And so imagine taking some of that function, which today resides all in the CPU and moving some of it into here. 5:52Today, the job of storage is to store. 5:56I know. 5:56That's amazing, isn't it? 5:58It stores data. 5:59It manages that data. 6:01It does data reduction on that data. 6:03It does encryption, but it largely just stores the data. 6:08But it's got these large processors in it. 6:11And what's more, if you looked at all the bandwidth inside for these systems, it can be 168 gigabytes per second. 6:22But the maximum you're going to get here is on the order of 50 gigabytes per second. 6:28What does that mean? 6:29There's extra bandwidth to use to scan things, to look for customers who bought handbags in 2022 6:36and only report to you those who are relevant for what you want to do with the data. 6:44Now we can go one step further. 6:46So we could do computation here by running a VM or a container and do some of that offload that today is here. 6:55But let's also look at one of these SSDs. 7:03In a modern SSD, there is also a lot of compute power. 7:08There are ARM cores. 7:11There's an ASIC or an FPGA. 7:15And inside here, like in the IBM FlashCore module, there's a compressor 7:21and that compressor is transparently doing data reduction, 7:25offloading what is done here so that the CPU can do other valuable analytics. 7:33Now there's also, of course, a bank of NAND flash, which is where the data is stored. 7:39And today the SSD job, you guessed it, it's to store data and to retrieve data. 7:45But what about using some of the extra bandwidth? 7:49This can have 16 gigabytes per second, 7:52and yet you're not going to get much more out of that than six gigabytes per second. 7:57So imagine all of the filtering, the scanning... 8:01You could do things like looking for ransomware indications, 8:05looking for filtering again and sorting and searching, 8:09offloading some of that so that each of these units is doing and distributing the load 8:18so that it doesn't all have to fall on the CPU. 8:21That's computational storage at its finest. 8:25I'm often asked, "Well, where is computational storage going to happen? 8:30Is it going to be in the smart NIC? 8:34Is it going to be in the storage box? 8:37Is it going to be in the SSD?" 8:40My brilliant answer is "yes". 8:43It's going to be in all of those because each one has a unique benefit for certain offload. 8:52So what is happening now to move computational storage to the next level 8:57is that OCP, SNIA and NVMe are working on standards 9:01so that you can do this offload in a in a standardized way 9:07so that it can be brought then to the applications, to the database, to the search engines 9:13so that we can get this across a wide variety of applications. 9:20What we're going to see then is that although Moore's Law is slowing down, 9:27we are still going to have a brilliant future of being able to optimize what we have, take advantage of it, 9:37and still further the fantastic revolution that we've had in technical advancements. 9:46Thank you so much for watching. 9:48If you like this video and want to see more like it, please like and subscribe. 9:54If you have questions, please drop them in the comments below.