Cadence’s Brad Griffin Digs Deep Into DDR


Reading time ( words)

Guest Editor Kelly Dack stopped by the Cadence Design Systems booth at DesignCon 2015, where he sat down with Product Marketing Manager Brad Griffin to discuss Cadence’s advanced PCB design and signal integrity tools, and the company’s focus on DDR.

Kelly Dack: Brad, since you’re the product marketing director for Cadence Design Systems, I’d like to ask a few questions about your DDR products. But first, please give us a brief overview of DDR.

Brad Griffin: I’d be happy to. One of the main things with a computer is that it has memory and you can store data in that memory—that’s kind of what makes it a computing device. So they’ve been finding ways over the life of electronics to store and retrieve data faster out of memory. Somewhere around 2002, we came up with this idea of doubling the data rate in DDR memory, or double data rate memory. That was unique because basically, we clocked the data into the memory, both on the rising edge and on the falling edge of the clock. It was a clever way with the same sort of signaling to basically double the data rate speeds.

KD: Was there an organization involved? Was it standardized? 

BG: That’s really good question. As of right now, there's a standard committee called JEDEC, and I'm going to assume they were in place back in the 2002 timeframe, but I’d have to go back and check. But obviously there's memory companies and they have to be able to plug-and-play with different controllers as they’re driving the memory, so there's probably always been a standard they’ve been marching toward. That process used to be a lot simpler. You’d be transferring data at maybe 100 megabits per second. You would send the data, clock it in, and it wasn’t nearly as complicated as it is now.

kelly_brad2.jpg

KD: So where has DDR come from, and where is it now?

BG: There was DDR2 and then DDR3, and probably 2015 is going to be the transition where most DDR3 designs go over to DDR4. Typically, this happens because the DDR4 memory will actually become less expensive than some of the DDR3 memory. 

KD: What does that mean as far as the technology from a power standpoint as well as a data standpoint?

BG: The main difference from a technology standpoint from DDR3 to DDR4 is the speed. It basically just gets faster. So any application you have in the computer that’s run with DDR4 memory will make for a faster computer than one running with DDR3. One of the exciting things that has migrated probably over the last five to seven years is this new version of DDR called LPDDR, which stands for low power. That’s been something primarily used in mobile devices because you certainly don’t want your cell phone to run out of power in the middle of the day.

KD: With this reference to power, if I understand correctly, DDR came from a 2.5 V system and shrunk to 1.8 V and 1.5 V, and DDR4 is down at a little over 1 V. That seems really low already, so where will the LPDDR take us? 

BG: If you can believe it, the LPDDR4 specification only has a 300 mV swing, so it's really low. That means that for signal integrity and power integrity engineers, there's really very little margin left. We said there was very little margin left when it was 1.5 V, and now we’re down to 300 mV; this very small swing of data means that your signals have to be clean and your power planes have to basically be stable. Because then you have to have a power/ground bounce associated with simultaneous switching signals. It’s going to basically make it so that you're not going to meet the signal quality requirements that JEDEC puts in place for LPDDR4. So designs are getting really interesting. What we’re excited about this year at DesignCon are the things we’ve been putting into our tools to enable designers to validate that they've done everything they need to do to meet the LPDDR4 requirements.

Share


Suggested Items

Artificial Intelligence: The Future of EDA?

10/10/2018 | Andy Shaughnessy, Design007 Magazine
Artificial intelligence (AI) has been making inroads into a variety of industries in the past decade or so, from automobiles to medical devices. Naturally, EDA tool companies are taking a look at AI. Does AI offer a way forward for PCB design tool developers? I recently interviewed Paul Musto, director of marketing for the Board Systems Division of Mentor. We discussed Mentor’s plans for integrating AI into EDA tools, and why we may be at the very beginning of understanding the pros and cons of this new technology.

Front-End Expert Mark Thompson of Prototron Circuits Publishes Book on Producing the Perfect Data Package

10/03/2018 | Dan Beaulieu
Errors and inaccuracies slow the process down because the CAM department has to correct their data package or ask customers to clarify their intent. In the quick-turn prototype business where people pay for time, a slowdown on a three-day turn can be a disaster. I recently sat down with Thompson to find out more about his new book and discuss the quest for the perfect data package.

Moving From 28 Gbps NRZ to 56 Gbps PAM-4: Is it a Free Lunch?

09/24/2018 | Yuriy Shlepnev, Simberian
The usual way of signaling through PCB interconnects is a two-level pulse, an encoding of 1s and 0s or bits, named NRZ (non-return-to-zero) or PAM-2 line code type. Increasing the data rate with the NRZ code type presents some obstacles. For a 28 Gbps NRZ signal, the bit time is about 35.7 ps with the main spectral lobe below 28 GHz. For a 56 Gbps NRZ signal, the bit time is about 17. 86 ps, with the main spectral lobe below 56 GHz. One can feel the problem already: Getting PCB interconnect analysis and measurements up to 56 GHz and beyond is very challenging, to say the least.



Copyright © 2018 I-Connect007. All rights reserved.