Memory Systems in Modern ASIC/SoCs

Memory Systems in Modern ASIC/SoCs

Arithmetic and logical operations in computing systems are executed by the Central Processing Unit (CPU). Modern Systems-on-Chip (SoCs) integrate CPU cores from architectures such as ARM, Intel, or MIPS. These cores process digital data stored in memory and generate results that are ultimately saved to another memory location. To understand the types of memory required in a computing system, consider this analogy:

The Analogy of Memory as a Scratchpad Solving a complex math problem typically requires intermediate steps on paper, even if only the final answer matters. Similarly, a processor needs volatile memory (temporary workspace) to execute operations and derive results. Once finalized, data is stored in non-volatile memory (e.g., SSDs, hard drives) for long-term retention.

Hierarchy of Volatile Memory 1. Registers (Flip-Flops):

  • The fastest and smallest memory units, directly embedded within the CPU.
  • Built using flip-flops (6-8 transistors per bit), they store operands, instructions, or intermediate results during execution.
  • Operate at CPU clock speeds (nanosecond access times) but are limited in quantity (e.g., 16–64 general-purpose registers in modern CPUs).
  • Managed explicitly by the compiler or assembly code to hold critical data for arithmetic/logic operations.

Article content

2. On-Chip Cache (SRAM):

  • Static Random-Access Memory (SRAM) acts as a secondary layer, using 6T cells to store bits without refresh cycles.
  • Divided into L1, L2, L3 caches, with L1 being the fastest (closest to registers) and L3 the largest but slowest.
  • Balances speed and capacity: L1 cache may be 32–64 KB per core, while L3 can reach tens of MB in high-end CPUs.

Article content

3. Off-Chip DRAM (DDR Modules):

  • Dynamic RAM (DRAM) provides high-density, cost-effective volatile memory but requires periodic refreshing.
  • Accessed via a DDR controller and PHY interface, with latencies in the 10–100 nanosecond range.

Article content

Cache Architecture and Memory Access When the CPU requests data, it first checks registers, then the L1 cache, followed by L2/L3 caches. If the data is absent, the DDR controller accesses off-chip DRAM. The memory hierarchy ensures that frequently used data resides in faster, closer storage (registers → cache → DRAM).

Article content

System-Level Optimization Optimal memory performance in SoCs hinges on:

  • Efficient use of registers to minimize cache/memory access.
  • Cache sizing and placement (L1/L2/L3) to balance speed and capacity.
  • Intelligent DDR controller algorithms for row management, request scheduling, and power efficiency.
  • Balancing cost, speed, and density trade-offs between SRAM (on-chip) and DRAM (off-chip).

In summary, modern computing systems rely on a synergistic combination of registers, on-chip SRAM caches, sophisticated DDR controllers, and high-density DRAM modules to deliver the speed and capacity required for complex applications.


AI is dominating ASIC manufacturing. AI is eager to see, a real, In Memory Computing emergence. Some words on MRAM, ReRAM expectation and limits?

To view or add a comment, sign in

More articles by Ajazul Haque

  • DDR2 SDRAM Technology Guide:

    Overview DDR2 SDRAM (Double Data Rate 2 Synchronous Dynamic Random-Access Memory) delivers higher bandwidth, lower…

  • RISCV ISA (Instruction Set Architecture)

    The RV32I base integer ISA provides essential 32-bit integer operations with 47 instructions, while RV64I extends this…

  • The Rise of RISC-V: A Flexible Solution for Modern SoC Design

    Growing Complexity of SoCs As computing demands surge—driven by applications in Machine Learning, Multimedia, and…

    1 Comment
  • DRAM Architecture and Standards:

    1. What is DRAM? DRAM (Dynamic Random Access Memory) is a type of volatile memory used in computers to temporarily…

  • All my UVM/SV articles link path:

    https://www.linkedin.

  • PCIe: All my articles links

    PCI/PCIe Device Type and Topology: PCIe System Architecture: Transaction Layer in PCIe Part -1: Overview and TLP PCIe…

  • Rethinking Education for a Changing World

    One of the defining challenges of our time is the environment in which our children are growing up and the uncertain…

  • PCIe Evolution, Questions and Answers

    Evolution of PCIe Generations The development of PCIe from its inception in 2003 to the latest generation, focusing on…

  • PCIe : Overview of Address Translation Service (ATS)

    The Address Translation Service (ATS) is a key feature in the PCIe architecture designed to manage memory access for…

  • PCIe Transaction Layer issues and TLP Errors:

    Common Issues: Transaction Layer: 1. Interpretation of the Read Completion Boundary (RCB) Parameter: · The RCB defines…

    1 Comment

Insights from the community

Others also viewed

Explore topics