Somedays it’s fun to look at the surface level of the computing experience, and other days it’s fun to delve right into the inner workings. Today we’re taking a look at the structure of computer memory and just how much stuff you can pack into a stick of RAM.
Today’s Question & Answer session comes to us courtesy of SuperUser—a subdivision of Stack Exchange, a community-driven grouping of Q&A web sites.
The Question
SuperUser reader Johan Smohan is grappling with how processor type and memory size work together to yield a total number of addresses. He writes:
These are the kinds of questions that can keep a curious geek up at night. How many addresses are available under each of Johan’s hypothetical systems?
I think that it’s something like this:
1GB of ram divided by either 32 bits 4 bits (?) to get the number of memory addresses?
I read on Wikipedia that 1 memory addresses is 32 bits wide or 4 octets (1 octet = 8 bits), compared to a 64 bit processor where 1 memory addresses or 1 integer is 64 bits wide or 8 octets. But don’t know if I understood it correctly either.
The Answer
SuperUser contributor Gronostaj offers some insight into how the RAM is divided and utilized:
Have something to add to the explanation? Sound off in the the comments. Want to read more answers from other tech-savvy Stack Exchange users? Check out the full discussion thread here.
Memory size in bytes Greatest unsigned integer that can be saved in CPU’s machine word
Long answer and explanation of the above:
Memory consists of bytes (B). Each byte consists of 8 bits (b).
1 GB of RAM is actually 1 GiB (gibibyte, not gigabyte). The difference is:
Every byte of memory has its own address, no matter how big the CPU machine word is. Eg. Intel 8086 CPU was 16-bit and it was addressing memory by bytes, so do modern 32-bit and 64-bit CPUs. That’s the cause of the first limit – you can’t have more addresses than memory bytes.
Memory address is just a number of bytes the CPU has to skip from the beginning of the memory to get to the one it’s looking for.
To access the first byte it has to skip 0 bytes, so first byte’s address is 0. To access the second byte it has to skip 1 byte, so its address is 1. (and so forth…) To access the last byte, CPU skips 1073741823 bytes, so its address is 1073741823.
Now you have to know what 32-bit actually means. As I mentioned before, it’s the size of a machine word.
Machine word is the amount of memory CPU uses to hold numbers (in RAM, cache or internal registers). 32-bit CPU uses 32 bits (4 bytes) to hold numbers. Memory addresses are numbers too, so on a 32-bit CPU the memory address consists of 32 bits.
Now think about this: if you have one bit, you can save two values on it: 0 or 1. Add one more bit and you have four values: 0, 1, 2, 3. On three bits, you can save eight values: 0, 1, 2… 6, 7. This is actually a binary system and it works like that:
It works exactly like usual addition, but the maximum digit is 1, not 9. Decimal 0 is 0000, then you add 1 and get 0001, add one once again and you have 0010. What happend here is like with having decimal 09 and adding one: you change 9 to 0 and increment next digit.
From the example above you can see that there’s always a maximum value you can keep in a number with constant number of bits – because when all bits are 1 and you try to increase the value by 1, all bits will become 0, thus breaking the number. It’s called integer overflow and causes many unpleasant problems, both for users and developers.
For 1 bit the greatest value is 1, 2 bits – 3, 3 bits – 7, 4 bits – 15
The greatest possible number is always 2^N-1, where N is the number of bits. As I said before, a memory address is a number and it also has a maximum value. That’s why machine word’s size is also a limit for the number of available memory addresses – sometimes your CPU just can’t process numbers big enough to address more memory.
So on 32 bits you can keep numbers from 0 to 2^32-1, and that’s 4 294 967 295. It’s more than the greatest address in 1 GB RAM, so in your specific case amount of RAM will be the limiting factor.
The RAM limit for 32-bit CPU is theoretically 4 GB (2^32) and for 64-bit CPU it’s 16 EB (exabytes, 1 EB = 2^30 GB). In other words, 64-bit CPU could address entire Internet… 200 times ;) (estimated byWolframAlpha).
However, in real-life operating systems 32-bit CPUs can address about 3 GiB of RAM. That’s because of operating system’s internal architecture – some addresses are reserved for other purposes. You can read more about this so-called 3 GB barrier on Wikipedia. You can lift this limit with Physical Address Extension.
Speaking about memory addressing, there are few things I should mention: virtual memory, segmentation and paging.
Virtual memory
As @Daniel R Hicks pointed out in another answer, OSes use virtual memory. What it means is that applications actually don’t operate on real memory addresses, but ones provided by OS.
This technique allows operating system to move some data from RAM to a so-called Pagefile (Windows) or Swap (*NIX). HDD is few magnitudes slower than RAM, but it’s not a serious problem for rarely accessed data and it allows OS to provide applications more RAM than you actually have installed.
Paging
What we were talking about so far is called flat addressing scheme.
Paging is an alternative addressing scheme that allows to address more memory that you normally could with one machine word in flat model.
Imagine a book filled with 4-letter words. Let’s say there are 1024 numbers on each page. To address a number, you have to know two things:
The number of page on which that word is printed. Which word on that page is the one you’re looking for.
Now that’s exactly how modern x86 CPUs handle memory. It’s divided into 4 KiB pages (1024 machine words each) and those pages have numbers. (actually pages can also be 4 MiB big or 2 MiB with PAE). When you want to address memory cell, you need the page number and address in that page. Note that each memory cell is referenced by exactly one pair of numbers, that won’t be the case for segmentation.
Segmentation
Well, this one is quite similar to paging. It was used in Intel 8086, just to name one example. Groups of addresses are now called memory segments, not pages. The difference is segments can overlap, and they do overlap a lot. For example on 8086 most memory cells were available from 4096 different segments.
An example:
Let’s say we have 8 bytes of memory, all holding zeros except for 4th byte which is equal to 255.
Illustration for flat memory model:
Illustration for paged memory with 4-byte pages:
Illustration for segmented memory with 4-byte segments shifted by 1:
As you can see, 4th byte can be addressed in four ways: (addressing from 0)
Segment 0, offset 3 Segment 1, offset 2 Segment 2, offset 1 Segment 3, offset 0
It’s always the same memory cell.
In real-life implementations segments are shifted by more than 1 byte (for 8086 it was 16 bytes).
What’s bad about segmentation is that it’s complicated (but I think you already know that ;) What’s good, is that you can use some clever techniques to create modular programs.
For example, you can load some module into a segment, then pretend the segment is smaller than it really is (just small enough to hold the module), then choose first segment that doesn’t overlap with that pseudo-smaller one and load next module, and so on. Basically, what you get this way is pages of variable size.