Hello.
Suppose I have 4 addresses going to cache from cpu, A1, A2, A3 and A4. All are 
read requests.
Assume, my L1 cache is blocking.
Now, load store unit will first sent A1.
if it is a miss, the L1 will send it to L2. If it is a L2 hit, it will send the 
data back to L1, then L1 will send it back to cpu.

My question is, when will  A2 be sent ?
1) When the memsideport of cpu becomes available, meaning once the L1 has 
received the A1 (and trying to determine hit/miss etc)  ?
In this case, what will be the delay ? The latency of the crossbar/bus 
connecting L1 and L2 ?

2) Or will the cpu wait until it has received the data for A1 from the L1 ? 
Meaning at any given cycle, that cache object can deal with one single request ?

Also, can the cache in gem5 send multiple misses to lower level memories in the 
same cycle (assuming all MSHR misses)?

_______________________________________________
gem5-users mailing list -- gem5-users@gem5.org
To unsubscribe send an email to gem5-users-le...@gem5.org

Reply via email to