mem: Add cache clusivity
This patch adds a parameter to control the cache clusivity, that is if the cache is mostly inclusive or exclusive. At the moment there is no intention to support strict policies, and thus the options are: 1) mostly inclusive, or 2) mostly exclusive. The choice of policy guides the behaviuor on a cache fill, and a new helper function, allocOnFill, is created to encapsulate the decision making process. For the timing mode, the decision is annotated on the MSHR on sending out the downstream packet, and in atomic we directly pass the decision to handleFill. We (ab)use the tempBlock in cases where we are not allocating on fill, leaving the rest of the cache unaffected. Simple and effective. This patch also makes it more explicit that multiple caches are allowed to consider a block writable (this is the case also before this patch). That is, for a mostly inclusive cache, multiple caches upstream may also consider the block exclusive. The caches considering the block writable/exclusive all appear along the same path to memory, and from a coherency protocol point of view it works due to the fact that we always snoop upwards in zero time before querying any downstream cache. Note that this patch does not introduce clean writebacks. Thus, for clean lines we are essentially removing a cache level if it is made mostly exclusive. For example, lines from the read-only L1 instruction cache or table-walker cache are always clean, and simply get dropped rather than being passed to the L2. If the L2 is mostly exclusive and does not allocate on fill it will thus never hold the line. A follow on patch adds the clean writebacks. The patch changes the L2 of the O3_ARM_v7a CPU configuration to be mostly exclusive (and stats are affected accordingly).
This commit is contained in:
parent
f02a9338c1
commit
654266f39c
9 changed files with 207 additions and 37 deletions
|
@ -185,6 +185,7 @@ class O3_ARM_v7aL2(Cache):
|
||||||
assoc = 16
|
assoc = 16
|
||||||
write_buffers = 8
|
write_buffers = 8
|
||||||
prefetch_on_access = True
|
prefetch_on_access = True
|
||||||
|
clusivity = 'mostly_excl'
|
||||||
# Simple stride prefetcher
|
# Simple stride prefetcher
|
||||||
prefetcher = StridePrefetcher(degree=8, latency = 1)
|
prefetcher = StridePrefetcher(degree=8, latency = 1)
|
||||||
tags = RandomRepl()
|
tags = RandomRepl()
|
||||||
|
|
16
src/mem/cache/Cache.py
vendored
16
src/mem/cache/Cache.py
vendored
|
@ -84,6 +84,22 @@ class BaseCache(MemObject):
|
||||||
|
|
||||||
system = Param.System(Parent.any, "System we belong to")
|
system = Param.System(Parent.any, "System we belong to")
|
||||||
|
|
||||||
|
# Enum for cache clusivity, currently mostly inclusive or mostly
|
||||||
|
# exclusive.
|
||||||
|
class Clusivity(Enum): vals = ['mostly_incl', 'mostly_excl']
|
||||||
|
|
||||||
class Cache(BaseCache):
|
class Cache(BaseCache):
|
||||||
type = 'Cache'
|
type = 'Cache'
|
||||||
cxx_header = 'mem/cache/cache.hh'
|
cxx_header = 'mem/cache/cache.hh'
|
||||||
|
|
||||||
|
# Control whether this cache should be mostly inclusive or mostly
|
||||||
|
# exclusive with respect to upstream caches. The behaviour on a
|
||||||
|
# fill is determined accordingly. For a mostly inclusive cache,
|
||||||
|
# blocks are allocated on all fill operations. Thus, L1 caches
|
||||||
|
# should be set as mostly inclusive even if they have no upstream
|
||||||
|
# caches. In the case of a mostly exclusive cache, fills are not
|
||||||
|
# allocating unless they came directly from a non-caching source,
|
||||||
|
# e.g. a table walker. Additionally, on a hit from an upstream
|
||||||
|
# cache a line is dropped for a mostly exclusive cache.
|
||||||
|
clusivity = Param.Clusivity('mostly_incl',
|
||||||
|
"Clusivity with upstream cache")
|
||||||
|
|
12
src/mem/cache/base.hh
vendored
12
src/mem/cache/base.hh
vendored
|
@ -210,7 +210,8 @@ class BaseCache : public MemObject
|
||||||
// overlap
|
// overlap
|
||||||
assert(addr == blockAlign(addr));
|
assert(addr == blockAlign(addr));
|
||||||
|
|
||||||
MSHR *mshr = mq->allocate(addr, size, pkt, time, order++);
|
MSHR *mshr = mq->allocate(addr, size, pkt, time, order++,
|
||||||
|
allocOnFill(pkt->cmd));
|
||||||
|
|
||||||
if (mq->isFull()) {
|
if (mq->isFull()) {
|
||||||
setBlocked((BlockedCause)mq->index);
|
setBlocked((BlockedCause)mq->index);
|
||||||
|
@ -233,6 +234,15 @@ class BaseCache : public MemObject
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Determine if we should allocate on a fill or not.
|
||||||
|
*
|
||||||
|
* @param cmd Packet command being added as an MSHR target
|
||||||
|
*
|
||||||
|
* @return Whether we should allocate on a fill or not
|
||||||
|
*/
|
||||||
|
virtual bool allocOnFill(MemCmd cmd) const = 0;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Write back dirty blocks in the cache using functional accesses.
|
* Write back dirty blocks in the cache using functional accesses.
|
||||||
*/
|
*/
|
||||||
|
|
114
src/mem/cache/cache.cc
vendored
114
src/mem/cache/cache.cc
vendored
|
@ -68,7 +68,11 @@ Cache::Cache(const CacheParams *p)
|
||||||
tags(p->tags),
|
tags(p->tags),
|
||||||
prefetcher(p->prefetcher),
|
prefetcher(p->prefetcher),
|
||||||
doFastWrites(true),
|
doFastWrites(true),
|
||||||
prefetchOnAccess(p->prefetch_on_access)
|
prefetchOnAccess(p->prefetch_on_access),
|
||||||
|
clusivity(p->clusivity),
|
||||||
|
tempBlockWriteback(nullptr),
|
||||||
|
writebackTempBlockAtomicEvent(this, false,
|
||||||
|
EventBase::Delayed_Writeback_Pri)
|
||||||
{
|
{
|
||||||
tempBlock = new CacheBlk();
|
tempBlock = new CacheBlk();
|
||||||
tempBlock->data = new uint8_t[blkSize];
|
tempBlock->data = new uint8_t[blkSize];
|
||||||
|
@ -198,10 +202,10 @@ Cache::satisfyCpuSideRequest(PacketPtr pkt, CacheBlk *blk,
|
||||||
if (blk->isDirty()) {
|
if (blk->isDirty()) {
|
||||||
pkt->assertMemInhibit();
|
pkt->assertMemInhibit();
|
||||||
}
|
}
|
||||||
// on ReadExReq we give up our copy unconditionally
|
// on ReadExReq we give up our copy unconditionally,
|
||||||
if (blk != tempBlock)
|
// even if this cache is mostly inclusive, we may want
|
||||||
tags->invalidate(blk);
|
// to revisit this
|
||||||
blk->invalidate();
|
invalidateBlock(blk);
|
||||||
} else if (blk->isWritable() && !pending_downgrade &&
|
} else if (blk->isWritable() && !pending_downgrade &&
|
||||||
!pkt->sharedAsserted() &&
|
!pkt->sharedAsserted() &&
|
||||||
pkt->cmd != MemCmd::ReadCleanReq) {
|
pkt->cmd != MemCmd::ReadCleanReq) {
|
||||||
|
@ -220,9 +224,30 @@ Cache::satisfyCpuSideRequest(PacketPtr pkt, CacheBlk *blk,
|
||||||
if (!deferred_response) {
|
if (!deferred_response) {
|
||||||
// if we are responding immediately and can
|
// if we are responding immediately and can
|
||||||
// signal that we're transferring ownership
|
// signal that we're transferring ownership
|
||||||
// along with exclusivity, do so
|
// (inhibit set) along with exclusivity
|
||||||
|
// (shared not set), do so
|
||||||
pkt->assertMemInhibit();
|
pkt->assertMemInhibit();
|
||||||
|
|
||||||
|
// if this cache is mostly inclusive, we keep
|
||||||
|
// the block as writable (exclusive), and pass
|
||||||
|
// it upwards as writable and dirty
|
||||||
|
// (modified), hence we have multiple caches
|
||||||
|
// considering the same block writable,
|
||||||
|
// something that we get away with due to the
|
||||||
|
// fact that: 1) this cache has been
|
||||||
|
// considered the ordering points and
|
||||||
|
// responded to all snoops up till now, and 2)
|
||||||
|
// we always snoop upwards before consulting
|
||||||
|
// the local cache, both on a normal request
|
||||||
|
// (snooping done by the crossbar), and on a
|
||||||
|
// snoop
|
||||||
blk->status &= ~BlkDirty;
|
blk->status &= ~BlkDirty;
|
||||||
|
|
||||||
|
// if this cache is mostly exclusive with
|
||||||
|
// respect to the cache above, drop the block
|
||||||
|
if (clusivity == Enums::mostly_excl) {
|
||||||
|
invalidateBlock(blk);
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
// if we're responding after our own miss,
|
// if we're responding after our own miss,
|
||||||
// there's a window where the recipient didn't
|
// there's a window where the recipient didn't
|
||||||
|
@ -241,9 +266,10 @@ Cache::satisfyCpuSideRequest(PacketPtr pkt, CacheBlk *blk,
|
||||||
// Upgrade or Invalidate, since we have it Exclusively (E or
|
// Upgrade or Invalidate, since we have it Exclusively (E or
|
||||||
// M), we ack then invalidate.
|
// M), we ack then invalidate.
|
||||||
assert(pkt->isUpgrade() || pkt->isInvalidate());
|
assert(pkt->isUpgrade() || pkt->isInvalidate());
|
||||||
assert(blk != tempBlock);
|
|
||||||
tags->invalidate(blk);
|
// for invalidations we could be looking at the temp block
|
||||||
blk->invalidate();
|
// (for upgrades we always allocate)
|
||||||
|
invalidateBlock(blk);
|
||||||
DPRINTF(Cache, "%s for %s addr %#llx size %d (invalidation)\n",
|
DPRINTF(Cache, "%s for %s addr %#llx size %d (invalidation)\n",
|
||||||
__func__, pkt->cmdString(), pkt->getAddr(), pkt->getSize());
|
__func__, pkt->cmdString(), pkt->getAddr(), pkt->getSize());
|
||||||
}
|
}
|
||||||
|
@ -761,7 +787,8 @@ Cache::recvTimingReq(PacketPtr pkt)
|
||||||
// buffer and to schedule an event to the queued
|
// buffer and to schedule an event to the queued
|
||||||
// port and also takes into account the additional
|
// port and also takes into account the additional
|
||||||
// delay of the xbar.
|
// delay of the xbar.
|
||||||
mshr->allocateTarget(pkt, forward_time, order++);
|
mshr->allocateTarget(pkt, forward_time, order++,
|
||||||
|
allocOnFill(pkt->cmd));
|
||||||
if (mshr->getNumTargets() == numTarget) {
|
if (mshr->getNumTargets() == numTarget) {
|
||||||
noTargetMSHR = mshr;
|
noTargetMSHR = mshr;
|
||||||
setBlocked(Blocked_NoTargets);
|
setBlocked(Blocked_NoTargets);
|
||||||
|
@ -1027,13 +1054,15 @@ Cache::recvAtomic(PacketPtr pkt)
|
||||||
|
|
||||||
// write-line request to the cache that promoted
|
// write-line request to the cache that promoted
|
||||||
// the write to a whole line
|
// the write to a whole line
|
||||||
blk = handleFill(pkt, blk, writebacks);
|
blk = handleFill(pkt, blk, writebacks,
|
||||||
|
allocOnFill(pkt->cmd));
|
||||||
satisfyCpuSideRequest(pkt, blk);
|
satisfyCpuSideRequest(pkt, blk);
|
||||||
} else if (bus_pkt->isRead() ||
|
} else if (bus_pkt->isRead() ||
|
||||||
bus_pkt->cmd == MemCmd::UpgradeResp) {
|
bus_pkt->cmd == MemCmd::UpgradeResp) {
|
||||||
// we're updating cache state to allow us to
|
// we're updating cache state to allow us to
|
||||||
// satisfy the upstream request from the cache
|
// satisfy the upstream request from the cache
|
||||||
blk = handleFill(bus_pkt, blk, writebacks);
|
blk = handleFill(bus_pkt, blk, writebacks,
|
||||||
|
allocOnFill(pkt->cmd));
|
||||||
satisfyCpuSideRequest(pkt, blk);
|
satisfyCpuSideRequest(pkt, blk);
|
||||||
} else {
|
} else {
|
||||||
// we're satisfying the upstream request without
|
// we're satisfying the upstream request without
|
||||||
|
@ -1056,9 +1085,34 @@ Cache::recvAtomic(PacketPtr pkt)
|
||||||
// immediately rather than calling requestMemSideBus() as we do
|
// immediately rather than calling requestMemSideBus() as we do
|
||||||
// there).
|
// there).
|
||||||
|
|
||||||
// Handle writebacks (from the response handling) if needed
|
// do any writebacks resulting from the response handling
|
||||||
doWritebacksAtomic(writebacks);
|
doWritebacksAtomic(writebacks);
|
||||||
|
|
||||||
|
// if we used temp block, check to see if its valid and if so
|
||||||
|
// clear it out, but only do so after the call to recvAtomic is
|
||||||
|
// finished so that any downstream observers (such as a snoop
|
||||||
|
// filter), first see the fill, and only then see the eviction
|
||||||
|
if (blk == tempBlock && tempBlock->isValid()) {
|
||||||
|
// the atomic CPU calls recvAtomic for fetch and load/store
|
||||||
|
// sequentuially, and we may already have a tempBlock
|
||||||
|
// writeback from the fetch that we have not yet sent
|
||||||
|
if (tempBlockWriteback) {
|
||||||
|
// if that is the case, write the prevoius one back, and
|
||||||
|
// do not schedule any new event
|
||||||
|
writebackTempBlockAtomic();
|
||||||
|
} else {
|
||||||
|
// the writeback/clean eviction happens after the call to
|
||||||
|
// recvAtomic has finished (but before any successive
|
||||||
|
// calls), so that the response handling from the fill is
|
||||||
|
// allowed to happen first
|
||||||
|
schedule(writebackTempBlockAtomicEvent, curTick());
|
||||||
|
}
|
||||||
|
|
||||||
|
tempBlockWriteback = blk->isDirty() ? writebackBlk(blk) :
|
||||||
|
cleanEvictBlk(blk);
|
||||||
|
blk->invalidate();
|
||||||
|
}
|
||||||
|
|
||||||
if (pkt->needsResponse()) {
|
if (pkt->needsResponse()) {
|
||||||
pkt->makeAtomicResponse();
|
pkt->makeAtomicResponse();
|
||||||
}
|
}
|
||||||
|
@ -1214,7 +1268,7 @@ Cache::recvTimingResp(PacketPtr pkt)
|
||||||
DPRINTF(Cache, "Block for addr %#llx being updated in Cache\n",
|
DPRINTF(Cache, "Block for addr %#llx being updated in Cache\n",
|
||||||
pkt->getAddr());
|
pkt->getAddr());
|
||||||
|
|
||||||
blk = handleFill(pkt, blk, writebacks);
|
blk = handleFill(pkt, blk, writebacks, mshr->allocOnFill);
|
||||||
assert(blk != NULL);
|
assert(blk != NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1258,7 +1312,7 @@ Cache::recvTimingResp(PacketPtr pkt)
|
||||||
// deferred targets if possible
|
// deferred targets if possible
|
||||||
mshr->promoteExclusive();
|
mshr->promoteExclusive();
|
||||||
// NB: we use the original packet here and not the response!
|
// NB: we use the original packet here and not the response!
|
||||||
blk = handleFill(tgt_pkt, blk, writebacks);
|
blk = handleFill(tgt_pkt, blk, writebacks, mshr->allocOnFill);
|
||||||
assert(blk != NULL);
|
assert(blk != NULL);
|
||||||
|
|
||||||
// treat as a fill, and discard the invalidation
|
// treat as a fill, and discard the invalidation
|
||||||
|
@ -1362,9 +1416,7 @@ Cache::recvTimingResp(PacketPtr pkt)
|
||||||
// should not invalidate the block, so check if the
|
// should not invalidate the block, so check if the
|
||||||
// invalidation should be discarded
|
// invalidation should be discarded
|
||||||
if (is_invalidate || mshr->hasPostInvalidate()) {
|
if (is_invalidate || mshr->hasPostInvalidate()) {
|
||||||
assert(blk != tempBlock);
|
invalidateBlock(blk);
|
||||||
tags->invalidate(blk);
|
|
||||||
blk->invalidate();
|
|
||||||
} else if (mshr->hasPostDowngrade()) {
|
} else if (mshr->hasPostDowngrade()) {
|
||||||
blk->status &= ~BlkWritable;
|
blk->status &= ~BlkWritable;
|
||||||
}
|
}
|
||||||
|
@ -1588,6 +1640,13 @@ Cache::allocateBlock(Addr addr, bool is_secure, PacketList &writebacks)
|
||||||
return blk;
|
return blk;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void
|
||||||
|
Cache::invalidateBlock(CacheBlk *blk)
|
||||||
|
{
|
||||||
|
if (blk != tempBlock)
|
||||||
|
tags->invalidate(blk);
|
||||||
|
blk->invalidate();
|
||||||
|
}
|
||||||
|
|
||||||
// Note that the reason we return a list of writebacks rather than
|
// Note that the reason we return a list of writebacks rather than
|
||||||
// inserting them directly in the write buffer is that this function
|
// inserting them directly in the write buffer is that this function
|
||||||
|
@ -1595,7 +1654,8 @@ Cache::allocateBlock(Addr addr, bool is_secure, PacketList &writebacks)
|
||||||
// mode we don't mess with the write buffer (we just perform the
|
// mode we don't mess with the write buffer (we just perform the
|
||||||
// writebacks atomically once the original request is complete).
|
// writebacks atomically once the original request is complete).
|
||||||
CacheBlk*
|
CacheBlk*
|
||||||
Cache::handleFill(PacketPtr pkt, CacheBlk *blk, PacketList &writebacks)
|
Cache::handleFill(PacketPtr pkt, CacheBlk *blk, PacketList &writebacks,
|
||||||
|
bool allocate)
|
||||||
{
|
{
|
||||||
assert(pkt->isResponse() || pkt->cmd == MemCmd::WriteLineReq);
|
assert(pkt->isResponse() || pkt->cmd == MemCmd::WriteLineReq);
|
||||||
Addr addr = pkt->getAddr();
|
Addr addr = pkt->getAddr();
|
||||||
|
@ -1619,11 +1679,14 @@ Cache::handleFill(PacketPtr pkt, CacheBlk *blk, PacketList &writebacks)
|
||||||
// happens in the subsequent satisfyCpuSideRequest.
|
// happens in the subsequent satisfyCpuSideRequest.
|
||||||
assert(pkt->isRead() || pkt->cmd == MemCmd::WriteLineReq);
|
assert(pkt->isRead() || pkt->cmd == MemCmd::WriteLineReq);
|
||||||
|
|
||||||
// need to do a replacement
|
// need to do a replacement if allocating, otherwise we stick
|
||||||
blk = allocateBlock(addr, is_secure, writebacks);
|
// with the temporary storage
|
||||||
|
blk = allocate ? allocateBlock(addr, is_secure, writebacks) : NULL;
|
||||||
|
|
||||||
if (blk == NULL) {
|
if (blk == NULL) {
|
||||||
// No replaceable block... just use temporary storage to
|
// No replaceable block or a mostly exclusive
|
||||||
// complete the current request and then get rid of it
|
// cache... just use temporary storage to complete the
|
||||||
|
// current request and then get rid of it
|
||||||
assert(!tempBlock->isValid());
|
assert(!tempBlock->isValid());
|
||||||
blk = tempBlock;
|
blk = tempBlock;
|
||||||
tempBlock->set = tags->extractSet(addr);
|
tempBlock->set = tags->extractSet(addr);
|
||||||
|
@ -1877,6 +1940,7 @@ Cache::handleSnoop(PacketPtr pkt, CacheBlk *blk, bool is_timing,
|
||||||
// applies both to reads and writes and that for writes it
|
// applies both to reads and writes and that for writes it
|
||||||
// works thanks to the fact that we still have dirty data and
|
// works thanks to the fact that we still have dirty data and
|
||||||
// will write it back at a later point
|
// will write it back at a later point
|
||||||
|
assert(!pkt->memInhibitAsserted());
|
||||||
pkt->assertMemInhibit();
|
pkt->assertMemInhibit();
|
||||||
if (have_exclusive) {
|
if (have_exclusive) {
|
||||||
// in the case of an uncacheable request there is no point
|
// in the case of an uncacheable request there is no point
|
||||||
|
@ -1911,9 +1975,7 @@ Cache::handleSnoop(PacketPtr pkt, CacheBlk *blk, bool is_timing,
|
||||||
// Do this last in case it deallocates block data or something
|
// Do this last in case it deallocates block data or something
|
||||||
// like that
|
// like that
|
||||||
if (invalidate) {
|
if (invalidate) {
|
||||||
if (blk != tempBlock)
|
invalidateBlock(blk);
|
||||||
tags->invalidate(blk);
|
|
||||||
blk->invalidate();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
DPRINTF(Cache, "new state is %s\n", blk->print());
|
DPRINTF(Cache, "new state is %s\n", blk->print());
|
||||||
|
|
71
src/mem/cache/cache.hh
vendored
71
src/mem/cache/cache.hh
vendored
|
@ -1,5 +1,5 @@
|
||||||
/*
|
/*
|
||||||
* Copyright (c) 2012-2014 ARM Limited
|
* Copyright (c) 2012-2015 ARM Limited
|
||||||
* All rights reserved.
|
* All rights reserved.
|
||||||
*
|
*
|
||||||
* The license below extends only to copyright in the software and shall
|
* The license below extends only to copyright in the software and shall
|
||||||
|
@ -53,6 +53,7 @@
|
||||||
#define __MEM_CACHE_CACHE_HH__
|
#define __MEM_CACHE_CACHE_HH__
|
||||||
|
|
||||||
#include "base/misc.hh" // fatal, panic, and warn
|
#include "base/misc.hh" // fatal, panic, and warn
|
||||||
|
#include "enums/Clusivity.hh"
|
||||||
#include "mem/cache/base.hh"
|
#include "mem/cache/base.hh"
|
||||||
#include "mem/cache/blk.hh"
|
#include "mem/cache/blk.hh"
|
||||||
#include "mem/cache/mshr.hh"
|
#include "mem/cache/mshr.hh"
|
||||||
|
@ -194,12 +195,48 @@ class Cache : public BaseCache
|
||||||
*/
|
*/
|
||||||
const bool prefetchOnAccess;
|
const bool prefetchOnAccess;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Clusivity with respect to the upstream cache, determining if we
|
||||||
|
* fill into both this cache and the cache above on a miss. Note
|
||||||
|
* that we currently do not support strict clusivity policies.
|
||||||
|
*/
|
||||||
|
const Enums::Clusivity clusivity;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Upstream caches need this packet until true is returned, so
|
* Upstream caches need this packet until true is returned, so
|
||||||
* hold it for deletion until a subsequent call
|
* hold it for deletion until a subsequent call
|
||||||
*/
|
*/
|
||||||
std::unique_ptr<Packet> pendingDelete;
|
std::unique_ptr<Packet> pendingDelete;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Writebacks from the tempBlock, resulting on the response path
|
||||||
|
* in atomic mode, must happen after the call to recvAtomic has
|
||||||
|
* finished (for the right ordering of the packets). We therefore
|
||||||
|
* need to hold on to the packets, and have a method and an event
|
||||||
|
* to send them.
|
||||||
|
*/
|
||||||
|
PacketPtr tempBlockWriteback;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Send the outstanding tempBlock writeback. To be called after
|
||||||
|
* recvAtomic finishes in cases where the block we filled is in
|
||||||
|
* fact the tempBlock, and now needs to be written back.
|
||||||
|
*/
|
||||||
|
void writebackTempBlockAtomic() {
|
||||||
|
assert(tempBlockWriteback != nullptr);
|
||||||
|
PacketList writebacks{tempBlockWriteback};
|
||||||
|
doWritebacksAtomic(writebacks);
|
||||||
|
tempBlockWriteback = nullptr;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* An event to writeback the tempBlock after recvAtomic
|
||||||
|
* finishes. To avoid other calls to recvAtomic getting in
|
||||||
|
* between, we create this event with a higher priority.
|
||||||
|
*/
|
||||||
|
EventWrapper<Cache, &Cache::writebackTempBlockAtomic> \
|
||||||
|
writebackTempBlockAtomicEvent;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Does all the processing necessary to perform the provided request.
|
* Does all the processing necessary to perform the provided request.
|
||||||
* @param pkt The memory request to perform.
|
* @param pkt The memory request to perform.
|
||||||
|
@ -225,6 +262,13 @@ class Cache : public BaseCache
|
||||||
*/
|
*/
|
||||||
CacheBlk *allocateBlock(Addr addr, bool is_secure, PacketList &writebacks);
|
CacheBlk *allocateBlock(Addr addr, bool is_secure, PacketList &writebacks);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Invalidate a cache block.
|
||||||
|
*
|
||||||
|
* @param blk Block to invalidate
|
||||||
|
*/
|
||||||
|
void invalidateBlock(CacheBlk *blk);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Populates a cache block and handles all outstanding requests for the
|
* Populates a cache block and handles all outstanding requests for the
|
||||||
* satisfied fill request. This version takes two memory requests. One
|
* satisfied fill request. This version takes two memory requests. One
|
||||||
|
@ -232,11 +276,34 @@ class Cache : public BaseCache
|
||||||
* @param pkt The memory request with the fill data.
|
* @param pkt The memory request with the fill data.
|
||||||
* @param blk The cache block if it already exists.
|
* @param blk The cache block if it already exists.
|
||||||
* @param writebacks List for any writebacks that need to be performed.
|
* @param writebacks List for any writebacks that need to be performed.
|
||||||
|
* @param allocate Whether to allocate a block or use the temp block
|
||||||
* @return Pointer to the new cache block.
|
* @return Pointer to the new cache block.
|
||||||
*/
|
*/
|
||||||
CacheBlk *handleFill(PacketPtr pkt, CacheBlk *blk,
|
CacheBlk *handleFill(PacketPtr pkt, CacheBlk *blk,
|
||||||
PacketList &writebacks);
|
PacketList &writebacks, bool allocate);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Determine whether we should allocate on a fill or not. If this
|
||||||
|
* cache is mostly inclusive with regards to the upstream cache(s)
|
||||||
|
* we always allocate (for any non-forwarded and cacheable
|
||||||
|
* requests). In the case of a mostly exclusive cache, we allocate
|
||||||
|
* on fill if the packet did not come from a cache, thus if we:
|
||||||
|
* are dealing with a whole-line write (the latter behaves much
|
||||||
|
* like a writeback), the original target packet came from a
|
||||||
|
* non-caching source, or if we are performing a prefetch or LLSC.
|
||||||
|
*
|
||||||
|
* @param cmd Command of the incoming requesting packet
|
||||||
|
* @return Whether we should allocate on the fill
|
||||||
|
*/
|
||||||
|
inline bool allocOnFill(MemCmd cmd) const
|
||||||
|
{
|
||||||
|
return clusivity == Enums::mostly_incl ||
|
||||||
|
cmd == MemCmd::WriteLineReq ||
|
||||||
|
cmd == MemCmd::ReadReq ||
|
||||||
|
cmd == MemCmd::WriteReq ||
|
||||||
|
cmd.isPrefetch() ||
|
||||||
|
cmd.isLLSC();
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Performs the access specified by the request.
|
* Performs the access specified by the request.
|
||||||
|
|
14
src/mem/cache/mshr.cc
vendored
14
src/mem/cache/mshr.cc
vendored
|
@ -66,7 +66,8 @@ MSHR::MSHR() : readyTime(0), _isUncacheable(false), downstreamPending(false),
|
||||||
postInvalidate(false), postDowngrade(false),
|
postInvalidate(false), postDowngrade(false),
|
||||||
queue(NULL), order(0), blkAddr(0),
|
queue(NULL), order(0), blkAddr(0),
|
||||||
blkSize(0), isSecure(false), inService(false),
|
blkSize(0), isSecure(false), inService(false),
|
||||||
isForward(false), threadNum(InvalidThreadID), data(NULL)
|
isForward(false), allocOnFill(false),
|
||||||
|
threadNum(InvalidThreadID), data(NULL)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -202,7 +203,7 @@ MSHR::TargetList::print(std::ostream &os, int verbosity,
|
||||||
|
|
||||||
void
|
void
|
||||||
MSHR::allocate(Addr blk_addr, unsigned blk_size, PacketPtr target,
|
MSHR::allocate(Addr blk_addr, unsigned blk_size, PacketPtr target,
|
||||||
Tick when_ready, Counter _order)
|
Tick when_ready, Counter _order, bool alloc_on_fill)
|
||||||
{
|
{
|
||||||
blkAddr = blk_addr;
|
blkAddr = blk_addr;
|
||||||
blkSize = blk_size;
|
blkSize = blk_size;
|
||||||
|
@ -211,6 +212,7 @@ MSHR::allocate(Addr blk_addr, unsigned blk_size, PacketPtr target,
|
||||||
order = _order;
|
order = _order;
|
||||||
assert(target);
|
assert(target);
|
||||||
isForward = false;
|
isForward = false;
|
||||||
|
allocOnFill = alloc_on_fill;
|
||||||
_isUncacheable = target->req->isUncacheable();
|
_isUncacheable = target->req->isUncacheable();
|
||||||
inService = false;
|
inService = false;
|
||||||
downstreamPending = false;
|
downstreamPending = false;
|
||||||
|
@ -274,7 +276,8 @@ MSHR::deallocate()
|
||||||
* Adds a target to an MSHR
|
* Adds a target to an MSHR
|
||||||
*/
|
*/
|
||||||
void
|
void
|
||||||
MSHR::allocateTarget(PacketPtr pkt, Tick whenReady, Counter _order)
|
MSHR::allocateTarget(PacketPtr pkt, Tick whenReady, Counter _order,
|
||||||
|
bool alloc_on_fill)
|
||||||
{
|
{
|
||||||
// assume we'd never issue a prefetch when we've got an
|
// assume we'd never issue a prefetch when we've got an
|
||||||
// outstanding miss
|
// outstanding miss
|
||||||
|
@ -285,6 +288,10 @@ MSHR::allocateTarget(PacketPtr pkt, Tick whenReady, Counter _order)
|
||||||
// have targets addded if originally allocated uncacheable
|
// have targets addded if originally allocated uncacheable
|
||||||
assert(!_isUncacheable);
|
assert(!_isUncacheable);
|
||||||
|
|
||||||
|
// potentially re-evaluate whether we should allocate on a fill or
|
||||||
|
// not
|
||||||
|
allocOnFill = allocOnFill || alloc_on_fill;
|
||||||
|
|
||||||
// if there's a request already in service for this MSHR, we will
|
// if there's a request already in service for this MSHR, we will
|
||||||
// have to defer the new target until after the response if any of
|
// have to defer the new target until after the response if any of
|
||||||
// the following are true:
|
// the following are true:
|
||||||
|
@ -478,6 +485,7 @@ MSHR::print(std::ostream &os, int verbosity, const std::string &prefix) const
|
||||||
prefix, blkAddr, blkAddr + blkSize - 1,
|
prefix, blkAddr, blkAddr + blkSize - 1,
|
||||||
isSecure ? "s" : "ns",
|
isSecure ? "s" : "ns",
|
||||||
isForward ? "Forward" : "",
|
isForward ? "Forward" : "",
|
||||||
|
allocOnFill ? "AllocOnFill" : "",
|
||||||
isForwardNoResponse() ? "ForwNoResp" : "",
|
isForwardNoResponse() ? "ForwNoResp" : "",
|
||||||
needsExclusive() ? "Excl" : "",
|
needsExclusive() ? "Excl" : "",
|
||||||
_isUncacheable ? "Unc" : "",
|
_isUncacheable ? "Unc" : "",
|
||||||
|
|
9
src/mem/cache/mshr.hh
vendored
9
src/mem/cache/mshr.hh
vendored
|
@ -161,6 +161,9 @@ class MSHR : public Packet::SenderState, public Printable
|
||||||
/** True if the request is just a simple forward from an upper level */
|
/** True if the request is just a simple forward from an upper level */
|
||||||
bool isForward;
|
bool isForward;
|
||||||
|
|
||||||
|
/** Keep track of whether we should allocate on fill or not */
|
||||||
|
bool allocOnFill;
|
||||||
|
|
||||||
/** The pending* and post* flags are only valid if inService is
|
/** The pending* and post* flags are only valid if inService is
|
||||||
* true. Using the accessor functions lets us detect if these
|
* true. Using the accessor functions lets us detect if these
|
||||||
* flags are accessed improperly.
|
* flags are accessed improperly.
|
||||||
|
@ -218,9 +221,10 @@ class MSHR : public Packet::SenderState, public Printable
|
||||||
* @param pkt The original miss.
|
* @param pkt The original miss.
|
||||||
* @param when_ready When should the MSHR be ready to act upon.
|
* @param when_ready When should the MSHR be ready to act upon.
|
||||||
* @param _order The logical order of this MSHR
|
* @param _order The logical order of this MSHR
|
||||||
|
* @param alloc_on_fill Should the cache allocate a block on fill
|
||||||
*/
|
*/
|
||||||
void allocate(Addr blk_addr, unsigned blk_size, PacketPtr pkt,
|
void allocate(Addr blk_addr, unsigned blk_size, PacketPtr pkt,
|
||||||
Tick when_ready, Counter _order);
|
Tick when_ready, Counter _order, bool alloc_on_fill);
|
||||||
|
|
||||||
bool markInService(bool pending_dirty_resp);
|
bool markInService(bool pending_dirty_resp);
|
||||||
|
|
||||||
|
@ -235,7 +239,8 @@ class MSHR : public Packet::SenderState, public Printable
|
||||||
* Add a request to the list of targets.
|
* Add a request to the list of targets.
|
||||||
* @param target The target.
|
* @param target The target.
|
||||||
*/
|
*/
|
||||||
void allocateTarget(PacketPtr target, Tick when, Counter order);
|
void allocateTarget(PacketPtr target, Tick when, Counter order,
|
||||||
|
bool alloc_on_fill);
|
||||||
bool handleSnoop(PacketPtr target, Counter order);
|
bool handleSnoop(PacketPtr target, Counter order);
|
||||||
|
|
||||||
/** A simple constructor. */
|
/** A simple constructor. */
|
||||||
|
|
4
src/mem/cache/mshr_queue.cc
vendored
4
src/mem/cache/mshr_queue.cc
vendored
|
@ -146,14 +146,14 @@ MSHRQueue::addToReadyList(MSHR *mshr)
|
||||||
|
|
||||||
MSHR *
|
MSHR *
|
||||||
MSHRQueue::allocate(Addr blk_addr, unsigned blk_size, PacketPtr pkt,
|
MSHRQueue::allocate(Addr blk_addr, unsigned blk_size, PacketPtr pkt,
|
||||||
Tick when_ready, Counter order)
|
Tick when_ready, Counter order, bool alloc_on_fill)
|
||||||
{
|
{
|
||||||
assert(!freeList.empty());
|
assert(!freeList.empty());
|
||||||
MSHR *mshr = freeList.front();
|
MSHR *mshr = freeList.front();
|
||||||
assert(mshr->getNumTargets() == 0);
|
assert(mshr->getNumTargets() == 0);
|
||||||
freeList.pop_front();
|
freeList.pop_front();
|
||||||
|
|
||||||
mshr->allocate(blk_addr, blk_size, pkt, when_ready, order);
|
mshr->allocate(blk_addr, blk_size, pkt, when_ready, order, alloc_on_fill);
|
||||||
mshr->allocIter = allocatedList.insert(allocatedList.end(), mshr);
|
mshr->allocIter = allocatedList.insert(allocatedList.end(), mshr);
|
||||||
mshr->readyIter = addToReadyList(mshr);
|
mshr->readyIter = addToReadyList(mshr);
|
||||||
|
|
||||||
|
|
3
src/mem/cache/mshr_queue.hh
vendored
3
src/mem/cache/mshr_queue.hh
vendored
|
@ -152,13 +152,14 @@ class MSHRQueue : public Drainable
|
||||||
* @param pkt The original miss.
|
* @param pkt The original miss.
|
||||||
* @param when_ready When should the MSHR be ready to act upon.
|
* @param when_ready When should the MSHR be ready to act upon.
|
||||||
* @param order The logical order of this MSHR
|
* @param order The logical order of this MSHR
|
||||||
|
* @param alloc_on_fill Should the cache allocate a block on fill
|
||||||
*
|
*
|
||||||
* @return The a pointer to the MSHR allocated.
|
* @return The a pointer to the MSHR allocated.
|
||||||
*
|
*
|
||||||
* @pre There are free entries.
|
* @pre There are free entries.
|
||||||
*/
|
*/
|
||||||
MSHR *allocate(Addr blk_addr, unsigned blk_size, PacketPtr pkt,
|
MSHR *allocate(Addr blk_addr, unsigned blk_size, PacketPtr pkt,
|
||||||
Tick when_ready, Counter order);
|
Tick when_ready, Counter order, bool alloc_on_fill);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Removes the given MSHR from the queue. This places the MSHR on the
|
* Removes the given MSHR from the queue. This places the MSHR on the
|
||||||
|
|
Loading…
Reference in a new issue