mem: Move the point of coherency to the coherent crossbar

This patch introduces the ability of making the coherent crossbar the
point of coherency. If so, the crossbar does not forward packets where
a cache with ownership has already committed to responding, and also
does not forward any coherency-related packets that are not intended
for a downstream memory controller. Thus, invalidations and upgrades
are turned around in the crossbar, and the memory controller only sees
normal reads and writes.

In addition this patch moves the express snoop promotion of a packet
to the crossbar, thus allowing the downstream cache to check the
express snoop flag (as it should) for bypassing any blocking, rather
than relying on whether a cache is responding or not.
This commit is contained in:
Andreas Hansson 2016-02-10 04:08:25 -05:00
parent f84ee031cc
commit 92f021cbbe
8 changed files with 200 additions and 119 deletions

View file

@ -299,6 +299,7 @@ make_cache_level(cachespec, cache_proto, len(cachespec), None)
# Connect the lowest level crossbar to the memory # Connect the lowest level crossbar to the memory
last_subsys = getattr(system, 'l%dsubsys0' % len(cachespec)) last_subsys = getattr(system, 'l%dsubsys0' % len(cachespec))
last_subsys.xbar.master = system.physmem.port last_subsys.xbar.master = system.physmem.port
last_subsys.xbar.point_of_coherency = True
root = Root(full_system = False, system = system) root = Root(full_system = False, system = system)
if options.atomic: if options.atomic:

View file

@ -100,6 +100,12 @@ class CoherentXBar(BaseXBar):
# An optional snoop filter # An optional snoop filter
snoop_filter = Param.SnoopFilter(NULL, "Selected snoop filter") snoop_filter = Param.SnoopFilter(NULL, "Selected snoop filter")
# Determine how this crossbar handles packets where caches have
# already committed to responding, by establishing if the crossbar
# is the point of coherency or not.
point_of_coherency = Param.Bool(False, "Consider this crossbar the " \
"point of coherency")
system = Param.System(Parent.any, "System that the crossbar belongs to.") system = Param.System(Parent.any, "System that the crossbar belongs to.")
class SnoopFilter(SimObject): class SnoopFilter(SimObject):
@ -147,6 +153,11 @@ class SystemXBar(CoherentXBar):
response_latency = 2 response_latency = 2
snoop_response_latency = 4 snoop_response_latency = 4
# This specialisation of the coherent crossbar is to be considered
# the point of coherency, as there are no (coherent) downstream
# caches.
point_of_coherency = True
# In addition to the system interconnect, we typically also have one # In addition to the system interconnect, we typically also have one
# or more on-chip I/O crossbars. Note that at some point we might want # or more on-chip I/O crossbars. Note that at some point we might want
# to also define an off-chip I/O crossbar such as PCIe. # to also define an off-chip I/O crossbar such as PCIe.

View file

@ -154,14 +154,8 @@ Bridge::BridgeSlavePort::recvTimingReq(PacketPtr pkt)
DPRINTF(Bridge, "recvTimingReq: %s addr 0x%x\n", DPRINTF(Bridge, "recvTimingReq: %s addr 0x%x\n",
pkt->cmdString(), pkt->getAddr()); pkt->cmdString(), pkt->getAddr());
// if a cache is responding, sink the packet without further panic_if(pkt->cacheResponding(), "Should not see packets where cache "
// action, also discard any packet that is not a read or a write "is responding");
if (pkt->cacheResponding() ||
!(pkt->isWrite() || pkt->isRead())) {
assert(!pkt->needsResponse());
pendingDelete.reset(pkt);
return true;
}
// we should not get a new request after committing to retry the // we should not get a new request after committing to retry the
// current one, but unfortunately the CPU violates this rule, so // current one, but unfortunately the CPU violates this rule, so
@ -352,6 +346,9 @@ Bridge::BridgeSlavePort::recvRespRetry()
Tick Tick
Bridge::BridgeSlavePort::recvAtomic(PacketPtr pkt) Bridge::BridgeSlavePort::recvAtomic(PacketPtr pkt)
{ {
panic_if(pkt->cacheResponding(), "Should not see packets where cache "
"is responding");
return delay * bridge.clockPeriod() + masterPort.sendAtomic(pkt); return delay * bridge.clockPeriod() + masterPort.sendAtomic(pkt);
} }

View file

@ -630,57 +630,55 @@ Cache::recvTimingReq(PacketPtr pkt)
// flag) is not providing writable (it is in Owned rather than // flag) is not providing writable (it is in Owned rather than
// the Modified state), we know that there may be other Shared // the Modified state), we know that there may be other Shared
// copies in the system; go out and invalidate them all // copies in the system; go out and invalidate them all
if (pkt->needsWritable() && !pkt->responderHadWritable()) { assert(pkt->needsWritable() && !pkt->responderHadWritable());
// an upstream cache that had the line in Owned state
// (dirty, but not writable), is responding and thus
// transferring the dirty line from one branch of the
// cache hierarchy to another
// send out an express snoop and invalidate all other // an upstream cache that had the line in Owned state
// copies (snooping a packet that needs writable is the // (dirty, but not writable), is responding and thus
// same as an invalidation), thus turning the Owned line // transferring the dirty line from one branch of the
// into a Modified line, note that we don't invalidate the // cache hierarchy to another
// block in the current cache or any other cache on the
// path to memory
// create a downstream express snoop with cleared packet // send out an express snoop and invalidate all other
// flags, there is no need to allocate any data as the // copies (snooping a packet that needs writable is the
// packet is merely used to co-ordinate state transitions // same as an invalidation), thus turning the Owned line
Packet *snoop_pkt = new Packet(pkt, true, false); // into a Modified line, note that we don't invalidate the
// block in the current cache or any other cache on the
// path to memory
// also reset the bus time that the original packet has // create a downstream express snoop with cleared packet
// not yet paid for // flags, there is no need to allocate any data as the
snoop_pkt->headerDelay = snoop_pkt->payloadDelay = 0; // packet is merely used to co-ordinate state transitions
Packet *snoop_pkt = new Packet(pkt, true, false);
// make this an instantaneous express snoop, and let the // also reset the bus time that the original packet has
// other caches in the system know that the another cache // not yet paid for
// is responding, because we have found the authorative snoop_pkt->headerDelay = snoop_pkt->payloadDelay = 0;
// copy (Modified or Owned) that will supply the right
// data
snoop_pkt->setExpressSnoop();
snoop_pkt->setCacheResponding();
// this express snoop travels towards the memory, and at // make this an instantaneous express snoop, and let the
// every crossbar it is snooped upwards thus reaching // other caches in the system know that the another cache
// every cache in the system // is responding, because we have found the authorative
bool M5_VAR_USED success = memSidePort->sendTimingReq(snoop_pkt); // copy (Modified or Owned) that will supply the right
// express snoops always succeed // data
assert(success); snoop_pkt->setExpressSnoop();
snoop_pkt->setCacheResponding();
// main memory will delete the snoop packet // this express snoop travels towards the memory, and at
} // every crossbar it is snooped upwards thus reaching
// every cache in the system
bool M5_VAR_USED success = memSidePort->sendTimingReq(snoop_pkt);
// express snoops always succeed
assert(success);
// main memory will delete the snoop packet
// queue for deletion, as opposed to immediate deletion, as // queue for deletion, as opposed to immediate deletion, as
// the sending cache is still relying on the packet // the sending cache is still relying on the packet
pendingDelete.reset(pkt); pendingDelete.reset(pkt);
// no need to take any action in this particular cache as an // no need to take any further action in this particular cache
// upstream cache has already committed to responding, and // as an upstram cache has already committed to responding,
// either the packet does not need writable (and we can let // and we have already sent out any express snoops in the
// the cache that set the cache responding flag pass on the // section above to ensure all other copies in the system are
// line without any need for intervention), or if the packet // invalidated
// needs writable it is provided, or we have already sent out
// any express snoops in the section above
return true; return true;
} }
@ -1028,9 +1026,8 @@ Cache::recvAtomic(PacketPtr pkt)
// if a cache is responding, and it had the line in Owned // if a cache is responding, and it had the line in Owned
// rather than Modified state, we need to invalidate any // rather than Modified state, we need to invalidate any
// copies that are not on the same path to memory // copies that are not on the same path to memory
if (pkt->needsWritable() && !pkt->responderHadWritable()) { assert(pkt->needsWritable() && !pkt->responderHadWritable());
lat += ticksToCycles(memSidePort->sendAtomic(pkt)); lat += ticksToCycles(memSidePort->sendAtomic(pkt));
}
return lat * clockPeriod(); return lat * clockPeriod();
} }
@ -2493,11 +2490,8 @@ Cache::CpuSidePort::recvTimingReq(PacketPtr pkt)
bool success = false; bool success = false;
// always let packets through if an upstream cache has committed // always let express snoop packets through if even if blocked
// to responding, even if blocked (we should technically look at if (pkt->isExpressSnoop()) {
// the isExpressSnoop flag, but it is set by the cache itself, and
// consequently we have to rely on the cacheResponding flag)
if (pkt->cacheResponding()) {
// do not change the current retry state // do not change the current retry state
bool M5_VAR_USED bypass_success = cache->recvTimingReq(pkt); bool M5_VAR_USED bypass_success = cache->recvTimingReq(pkt);
assert(bypass_success); assert(bypass_success);

View file

@ -56,7 +56,8 @@
CoherentXBar::CoherentXBar(const CoherentXBarParams *p) CoherentXBar::CoherentXBar(const CoherentXBarParams *p)
: BaseXBar(p), system(p->system), snoopFilter(p->snoop_filter), : BaseXBar(p), system(p->system), snoopFilter(p->snoop_filter),
snoopResponseLatency(p->snoop_response_latency) snoopResponseLatency(p->snoop_response_latency),
pointOfCoherency(p->point_of_coherency)
{ {
// create the ports based on the size of the master and slave // create the ports based on the size of the master and slave
// vector ports, and the presence of the default port, the ports // vector ports, and the presence of the default port, the ports
@ -219,32 +220,48 @@ CoherentXBar::recvTimingReq(PacketPtr pkt, PortID slave_port_id)
pkt->snoopDelay = 0; pkt->snoopDelay = 0;
} }
// forwardTiming snooped into peer caches of the sender, and if // set up a sensible starting point
// this is a clean evict or clean writeback, but the packet is bool success = true;
// found in a cache, do not forward it
if ((pkt->cmd == MemCmd::CleanEvict ||
pkt->cmd == MemCmd::WritebackClean) && pkt->isBlockCached()) {
DPRINTF(CoherentXBar, "Clean evict/writeback %#llx still cached, "
"not forwarding\n", pkt->getAddr());
// update the layer state and schedule an idle event
reqLayers[master_port_id]->succeededTiming(packetFinishTime);
// queue the packet for deletion
pendingDelete.reset(pkt);
return true;
}
// remember if the packet will generate a snoop response by // remember if the packet will generate a snoop response by
// checking if a cache set the cacheResponding flag during the // checking if a cache set the cacheResponding flag during the
// snooping above // snooping above
const bool expect_snoop_resp = !cache_responding && pkt->cacheResponding(); const bool expect_snoop_resp = !cache_responding && pkt->cacheResponding();
const bool expect_response = pkt->needsResponse() && bool expect_response = pkt->needsResponse() && !pkt->cacheResponding();
!pkt->cacheResponding();
// since it is a normal request, attempt to send the packet const bool sink_packet = sinkPacket(pkt);
bool success = masterPorts[master_port_id]->sendTimingReq(pkt);
// in certain cases the crossbar is responsible for responding
bool respond_directly = false;
if (sink_packet) {
DPRINTF(CoherentXBar, "Not forwarding %s to %#llx\n",
pkt->cmdString(), pkt->getAddr());
} else {
// determine if we are forwarding the packet, or responding to
// it
if (!pointOfCoherency || pkt->isRead() || pkt->isWrite()) {
// if we are passing on, rather than sinking, a packet to
// which an upstream cache has committed to responding,
// the line was needs writable, and the responding only
// had an Owned copy, so we need to immidiately let the
// downstream caches know, bypass any flow control
if (pkt->cacheResponding()) {
pkt->setExpressSnoop();
}
// since it is a normal request, attempt to send the packet
success = masterPorts[master_port_id]->sendTimingReq(pkt);
} else {
// no need to forward, turn this packet around and respond
// directly
assert(pkt->needsResponse());
respond_directly = true;
assert(!expect_snoop_resp);
expect_response = false;
}
}
if (snoopFilter && !system->bypassCaches()) { if (snoopFilter && !system->bypassCaches()) {
// Let the snoop filter know about the success of the send operation // Let the snoop filter know about the success of the send operation
@ -303,6 +320,27 @@ CoherentXBar::recvTimingReq(PacketPtr pkt, PortID slave_port_id)
snoops++; snoops++;
} }
if (sink_packet)
// queue the packet for deletion
pendingDelete.reset(pkt);
if (respond_directly) {
assert(pkt->needsResponse());
assert(success);
pkt->makeResponse();
if (snoopFilter && !system->bypassCaches()) {
// let the snoop filter inspect the response and update its state
snoopFilter->updateResponse(pkt, *slavePorts[slave_port_id]);
}
Tick response_time = clockEdge() + pkt->headerDelay;
pkt->headerDelay = 0;
slavePorts[slave_port_id]->schedTimingResp(pkt, response_time);
}
return success; return success;
} }
@ -633,27 +671,35 @@ CoherentXBar::recvAtomic(PacketPtr pkt, PortID slave_port_id)
snoop_response_latency += snoop_result.second; snoop_response_latency += snoop_result.second;
} }
// forwardAtomic snooped into peer caches of the sender, and if // set up a sensible default value
// this is a clean evict, but the packet is found in a cache, do Tick response_latency = 0;
// not forward it
if ((pkt->cmd == MemCmd::CleanEvict || const bool sink_packet = sinkPacket(pkt);
pkt->cmd == MemCmd::WritebackClean) && pkt->isBlockCached()) {
DPRINTF(CoherentXBar, "Clean evict/writeback %#llx still cached, "
"not forwarding\n", pkt->getAddr());
return 0;
}
// even if we had a snoop response, we must continue and also // even if we had a snoop response, we must continue and also
// perform the actual request at the destination // perform the actual request at the destination
PortID master_port_id = findPort(pkt->getAddr()); PortID master_port_id = findPort(pkt->getAddr());
if (sink_packet) {
DPRINTF(CoherentXBar, "Not forwarding %s to %#llx\n",
pkt->cmdString(), pkt->getAddr());
} else {
if (!pointOfCoherency || pkt->isRead() || pkt->isWrite()) {
// forward the request to the appropriate destination
response_latency = masterPorts[master_port_id]->sendAtomic(pkt);
} else {
// if it does not need a response we sink the packet above
assert(pkt->needsResponse());
pkt->makeResponse();
}
}
// stats updates for the request // stats updates for the request
pktCount[slave_port_id][master_port_id]++; pktCount[slave_port_id][master_port_id]++;
pktSize[slave_port_id][master_port_id] += pkt_size; pktSize[slave_port_id][master_port_id] += pkt_size;
transDist[pkt_cmd]++; transDist[pkt_cmd]++;
// forward the request to the appropriate destination
Tick response_latency = masterPorts[master_port_id]->sendAtomic(pkt);
// if lower levels have replied, tell the snoop filter // if lower levels have replied, tell the snoop filter
if (!system->bypassCaches() && snoopFilter && pkt->isResponse()) { if (!system->bypassCaches() && snoopFilter && pkt->isResponse()) {
@ -877,6 +923,30 @@ CoherentXBar::forwardFunctional(PacketPtr pkt, PortID exclude_slave_port_id)
} }
} }
bool
CoherentXBar::sinkPacket(const PacketPtr pkt) const
{
// we can sink the packet if:
// 1) the crossbar is the point of coherency, and a cache is
// responding after being snooped
// 2) the crossbar is the point of coherency, and the packet is a
// coherency packet (not a read or a write) that does not
// require a response
// 3) this is a clean evict or clean writeback, but the packet is
// found in a cache above this crossbar
// 4) a cache is responding after being snooped, and the packet
// either does not need the block to be writable, or the cache
// that has promised to respond (setting the cache responding
// flag) is providing writable and thus had a Modified block,
// and no further action is needed
return (pointOfCoherency && pkt->cacheResponding()) ||
(pointOfCoherency && !(pkt->isRead() || pkt->isWrite()) &&
!pkt->needsResponse()) ||
(pkt->isCleanEviction() && pkt->isBlockCached()) ||
(pkt->cacheResponding() &&
(!pkt->needsWritable() || pkt->responderHadWritable()));
}
void void
CoherentXBar::regStats() CoherentXBar::regStats()
{ {

View file

@ -273,6 +273,9 @@ class CoherentXBar : public BaseXBar
/** Cycles of snoop response latency.*/ /** Cycles of snoop response latency.*/
const Cycles snoopResponseLatency; const Cycles snoopResponseLatency;
/** Is this crossbar the point of coherency? **/
const bool pointOfCoherency;
/** /**
* Upstream caches need this packet until true is returned, so * Upstream caches need this packet until true is returned, so
* hold it for deletion until a subsequent call * hold it for deletion until a subsequent call
@ -384,6 +387,12 @@ class CoherentXBar : public BaseXBar
*/ */
void forwardFunctional(PacketPtr pkt, PortID exclude_slave_port_id); void forwardFunctional(PacketPtr pkt, PortID exclude_slave_port_id);
/**
* Determine if the crossbar should sink the packet, as opposed to
* forwarding it, or responding.
*/
bool sinkPacket(const PacketPtr pkt) const;
Stats::Scalar snoops; Stats::Scalar snoops;
Stats::Distribution snoopFanout; Stats::Distribution snoopFanout;

View file

@ -273,11 +273,14 @@ DRAMCtrl::recvAtomic(PacketPtr pkt)
{ {
DPRINTF(DRAM, "recvAtomic: %s 0x%x\n", pkt->cmdString(), pkt->getAddr()); DPRINTF(DRAM, "recvAtomic: %s 0x%x\n", pkt->cmdString(), pkt->getAddr());
panic_if(pkt->cacheResponding(), "Should not see packets where cache "
"is responding");
// do the actual memory access and turn the packet into a response // do the actual memory access and turn the packet into a response
access(pkt); access(pkt);
Tick latency = 0; Tick latency = 0;
if (!pkt->cacheResponding() && pkt->hasData()) { if (pkt->hasData()) {
// this value is not supposed to be accurate, just enough to // this value is not supposed to be accurate, just enough to
// keep things going, mimic a closed page // keep things going, mimic a closed page
latency = tRP + tRCD + tCL; latency = tRP + tRCD + tCL;
@ -590,11 +593,11 @@ DRAMCtrl::recvTimingReq(PacketPtr pkt)
DPRINTF(DRAM, "recvTimingReq: request %s addr %lld size %d\n", DPRINTF(DRAM, "recvTimingReq: request %s addr %lld size %d\n",
pkt->cmdString(), pkt->getAddr(), pkt->getSize()); pkt->cmdString(), pkt->getAddr(), pkt->getSize());
// if a cache is responding, sink the packet without further action panic_if(pkt->cacheResponding(), "Should not see packets where cache "
if (pkt->cacheResponding()) { "is responding");
pendingDelete.reset(pkt);
return true; panic_if(!(pkt->isRead() || pkt->isWrite()),
} "Should only see read and writes at memory controller\n");
// Calc avg gap between requests // Calc avg gap between requests
if (prevArrival != 0) { if (prevArrival != 0) {
@ -625,7 +628,8 @@ DRAMCtrl::recvTimingReq(PacketPtr pkt)
readReqs++; readReqs++;
bytesReadSys += size; bytesReadSys += size;
} }
} else if (pkt->isWrite()) { } else {
assert(pkt->isWrite());
assert(size != 0); assert(size != 0);
if (writeQueueFull(dram_pkt_count)) { if (writeQueueFull(dram_pkt_count)) {
DPRINTF(DRAM, "Write queue full, not accepting\n"); DPRINTF(DRAM, "Write queue full, not accepting\n");
@ -638,10 +642,6 @@ DRAMCtrl::recvTimingReq(PacketPtr pkt)
writeReqs++; writeReqs++;
bytesWrittenSys += size; bytesWrittenSys += size;
} }
} else {
DPRINTF(DRAM,"Neither read nor write, ignore timing\n");
neitherReadNorWrite++;
accessAndRespond(pkt, 1);
} }
return true; return true;

View file

@ -72,8 +72,11 @@ SimpleMemory::init()
Tick Tick
SimpleMemory::recvAtomic(PacketPtr pkt) SimpleMemory::recvAtomic(PacketPtr pkt)
{ {
panic_if(pkt->cacheResponding(), "Should not see packets where cache "
"is responding");
access(pkt); access(pkt);
return pkt->cacheResponding() ? 0 : getLatency(); return getLatency();
} }
void void
@ -97,11 +100,12 @@ SimpleMemory::recvFunctional(PacketPtr pkt)
bool bool
SimpleMemory::recvTimingReq(PacketPtr pkt) SimpleMemory::recvTimingReq(PacketPtr pkt)
{ {
// if a cache is responding, sink the packet without further action panic_if(pkt->cacheResponding(), "Should not see packets where cache "
if (pkt->cacheResponding()) { "is responding");
pendingDelete.reset(pkt);
return true; panic_if(!(pkt->isRead() || pkt->isWrite()),
} "Should only see read and writes at memory controller, "
"saw %s to %#llx\n", pkt->cmdString(), pkt->getAddr());
// we should not get a new request after committing to retry the // we should not get a new request after committing to retry the
// current one, but unfortunately the CPU violates this rule, so // current one, but unfortunately the CPU violates this rule, so
@ -127,21 +131,16 @@ SimpleMemory::recvTimingReq(PacketPtr pkt)
// rather than long term as it is the short term data rate that is // rather than long term as it is the short term data rate that is
// limited for any real memory // limited for any real memory
// only look at reads and writes when determining if we are busy, // calculate an appropriate tick to release to not exceed
// and for how long, as it is not clear what to regulate for the // the bandwidth limit
// other types of commands Tick duration = pkt->getSize() * bandwidth;
if (pkt->isRead() || pkt->isWrite()) {
// calculate an appropriate tick to release to not exceed
// the bandwidth limit
Tick duration = pkt->getSize() * bandwidth;
// only consider ourselves busy if there is any need to wait // only consider ourselves busy if there is any need to wait
// to avoid extra events being scheduled for (infinitely) fast // to avoid extra events being scheduled for (infinitely) fast
// memories // memories
if (duration != 0) { if (duration != 0) {
schedule(releaseEvent, curTick() + duration); schedule(releaseEvent, curTick() + duration);
isBusy = true; isBusy = true;
}
} }
// go ahead and deal with the packet and put the response in the // go ahead and deal with the packet and put the response in the