2006-06-28 17:02:14 +02:00
|
|
|
/*
|
2012-01-17 19:55:07 +01:00
|
|
|
* Copyright (c) 2010-2012 ARM Limited
|
2010-08-23 18:18:41 +02:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* The license below extends only to copyright in the software and shall
|
|
|
|
* not be construed as granting a license to any other intellectual
|
|
|
|
* property including but not limited to intellectual property relating
|
|
|
|
* to a hardware implementation of the functionality of the software
|
|
|
|
* licensed hereunder. You may use the software subject to the license
|
|
|
|
* terms below provided that you ensure that this notice is replicated
|
|
|
|
* unmodified and in its entirety in all distributions of the software,
|
|
|
|
* modified or unmodified, in source code or in binary form.
|
|
|
|
*
|
2006-06-28 17:02:14 +02:00
|
|
|
* Copyright (c) 2002-2005 The Regents of The University of Michigan
|
2010-06-17 00:25:57 +02:00
|
|
|
* Copyright (c) 2010 Advanced Micro Devices, Inc.
|
2006-06-28 17:02:14 +02:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions are
|
|
|
|
* met: redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer;
|
|
|
|
* redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution;
|
|
|
|
* neither the name of the copyright holders nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived from
|
|
|
|
* this software without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*
|
|
|
|
* Authors: Erik Hallnor
|
|
|
|
* Dave Greene
|
|
|
|
* Nathan Binkert
|
2007-05-19 07:35:04 +02:00
|
|
|
* Steve Reinhardt
|
|
|
|
* Ron Dreslinski
|
2012-11-02 17:32:02 +01:00
|
|
|
* Andreas Sandberg
|
2006-06-28 17:02:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
* @file
|
|
|
|
* Cache definitions.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "base/misc.hh"
|
2009-05-17 23:34:52 +02:00
|
|
|
#include "base/types.hh"
|
2011-04-15 19:44:32 +02:00
|
|
|
#include "debug/Cache.hh"
|
|
|
|
#include "debug/CachePort.hh"
|
2011-04-15 19:44:06 +02:00
|
|
|
#include "mem/cache/prefetch/base.hh"
|
2008-02-10 23:45:25 +01:00
|
|
|
#include "mem/cache/blk.hh"
|
2009-05-17 23:34:52 +02:00
|
|
|
#include "mem/cache/cache.hh"
|
2008-02-10 23:45:25 +01:00
|
|
|
#include "mem/cache/mshr.hh"
|
2009-05-17 23:34:52 +02:00
|
|
|
#include "sim/sim_exit.hh"
|
2006-11-12 15:06:15 +01:00
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2012-02-12 23:07:38 +01:00
|
|
|
Cache<TagStore>::Cache(const Params *p, TagStore *tags)
|
2007-08-30 21:16:59 +02:00
|
|
|
: BaseCache(p),
|
|
|
|
tags(tags),
|
2012-02-12 23:07:38 +01:00
|
|
|
prefetcher(p->prefetcher),
|
2007-08-30 21:16:59 +02:00
|
|
|
doFastWrites(true),
|
2009-02-16 17:56:40 +01:00
|
|
|
prefetchOnAccess(p->prefetch_on_access)
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2007-06-27 08:30:30 +02:00
|
|
|
tempBlock = new BlkType();
|
|
|
|
tempBlock->data = new uint8_t[blkSize];
|
|
|
|
|
2012-07-09 18:35:39 +02:00
|
|
|
cpuSidePort = new CpuSidePort(p->name + ".cpu_side", this,
|
2008-07-16 20:10:33 +02:00
|
|
|
"CpuSidePort");
|
2012-07-09 18:35:39 +02:00
|
|
|
memSidePort = new MemSidePort(p->name + ".mem_side", this,
|
2008-07-16 20:10:33 +02:00
|
|
|
"MemSidePort");
|
2007-05-19 07:35:04 +02:00
|
|
|
|
2006-10-12 20:21:25 +02:00
|
|
|
tags->setCache(this);
|
2009-02-16 17:56:40 +01:00
|
|
|
if (prefetcher)
|
|
|
|
prefetcher->setCache(this);
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-06-28 17:02:14 +02:00
|
|
|
void
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::regStats()
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
|
|
|
BaseCache::regStats();
|
|
|
|
tags->regStats(name());
|
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
void
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::cmpAndSwap(BlkType *blk, PacketPtr pkt)
|
2006-12-19 05:47:12 +01:00
|
|
|
{
|
2007-06-18 02:27:53 +02:00
|
|
|
uint64_t overwrite_val;
|
|
|
|
bool overwrite_mem;
|
|
|
|
uint64_t condition_val64;
|
|
|
|
uint32_t condition_val32;
|
|
|
|
|
2006-12-19 05:47:12 +01:00
|
|
|
int offset = tags->extractBlkOffset(pkt->getAddr());
|
2007-06-18 02:27:53 +02:00
|
|
|
uint8_t *blk_data = blk->data + offset;
|
|
|
|
|
|
|
|
assert(sizeof(uint64_t) >= pkt->getSize());
|
|
|
|
|
|
|
|
overwrite_mem = true;
|
|
|
|
// keep a copy of our possible write value, and copy what is at the
|
|
|
|
// memory address into the packet
|
|
|
|
pkt->writeData((uint8_t *)&overwrite_val);
|
|
|
|
pkt->setData(blk_data);
|
|
|
|
|
|
|
|
if (pkt->req->isCondSwap()) {
|
|
|
|
if (pkt->getSize() == sizeof(uint64_t)) {
|
|
|
|
condition_val64 = pkt->req->getExtraData();
|
|
|
|
overwrite_mem = !std::memcmp(&condition_val64, blk_data,
|
|
|
|
sizeof(uint64_t));
|
|
|
|
} else if (pkt->getSize() == sizeof(uint32_t)) {
|
|
|
|
condition_val32 = (uint32_t)pkt->req->getExtraData();
|
|
|
|
overwrite_mem = !std::memcmp(&condition_val32, blk_data,
|
|
|
|
sizeof(uint32_t));
|
|
|
|
} else
|
|
|
|
panic("Invalid size for conditional read/write\n");
|
|
|
|
}
|
|
|
|
|
2009-03-12 07:05:26 +01:00
|
|
|
if (overwrite_mem) {
|
2007-06-18 02:27:53 +02:00
|
|
|
std::memcpy(blk_data, &overwrite_val, pkt->getSize());
|
2009-03-12 07:05:26 +01:00
|
|
|
blk->status |= BlkDirty;
|
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-30 22:56:25 +02:00
|
|
|
template<class TagStore>
|
|
|
|
void
|
2010-09-09 20:40:18 +02:00
|
|
|
Cache<TagStore>::satisfyCpuSideRequest(PacketPtr pkt, BlkType *blk,
|
|
|
|
bool deferred_response,
|
|
|
|
bool pending_downgrade)
|
2007-06-30 22:56:25 +02:00
|
|
|
{
|
2010-09-09 20:40:18 +02:00
|
|
|
assert(blk && blk->isValid());
|
2007-07-24 07:28:40 +02:00
|
|
|
// Occasionally this is not true... if we are a lower-level cache
|
|
|
|
// satisfying a string of Read and ReadEx requests from
|
|
|
|
// upper-level caches, a Read will mark the block as shared but we
|
|
|
|
// can satisfy a following ReadEx anyway since we can rely on the
|
|
|
|
// Read requester(s) to have buffered the ReadEx snoop and to
|
|
|
|
// invalidate their blocks after receiving them.
|
2010-09-09 20:40:18 +02:00
|
|
|
// assert(!pkt->needsExclusive() || blk->isWritable());
|
2007-06-30 22:56:25 +02:00
|
|
|
assert(pkt->getOffset(blkSize) + pkt->getSize() <= blkSize);
|
|
|
|
|
|
|
|
// Check RMW operations first since both isRead() and
|
|
|
|
// isWrite() will be true for them
|
|
|
|
if (pkt->cmd == MemCmd::SwapReq) {
|
|
|
|
cmpAndSwap(blk, pkt);
|
|
|
|
} else if (pkt->isWrite()) {
|
|
|
|
if (blk->checkWrite(pkt)) {
|
|
|
|
pkt->writeDataToBlock(blk->data, blkSize);
|
2010-06-17 00:25:57 +02:00
|
|
|
blk->status |= BlkDirty;
|
2007-06-30 22:56:25 +02:00
|
|
|
}
|
|
|
|
} else if (pkt->isRead()) {
|
2009-04-20 06:44:15 +02:00
|
|
|
if (pkt->isLLSC()) {
|
2007-06-30 22:56:25 +02:00
|
|
|
blk->trackLoadLocked(pkt);
|
|
|
|
}
|
|
|
|
pkt->setDataFromBlock(blk->data, blkSize);
|
2007-07-16 05:11:06 +02:00
|
|
|
if (pkt->getSize() == blkSize) {
|
|
|
|
// special handling for coherent block requests from
|
|
|
|
// upper-level caches
|
|
|
|
if (pkt->needsExclusive()) {
|
2010-09-09 20:40:18 +02:00
|
|
|
// if we have a dirty copy, make sure the recipient
|
|
|
|
// keeps it marked dirty
|
|
|
|
if (blk->isDirty()) {
|
|
|
|
pkt->assertMemInhibit();
|
|
|
|
}
|
|
|
|
// on ReadExReq we give up our copy unconditionally
|
2012-09-11 20:14:49 +02:00
|
|
|
assert(blk != tempBlock);
|
|
|
|
tags->invalidate(blk);
|
|
|
|
blk->invalidate();
|
2010-09-09 20:40:18 +02:00
|
|
|
} else if (blk->isWritable() && !pending_downgrade
|
2012-07-27 22:08:04 +02:00
|
|
|
&& !pkt->sharedAsserted() && !pkt->req->isInstFetch()) {
|
2010-09-09 20:40:18 +02:00
|
|
|
// we can give the requester an exclusive copy (by not
|
|
|
|
// asserting shared line) on a read request if:
|
|
|
|
// - we have an exclusive copy at this level (& below)
|
|
|
|
// - we don't have a pending snoop from below
|
|
|
|
// signaling another read request
|
|
|
|
// - no other cache above has a copy (otherwise it
|
|
|
|
// would have asseretd shared line on request)
|
2012-07-27 22:08:04 +02:00
|
|
|
// - we are not satisfying an instruction fetch (this
|
|
|
|
// prevents dirty data in the i-cache)
|
|
|
|
|
2010-09-09 20:40:18 +02:00
|
|
|
if (blk->isDirty()) {
|
|
|
|
// special considerations if we're owner:
|
2011-03-18 01:20:19 +01:00
|
|
|
if (!deferred_response && !isTopLevel) {
|
2010-09-09 20:40:18 +02:00
|
|
|
// if we are responding immediately and can
|
|
|
|
// signal that we're transferring ownership
|
|
|
|
// along with exclusivity, do so
|
|
|
|
pkt->assertMemInhibit();
|
|
|
|
blk->status &= ~BlkDirty;
|
|
|
|
} else {
|
|
|
|
// if we're responding after our own miss,
|
|
|
|
// there's a window where the recipient didn't
|
|
|
|
// know it was getting ownership and may not
|
|
|
|
// have responded to snoops correctly, so we
|
|
|
|
// can't pass off ownership *or* exclusivity
|
|
|
|
pkt->assertShared();
|
|
|
|
}
|
|
|
|
}
|
2007-07-16 05:11:06 +02:00
|
|
|
} else {
|
2010-09-09 20:40:18 +02:00
|
|
|
// otherwise only respond with a shared copy
|
2007-07-16 05:11:06 +02:00
|
|
|
pkt->assertShared();
|
|
|
|
}
|
|
|
|
}
|
2007-06-30 22:56:25 +02:00
|
|
|
} else {
|
|
|
|
// Not a read or write... must be an upgrade. it's OK
|
|
|
|
// to just ack those as long as we have an exclusive
|
|
|
|
// copy at this level.
|
2010-06-17 00:25:57 +02:00
|
|
|
assert(pkt->isUpgrade());
|
2012-09-11 20:14:49 +02:00
|
|
|
assert(blk != tempBlock);
|
|
|
|
tags->invalidate(blk);
|
|
|
|
blk->invalidate();
|
2007-06-30 22:56:25 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
//
|
|
|
|
// MSHR helper functions
|
|
|
|
//
|
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
void
|
2010-09-09 20:40:18 +02:00
|
|
|
Cache<TagStore>::markInService(MSHR *mshr, PacketPtr pkt)
|
2007-06-18 02:27:53 +02:00
|
|
|
{
|
2010-09-09 20:40:18 +02:00
|
|
|
markInServiceInternal(mshr, pkt);
|
2007-06-18 02:27:53 +02:00
|
|
|
#if 0
|
|
|
|
if (mshr->originalCmd == MemCmd::HardPFReq) {
|
|
|
|
DPRINTF(HWPrefetch, "%s:Marking a HW_PF in service\n",
|
|
|
|
name());
|
|
|
|
//Also clear pending if need be
|
|
|
|
if (!prefetcher->havePending())
|
|
|
|
{
|
|
|
|
deassertMemSideBusRequest(Request_PF);
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
void
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::squash(int threadNum)
|
2007-06-18 02:27:53 +02:00
|
|
|
{
|
|
|
|
bool unblock = false;
|
|
|
|
BlockedCause cause = NUM_BLOCKED_CAUSES;
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
if (noTargetMSHR && noTargetMSHR->threadNum == threadNum) {
|
|
|
|
noTargetMSHR = NULL;
|
|
|
|
unblock = true;
|
|
|
|
cause = Blocked_NoTargets;
|
|
|
|
}
|
|
|
|
if (mshrQueue.isFull()) {
|
|
|
|
unblock = true;
|
|
|
|
cause = Blocked_NoMSHRs;
|
|
|
|
}
|
|
|
|
mshrQueue.squash(threadNum);
|
|
|
|
if (unblock && !mshrQueue.isFull()) {
|
|
|
|
clearBlocked(cause);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
//
|
|
|
|
// Access path: requests coming in from the CPU side
|
|
|
|
//
|
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
bool
|
2008-03-25 15:01:21 +01:00
|
|
|
Cache<TagStore>::access(PacketPtr pkt, BlkType *&blk,
|
2012-10-15 14:10:54 +02:00
|
|
|
Cycles &lat, PacketList &writebacks)
|
2007-06-18 02:27:53 +02:00
|
|
|
{
|
2010-08-23 18:18:41 +02:00
|
|
|
if (pkt->req->isUncacheable()) {
|
2013-01-07 19:05:47 +01:00
|
|
|
uncacheableFlush(pkt);
|
2007-06-21 20:59:17 +02:00
|
|
|
blk = NULL;
|
|
|
|
lat = hitLatency;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2010-08-23 18:18:41 +02:00
|
|
|
int id = pkt->req->hasContextId() ? pkt->req->contextId() : -1;
|
|
|
|
blk = tags->accessBlock(pkt->getAddr(), lat, id);
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2009-08-02 07:50:14 +02:00
|
|
|
DPRINTF(Cache, "%s%s %x %s\n", pkt->cmdString(),
|
|
|
|
pkt->req->isInstFetch() ? " (ifetch)" : "",
|
|
|
|
pkt->getAddr(), (blk) ? "hit" : "miss");
|
2007-06-18 02:27:53 +02:00
|
|
|
|
|
|
|
if (blk != NULL) {
|
2008-02-16 20:58:03 +01:00
|
|
|
|
2008-02-27 07:03:28 +01:00
|
|
|
if (pkt->needsExclusive() ? blk->isWritable() : blk->isReadable()) {
|
2007-06-18 02:27:53 +02:00
|
|
|
// OK to satisfy access
|
2012-02-12 23:07:39 +01:00
|
|
|
incHitCount(pkt);
|
2007-06-30 22:56:25 +02:00
|
|
|
satisfyCpuSideRequest(pkt, blk);
|
2008-02-16 20:58:03 +01:00
|
|
|
return true;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
2008-02-16 20:58:03 +01:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2008-02-16 20:58:03 +01:00
|
|
|
// Can't satisfy access normally... either no block (blk == NULL)
|
|
|
|
// or have block but need exclusive & only have shared.
|
|
|
|
|
|
|
|
// Writeback handling is special case. We can write the block
|
|
|
|
// into the cache without having a writeable copy (or any copy at
|
|
|
|
// all).
|
|
|
|
if (pkt->cmd == MemCmd::Writeback) {
|
|
|
|
assert(blkSize == pkt->getSize());
|
|
|
|
if (blk == NULL) {
|
|
|
|
// need to do a replacement
|
|
|
|
blk = allocateBlock(pkt->getAddr(), writebacks);
|
|
|
|
if (blk == NULL) {
|
|
|
|
// no replaceable block available, give up.
|
|
|
|
// writeback will be forwarded to next level.
|
2012-02-12 23:07:39 +01:00
|
|
|
incMissCount(pkt);
|
2008-02-16 20:58:03 +01:00
|
|
|
return false;
|
|
|
|
}
|
2012-02-12 23:07:39 +01:00
|
|
|
int id = pkt->req->masterId();
|
2010-01-12 19:53:02 +01:00
|
|
|
tags->insertBlock(pkt->getAddr(), blk, id);
|
2008-02-27 07:03:28 +01:00
|
|
|
blk->status = BlkValid | BlkReadable;
|
2008-02-16 20:58:03 +01:00
|
|
|
}
|
|
|
|
std::memcpy(blk->data, pkt->getPtr<uint8_t>(), blkSize);
|
|
|
|
blk->status |= BlkDirty;
|
2010-09-22 08:07:34 +02:00
|
|
|
if (pkt->isSupplyExclusive()) {
|
|
|
|
blk->status |= BlkWritable;
|
|
|
|
}
|
2008-02-16 20:58:03 +01:00
|
|
|
// nothing else to do; writeback doesn't expect response
|
|
|
|
assert(!pkt->needsResponse());
|
2012-02-12 23:07:39 +01:00
|
|
|
incHitCount(pkt);
|
2008-02-16 20:58:03 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2012-02-12 23:07:39 +01:00
|
|
|
incMissCount(pkt);
|
2008-02-16 20:58:03 +01:00
|
|
|
|
2009-04-20 06:44:15 +02:00
|
|
|
if (blk == NULL && pkt->isLLSC() && pkt->isWrite()) {
|
2008-02-16 20:58:03 +01:00
|
|
|
// complete miss on store conditional... just give up now
|
|
|
|
pkt->req->setExtraData(0);
|
|
|
|
return true;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
|
2008-02-16 20:58:03 +01:00
|
|
|
return false;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
|
2007-03-28 00:05:25 +02:00
|
|
|
|
2012-06-05 07:23:08 +02:00
|
|
|
class ForwardResponseRecord : public Packet::SenderState
|
2007-07-17 15:33:28 +02:00
|
|
|
{
|
|
|
|
Packet::SenderState *prevSenderState;
|
2012-05-30 11:29:42 +02:00
|
|
|
PortID prevSrc;
|
2007-07-17 15:33:28 +02:00
|
|
|
#ifndef NDEBUG
|
|
|
|
BaseCache *cache;
|
|
|
|
#endif
|
|
|
|
public:
|
|
|
|
ForwardResponseRecord(Packet *pkt, BaseCache *_cache)
|
|
|
|
: prevSenderState(pkt->senderState), prevSrc(pkt->getSrc())
|
|
|
|
#ifndef NDEBUG
|
|
|
|
, cache(_cache)
|
|
|
|
#endif
|
|
|
|
{}
|
|
|
|
void restore(Packet *pkt, BaseCache *_cache)
|
|
|
|
{
|
|
|
|
assert(_cache == cache);
|
|
|
|
pkt->senderState = prevSenderState;
|
|
|
|
pkt->setDest(prevSrc);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
bool
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::timingAccess(PacketPtr pkt)
|
2006-12-19 05:47:12 +01:00
|
|
|
{
|
2007-06-18 02:27:53 +02:00
|
|
|
//@todo Add back in MemDebug Calls
|
|
|
|
// MemDebug::cacheAccess(pkt);
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2012-06-07 16:59:03 +02:00
|
|
|
|
|
|
|
/// @todo temporary hack to deal with memory corruption issue until
|
|
|
|
/// 4-phase transactions are complete
|
|
|
|
for (int x = 0; x < pendingDelete.size(); x++)
|
|
|
|
delete pendingDelete[x];
|
|
|
|
pendingDelete.clear();
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// we charge hitLatency for doing just about anything here
|
2012-10-15 14:10:54 +02:00
|
|
|
Tick time = clockEdge(hitLatency);
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2007-07-17 15:33:28 +02:00
|
|
|
if (pkt->isResponse()) {
|
|
|
|
// must be cache-to-cache response from upper to lower level
|
|
|
|
ForwardResponseRecord *rec =
|
|
|
|
dynamic_cast<ForwardResponseRecord *>(pkt->senderState);
|
2011-07-15 18:53:35 +02:00
|
|
|
|
|
|
|
if (rec == NULL) {
|
|
|
|
assert(pkt->cmd == MemCmd::HardPFResp);
|
|
|
|
// Check if it's a prefetch response and handle it. We shouldn't
|
|
|
|
// get any other kinds of responses without FRRs.
|
|
|
|
DPRINTF(Cache, "Got prefetch response from above for addr %#x\n",
|
|
|
|
pkt->getAddr());
|
|
|
|
handleResponse(pkt);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2007-07-17 15:33:28 +02:00
|
|
|
rec->restore(pkt, this);
|
|
|
|
delete rec;
|
2012-08-22 17:39:56 +02:00
|
|
|
memSidePort->schedTimingSnoopResp(pkt, time);
|
2007-07-17 15:33:28 +02:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(pkt->isRequest());
|
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
if (pkt->memInhibitAsserted()) {
|
|
|
|
DPRINTF(Cache, "mem inhibited on 0x%x: not responding\n",
|
|
|
|
pkt->getAddr());
|
|
|
|
assert(!pkt->req->isUncacheable());
|
2007-07-27 02:04:17 +02:00
|
|
|
// Special tweak for multilevel coherence: snoop downward here
|
|
|
|
// on invalidates since there may be other caches below here
|
|
|
|
// that have shared copies. Not necessary if we know that
|
|
|
|
// supplier had exclusive copy to begin with.
|
|
|
|
if (pkt->needsExclusive() && !pkt->isSupplyExclusive()) {
|
|
|
|
Packet *snoopPkt = new Packet(pkt, true); // clear flags
|
|
|
|
snoopPkt->setExpressSnoop();
|
|
|
|
snoopPkt->assertMemInhibit();
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
memSidePort->sendTimingReq(snoopPkt);
|
2007-07-27 02:04:17 +02:00
|
|
|
// main memory will delete snoopPkt
|
|
|
|
}
|
2008-03-17 08:08:28 +01:00
|
|
|
// since we're the official target but we aren't responding,
|
|
|
|
// delete the packet now.
|
2012-06-07 16:59:03 +02:00
|
|
|
|
|
|
|
/// @todo nominally we should just delete the packet here,
|
|
|
|
/// however, until 4-phase stuff we can't because sending
|
|
|
|
/// cache is still relying on it
|
|
|
|
pendingDelete.push_back(pkt);
|
2007-06-21 20:59:17 +02:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
if (pkt->req->isUncacheable()) {
|
2013-01-07 19:05:47 +01:00
|
|
|
uncacheableFlush(pkt);
|
2010-08-23 18:18:41 +02:00
|
|
|
|
2007-07-21 22:45:17 +02:00
|
|
|
// writes go in write buffer, reads use MSHR
|
|
|
|
if (pkt->isWrite() && !pkt->isRead()) {
|
|
|
|
allocateWriteBuffer(pkt, time, true);
|
|
|
|
} else {
|
|
|
|
allocateUncachedReadBuffer(pkt, time, true);
|
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
assert(pkt->needsResponse()); // else we should delete it here??
|
|
|
|
return true;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
|
2012-10-15 14:10:54 +02:00
|
|
|
Cycles lat = hitLatency;
|
2008-02-27 07:03:28 +01:00
|
|
|
BlkType *blk = NULL;
|
2008-03-25 15:01:21 +01:00
|
|
|
PacketList writebacks;
|
|
|
|
|
|
|
|
bool satisfied = access(pkt, blk, lat, writebacks);
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
#if 0
|
2008-02-27 05:17:26 +01:00
|
|
|
/** @todo make the fast write alloc (wh64) work with coherence. */
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// If this is a block size write/hint (WH64) allocate the block here
|
|
|
|
// if the coherence protocol allows it.
|
|
|
|
if (!blk && pkt->getSize() >= blkSize && coherence->allowFastWrites() &&
|
|
|
|
(pkt->cmd == MemCmd::WriteReq
|
|
|
|
|| pkt->cmd == MemCmd::WriteInvalidateReq) ) {
|
|
|
|
// not outstanding misses, can do this
|
|
|
|
MSHR *outstanding_miss = mshrQueue.findMatch(pkt->getAddr());
|
|
|
|
if (pkt->cmd == MemCmd::WriteInvalidateReq || !outstanding_miss) {
|
|
|
|
if (outstanding_miss) {
|
|
|
|
warn("WriteInv doing a fastallocate"
|
|
|
|
"with an outstanding miss to the same address\n");
|
|
|
|
}
|
|
|
|
blk = handleFill(NULL, pkt, BlkValid | BlkWritable,
|
|
|
|
writebacks);
|
|
|
|
++fastWrites;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
2007-06-26 23:53:15 +02:00
|
|
|
#endif
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2009-02-16 17:56:40 +01:00
|
|
|
// track time of availability of next prefetch, if any
|
|
|
|
Tick next_pf_time = 0;
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
bool needsResponse = pkt->needsResponse();
|
|
|
|
|
|
|
|
if (satisfied) {
|
2011-08-19 22:08:08 +02:00
|
|
|
if (prefetcher && (prefetchOnAccess || (blk && blk->wasPrefetched()))) {
|
|
|
|
if (blk)
|
|
|
|
blk->status &= ~BlkHWPrefetched;
|
|
|
|
next_pf_time = prefetcher->notify(pkt, time);
|
|
|
|
}
|
|
|
|
|
2007-07-16 05:11:06 +02:00
|
|
|
if (needsResponse) {
|
|
|
|
pkt->makeTimingResponse();
|
2012-10-15 14:10:54 +02:00
|
|
|
cpuSidePort->schedTimingResp(pkt, clockEdge(lat));
|
2007-07-17 15:33:28 +02:00
|
|
|
} else {
|
2012-06-07 16:59:03 +02:00
|
|
|
/// @todo nominally we should just delete the packet here,
|
|
|
|
/// however, until 4-phase stuff we can't because sending
|
|
|
|
/// cache is still relying on it
|
|
|
|
pendingDelete.push_back(pkt);
|
2007-07-16 05:11:06 +02:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
} else {
|
|
|
|
// miss
|
|
|
|
|
2009-09-26 19:50:50 +02:00
|
|
|
Addr blk_addr = blockAlign(pkt->getAddr());
|
2008-02-27 07:03:28 +01:00
|
|
|
MSHR *mshr = mshrQueue.findMatch(blk_addr);
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
if (mshr) {
|
2013-01-07 19:05:32 +01:00
|
|
|
/// MSHR hit
|
|
|
|
/// @note writebacks will be checked in getNextMSHR()
|
|
|
|
/// for any conflicting requests to the same block
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
//@todo remove hw_pf here
|
2012-02-12 23:07:39 +01:00
|
|
|
assert(pkt->req->masterId() < system->maxMasters());
|
|
|
|
mshr_hits[pkt->cmdToIndex()][pkt->req->masterId()]++;
|
2008-11-04 17:35:42 +01:00
|
|
|
if (mshr->threadNum != 0/*pkt->req->threadId()*/) {
|
2007-06-18 02:27:53 +02:00
|
|
|
mshr->threadNum = -1;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
2007-06-25 15:47:05 +02:00
|
|
|
mshr->allocateTarget(pkt, time, order++);
|
2007-06-18 02:27:53 +02:00
|
|
|
if (mshr->getNumTargets() == numTarget) {
|
|
|
|
noTargetMSHR = mshr;
|
|
|
|
setBlocked(Blocked_NoTargets);
|
2007-06-25 15:47:05 +02:00
|
|
|
// need to be careful with this... if this mshr isn't
|
2011-01-08 06:50:29 +01:00
|
|
|
// ready yet (i.e. time > curTick()_, we don't want to
|
2007-06-25 15:47:05 +02:00
|
|
|
// move it ahead of mshrs that are ready
|
|
|
|
// mshrQueue.moveToFront(mshr);
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// no MSHR
|
2012-02-12 23:07:39 +01:00
|
|
|
assert(pkt->req->masterId() < system->maxMasters());
|
|
|
|
mshr_misses[pkt->cmdToIndex()][pkt->req->masterId()]++;
|
2007-06-18 02:27:53 +02:00
|
|
|
// always mark as cache fill for now... if we implement
|
|
|
|
// no-write-allocate or bypass accesses this will have to
|
|
|
|
// be changed.
|
2007-07-21 22:45:17 +02:00
|
|
|
if (pkt->cmd == MemCmd::Writeback) {
|
|
|
|
allocateWriteBuffer(pkt, time, true);
|
|
|
|
} else {
|
2008-02-27 07:03:28 +01:00
|
|
|
if (blk && blk->isValid()) {
|
|
|
|
// If we have a write miss to a valid block, we
|
|
|
|
// need to mark the block non-readable. Otherwise
|
|
|
|
// if we allow reads while there's an outstanding
|
|
|
|
// write miss, the read could return stale data
|
|
|
|
// out of the cache block... a more aggressive
|
|
|
|
// system could detect the overlap (if any) and
|
|
|
|
// forward data out of the MSHRs, but we don't do
|
|
|
|
// that yet. Note that we do need to leave the
|
|
|
|
// block valid so that it stays in the cache, in
|
|
|
|
// case we get an upgrade response (and hence no
|
|
|
|
// new data) when the write miss completes.
|
|
|
|
// As long as CPUs do proper store/load forwarding
|
|
|
|
// internally, and have a sufficiently weak memory
|
|
|
|
// model, this is probably unnecessary, but at some
|
|
|
|
// point it must have seemed like we needed it...
|
|
|
|
assert(pkt->needsExclusive() && !blk->isWritable());
|
|
|
|
blk->status &= ~BlkReadable;
|
|
|
|
}
|
|
|
|
|
2007-07-21 22:45:17 +02:00
|
|
|
allocateMissBuffer(pkt, time, true);
|
|
|
|
}
|
2009-02-16 17:56:40 +01:00
|
|
|
|
|
|
|
if (prefetcher) {
|
|
|
|
next_pf_time = prefetcher->notify(pkt, time);
|
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2009-02-16 17:56:40 +01:00
|
|
|
if (next_pf_time != 0)
|
|
|
|
requestMemSideBus(Request_PF, std::max(time, next_pf_time));
|
|
|
|
|
2008-03-25 15:01:21 +01:00
|
|
|
// copy writebacks to write buffer
|
|
|
|
while (!writebacks.empty()) {
|
|
|
|
PacketPtr wbPkt = writebacks.front();
|
|
|
|
allocateWriteBuffer(wbPkt, time, true);
|
|
|
|
writebacks.pop_front();
|
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
return true;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2008-02-27 05:17:26 +01:00
|
|
|
// See comment in cache.hh.
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-21 20:59:17 +02:00
|
|
|
PacketPtr
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::getBusPacket(PacketPtr cpu_pkt, BlkType *blk,
|
|
|
|
bool needsExclusive)
|
2007-06-21 20:59:17 +02:00
|
|
|
{
|
|
|
|
bool blkValid = blk && blk->isValid();
|
|
|
|
|
|
|
|
if (cpu_pkt->req->isUncacheable()) {
|
2010-08-23 18:18:41 +02:00
|
|
|
//assert(blk == NULL);
|
2007-06-21 20:59:17 +02:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2010-06-17 00:25:57 +02:00
|
|
|
if (!blkValid &&
|
|
|
|
(cpu_pkt->cmd == MemCmd::Writeback || cpu_pkt->isUpgrade())) {
|
2008-02-27 05:17:26 +01:00
|
|
|
// Writebacks that weren't allocated in access() and upgrades
|
|
|
|
// from upper-level caches that missed completely just go
|
|
|
|
// through.
|
2007-06-21 20:59:17 +02:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2007-06-22 18:24:07 +02:00
|
|
|
assert(cpu_pkt->needsResponse());
|
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
MemCmd cmd;
|
2007-07-21 22:45:17 +02:00
|
|
|
// @TODO make useUpgrades a parameter.
|
|
|
|
// Note that ownership protocols require upgrade, otherwise a
|
|
|
|
// write miss on a shared owned block will generate a ReadExcl,
|
|
|
|
// which will clobber the owned copy.
|
2007-06-21 20:59:17 +02:00
|
|
|
const bool useUpgrades = true;
|
|
|
|
if (blkValid && useUpgrades) {
|
|
|
|
// only reason to be here is that blk is shared
|
|
|
|
// (read-only) and we need exclusive
|
|
|
|
assert(needsExclusive && !blk->isWritable());
|
2010-06-17 00:25:57 +02:00
|
|
|
cmd = cpu_pkt->isLLSC() ? MemCmd::SCUpgradeReq : MemCmd::UpgradeReq;
|
2007-06-21 20:59:17 +02:00
|
|
|
} else {
|
|
|
|
// block is invalid
|
|
|
|
cmd = needsExclusive ? MemCmd::ReadExReq : MemCmd::ReadReq;
|
|
|
|
}
|
MEM: Remove the Broadcast destination from the packet
This patch simplifies the packet by removing the broadcast flag and
instead more firmly relying on (and enforcing) the semantics of
transactions in the classic memory system, i.e. request packets are
routed from a master to a slave based on the address, and when they
are created they have neither a valid source, nor destination. On
their way to the slave, the request packet is updated with a source
field for all modules that multiplex packets from multiple master
(e.g. a bus). When a request packet is turned into a response packet
(at the final slave), it moves the potentially populated source field
to the destination field, and the response packet is routed through
any multiplexing components back to the master based on the
destination field.
Modules that connect multiplexing components, such as caches and
bridges store any existing source and destination field in the sender
state as a stack (just as before).
The packet constructor is simplified in that there is no longer a need
to pass the Packet::Broadcast as the destination (this was always the
case for the classic memory system). In the case of Ruby, rather than
using the parameter to the constructor we now rely on setDest, as
there is already another three-argument constructor in the packet
class.
In many places where the packet information was printed as part of
DPRINTFs, request packets would be printed with a numeric "dest" that
would always be -1 (Broadcast) and that field is now removed from the
printing.
2012-04-14 11:45:55 +02:00
|
|
|
PacketPtr pkt = new Packet(cpu_pkt->req, cmd, blkSize);
|
2007-06-21 20:59:17 +02:00
|
|
|
|
|
|
|
pkt->allocate();
|
|
|
|
return pkt;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
Tick
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::atomicAccess(PacketPtr pkt)
|
2006-12-19 05:47:12 +01:00
|
|
|
{
|
2012-10-15 14:10:54 +02:00
|
|
|
Cycles lat = hitLatency;
|
2007-06-21 20:59:17 +02:00
|
|
|
|
2007-07-16 05:11:06 +02:00
|
|
|
// @TODO: make this a parameter
|
|
|
|
bool last_level_cache = false;
|
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
if (pkt->memInhibitAsserted()) {
|
|
|
|
assert(!pkt->req->isUncacheable());
|
2007-07-16 05:11:06 +02:00
|
|
|
// have to invalidate ourselves and any lower caches even if
|
|
|
|
// upper cache will be responding
|
|
|
|
if (pkt->isInvalidate()) {
|
|
|
|
BlkType *blk = tags->findBlock(pkt->getAddr());
|
|
|
|
if (blk && blk->isValid()) {
|
2012-09-11 20:14:49 +02:00
|
|
|
tags->invalidate(blk);
|
|
|
|
blk->invalidate();
|
2007-07-16 05:11:06 +02:00
|
|
|
DPRINTF(Cache, "rcvd mem-inhibited %s on 0x%x: invalidating\n",
|
|
|
|
pkt->cmdString(), pkt->getAddr());
|
|
|
|
}
|
|
|
|
if (!last_level_cache) {
|
|
|
|
DPRINTF(Cache, "forwarding mem-inhibited %s on 0x%x\n",
|
|
|
|
pkt->cmdString(), pkt->getAddr());
|
2012-10-15 14:10:54 +02:00
|
|
|
lat += ticksToCycles(memSidePort->sendAtomic(pkt));
|
2007-07-16 05:11:06 +02:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
DPRINTF(Cache, "rcvd mem-inhibited %s on 0x%x: not responding\n",
|
|
|
|
pkt->cmdString(), pkt->getAddr());
|
|
|
|
}
|
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
return lat;
|
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// should assert here that there are no outstanding MSHRs or
|
|
|
|
// writebacks... that would mean that someone used an atomic
|
|
|
|
// access in timing mode
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
BlkType *blk = NULL;
|
2008-03-25 15:01:21 +01:00
|
|
|
PacketList writebacks;
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2008-03-25 15:01:21 +01:00
|
|
|
if (!access(pkt, blk, lat, writebacks)) {
|
2007-06-18 02:27:53 +02:00
|
|
|
// MISS
|
2008-11-10 23:10:28 +01:00
|
|
|
PacketPtr bus_pkt = getBusPacket(pkt, blk, pkt->needsExclusive());
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2008-11-10 23:10:28 +01:00
|
|
|
bool is_forward = (bus_pkt == NULL);
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2008-11-10 23:10:28 +01:00
|
|
|
if (is_forward) {
|
2007-06-21 20:59:17 +02:00
|
|
|
// just forwarding the same request to the next level
|
|
|
|
// no local cache operation involved
|
2008-11-10 23:10:28 +01:00
|
|
|
bus_pkt = pkt;
|
2007-06-21 20:59:17 +02:00
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
DPRINTF(Cache, "Sending an atomic %s for %x\n",
|
2008-11-10 23:10:28 +01:00
|
|
|
bus_pkt->cmdString(), bus_pkt->getAddr());
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
#if TRACING_ON
|
|
|
|
CacheBlk::State old_state = blk ? blk->status : 0;
|
|
|
|
#endif
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2012-10-15 14:10:54 +02:00
|
|
|
lat += ticksToCycles(memSidePort->sendAtomic(bus_pkt));
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
DPRINTF(Cache, "Receive response: %s for addr %x in state %i\n",
|
2008-11-10 23:10:28 +01:00
|
|
|
bus_pkt->cmdString(), bus_pkt->getAddr(), old_state);
|
|
|
|
|
|
|
|
// If packet was a forward, the response (if any) is already
|
|
|
|
// in place in the bus_pkt == pkt structure, so we don't need
|
|
|
|
// to do anything. Otherwise, use the separate bus_pkt to
|
|
|
|
// generate response to pkt and then delete it.
|
|
|
|
if (!is_forward) {
|
|
|
|
if (pkt->needsResponse()) {
|
|
|
|
assert(bus_pkt->isResponse());
|
|
|
|
if (bus_pkt->isError()) {
|
|
|
|
pkt->makeAtomicResponse();
|
|
|
|
pkt->copyError(bus_pkt);
|
|
|
|
} else if (bus_pkt->isRead() ||
|
|
|
|
bus_pkt->cmd == MemCmd::UpgradeResp) {
|
|
|
|
// we're updating cache state to allow us to
|
|
|
|
// satisfy the upstream request from the cache
|
|
|
|
blk = handleFill(bus_pkt, blk, writebacks);
|
|
|
|
satisfyCpuSideRequest(pkt, blk);
|
|
|
|
} else {
|
|
|
|
// we're satisfying the upstream request without
|
|
|
|
// modifying cache state, e.g., a write-through
|
|
|
|
pkt->makeAtomicResponse();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
delete bus_pkt;
|
2007-06-21 20:59:17 +02:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2009-02-16 17:56:40 +01:00
|
|
|
// Note that we don't invoke the prefetcher at all in atomic mode.
|
|
|
|
// It's not clear how to do it properly, particularly for
|
|
|
|
// prefetchers that aggressively generate prefetch candidates and
|
|
|
|
// rely on bandwidth contention to throttle them; these will tend
|
|
|
|
// to pollute the cache in atomic mode since there is no bandwidth
|
|
|
|
// contention. If we ever do want to enable prefetching in atomic
|
|
|
|
// mode, though, this is the place to do it... see timingAccess()
|
|
|
|
// for an example (though we'd want to issue the prefetch(es)
|
|
|
|
// immediately rather than calling requestMemSideBus() as we do
|
|
|
|
// there).
|
|
|
|
|
2008-03-25 15:01:21 +01:00
|
|
|
// Handle writebacks if needed
|
|
|
|
while (!writebacks.empty()){
|
|
|
|
PacketPtr wbPkt = writebacks.front();
|
|
|
|
memSidePort->sendAtomic(wbPkt);
|
|
|
|
writebacks.pop_front();
|
|
|
|
delete wbPkt;
|
|
|
|
}
|
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
// We now have the block one way or another (hit or completed miss)
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
if (pkt->needsResponse()) {
|
|
|
|
pkt->makeAtomicResponse();
|
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
return lat;
|
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
void
|
2012-01-17 19:55:07 +01:00
|
|
|
Cache<TagStore>::functionalAccess(PacketPtr pkt, bool fromCpuSide)
|
2007-06-18 02:27:53 +02:00
|
|
|
{
|
2009-09-26 19:50:50 +02:00
|
|
|
Addr blk_addr = blockAlign(pkt->getAddr());
|
2007-06-18 02:27:53 +02:00
|
|
|
BlkType *blk = tags->findBlock(pkt->getAddr());
|
2010-09-09 20:40:19 +02:00
|
|
|
MSHR *mshr = mshrQueue.findMatch(blk_addr);
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2008-01-02 21:20:15 +01:00
|
|
|
pkt->pushLabel(name());
|
|
|
|
|
|
|
|
CacheBlkPrintWrapper cbpw(blk);
|
2010-09-09 20:40:19 +02:00
|
|
|
|
|
|
|
// Note that just because an L2/L3 has valid data doesn't mean an
|
|
|
|
// L1 doesn't have a more up-to-date modified copy that still
|
|
|
|
// needs to be found. As a result we always update the request if
|
|
|
|
// we have it, but only declare it satisfied if we are the owner.
|
|
|
|
|
|
|
|
// see if we have data at all (owned or otherwise)
|
|
|
|
bool have_data = blk && blk->isValid()
|
|
|
|
&& pkt->checkFunctional(&cbpw, blk_addr, blkSize, blk->data);
|
|
|
|
|
|
|
|
// data we have is dirty if marked as such or if valid & ownership
|
|
|
|
// pending due to outstanding UpgradeReq
|
|
|
|
bool have_dirty =
|
|
|
|
have_data && (blk->isDirty() ||
|
|
|
|
(mshr && mshr->inService && mshr->isPendingDirty()));
|
|
|
|
|
|
|
|
bool done = have_dirty
|
2012-01-17 19:55:07 +01:00
|
|
|
|| cpuSidePort->checkFunctional(pkt)
|
2008-01-02 21:20:15 +01:00
|
|
|
|| mshrQueue.checkFunctional(pkt, blk_addr)
|
|
|
|
|| writeBuffer.checkFunctional(pkt, blk_addr)
|
2012-01-17 19:55:07 +01:00
|
|
|
|| memSidePort->checkFunctional(pkt);
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2010-09-09 20:40:19 +02:00
|
|
|
DPRINTF(Cache, "functional %s %x %s%s%s\n",
|
|
|
|
pkt->cmdString(), pkt->getAddr(),
|
|
|
|
(blk && blk->isValid()) ? "valid " : "",
|
|
|
|
have_data ? "data " : "", done ? "done " : "");
|
|
|
|
|
2008-01-02 21:20:15 +01:00
|
|
|
// We're leaving the cache, so pop cache->name() label
|
|
|
|
pkt->popLabel();
|
|
|
|
|
2010-09-09 20:40:19 +02:00
|
|
|
if (done) {
|
|
|
|
pkt->makeResponse();
|
|
|
|
} else {
|
2012-01-17 19:55:07 +01:00
|
|
|
// if it came as a request from the CPU side then make sure it
|
|
|
|
// continues towards the memory side
|
|
|
|
if (fromCpuSide) {
|
|
|
|
memSidePort->sendFunctional(pkt);
|
2012-07-09 18:35:32 +02:00
|
|
|
} else if (forwardSnoops && cpuSidePort->isSnooping()) {
|
2012-01-17 19:55:07 +01:00
|
|
|
// if it came from the memory side, it must be a snoop request
|
|
|
|
// and we should only forward it if we are forwarding snoops
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
cpuSidePort->sendFunctionalSnoop(pkt);
|
2012-01-17 19:55:07 +01:00
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
//
|
|
|
|
// Response handling: responses from the memory side
|
|
|
|
//
|
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-12-19 05:47:12 +01:00
|
|
|
void
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::handleResponse(PacketPtr pkt)
|
2006-12-19 05:47:12 +01:00
|
|
|
{
|
2012-10-15 14:10:54 +02:00
|
|
|
Tick time = clockEdge(hitLatency);
|
2007-06-18 02:27:53 +02:00
|
|
|
MSHR *mshr = dynamic_cast<MSHR*>(pkt->senderState);
|
2007-08-27 06:45:40 +02:00
|
|
|
bool is_error = pkt->isError();
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
assert(mshr);
|
2007-06-25 02:32:31 +02:00
|
|
|
|
2007-08-27 06:45:40 +02:00
|
|
|
if (is_error) {
|
|
|
|
DPRINTF(Cache, "Cache received packet with error for address %x, "
|
|
|
|
"cmd: %s\n", pkt->getAddr(), pkt->cmdString());
|
|
|
|
}
|
|
|
|
|
2007-06-25 02:32:31 +02:00
|
|
|
DPRINTF(Cache, "Handling response to %x\n", pkt->getAddr());
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2007-06-21 20:59:17 +02:00
|
|
|
MSHRQueue *mq = mshr->queue;
|
|
|
|
bool wasFull = mq->isFull();
|
|
|
|
|
|
|
|
if (mshr == noTargetMSHR) {
|
|
|
|
// we always clear at least one target
|
|
|
|
clearBlocked(Blocked_NoTargets);
|
|
|
|
noTargetMSHR = NULL;
|
|
|
|
}
|
|
|
|
|
2007-06-30 22:34:16 +02:00
|
|
|
// Initial target is used just for stats
|
|
|
|
MSHR::Target *initial_tgt = mshr->getTarget();
|
2007-07-21 22:45:17 +02:00
|
|
|
BlkType *blk = tags->findBlock(pkt->getAddr());
|
2007-06-30 22:34:16 +02:00
|
|
|
int stats_cmd_idx = initial_tgt->pkt->cmdToIndex();
|
2011-01-08 06:50:29 +01:00
|
|
|
Tick miss_latency = curTick() - initial_tgt->recvTime;
|
2007-07-21 22:45:17 +02:00
|
|
|
PacketList writebacks;
|
2007-06-30 22:34:16 +02:00
|
|
|
|
2007-07-21 22:45:17 +02:00
|
|
|
if (pkt->req->isUncacheable()) {
|
2012-02-12 23:07:39 +01:00
|
|
|
assert(pkt->req->masterId() < system->maxMasters());
|
|
|
|
mshr_uncacheable_lat[stats_cmd_idx][pkt->req->masterId()] +=
|
2007-07-21 22:45:17 +02:00
|
|
|
miss_latency;
|
|
|
|
} else {
|
2012-02-12 23:07:39 +01:00
|
|
|
assert(pkt->req->masterId() < system->maxMasters());
|
|
|
|
mshr_miss_latency[stats_cmd_idx][pkt->req->masterId()] +=
|
2007-06-30 22:34:16 +02:00
|
|
|
miss_latency;
|
2007-07-21 22:45:17 +02:00
|
|
|
}
|
|
|
|
|
2008-11-10 23:10:28 +01:00
|
|
|
bool is_fill = !mshr->isForward &&
|
|
|
|
(pkt->isRead() || pkt->cmd == MemCmd::UpgradeResp);
|
|
|
|
|
|
|
|
if (is_fill && !is_error) {
|
2007-06-18 02:27:53 +02:00
|
|
|
DPRINTF(Cache, "Block for addr %x being updated in Cache\n",
|
|
|
|
pkt->getAddr());
|
2007-06-26 23:53:15 +02:00
|
|
|
|
2007-06-27 08:30:30 +02:00
|
|
|
// give mshr a chance to do some dirty work
|
|
|
|
mshr->handleFill(pkt, blk);
|
2007-06-26 23:53:15 +02:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
blk = handleFill(pkt, blk, writebacks);
|
2007-07-21 22:45:17 +02:00
|
|
|
assert(blk != NULL);
|
|
|
|
}
|
2007-06-21 20:59:17 +02:00
|
|
|
|
2007-07-21 22:45:17 +02:00
|
|
|
// First offset for critical word first calculations
|
|
|
|
int initial_offset = 0;
|
|
|
|
|
|
|
|
if (mshr->hasTargets()) {
|
|
|
|
initial_offset = mshr->getTarget()->pkt->getOffset(blkSize);
|
|
|
|
}
|
|
|
|
|
|
|
|
while (mshr->hasTargets()) {
|
|
|
|
MSHR::Target *target = mshr->getTarget();
|
|
|
|
|
2009-02-16 17:56:40 +01:00
|
|
|
switch (target->source) {
|
|
|
|
case MSHR::Target::FromCPU:
|
2007-07-21 22:45:17 +02:00
|
|
|
Tick completion_time;
|
2008-11-10 23:10:28 +01:00
|
|
|
if (is_fill) {
|
2010-09-09 20:40:18 +02:00
|
|
|
satisfyCpuSideRequest(target->pkt, blk,
|
|
|
|
true, mshr->hasPostDowngrade());
|
2007-07-24 07:28:40 +02:00
|
|
|
// How many bytes past the first request is this one
|
2007-07-21 22:45:17 +02:00
|
|
|
int transfer_offset =
|
|
|
|
target->pkt->getOffset(blkSize) - initial_offset;
|
|
|
|
if (transfer_offset < 0) {
|
|
|
|
transfer_offset += blkSize;
|
|
|
|
}
|
|
|
|
|
2012-09-25 18:49:41 +02:00
|
|
|
// If critical word (no offset) return first word time.
|
|
|
|
// responseLatency is the latency of the return path
|
|
|
|
// from lower level caches/memory to an upper level cache or
|
|
|
|
// the core.
|
2012-10-15 14:10:54 +02:00
|
|
|
completion_time = responseLatency * clock +
|
2007-08-13 01:43:54 +02:00
|
|
|
(transfer_offset ? pkt->finishTime : pkt->firstWordTime);
|
2007-07-21 22:45:17 +02:00
|
|
|
|
2007-07-24 07:28:40 +02:00
|
|
|
assert(!target->pkt->req->isUncacheable());
|
2012-02-12 23:07:39 +01:00
|
|
|
|
2012-06-29 17:19:07 +02:00
|
|
|
assert(target->pkt->req->masterId() < system->maxMasters());
|
2012-02-12 23:07:39 +01:00
|
|
|
missLatency[target->pkt->cmdToIndex()][target->pkt->req->masterId()] +=
|
2007-07-24 07:28:40 +02:00
|
|
|
completion_time - target->recvTime;
|
2010-09-09 20:40:19 +02:00
|
|
|
} else if (pkt->cmd == MemCmd::UpgradeFailResp) {
|
2010-06-17 00:25:57 +02:00
|
|
|
// failed StoreCond upgrade
|
2010-09-09 20:40:19 +02:00
|
|
|
assert(target->pkt->cmd == MemCmd::StoreCondReq ||
|
2010-10-18 22:05:15 +02:00
|
|
|
target->pkt->cmd == MemCmd::StoreCondFailReq ||
|
|
|
|
target->pkt->cmd == MemCmd::SCUpgradeFailReq);
|
2012-09-25 18:49:41 +02:00
|
|
|
// responseLatency is the latency of the return path
|
|
|
|
// from lower level caches/memory to an upper level cache or
|
|
|
|
// the core.
|
2012-10-15 14:10:54 +02:00
|
|
|
completion_time = responseLatency * clock + pkt->finishTime;
|
2010-06-17 00:25:57 +02:00
|
|
|
target->pkt->req->setExtraData(0);
|
2007-07-21 22:45:17 +02:00
|
|
|
} else {
|
|
|
|
// not a cache fill, just forwarding response
|
2012-09-25 18:49:41 +02:00
|
|
|
// responseLatency is the latency of the return path
|
|
|
|
// from lower level cahces/memory to the core.
|
2012-10-15 14:10:54 +02:00
|
|
|
completion_time = responseLatency * clock + pkt->finishTime;
|
2007-08-27 06:45:40 +02:00
|
|
|
if (pkt->isRead() && !is_error) {
|
2007-07-21 22:45:17 +02:00
|
|
|
target->pkt->setData(pkt->getPtr<uint8_t>());
|
|
|
|
}
|
2007-06-21 20:59:17 +02:00
|
|
|
}
|
2007-07-02 22:57:45 +02:00
|
|
|
target->pkt->makeTimingResponse();
|
2007-08-27 06:45:40 +02:00
|
|
|
// if this packet is an error copy that to the new packet
|
|
|
|
if (is_error)
|
|
|
|
target->pkt->copyError(pkt);
|
2010-09-09 20:40:18 +02:00
|
|
|
if (target->pkt->cmd == MemCmd::ReadResp &&
|
|
|
|
(pkt->isInvalidate() || mshr->hasPostInvalidate())) {
|
2008-01-03 00:22:38 +01:00
|
|
|
// If intermediate cache got ReadRespWithInvalidate,
|
|
|
|
// propagate that. Response should not have
|
|
|
|
// isInvalidate() set otherwise.
|
|
|
|
target->pkt->cmd = MemCmd::ReadRespWithInvalidate;
|
|
|
|
}
|
2012-08-22 17:39:56 +02:00
|
|
|
cpuSidePort->schedTimingResp(target->pkt, completion_time);
|
2009-02-16 17:56:40 +01:00
|
|
|
break;
|
|
|
|
|
|
|
|
case MSHR::Target::FromPrefetcher:
|
|
|
|
assert(target->pkt->cmd == MemCmd::HardPFReq);
|
|
|
|
if (blk)
|
|
|
|
blk->status |= BlkHWPrefetched;
|
|
|
|
delete target->pkt->req;
|
|
|
|
delete target->pkt;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case MSHR::Target::FromSnoop:
|
2007-08-27 06:45:40 +02:00
|
|
|
// I don't believe that a snoop can be in an error state
|
|
|
|
assert(!is_error);
|
2007-07-21 22:45:17 +02:00
|
|
|
// response to snoop request
|
|
|
|
DPRINTF(Cache, "processing deferred snoop...\n");
|
2010-09-09 20:40:18 +02:00
|
|
|
assert(!(pkt->isInvalidate() && !mshr->hasPostInvalidate()));
|
2008-01-03 00:22:38 +01:00
|
|
|
handleSnoop(target->pkt, blk, true, true,
|
2010-09-09 20:40:18 +02:00
|
|
|
mshr->hasPostInvalidate());
|
2009-02-16 17:56:40 +01:00
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
panic("Illegal target->source enum %d\n", target->source);
|
2007-06-21 20:59:17 +02:00
|
|
|
}
|
|
|
|
|
2007-07-21 22:45:17 +02:00
|
|
|
mshr->popTarget();
|
|
|
|
}
|
2007-07-03 06:40:31 +02:00
|
|
|
|
2012-09-11 20:14:49 +02:00
|
|
|
if (blk && blk->isValid()) {
|
2010-09-09 20:40:18 +02:00
|
|
|
if (pkt->isInvalidate() || mshr->hasPostInvalidate()) {
|
2012-09-11 20:14:49 +02:00
|
|
|
assert(blk != tempBlock);
|
|
|
|
tags->invalidate(blk);
|
|
|
|
blk->invalidate();
|
2010-09-09 20:40:18 +02:00
|
|
|
} else if (mshr->hasPostDowngrade()) {
|
|
|
|
blk->status &= ~BlkWritable;
|
|
|
|
}
|
2008-01-03 00:22:38 +01:00
|
|
|
}
|
|
|
|
|
2007-07-21 22:45:17 +02:00
|
|
|
if (mshr->promoteDeferredTargets()) {
|
2008-03-15 13:03:55 +01:00
|
|
|
// avoid later read getting stale data while write miss is
|
|
|
|
// outstanding.. see comment in timingAccess()
|
2010-09-09 20:40:18 +02:00
|
|
|
if (blk) {
|
|
|
|
blk->status &= ~BlkReadable;
|
|
|
|
}
|
2007-07-21 22:45:17 +02:00
|
|
|
MSHRQueue *mq = mshr->queue;
|
|
|
|
mq->markPending(mshr);
|
|
|
|
requestMemSideBus((RequestCause)mq->index, pkt->finishTime);
|
|
|
|
} else {
|
2007-06-21 20:59:17 +02:00
|
|
|
mq->deallocate(mshr);
|
|
|
|
if (wasFull && !mq->isFull()) {
|
|
|
|
clearBlocked((BlockedCause)mq->index);
|
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
2007-07-21 22:45:17 +02:00
|
|
|
|
|
|
|
// copy writebacks to write buffer
|
|
|
|
while (!writebacks.empty()) {
|
|
|
|
PacketPtr wbPkt = writebacks.front();
|
|
|
|
allocateWriteBuffer(wbPkt, time, true);
|
|
|
|
writebacks.pop_front();
|
|
|
|
}
|
|
|
|
// if we used temp block, clear it out
|
|
|
|
if (blk == tempBlock) {
|
|
|
|
if (blk->isDirty()) {
|
|
|
|
allocateWriteBuffer(writebackBlk(blk), time, true);
|
|
|
|
}
|
2012-09-11 20:14:49 +02:00
|
|
|
blk->invalidate();
|
2007-07-21 22:45:17 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
delete pkt;
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
|
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-12-19 05:47:12 +01:00
|
|
|
PacketPtr
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::writebackBlk(BlkType *blk)
|
2006-12-19 05:47:12 +01:00
|
|
|
{
|
2007-06-18 02:27:53 +02:00
|
|
|
assert(blk && blk->isValid() && blk->isDirty());
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2012-02-12 23:07:39 +01:00
|
|
|
writebacks[Request::wbMasterId]++;
|
2006-12-19 05:47:12 +01:00
|
|
|
|
|
|
|
Request *writebackReq =
|
2012-02-12 23:07:38 +01:00
|
|
|
new Request(tags->regenerateBlkAddr(blk->tag, blk->set), blkSize, 0,
|
|
|
|
Request::wbMasterId);
|
MEM: Remove the Broadcast destination from the packet
This patch simplifies the packet by removing the broadcast flag and
instead more firmly relying on (and enforcing) the semantics of
transactions in the classic memory system, i.e. request packets are
routed from a master to a slave based on the address, and when they
are created they have neither a valid source, nor destination. On
their way to the slave, the request packet is updated with a source
field for all modules that multiplex packets from multiple master
(e.g. a bus). When a request packet is turned into a response packet
(at the final slave), it moves the potentially populated source field
to the destination field, and the response packet is routed through
any multiplexing components back to the master based on the
destination field.
Modules that connect multiplexing components, such as caches and
bridges store any existing source and destination field in the sender
state as a stack (just as before).
The packet constructor is simplified in that there is no longer a need
to pass the Packet::Broadcast as the destination (this was always the
case for the classic memory system). In the case of Ruby, rather than
using the parameter to the constructor we now rely on setDest, as
there is already another three-argument constructor in the packet
class.
In many places where the packet information was printed as part of
DPRINTFs, request packets would be printed with a numeric "dest" that
would always be -1 (Broadcast) and that field is now removed from the
printing.
2012-04-14 11:45:55 +02:00
|
|
|
PacketPtr writeback = new Packet(writebackReq, MemCmd::Writeback);
|
2010-09-22 08:07:34 +02:00
|
|
|
if (blk->isWritable()) {
|
|
|
|
writeback->setSupplyExclusive();
|
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
writeback->allocate();
|
2007-06-18 02:27:53 +02:00
|
|
|
std::memcpy(writeback->getPtr<uint8_t>(), blk->data, blkSize);
|
2006-12-19 05:47:12 +01:00
|
|
|
|
|
|
|
blk->status &= ~BlkDirty;
|
|
|
|
return writeback;
|
|
|
|
}
|
|
|
|
|
2012-11-02 17:32:02 +01:00
|
|
|
template<class TagStore>
|
|
|
|
void
|
|
|
|
Cache<TagStore>::memWriteback()
|
|
|
|
{
|
|
|
|
WrappedBlkVisitor visitor(*this, &Cache<TagStore>::writebackVisitor);
|
|
|
|
tags->forEachBlk(visitor);
|
|
|
|
}
|
|
|
|
|
|
|
|
template<class TagStore>
|
|
|
|
void
|
|
|
|
Cache<TagStore>::memInvalidate()
|
|
|
|
{
|
|
|
|
WrappedBlkVisitor visitor(*this, &Cache<TagStore>::invalidateVisitor);
|
|
|
|
tags->forEachBlk(visitor);
|
|
|
|
}
|
|
|
|
|
|
|
|
template<class TagStore>
|
|
|
|
bool
|
|
|
|
Cache<TagStore>::isDirty() const
|
|
|
|
{
|
|
|
|
CacheBlkIsDirtyVisitor<BlkType> visitor;
|
|
|
|
tags->forEachBlk(visitor);
|
|
|
|
|
|
|
|
return visitor.isDirty();
|
|
|
|
}
|
|
|
|
|
|
|
|
template<class TagStore>
|
|
|
|
bool
|
|
|
|
Cache<TagStore>::writebackVisitor(BlkType &blk)
|
|
|
|
{
|
|
|
|
if (blk.isDirty()) {
|
|
|
|
assert(blk.isValid());
|
|
|
|
|
|
|
|
Request request(tags->regenerateBlkAddr(blk.tag, blk.set),
|
|
|
|
blkSize, 0, Request::funcMasterId);
|
|
|
|
|
|
|
|
Packet packet(&request, MemCmd::WriteReq);
|
|
|
|
packet.dataStatic(blk.data);
|
|
|
|
|
|
|
|
memSidePort->sendFunctional(&packet);
|
|
|
|
|
|
|
|
blk.status &= ~BlkDirty;
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
template<class TagStore>
|
|
|
|
bool
|
|
|
|
Cache<TagStore>::invalidateVisitor(BlkType &blk)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (blk.isDirty())
|
|
|
|
warn_once("Invalidating dirty cache lines. Expect things to break.\n");
|
|
|
|
|
|
|
|
if (blk.isValid()) {
|
|
|
|
assert(!blk.isDirty());
|
|
|
|
tags->invalidate(dynamic_cast< BlkType *>(&blk));
|
|
|
|
blk.invalidate();
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
2006-12-19 05:47:12 +01:00
|
|
|
|
2013-01-07 19:05:47 +01:00
|
|
|
template<class TagStore>
|
|
|
|
void
|
|
|
|
Cache<TagStore>::uncacheableFlush(PacketPtr pkt)
|
|
|
|
{
|
|
|
|
DPRINTF(Cache, "%s%s %x uncacheable\n", pkt->cmdString(),
|
|
|
|
pkt->req->isInstFetch() ? " (ifetch)" : "",
|
|
|
|
pkt->getAddr());
|
|
|
|
|
|
|
|
if (pkt->req->isClearLL())
|
|
|
|
tags->clearLocks();
|
|
|
|
|
|
|
|
BlkType *blk(tags->findBlock(pkt->getAddr()));
|
|
|
|
if (blk) {
|
|
|
|
writebackVisitor(*blk);
|
|
|
|
invalidateVisitor(*blk);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2008-02-16 20:58:03 +01:00
|
|
|
template<class TagStore>
|
|
|
|
typename Cache<TagStore>::BlkType*
|
|
|
|
Cache<TagStore>::allocateBlock(Addr addr, PacketList &writebacks)
|
|
|
|
{
|
2008-11-04 17:35:58 +01:00
|
|
|
BlkType *blk = tags->findVictim(addr, writebacks);
|
2008-02-16 20:58:03 +01:00
|
|
|
|
|
|
|
if (blk->isValid()) {
|
|
|
|
Addr repl_addr = tags->regenerateBlkAddr(blk->tag, blk->set);
|
|
|
|
MSHR *repl_mshr = mshrQueue.findMatch(repl_addr);
|
|
|
|
if (repl_mshr) {
|
|
|
|
// must be an outstanding upgrade request on block
|
|
|
|
// we're about to replace...
|
|
|
|
assert(!blk->isWritable());
|
|
|
|
assert(repl_mshr->needsExclusive());
|
|
|
|
// too hard to replace block with transient state
|
2008-11-04 17:35:58 +01:00
|
|
|
// allocation failed, block not inserted
|
2008-02-16 20:58:03 +01:00
|
|
|
return NULL;
|
|
|
|
} else {
|
|
|
|
DPRINTF(Cache, "replacement: replacing %x with %x: %s\n",
|
|
|
|
repl_addr, addr,
|
|
|
|
blk->isDirty() ? "writeback" : "clean");
|
|
|
|
|
|
|
|
if (blk->isDirty()) {
|
|
|
|
// Save writeback packet for handling by caller
|
|
|
|
writebacks.push_back(writebackBlk(blk));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return blk;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// Note that the reason we return a list of writebacks rather than
|
|
|
|
// inserting them directly in the write buffer is that this function
|
|
|
|
// is called by both atomic and timing-mode accesses, and in atomic
|
|
|
|
// mode we don't mess with the write buffer (we just perform the
|
|
|
|
// writebacks atomically once the original request is complete).
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
|
|
|
typename Cache<TagStore>::BlkType*
|
|
|
|
Cache<TagStore>::handleFill(PacketPtr pkt, BlkType *blk,
|
|
|
|
PacketList &writebacks)
|
2006-12-19 05:47:12 +01:00
|
|
|
{
|
2007-06-18 02:27:53 +02:00
|
|
|
Addr addr = pkt->getAddr();
|
2007-07-22 09:07:26 +02:00
|
|
|
#if TRACING_ON
|
2007-06-28 05:54:13 +02:00
|
|
|
CacheBlk::State old_state = blk ? blk->status : 0;
|
2007-07-22 09:07:26 +02:00
|
|
|
#endif
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2006-12-19 05:47:12 +01:00
|
|
|
if (blk == NULL) {
|
2007-06-26 23:53:15 +02:00
|
|
|
// better have read new data...
|
2008-01-03 00:22:38 +01:00
|
|
|
assert(pkt->hasData());
|
2006-12-19 05:47:12 +01:00
|
|
|
// need to do a replacement
|
2008-02-16 20:58:03 +01:00
|
|
|
blk = allocateBlock(addr, writebacks);
|
|
|
|
if (blk == NULL) {
|
|
|
|
// No replaceable block... just use temporary storage to
|
|
|
|
// complete the current request and then get rid of it
|
|
|
|
assert(!tempBlock->isValid());
|
|
|
|
blk = tempBlock;
|
|
|
|
tempBlock->set = tags->extractSet(addr);
|
2008-03-23 03:17:15 +01:00
|
|
|
tempBlock->tag = tags->extractTag(addr);
|
2008-02-16 20:58:03 +01:00
|
|
|
DPRINTF(Cache, "using temp block for %x\n", addr);
|
2008-11-04 17:35:58 +01:00
|
|
|
} else {
|
2012-02-12 23:07:39 +01:00
|
|
|
int id = pkt->req->masterId();
|
2010-01-12 19:53:02 +01:00
|
|
|
tags->insertBlock(pkt->getAddr(), blk, id);
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
2010-09-09 20:40:18 +02:00
|
|
|
|
2012-09-11 20:14:49 +02:00
|
|
|
// we should never be overwriting a valid block
|
|
|
|
assert(!blk->isValid());
|
2006-12-19 05:47:12 +01:00
|
|
|
} else {
|
2007-06-18 02:27:53 +02:00
|
|
|
// existing block... probably an upgrade
|
|
|
|
assert(blk->tag == tags->extractTag(addr));
|
|
|
|
// either we're getting new data or the block should already be valid
|
2008-01-03 00:22:38 +01:00
|
|
|
assert(pkt->hasData() || blk->isValid());
|
2010-09-09 20:40:18 +02:00
|
|
|
// don't clear block status... if block is already dirty we
|
|
|
|
// don't want to lose that
|
2006-12-19 05:47:12 +01:00
|
|
|
}
|
|
|
|
|
2010-09-09 20:40:18 +02:00
|
|
|
blk->status |= BlkValid | BlkReadable;
|
2010-06-17 00:25:57 +02:00
|
|
|
|
2007-11-17 05:10:33 +01:00
|
|
|
if (!pkt->sharedAsserted()) {
|
2010-06-17 00:25:57 +02:00
|
|
|
blk->status |= BlkWritable;
|
|
|
|
// If we got this via cache-to-cache transfer (i.e., from a
|
|
|
|
// cache that was an owner) and took away that owner's copy,
|
|
|
|
// then we need to write it back. Normally this happens
|
|
|
|
// anyway as a side effect of getting a copy to write it, but
|
|
|
|
// there are cases (such as failed store conditionals or
|
|
|
|
// compare-and-swaps) where we'll demand an exclusive copy but
|
|
|
|
// end up not writing it.
|
|
|
|
if (pkt->memInhibitAsserted())
|
|
|
|
blk->status |= BlkDirty;
|
2007-06-28 05:54:13 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
DPRINTF(Cache, "Block addr %x moving from state %i to %i\n",
|
|
|
|
addr, old_state, blk->status);
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// if we got new data, copy it in
|
|
|
|
if (pkt->isRead()) {
|
|
|
|
std::memcpy(blk->data, pkt->getPtr<uint8_t>(), blkSize);
|
|
|
|
}
|
|
|
|
|
|
|
|
blk->whenReady = pkt->finishTime;
|
|
|
|
|
2006-12-19 05:47:12 +01:00
|
|
|
return blk;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
//
|
|
|
|
// Snoop path: requests coming in from the memory side
|
|
|
|
//
|
|
|
|
/////////////////////////////////////////////////////
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-06-28 17:02:14 +02:00
|
|
|
void
|
2008-01-03 00:22:38 +01:00
|
|
|
Cache<TagStore>::
|
|
|
|
doTimingSupplyResponse(PacketPtr req_pkt, uint8_t *blk_data,
|
|
|
|
bool already_copied, bool pending_inval)
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2007-06-27 07:23:10 +02:00
|
|
|
// timing-mode snoop responses require a new packet, unless we
|
|
|
|
// already made a copy...
|
2010-06-17 00:25:57 +02:00
|
|
|
PacketPtr pkt = already_copied ? req_pkt : new Packet(req_pkt);
|
|
|
|
assert(req_pkt->isInvalidate() || pkt->sharedAsserted());
|
2007-06-18 02:27:53 +02:00
|
|
|
pkt->allocate();
|
|
|
|
pkt->makeTimingResponse();
|
2007-07-27 02:04:17 +02:00
|
|
|
if (pkt->isRead()) {
|
|
|
|
pkt->setDataFromBlock(blk_data, blkSize);
|
|
|
|
}
|
2008-01-03 00:22:38 +01:00
|
|
|
if (pkt->cmd == MemCmd::ReadResp && pending_inval) {
|
|
|
|
// Assume we defer a response to a read from a far-away cache
|
|
|
|
// A, then later defer a ReadExcl from a cache B on the same
|
|
|
|
// bus as us. We'll assert MemInhibit in both cases, but in
|
|
|
|
// the latter case MemInhibit will keep the invalidation from
|
|
|
|
// reaching cache A. This special response tells cache A that
|
|
|
|
// it gets the block to satisfy its read, but must immediately
|
|
|
|
// invalidate it.
|
|
|
|
pkt->cmd = MemCmd::ReadRespWithInvalidate;
|
|
|
|
}
|
2012-10-15 14:10:54 +02:00
|
|
|
memSidePort->schedTimingSnoopResp(pkt, clockEdge(hitLatency));
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-06-28 17:02:14 +02:00
|
|
|
void
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::handleSnoop(PacketPtr pkt, BlkType *blk,
|
2008-01-03 00:22:38 +01:00
|
|
|
bool is_timing, bool is_deferred,
|
|
|
|
bool pending_inval)
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2008-01-03 00:22:38 +01:00
|
|
|
// deferred snoops can only happen in timing mode
|
|
|
|
assert(!(is_deferred && !is_timing));
|
|
|
|
// pending_inval only makes sense on deferred snoops
|
|
|
|
assert(!(pending_inval && !is_deferred));
|
2007-07-16 05:11:06 +02:00
|
|
|
assert(pkt->isRequest());
|
|
|
|
|
2010-08-26 06:55:55 +02:00
|
|
|
// the packet may get modified if we or a forwarded snooper
|
|
|
|
// responds in atomic mode, so remember a few things about the
|
|
|
|
// original packet up front
|
|
|
|
bool invalidate = pkt->isInvalidate();
|
2010-08-26 17:03:20 +02:00
|
|
|
bool M5_VAR_USED needs_exclusive = pkt->needsExclusive();
|
2010-08-26 06:55:55 +02:00
|
|
|
|
2008-07-16 20:10:33 +02:00
|
|
|
if (forwardSnoops) {
|
|
|
|
// first propagate snoop upward to see if anyone above us wants to
|
|
|
|
// handle it. save & restore packet src since it will get
|
|
|
|
// rewritten to be relative to cpu-side bus (if any)
|
|
|
|
bool alreadyResponded = pkt->memInhibitAsserted();
|
|
|
|
if (is_timing) {
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
Packet snoopPkt(pkt, true); // clear flags
|
|
|
|
snoopPkt.setExpressSnoop();
|
|
|
|
snoopPkt.senderState = new ForwardResponseRecord(pkt, this);
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
cpuSidePort->sendTimingSnoopReq(&snoopPkt);
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
if (snoopPkt.memInhibitAsserted()) {
|
2008-07-16 20:10:33 +02:00
|
|
|
// cache-to-cache response from some upper cache
|
|
|
|
assert(!alreadyResponded);
|
|
|
|
pkt->assertMemInhibit();
|
|
|
|
} else {
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
delete snoopPkt.senderState;
|
2008-07-16 20:10:33 +02:00
|
|
|
}
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
if (snoopPkt.sharedAsserted()) {
|
2008-07-16 20:10:33 +02:00
|
|
|
pkt->assertShared();
|
|
|
|
}
|
2007-07-17 15:33:28 +02:00
|
|
|
} else {
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
cpuSidePort->sendAtomicSnoop(pkt);
|
2008-07-16 20:10:33 +02:00
|
|
|
if (!alreadyResponded && pkt->memInhibitAsserted()) {
|
|
|
|
// cache-to-cache response from some upper cache:
|
|
|
|
// forward response to original requester
|
|
|
|
assert(pkt->isResponse());
|
|
|
|
}
|
2007-07-16 05:11:06 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
if (!blk || !blk->isValid()) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// we may end up modifying both the block state and the packet (if
|
|
|
|
// we respond in atomic mode), so just figure out what to do now
|
|
|
|
// and then do it later
|
2007-07-27 02:04:17 +02:00
|
|
|
bool respond = blk->isDirty() && pkt->needsResponse();
|
2007-07-27 02:04:17 +02:00
|
|
|
bool have_exclusive = blk->isWritable();
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2010-08-26 06:55:55 +02:00
|
|
|
if (pkt->isRead() && !invalidate) {
|
|
|
|
assert(!needs_exclusive);
|
2007-06-18 02:27:53 +02:00
|
|
|
pkt->assertShared();
|
|
|
|
int bits_to_clear = BlkWritable;
|
|
|
|
const bool haveOwnershipState = true; // for now
|
|
|
|
if (!haveOwnershipState) {
|
|
|
|
// if we don't support pure ownership (dirty && !writable),
|
|
|
|
// have to clear dirty bit here, assume memory snarfs data
|
|
|
|
// on cache-to-cache xfer
|
|
|
|
bits_to_clear |= BlkDirty;
|
2006-10-10 00:52:20 +02:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
blk->status &= ~bits_to_clear;
|
|
|
|
}
|
|
|
|
|
2007-09-17 01:46:38 +02:00
|
|
|
DPRINTF(Cache, "snooped a %s request for addr %x, %snew state is %i\n",
|
|
|
|
pkt->cmdString(), blockAlign(pkt->getAddr()),
|
|
|
|
respond ? "responding, " : "", invalidate ? 0 : blk->status);
|
|
|
|
|
2007-07-27 02:04:17 +02:00
|
|
|
if (respond) {
|
2007-06-27 07:23:10 +02:00
|
|
|
assert(!pkt->memInhibitAsserted());
|
2007-06-18 02:27:53 +02:00
|
|
|
pkt->assertMemInhibit();
|
2007-07-27 02:04:17 +02:00
|
|
|
if (have_exclusive) {
|
|
|
|
pkt->setSupplyExclusive();
|
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
if (is_timing) {
|
2008-01-03 00:22:38 +01:00
|
|
|
doTimingSupplyResponse(pkt, blk->data, is_deferred, pending_inval);
|
2007-06-18 02:27:53 +02:00
|
|
|
} else {
|
|
|
|
pkt->makeAtomicResponse();
|
|
|
|
pkt->setDataFromBlock(blk->data, blkSize);
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
2008-03-17 08:08:28 +01:00
|
|
|
} else if (is_timing && is_deferred) {
|
|
|
|
// if it's a deferred timing snoop then we've made a copy of
|
|
|
|
// the packet, and so if we're not using that copy to respond
|
|
|
|
// then we need to delete it here.
|
|
|
|
delete pkt;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
|
|
|
|
// Do this last in case it deallocates block data or something
|
|
|
|
// like that
|
|
|
|
if (invalidate) {
|
2012-09-11 20:14:49 +02:00
|
|
|
assert(blk != tempBlock);
|
|
|
|
tags->invalidate(blk);
|
|
|
|
blk->invalidate();
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-06-28 17:02:14 +02:00
|
|
|
void
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::snoopTiming(PacketPtr pkt)
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2007-07-24 07:28:40 +02:00
|
|
|
// Note that some deferred snoops don't have requests, since the
|
|
|
|
// original access may have already completed
|
2007-07-27 02:04:17 +02:00
|
|
|
if ((pkt->req && pkt->req->isUncacheable()) ||
|
|
|
|
pkt->cmd == MemCmd::Writeback) {
|
2006-10-12 19:33:21 +02:00
|
|
|
//Can't get a hit on an uncacheable address
|
|
|
|
//Revisit this for multi level coherence
|
|
|
|
return;
|
|
|
|
}
|
2006-10-20 02:02:57 +02:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
BlkType *blk = tags->findBlock(pkt->getAddr());
|
2006-10-20 02:02:57 +02:00
|
|
|
|
2009-09-26 19:50:50 +02:00
|
|
|
Addr blk_addr = blockAlign(pkt->getAddr());
|
2007-06-18 02:27:53 +02:00
|
|
|
MSHR *mshr = mshrQueue.findMatch(blk_addr);
|
2007-07-22 17:09:24 +02:00
|
|
|
|
|
|
|
// Let the MSHR itself track the snoop and decide whether we want
|
|
|
|
// to go ahead and do the regular cache snoop
|
2007-07-21 22:45:17 +02:00
|
|
|
if (mshr && mshr->handleSnoop(pkt, order++)) {
|
2007-06-27 07:23:10 +02:00
|
|
|
DPRINTF(Cache, "Deferring snoop on in-service MSHR to blk %x\n",
|
|
|
|
blk_addr);
|
2007-06-27 00:01:22 +02:00
|
|
|
if (mshr->getNumTargets() > numTarget)
|
2007-07-21 22:45:17 +02:00
|
|
|
warn("allocating bonus target for snoop"); //handle later
|
2007-06-25 02:32:31 +02:00
|
|
|
return;
|
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
|
|
|
|
//We also need to check the writeback buffers and handle those
|
|
|
|
std::vector<MSHR *> writebacks;
|
|
|
|
if (writeBuffer.findMatches(blk_addr, writebacks)) {
|
|
|
|
DPRINTF(Cache, "Snoop hit in writeback to addr: %x\n",
|
|
|
|
pkt->getAddr());
|
|
|
|
|
|
|
|
//Look through writebacks for any non-uncachable writes, use that
|
2012-05-11 01:04:27 +02:00
|
|
|
if (writebacks.size()) {
|
|
|
|
// We should only ever find a single match
|
|
|
|
assert(writebacks.size() == 1);
|
|
|
|
mshr = writebacks[0];
|
2007-06-18 02:27:53 +02:00
|
|
|
assert(!mshr->isUncacheable());
|
2007-06-25 02:32:31 +02:00
|
|
|
assert(mshr->getNumTargets() == 1);
|
|
|
|
PacketPtr wb_pkt = mshr->getTarget()->pkt;
|
|
|
|
assert(wb_pkt->cmd == MemCmd::Writeback);
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2007-07-27 02:04:17 +02:00
|
|
|
assert(!pkt->memInhibitAsserted());
|
|
|
|
pkt->assertMemInhibit();
|
|
|
|
if (!pkt->needsExclusive()) {
|
|
|
|
pkt->assertShared();
|
2010-09-22 08:07:34 +02:00
|
|
|
// the writeback is no longer the exclusive copy in the system
|
|
|
|
wb_pkt->clearSupplyExclusive();
|
2007-07-27 02:04:17 +02:00
|
|
|
} else {
|
|
|
|
// if we're not asserting the shared line, we need to
|
|
|
|
// invalidate our copy. we'll do that below as long as
|
|
|
|
// the packet's invalidate flag is set...
|
|
|
|
assert(pkt->isInvalidate());
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
2008-01-03 00:22:38 +01:00
|
|
|
doTimingSupplyResponse(pkt, wb_pkt->getPtr<uint8_t>(),
|
|
|
|
false, false);
|
2007-06-18 02:27:53 +02:00
|
|
|
|
|
|
|
if (pkt->isInvalidate()) {
|
|
|
|
// Invalidation trumps our writeback... discard here
|
|
|
|
markInService(mshr);
|
2008-03-17 08:08:28 +01:00
|
|
|
delete wb_pkt;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
2012-05-11 01:04:27 +02:00
|
|
|
} // writebacks.size()
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
2006-11-12 12:44:05 +01:00
|
|
|
|
2012-05-11 01:04:27 +02:00
|
|
|
// If this was a shared writeback, there may still be
|
|
|
|
// other shared copies above that require invalidation.
|
|
|
|
// We could be more selective and return here if the
|
|
|
|
// request is non-exclusive or if the writeback is
|
|
|
|
// exclusive.
|
2008-01-03 00:22:38 +01:00
|
|
|
handleSnoop(pkt, blk, true, false, false);
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
template<class TagStore>
|
|
|
|
bool
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
Cache<TagStore>::CpuSidePort::recvTimingSnoopResp(PacketPtr pkt)
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
{
|
|
|
|
// Express snoop responses from master to slave, e.g., from L1 to L2
|
|
|
|
cache->timingAccess(pkt);
|
|
|
|
return true;
|
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2012-10-15 14:10:54 +02:00
|
|
|
Cycles
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::snoopAtomic(PacketPtr pkt)
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2007-07-27 02:04:17 +02:00
|
|
|
if (pkt->req->isUncacheable() || pkt->cmd == MemCmd::Writeback) {
|
2007-06-18 02:27:53 +02:00
|
|
|
// Can't get a hit on an uncacheable address
|
|
|
|
// Revisit this for multi level coherence
|
|
|
|
return hitLatency;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
|
|
|
|
BlkType *blk = tags->findBlock(pkt->getAddr());
|
2008-01-03 00:22:38 +01:00
|
|
|
handleSnoop(pkt, blk, false, false, false);
|
2007-06-18 02:27:53 +02:00
|
|
|
return hitLatency;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
MSHR *
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::getNextMSHR()
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2007-06-18 02:27:53 +02:00
|
|
|
// Check both MSHR queue and write buffer for potential requests
|
|
|
|
MSHR *miss_mshr = mshrQueue.getNextMSHR();
|
|
|
|
MSHR *write_mshr = writeBuffer.getNextMSHR();
|
2006-06-28 17:02:14 +02:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// Now figure out which one to send... some cases are easy
|
|
|
|
if (miss_mshr && !write_mshr) {
|
|
|
|
return miss_mshr;
|
|
|
|
}
|
|
|
|
if (write_mshr && !miss_mshr) {
|
|
|
|
return write_mshr;
|
2006-10-20 01:00:27 +02:00
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
if (miss_mshr && write_mshr) {
|
|
|
|
// We have one of each... normally we favor the miss request
|
|
|
|
// unless the write buffer is full
|
|
|
|
if (writeBuffer.isFull() && writeBuffer.inServiceEntries == 0) {
|
|
|
|
// Write buffer is full, so we'd like to issue a write;
|
|
|
|
// need to search MSHR queue for conflicting earlier miss.
|
|
|
|
MSHR *conflict_mshr =
|
|
|
|
mshrQueue.findPending(write_mshr->addr, write_mshr->size);
|
|
|
|
|
|
|
|
if (conflict_mshr && conflict_mshr->order < write_mshr->order) {
|
|
|
|
// Service misses in order until conflict is cleared.
|
|
|
|
return conflict_mshr;
|
2006-10-20 01:00:27 +02:00
|
|
|
}
|
2006-10-06 05:28:03 +02:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// No conflicts; issue write
|
|
|
|
return write_mshr;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// Write buffer isn't full, but need to check it for
|
|
|
|
// conflicting earlier writeback
|
|
|
|
MSHR *conflict_mshr =
|
|
|
|
writeBuffer.findPending(miss_mshr->addr, miss_mshr->size);
|
|
|
|
if (conflict_mshr) {
|
|
|
|
// not sure why we don't check order here... it was in the
|
|
|
|
// original code but commented out.
|
|
|
|
|
|
|
|
// The only way this happens is if we are
|
|
|
|
// doing a write and we didn't have permissions
|
|
|
|
// then subsequently saw a writeback (owned got evicted)
|
|
|
|
// We need to make sure to perform the writeback first
|
|
|
|
// To preserve the dirty data, then we can issue the write
|
|
|
|
|
|
|
|
// should we return write_mshr here instead? I.e. do we
|
|
|
|
// have to flush writes in order? I don't think so... not
|
|
|
|
// for Alpha anyway. Maybe for x86?
|
|
|
|
return conflict_mshr;
|
2006-10-22 02:19:33 +02:00
|
|
|
}
|
2006-10-20 01:00:27 +02:00
|
|
|
|
2007-06-26 23:53:15 +02:00
|
|
|
// No conflicts; issue read
|
2007-06-18 02:27:53 +02:00
|
|
|
return miss_mshr;
|
|
|
|
}
|
|
|
|
|
|
|
|
// fall through... no pending requests. Try a prefetch.
|
|
|
|
assert(!miss_mshr && !write_mshr);
|
2010-06-23 06:29:43 +02:00
|
|
|
if (prefetcher && !mshrQueue.isFull()) {
|
2007-06-18 02:27:53 +02:00
|
|
|
// If we have a miss queue slot, we can try a prefetch
|
|
|
|
PacketPtr pkt = prefetcher->getPacket();
|
|
|
|
if (pkt) {
|
2009-09-26 19:50:50 +02:00
|
|
|
Addr pf_addr = blockAlign(pkt->getAddr());
|
2011-09-13 19:06:13 +02:00
|
|
|
if (!tags->findBlock(pf_addr) && !mshrQueue.findMatch(pf_addr) &&
|
|
|
|
!writeBuffer.findMatch(pf_addr)) {
|
2009-09-26 19:50:50 +02:00
|
|
|
// Update statistic on number of prefetches issued
|
|
|
|
// (hwpf_mshr_misses)
|
2012-02-12 23:07:39 +01:00
|
|
|
assert(pkt->req->masterId() < system->maxMasters());
|
|
|
|
mshr_misses[pkt->cmdToIndex()][pkt->req->masterId()]++;
|
2009-09-26 19:50:50 +02:00
|
|
|
// Don't request bus, since we already have it
|
2011-01-08 06:50:29 +01:00
|
|
|
return allocateMissBuffer(pkt, curTick(), false);
|
2011-08-19 22:08:05 +02:00
|
|
|
} else {
|
|
|
|
// free the request and packet
|
|
|
|
delete pkt->req;
|
|
|
|
delete pkt;
|
2009-09-26 19:50:50 +02:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
2006-10-22 02:19:33 +02:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
return NULL;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
PacketPtr
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::getTimingPacket()
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2007-06-18 02:27:53 +02:00
|
|
|
MSHR *mshr = getNextMSHR();
|
2006-10-20 02:02:57 +02:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
if (mshr == NULL) {
|
|
|
|
return NULL;
|
2006-10-20 02:02:57 +02:00
|
|
|
}
|
2006-06-28 17:02:14 +02:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
// use request from 1st target
|
2007-06-21 20:59:17 +02:00
|
|
|
PacketPtr tgt_pkt = mshr->getTarget()->pkt;
|
2007-06-22 18:24:07 +02:00
|
|
|
PacketPtr pkt = NULL;
|
2007-06-18 02:27:53 +02:00
|
|
|
|
2010-09-09 20:40:19 +02:00
|
|
|
if (tgt_pkt->cmd == MemCmd::SCUpgradeFailReq ||
|
|
|
|
tgt_pkt->cmd == MemCmd::StoreCondFailReq) {
|
|
|
|
// SCUpgradeReq or StoreCondReq saw invalidation while queued
|
|
|
|
// in MSHR, so now that we are getting around to processing
|
|
|
|
// it, just treat it as if we got a failure response
|
2010-06-17 00:25:57 +02:00
|
|
|
pkt = new Packet(tgt_pkt);
|
|
|
|
pkt->cmd = MemCmd::UpgradeFailResp;
|
2010-07-09 02:56:13 +02:00
|
|
|
pkt->senderState = mshr;
|
2011-01-08 06:50:29 +01:00
|
|
|
pkt->firstWordTime = pkt->finishTime = curTick();
|
2010-06-17 00:25:57 +02:00
|
|
|
handleResponse(pkt);
|
|
|
|
return NULL;
|
|
|
|
} else if (mshr->isForwardNoResponse()) {
|
2007-06-22 18:24:07 +02:00
|
|
|
// no response expected, just forward packet as it is
|
|
|
|
assert(tags->findBlock(mshr->addr) == NULL);
|
|
|
|
pkt = tgt_pkt;
|
|
|
|
} else {
|
|
|
|
BlkType *blk = tags->findBlock(mshr->addr);
|
2011-07-15 18:53:35 +02:00
|
|
|
|
|
|
|
if (tgt_pkt->cmd == MemCmd::HardPFReq) {
|
|
|
|
// It might be possible for a writeback to arrive between
|
|
|
|
// the time the prefetch is placed in the MSHRs and when
|
|
|
|
// it's selected to send... if so, this assert will catch
|
|
|
|
// that, and then we'll have to figure out what to do.
|
|
|
|
assert(blk == NULL);
|
|
|
|
|
|
|
|
// We need to check the caches above us to verify that they don't have
|
|
|
|
// a copy of this block in the dirty state at the moment. Without this
|
|
|
|
// check we could get a stale copy from memory that might get used
|
|
|
|
// in place of the dirty one.
|
|
|
|
PacketPtr snoop_pkt = new Packet(tgt_pkt, true);
|
|
|
|
snoop_pkt->setExpressSnoop();
|
|
|
|
snoop_pkt->senderState = mshr;
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
cpuSidePort->sendTimingSnoopReq(snoop_pkt);
|
2011-07-15 18:53:35 +02:00
|
|
|
|
|
|
|
if (snoop_pkt->memInhibitAsserted()) {
|
|
|
|
markInService(mshr, snoop_pkt);
|
|
|
|
DPRINTF(Cache, "Upward snoop of prefetch for addr %#x hit\n",
|
|
|
|
tgt_pkt->getAddr());
|
|
|
|
delete snoop_pkt;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
delete snoop_pkt;
|
|
|
|
}
|
|
|
|
|
2007-07-22 03:18:42 +02:00
|
|
|
pkt = getBusPacket(tgt_pkt, blk, mshr->needsExclusive());
|
2007-06-22 18:24:07 +02:00
|
|
|
|
2008-11-10 23:10:28 +01:00
|
|
|
mshr->isForward = (pkt == NULL);
|
2007-06-22 18:24:07 +02:00
|
|
|
|
2008-11-10 23:10:28 +01:00
|
|
|
if (mshr->isForward) {
|
2007-06-22 18:24:07 +02:00
|
|
|
// not a cache block request, but a response is expected
|
|
|
|
// make copy of current packet to forward, keep current
|
|
|
|
// copy for response handling
|
|
|
|
pkt = new Packet(tgt_pkt);
|
|
|
|
pkt->allocate();
|
|
|
|
if (pkt->isWrite()) {
|
|
|
|
pkt->setData(tgt_pkt->getPtr<uint8_t>());
|
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
2006-12-14 07:04:36 +01:00
|
|
|
}
|
2007-04-04 19:56:38 +02:00
|
|
|
|
2007-06-22 18:24:07 +02:00
|
|
|
assert(pkt != NULL);
|
2007-06-18 02:27:53 +02:00
|
|
|
pkt->senderState = mshr;
|
|
|
|
return pkt;
|
2007-03-09 00:57:15 +01:00
|
|
|
}
|
|
|
|
|
2006-12-14 07:04:36 +01:00
|
|
|
|
2009-02-16 17:56:40 +01:00
|
|
|
template<class TagStore>
|
|
|
|
Tick
|
|
|
|
Cache<TagStore>::nextMSHRReadyTime()
|
|
|
|
{
|
|
|
|
Tick nextReady = std::min(mshrQueue.nextMSHRReadyTime(),
|
|
|
|
writeBuffer.nextMSHRReadyTime());
|
|
|
|
|
|
|
|
if (prefetcher) {
|
|
|
|
nextReady = std::min(nextReady,
|
|
|
|
prefetcher->nextPrefetchReadyTime());
|
|
|
|
}
|
|
|
|
|
|
|
|
return nextReady;
|
|
|
|
}
|
|
|
|
|
2012-05-11 01:04:26 +02:00
|
|
|
template<class TagStore>
|
|
|
|
void
|
|
|
|
Cache<TagStore>::serialize(std::ostream &os)
|
|
|
|
{
|
2012-11-02 17:32:02 +01:00
|
|
|
bool dirty(isDirty());
|
2012-05-11 01:04:26 +02:00
|
|
|
|
2012-11-02 17:32:02 +01:00
|
|
|
if (dirty) {
|
|
|
|
warn("*** The cache still contains dirty data. ***\n");
|
|
|
|
warn(" Make sure to drain the system using the correct flags.\n");
|
|
|
|
warn(" This checkpoint will not restore correctly and dirty data in "
|
|
|
|
"the cache will be lost!\n");
|
|
|
|
}
|
2012-05-11 01:04:26 +02:00
|
|
|
|
2012-11-02 17:32:02 +01:00
|
|
|
// Since we don't checkpoint the data in the cache, any dirty data
|
|
|
|
// will be lost when restoring from a checkpoint of a system that
|
|
|
|
// wasn't drained properly. Flag the checkpoint as invalid if the
|
|
|
|
// cache contains dirty data.
|
|
|
|
bool bad_checkpoint(dirty);
|
2012-05-11 01:04:26 +02:00
|
|
|
SERIALIZE_SCALAR(bad_checkpoint);
|
|
|
|
}
|
|
|
|
|
|
|
|
template<class TagStore>
|
|
|
|
void
|
|
|
|
Cache<TagStore>::unserialize(Checkpoint *cp, const std::string §ion)
|
|
|
|
{
|
|
|
|
bool bad_checkpoint;
|
|
|
|
UNSERIALIZE_SCALAR(bad_checkpoint);
|
|
|
|
if (bad_checkpoint) {
|
2012-11-02 17:32:02 +01:00
|
|
|
fatal("Restoring from checkpoints with dirty caches is not supported "
|
|
|
|
"in the classic memory system. Please remove any caches or "
|
|
|
|
" drain them properly before taking checkpoints.\n");
|
2012-05-11 01:04:26 +02:00
|
|
|
}
|
|
|
|
}
|
2009-02-16 17:56:40 +01:00
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
///////////////
|
|
|
|
//
|
|
|
|
// CpuSidePort
|
|
|
|
//
|
|
|
|
///////////////
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2012-01-17 19:55:09 +01:00
|
|
|
AddrRangeList
|
2012-07-09 18:35:34 +02:00
|
|
|
Cache<TagStore>::CpuSidePort::getAddrRanges() const
|
2007-05-22 16:30:55 +02:00
|
|
|
{
|
2012-03-09 15:59:25 +01:00
|
|
|
return cache->getAddrRanges();
|
2007-05-22 16:30:55 +02:00
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-12-14 07:04:36 +01:00
|
|
|
bool
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
Cache<TagStore>::CpuSidePort::recvTimingReq(PacketPtr pkt)
|
2006-12-14 07:04:36 +01:00
|
|
|
{
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
// always let inhibited requests through even if blocked
|
|
|
|
if (!pkt->memInhibitAsserted() && blocked) {
|
2006-12-14 07:04:36 +01:00
|
|
|
DPRINTF(Cache,"Scheduling a retry while blocked\n");
|
|
|
|
mustSendRetry = true;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2012-02-24 17:52:49 +01:00
|
|
|
cache->timingAccess(pkt);
|
2006-12-14 07:04:36 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
Tick
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::CpuSidePort::recvAtomic(PacketPtr pkt)
|
2007-05-19 07:35:04 +02:00
|
|
|
{
|
2012-02-24 17:52:49 +01:00
|
|
|
// atomic request
|
|
|
|
return cache->atomicAccess(pkt);
|
2007-05-19 07:35:04 +02:00
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-05-19 07:35:04 +02:00
|
|
|
void
|
2007-06-28 05:54:13 +02:00
|
|
|
Cache<TagStore>::CpuSidePort::recvFunctional(PacketPtr pkt)
|
2007-05-19 07:35:04 +02:00
|
|
|
{
|
2012-02-24 17:52:49 +01:00
|
|
|
// functional request
|
|
|
|
cache->functionalAccess(pkt, true);
|
2007-05-19 07:35:04 +02:00
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
|
|
|
Cache<TagStore>::
|
2008-01-02 21:20:15 +01:00
|
|
|
CpuSidePort::CpuSidePort(const std::string &_name, Cache<TagStore> *_cache,
|
2008-07-16 20:10:33 +02:00
|
|
|
const std::string &_label)
|
2012-02-24 17:52:49 +01:00
|
|
|
: BaseCache::CacheSlavePort(_name, _cache, _label), cache(_cache)
|
2006-12-14 07:04:36 +01:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2007-06-18 02:27:53 +02:00
|
|
|
///////////////
|
|
|
|
//
|
|
|
|
// MemSidePort
|
|
|
|
//
|
|
|
|
///////////////
|
2006-12-14 07:04:36 +01:00
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2006-12-14 07:04:36 +01:00
|
|
|
bool
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
Cache<TagStore>::MemSidePort::recvTimingResp(PacketPtr pkt)
|
2006-12-14 07:04:36 +01:00
|
|
|
{
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
cache->handleResponse(pkt);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Express snooping requests to memside port
|
|
|
|
template<class TagStore>
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
void
|
|
|
|
Cache<TagStore>::MemSidePort::recvTimingSnoopReq(PacketPtr pkt)
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
{
|
|
|
|
// handle snooping requests
|
|
|
|
cache->snoopTiming(pkt);
|
2006-12-14 07:04:36 +01:00
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
Tick
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
Cache<TagStore>::MemSidePort::recvAtomicSnoop(PacketPtr pkt)
|
2007-05-19 07:35:04 +02:00
|
|
|
{
|
2012-02-24 17:52:49 +01:00
|
|
|
// atomic snoop
|
|
|
|
return cache->snoopAtomic(pkt);
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
2007-05-19 07:35:04 +02:00
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-06-18 02:27:53 +02:00
|
|
|
void
|
MEM: Separate snoops and normal memory requests/responses
This patch introduces port access methods that separates snoop
request/responses from normal memory request/responses. The
differentiation is made for functional, atomic and timing accesses and
builds on the introduction of master and slave ports.
Before the introduction of this patch, the packets belonging to the
different phases of the protocol (request -> [forwarded snoop request
-> snoop response]* -> response) all use the same port access
functions, even though the snoop packets flow in the opposite
direction to the normal packet. That is, a coherent master sends
normal request and receives responses, but receives snoop requests and
sends snoop responses (vice versa for the slave). These two distinct
phases now use different access functions, as described below.
Starting with the functional access, a master sends a request to a
slave through sendFunctional, and the request packet is turned into a
response before the call returns. In a system without cache coherence,
this is all that is needed from the functional interface. For the
cache-coherent scenario, a slave also sends snoop requests to coherent
masters through sendFunctionalSnoop, with responses returned within
the same packet pointer. This is currently used by the bus and caches,
and the LSQ of the O3 CPU. The send/recvFunctional and
send/recvFunctionalSnoop are moved from the Port super class to the
appropriate subclass.
Atomic accesses follow the same flow as functional accesses, with
request being sent from master to slave through sendAtomic. In the
case of cache-coherent ports, a slave can send snoop requests to a
master through sendAtomicSnoop. Just as for the functional access
methods, the atomic send and receive member functions are moved to the
appropriate subclasses.
The timing access methods are different from the functional and atomic
in that requests and responses are separated in time and
send/recvTiming are used for both directions. Hence, a master uses
sendTiming to send a request to a slave, and a slave uses sendTiming
to send a response back to a master, at a later point in time. Snoop
requests and responses travel in the opposite direction, similar to
what happens in functional and atomic accesses. With the introduction
of this patch, it is possible to determine the direction of packets in
the bus, and no longer necessary to look for both a master and a slave
port with the requested port id.
In contrast to the normal recvFunctional, recvAtomic and recvTiming
that are pure virtual functions, the recvFunctionalSnoop,
recvAtomicSnoop and recvTimingSnoop have a default implementation that
calls panic. This is to allow non-coherent master and slave ports to
not implement these functions.
2012-04-14 11:45:07 +02:00
|
|
|
Cache<TagStore>::MemSidePort::recvFunctionalSnoop(PacketPtr pkt)
|
2007-06-18 02:27:53 +02:00
|
|
|
{
|
2012-02-24 17:52:49 +01:00
|
|
|
// functional snoop (note that in contrast to atomic we don't have
|
|
|
|
// a specific functionalSnoop method, as they have the same
|
|
|
|
// behaviour regardless)
|
|
|
|
cache->functionalAccess(pkt, false);
|
2007-05-19 07:35:04 +02:00
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
2007-05-19 07:35:04 +02:00
|
|
|
void
|
2012-03-22 11:36:27 +01:00
|
|
|
Cache<TagStore>::MemSidePacketQueue::sendDeferredPacket()
|
2007-05-19 07:35:04 +02:00
|
|
|
{
|
2012-02-24 17:52:49 +01:00
|
|
|
// if we have a response packet waiting we have to start with that
|
2007-06-18 02:27:53 +02:00
|
|
|
if (deferredPacketReady()) {
|
2012-02-24 17:52:49 +01:00
|
|
|
// use the normal approach from the timing port
|
|
|
|
trySendTiming();
|
2007-06-18 02:27:53 +02:00
|
|
|
} else {
|
2012-02-24 17:52:49 +01:00
|
|
|
// check for request packets (requests & writebacks)
|
2012-03-22 11:36:27 +01:00
|
|
|
PacketPtr pkt = cache.getTimingPacket();
|
2007-06-26 07:23:29 +02:00
|
|
|
if (pkt == NULL) {
|
|
|
|
// can happen if e.g. we attempt a writeback and fail, but
|
|
|
|
// before the retry, the writeback is eliminated because
|
|
|
|
// we snoop another cache's ReadEx.
|
|
|
|
waitingOnRetry = false;
|
|
|
|
} else {
|
|
|
|
MSHR *mshr = dynamic_cast<MSHR*>(pkt->senderState);
|
2007-05-19 07:35:04 +02:00
|
|
|
|
MEM: Separate requests and responses for timing accesses
This patch moves send/recvTiming and send/recvTimingSnoop from the
Port base class to the MasterPort and SlavePort, and also splits them
into separate member functions for requests and responses:
send/recvTimingReq, send/recvTimingResp, and send/recvTimingSnoopReq,
send/recvTimingSnoopResp. A master port sends requests and receives
responses, and also receives snoop requests and sends snoop
responses. A slave port has the reciprocal behaviour as it receives
requests and sends responses, and sends snoop requests and receives
snoop responses.
For all MemObjects that have only master ports or slave ports (but not
both), e.g. a CPU, or a PIO device, this patch merely adds more
clarity to what kind of access is taking place. For example, a CPU
port used to call sendTiming, and will now call
sendTimingReq. Similarly, a response previously came back through
recvTiming, which is now recvTimingResp. For the modules that have
both master and slave ports, e.g. the bus, the behaviour was
previously relying on branches based on pkt->isRequest(), and this is
now replaced with a direct call to the apprioriate member function
depending on the type of access. Please note that send/recvRetry is
still shared by all the timing accessors and remains in the Port base
class for now (to maintain the current bus functionality and avoid
changing the statistics of all regressions).
The packet queue is split into a MasterPort and SlavePort version to
facilitate the use of the new timing accessors. All uses of the
PacketQueue are updated accordingly.
With this patch, the type of packet (request or response) is now well
defined for each type of access, and asserts on pkt->isRequest() and
pkt->isResponse() are now moved to the appropriate send member
functions. It is also worth noting that sendTimingSnoopReq no longer
returns a boolean, as the semantics do not alow snoop requests to be
rejected or stalled. All these assumptions are now excplicitly part of
the port interface itself.
2012-05-01 19:40:42 +02:00
|
|
|
waitingOnRetry = !masterPort.sendTimingReq(pkt);
|
2007-05-19 07:35:04 +02:00
|
|
|
|
2007-06-26 07:23:29 +02:00
|
|
|
if (waitingOnRetry) {
|
|
|
|
DPRINTF(CachePort, "now waiting on a retry\n");
|
2008-11-10 23:10:28 +01:00
|
|
|
if (!mshr->isForwardNoResponse()) {
|
2012-02-24 17:52:49 +01:00
|
|
|
// we are awaiting a retry, but we
|
|
|
|
// delete the packet and will be creating a new packet
|
|
|
|
// when we get the opportunity
|
2007-07-03 06:40:31 +02:00
|
|
|
delete pkt;
|
|
|
|
}
|
2012-02-24 17:52:49 +01:00
|
|
|
// note that we have now masked any requestBus and
|
|
|
|
// schedSendEvent (we will wait for a retry before
|
|
|
|
// doing anything), and this is so even if we do not
|
|
|
|
// care about this packet and might override it before
|
|
|
|
// it gets retried
|
2007-06-26 07:23:29 +02:00
|
|
|
} else {
|
2012-03-22 11:36:27 +01:00
|
|
|
cache.markInService(mshr, pkt);
|
2007-06-26 07:23:29 +02:00
|
|
|
}
|
2007-06-18 02:27:53 +02:00
|
|
|
}
|
2007-05-19 07:35:04 +02:00
|
|
|
}
|
|
|
|
|
2012-02-24 17:52:49 +01:00
|
|
|
// if we succeeded and are not waiting for a retry, schedule the
|
|
|
|
// next send, not only looking at the response transmit list, but
|
|
|
|
// also considering when the next MSHR is ready
|
2007-06-18 02:27:53 +02:00
|
|
|
if (!waitingOnRetry) {
|
2012-03-22 11:36:27 +01:00
|
|
|
scheduleSend(cache.nextMSHRReadyTime());
|
2007-05-19 07:35:04 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-06-28 05:54:13 +02:00
|
|
|
template<class TagStore>
|
|
|
|
Cache<TagStore>::
|
2007-08-10 22:14:01 +02:00
|
|
|
MemSidePort::MemSidePort(const std::string &_name, Cache<TagStore> *_cache,
|
2008-07-16 20:10:33 +02:00
|
|
|
const std::string &_label)
|
2012-03-22 11:36:27 +01:00
|
|
|
: BaseCache::CacheMasterPort(_name, _cache, _queue),
|
|
|
|
_queue(*_cache, *this, _label), cache(_cache)
|
2006-12-14 07:04:36 +01:00
|
|
|
{
|
|
|
|
}
|