2006-06-28 17:02:14 +02:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2003-2005 The Regents of The University of Michigan
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions are
|
|
|
|
* met: redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer;
|
|
|
|
* redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution;
|
|
|
|
* neither the name of the copyright holders nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived from
|
|
|
|
* this software without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*
|
|
|
|
* Authors: Erik Hallnor
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
* @file
|
|
|
|
* Definition of BaseCache functions.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "mem/cache/base_cache.hh"
|
2006-10-18 06:16:17 +02:00
|
|
|
#include "mem/cache/miss/mshr.hh"
|
|
|
|
#include "mem/packet_impl.hh"
|
2006-06-28 17:02:14 +02:00
|
|
|
#include "cpu/smt.hh"
|
|
|
|
#include "cpu/base.hh"
|
|
|
|
|
|
|
|
using namespace std;
|
|
|
|
|
|
|
|
BaseCache::CachePort::CachePort(const std::string &_name, BaseCache *_cache,
|
|
|
|
bool _isCpuSide)
|
|
|
|
: Port(_name), cache(_cache), isCpuSide(_isCpuSide)
|
|
|
|
{
|
|
|
|
blocked = false;
|
2006-10-10 23:10:56 +02:00
|
|
|
waitingOnRetry = false;
|
2006-06-28 17:02:14 +02:00
|
|
|
//Start ports at null if more than one is created we should panic
|
2006-06-28 20:35:00 +02:00
|
|
|
//cpuSidePort = NULL;
|
|
|
|
//memSidePort = NULL;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
2006-06-28 20:35:00 +02:00
|
|
|
void
|
2006-06-28 17:02:14 +02:00
|
|
|
BaseCache::CachePort::recvStatusChange(Port::Status status)
|
|
|
|
{
|
|
|
|
cache->recvStatusChange(status, isCpuSide);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
BaseCache::CachePort::getDeviceAddressRanges(AddrRangeList &resp,
|
|
|
|
AddrRangeList &snoop)
|
|
|
|
{
|
2006-07-07 22:02:22 +02:00
|
|
|
cache->getAddressRanges(resp, snoop, isCpuSide);
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
BaseCache::CachePort::deviceBlockSize()
|
|
|
|
{
|
|
|
|
return cache->getBlockSize();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool
|
|
|
|
BaseCache::CachePort::recvTiming(Packet *pkt)
|
|
|
|
{
|
2006-10-10 21:53:25 +02:00
|
|
|
if (isCpuSide
|
|
|
|
&& !pkt->req->isUncacheable()
|
|
|
|
&& pkt->isInvalidate()
|
|
|
|
&& !pkt->isRead() && !pkt->isWrite()) {
|
|
|
|
//Upgrade or Invalidate
|
|
|
|
//Look into what happens if two slave caches on bus
|
|
|
|
DPRINTF(Cache, "%s %x ? blk_addr: %x\n", pkt->cmdString(),
|
|
|
|
pkt->getAddr() & (((ULL(1))<<48)-1),
|
|
|
|
pkt->getAddr() & ~((Addr)cache->blkSize - 1));
|
|
|
|
|
|
|
|
assert(!(pkt->flags & SATISFIED));
|
|
|
|
pkt->flags |= SATISFIED;
|
|
|
|
//Invalidates/Upgrades need no response if they get the bus
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2006-10-09 06:27:03 +02:00
|
|
|
if (pkt->isRequest() && blocked)
|
2006-08-15 20:24:49 +02:00
|
|
|
{
|
2006-08-16 21:54:02 +02:00
|
|
|
DPRINTF(Cache,"Scheduling a retry while blocked\n");
|
2006-08-15 20:24:49 +02:00
|
|
|
mustSendRetry = true;
|
|
|
|
return false;
|
|
|
|
}
|
2006-06-28 17:02:14 +02:00
|
|
|
return cache->doTimingAccess(pkt, this, isCpuSide);
|
|
|
|
}
|
|
|
|
|
|
|
|
Tick
|
|
|
|
BaseCache::CachePort::recvAtomic(Packet *pkt)
|
|
|
|
{
|
|
|
|
return cache->doAtomicAccess(pkt, isCpuSide);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
BaseCache::CachePort::recvFunctional(Packet *pkt)
|
|
|
|
{
|
2006-10-12 19:59:03 +02:00
|
|
|
//Check storage here first
|
|
|
|
list<Packet *>::iterator i = drainList.begin();
|
|
|
|
list<Packet *>::iterator end = drainList.end();
|
|
|
|
for (; i != end; ++i) {
|
|
|
|
Packet * target = *i;
|
|
|
|
// If the target contains data, and it overlaps the
|
|
|
|
// probed request, need to update data
|
|
|
|
if (target->intersect(pkt)) {
|
|
|
|
uint8_t* pkt_data;
|
|
|
|
uint8_t* write_data;
|
|
|
|
int data_size;
|
|
|
|
if (target->getAddr() < pkt->getAddr()) {
|
|
|
|
int offset = pkt->getAddr() - target->getAddr();
|
|
|
|
pkt_data = pkt->getPtr<uint8_t>();
|
|
|
|
write_data = target->getPtr<uint8_t>() + offset;
|
|
|
|
data_size = target->getSize() - offset;
|
|
|
|
assert(data_size > 0);
|
|
|
|
if (data_size > pkt->getSize())
|
|
|
|
data_size = pkt->getSize();
|
|
|
|
} else {
|
|
|
|
int offset = target->getAddr() - pkt->getAddr();
|
|
|
|
pkt_data = pkt->getPtr<uint8_t>() + offset;
|
|
|
|
write_data = target->getPtr<uint8_t>();
|
|
|
|
data_size = pkt->getSize() - offset;
|
2006-10-20 03:26:46 +02:00
|
|
|
assert(data_size >= pkt->getSize());
|
2006-10-12 19:59:03 +02:00
|
|
|
if (data_size > target->getSize())
|
|
|
|
data_size = target->getSize();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (pkt->isWrite()) {
|
|
|
|
memcpy(pkt_data, write_data, data_size);
|
|
|
|
} else {
|
|
|
|
memcpy(write_data, pkt_data, data_size);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2006-06-28 17:02:14 +02:00
|
|
|
cache->doFunctionalAccess(pkt, isCpuSide);
|
|
|
|
}
|
|
|
|
|
2006-08-16 21:54:02 +02:00
|
|
|
void
|
|
|
|
BaseCache::CachePort::recvRetry()
|
|
|
|
{
|
|
|
|
Packet *pkt;
|
2006-10-10 23:10:56 +02:00
|
|
|
assert(waitingOnRetry);
|
2006-10-07 18:02:59 +02:00
|
|
|
if (!drainList.empty()) {
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s attempting to send a retry for response\n", name());
|
2006-10-07 18:02:59 +02:00
|
|
|
//We have some responses to drain first
|
2006-10-10 23:10:56 +02:00
|
|
|
if (sendTiming(drainList.front())) {
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s sucessful in sending a retry for response\n", name());
|
2006-10-10 23:10:56 +02:00
|
|
|
drainList.pop_front();
|
|
|
|
if (!drainList.empty() ||
|
|
|
|
!isCpuSide && cache->doMasterRequest() ||
|
|
|
|
isCpuSide && cache->doSlaveRequest()) {
|
2006-10-11 04:50:36 +02:00
|
|
|
|
|
|
|
DPRINTF(CachePort, "%s has more responses/requests\n", name());
|
2006-10-10 23:10:56 +02:00
|
|
|
BaseCache::CacheEvent * reqCpu = new BaseCache::CacheEvent(this);
|
|
|
|
reqCpu->schedule(curTick + 1);
|
|
|
|
}
|
|
|
|
waitingOnRetry = false;
|
2006-10-07 18:02:59 +02:00
|
|
|
}
|
|
|
|
}
|
2006-10-09 00:45:21 +02:00
|
|
|
else if (!isCpuSide)
|
2006-08-16 21:54:02 +02:00
|
|
|
{
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s attempting to send a retry for MSHR\n", name());
|
2006-10-11 06:19:31 +02:00
|
|
|
if (!cache->doMasterRequest()) {
|
2006-10-11 06:13:53 +02:00
|
|
|
//This can happen if I am the owner of a block and see an upgrade
|
|
|
|
//while the block was in my WB Buffers. I just remove the
|
|
|
|
//wb and de-assert the masterRequest
|
2006-10-11 07:01:40 +02:00
|
|
|
waitingOnRetry = false;
|
2006-10-11 06:13:53 +02:00
|
|
|
return;
|
|
|
|
}
|
2006-08-16 21:54:02 +02:00
|
|
|
pkt = cache->getPacket();
|
2006-10-17 22:47:22 +02:00
|
|
|
MSHR* mshr = (MSHR*) pkt->senderState;
|
2006-10-17 21:05:21 +02:00
|
|
|
//Copy the packet, it may be modified/destroyed elsewhere
|
|
|
|
Packet * copyPkt = new Packet(*pkt);
|
|
|
|
copyPkt->dataStatic<uint8_t>(pkt->getPtr<uint8_t>());
|
|
|
|
mshr->pkt = copyPkt;
|
2006-10-17 22:47:22 +02:00
|
|
|
|
2006-08-16 21:54:02 +02:00
|
|
|
bool success = sendTiming(pkt);
|
|
|
|
DPRINTF(Cache, "Address %x was %s in sending the timing request\n",
|
|
|
|
pkt->getAddr(), success ? "succesful" : "unsuccesful");
|
2006-10-17 22:47:22 +02:00
|
|
|
|
2006-10-10 23:10:56 +02:00
|
|
|
waitingOnRetry = !success;
|
2006-10-17 22:47:22 +02:00
|
|
|
if (waitingOnRetry) {
|
|
|
|
DPRINTF(CachePort, "%s now waiting on a retry\n", name());
|
|
|
|
}
|
|
|
|
|
|
|
|
cache->sendResult(pkt, mshr, success);
|
|
|
|
|
2006-08-16 21:54:02 +02:00
|
|
|
if (success && cache->doMasterRequest())
|
|
|
|
{
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s has more requests\n", name());
|
2006-08-16 21:54:02 +02:00
|
|
|
//Still more to issue, rerequest in 1 cycle
|
|
|
|
BaseCache::CacheEvent * reqCpu = new BaseCache::CacheEvent(this);
|
|
|
|
reqCpu->schedule(curTick + 1);
|
|
|
|
}
|
|
|
|
}
|
2006-10-10 23:10:56 +02:00
|
|
|
else
|
2006-08-16 21:54:02 +02:00
|
|
|
{
|
2006-10-13 21:47:05 +02:00
|
|
|
assert(cache->doSlaveRequest());
|
2006-10-09 22:37:02 +02:00
|
|
|
//pkt = cache->getCoherencePacket();
|
|
|
|
//We save the packet, no reordering on CSHRS
|
2006-10-13 21:47:05 +02:00
|
|
|
pkt = cache->getCoherencePacket();
|
|
|
|
MSHR* cshr = (MSHR*)pkt->senderState;
|
2006-08-16 21:54:02 +02:00
|
|
|
bool success = sendTiming(pkt);
|
2006-10-13 21:47:05 +02:00
|
|
|
cache->sendCoherenceResult(pkt, cshr, success);
|
2006-10-10 23:10:56 +02:00
|
|
|
waitingOnRetry = !success;
|
2006-10-13 21:47:05 +02:00
|
|
|
if (success && cache->doSlaveRequest())
|
2006-08-16 21:54:02 +02:00
|
|
|
{
|
2006-10-13 21:47:05 +02:00
|
|
|
DPRINTF(CachePort, "%s has more requests\n", name());
|
|
|
|
//Still more to issue, rerequest in 1 cycle
|
|
|
|
BaseCache::CacheEvent * reqCpu = new BaseCache::CacheEvent(this);
|
|
|
|
reqCpu->schedule(curTick + 1);
|
2006-08-16 21:54:02 +02:00
|
|
|
}
|
|
|
|
}
|
2006-10-11 04:50:36 +02:00
|
|
|
if (waitingOnRetry) DPRINTF(CachePort, "%s STILL Waiting on retry\n", name());
|
|
|
|
else DPRINTF(CachePort, "%s no longer waiting on retry\n", name());
|
2006-08-16 21:54:02 +02:00
|
|
|
return;
|
|
|
|
}
|
2006-06-28 17:02:14 +02:00
|
|
|
void
|
|
|
|
BaseCache::CachePort::setBlocked()
|
|
|
|
{
|
2006-08-16 21:54:02 +02:00
|
|
|
assert(!blocked);
|
|
|
|
DPRINTF(Cache, "Cache Blocking\n");
|
2006-06-28 17:02:14 +02:00
|
|
|
blocked = true;
|
2006-08-16 21:54:02 +02:00
|
|
|
//Clear the retry flag
|
|
|
|
mustSendRetry = false;
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
BaseCache::CachePort::clearBlocked()
|
|
|
|
{
|
2006-08-16 21:54:02 +02:00
|
|
|
assert(blocked);
|
|
|
|
DPRINTF(Cache, "Cache Unblocking\n");
|
|
|
|
blocked = false;
|
2006-08-15 20:24:49 +02:00
|
|
|
if (mustSendRetry)
|
|
|
|
{
|
2006-08-16 21:54:02 +02:00
|
|
|
DPRINTF(Cache, "Cache Sending Retry\n");
|
2006-08-15 20:24:49 +02:00
|
|
|
mustSendRetry = false;
|
|
|
|
sendRetry();
|
|
|
|
}
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|
|
|
|
|
2006-07-06 21:15:37 +02:00
|
|
|
BaseCache::CacheEvent::CacheEvent(CachePort *_cachePort)
|
|
|
|
: Event(&mainEventQueue, CPU_Tick_Pri), cachePort(_cachePort)
|
|
|
|
{
|
|
|
|
this->setFlags(AutoDelete);
|
|
|
|
pkt = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
BaseCache::CacheEvent::CacheEvent(CachePort *_cachePort, Packet *_pkt)
|
|
|
|
: Event(&mainEventQueue, CPU_Tick_Pri), cachePort(_cachePort), pkt(_pkt)
|
|
|
|
{
|
|
|
|
this->setFlags(AutoDelete);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
BaseCache::CacheEvent::process()
|
|
|
|
{
|
|
|
|
if (!pkt)
|
|
|
|
{
|
2006-10-10 23:10:56 +02:00
|
|
|
if (cachePort->waitingOnRetry) return;
|
|
|
|
//We have some responses to drain first
|
|
|
|
if (!cachePort->drainList.empty()) {
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s trying to drain a response\n", cachePort->name());
|
2006-10-10 23:10:56 +02:00
|
|
|
if (cachePort->sendTiming(cachePort->drainList.front())) {
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s drains a response succesfully\n", cachePort->name());
|
2006-10-10 23:10:56 +02:00
|
|
|
cachePort->drainList.pop_front();
|
|
|
|
if (!cachePort->drainList.empty() ||
|
|
|
|
!cachePort->isCpuSide && cachePort->cache->doMasterRequest() ||
|
2006-10-11 04:50:36 +02:00
|
|
|
cachePort->isCpuSide && cachePort->cache->doSlaveRequest()) {
|
|
|
|
|
|
|
|
DPRINTF(CachePort, "%s still has outstanding bus reqs\n", cachePort->name());
|
2006-10-10 23:10:56 +02:00
|
|
|
this->schedule(curTick + 1);
|
2006-10-11 04:50:36 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
cachePort->waitingOnRetry = true;
|
|
|
|
DPRINTF(CachePort, "%s now waiting on a retry\n", cachePort->name());
|
2006-10-10 23:10:56 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (!cachePort->isCpuSide)
|
2006-10-11 06:13:53 +02:00
|
|
|
{ //MSHR
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s trying to send a MSHR request\n", cachePort->name());
|
2006-10-11 06:19:31 +02:00
|
|
|
if (!cachePort->cache->doMasterRequest()) {
|
2006-10-11 06:13:53 +02:00
|
|
|
//This can happen if I am the owner of a block and see an upgrade
|
|
|
|
//while the block was in my WB Buffers. I just remove the
|
|
|
|
//wb and de-assert the masterRequest
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2006-07-06 21:15:37 +02:00
|
|
|
pkt = cachePort->cache->getPacket();
|
2006-10-09 22:37:02 +02:00
|
|
|
MSHR* mshr = (MSHR*) pkt->senderState;
|
2006-10-17 21:05:21 +02:00
|
|
|
//Copy the packet, it may be modified/destroyed elsewhere
|
|
|
|
Packet * copyPkt = new Packet(*pkt);
|
|
|
|
copyPkt->dataStatic<uint8_t>(pkt->getPtr<uint8_t>());
|
|
|
|
mshr->pkt = copyPkt;
|
|
|
|
|
2006-07-10 23:16:15 +02:00
|
|
|
bool success = cachePort->sendTiming(pkt);
|
|
|
|
DPRINTF(Cache, "Address %x was %s in sending the timing request\n",
|
|
|
|
pkt->getAddr(), success ? "succesful" : "unsuccesful");
|
2006-10-17 22:47:22 +02:00
|
|
|
|
2006-10-10 23:10:56 +02:00
|
|
|
cachePort->waitingOnRetry = !success;
|
2006-10-17 22:47:22 +02:00
|
|
|
if (cachePort->waitingOnRetry) {
|
2006-10-13 21:47:05 +02:00
|
|
|
DPRINTF(CachePort, "%s now waiting on a retry\n", cachePort->name());
|
2006-10-17 22:47:22 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
cachePort->cache->sendResult(pkt, mshr, success);
|
2006-07-10 23:16:15 +02:00
|
|
|
if (success && cachePort->cache->doMasterRequest())
|
|
|
|
{
|
2006-10-13 21:47:05 +02:00
|
|
|
DPRINTF(CachePort, "%s still more MSHR requests to send\n",
|
|
|
|
cachePort->name());
|
2006-07-10 23:16:15 +02:00
|
|
|
//Still more to issue, rerequest in 1 cycle
|
|
|
|
pkt = NULL;
|
|
|
|
this->schedule(curTick+1);
|
|
|
|
}
|
|
|
|
}
|
2006-07-06 22:52:05 +02:00
|
|
|
else
|
2006-07-10 23:16:15 +02:00
|
|
|
{
|
2006-08-16 21:54:02 +02:00
|
|
|
//CSHR
|
2006-10-13 21:47:05 +02:00
|
|
|
assert(cachePort->cache->doSlaveRequest());
|
|
|
|
pkt = cachePort->cache->getCoherencePacket();
|
|
|
|
MSHR* cshr = (MSHR*) pkt->senderState;
|
2006-08-16 21:54:02 +02:00
|
|
|
bool success = cachePort->sendTiming(pkt);
|
2006-10-18 00:50:19 +02:00
|
|
|
cachePort->cache->sendCoherenceResult(pkt, cshr, success);
|
2006-10-13 21:47:05 +02:00
|
|
|
cachePort->waitingOnRetry = !success;
|
|
|
|
if (cachePort->waitingOnRetry)
|
|
|
|
DPRINTF(CachePort, "%s now waiting on a retry\n", cachePort->name());
|
|
|
|
if (success && cachePort->cache->doSlaveRequest())
|
2006-08-16 21:54:02 +02:00
|
|
|
{
|
2006-10-13 21:47:05 +02:00
|
|
|
DPRINTF(CachePort, "%s still more CSHR requests to send\n",
|
|
|
|
cachePort->name());
|
|
|
|
//Still more to issue, rerequest in 1 cycle
|
|
|
|
pkt = NULL;
|
|
|
|
this->schedule(curTick+1);
|
2006-08-16 21:54:02 +02:00
|
|
|
}
|
2006-07-10 23:16:15 +02:00
|
|
|
}
|
2006-07-06 22:52:05 +02:00
|
|
|
return;
|
2006-07-06 21:15:37 +02:00
|
|
|
}
|
2006-08-16 21:54:02 +02:00
|
|
|
//Response
|
2006-10-06 03:10:03 +02:00
|
|
|
//Know the packet to send
|
2006-10-10 00:52:20 +02:00
|
|
|
if (pkt->flags & NACKED_LINE)
|
|
|
|
pkt->result = Packet::Nacked;
|
|
|
|
else
|
|
|
|
pkt->result = Packet::Success;
|
2006-10-06 03:10:03 +02:00
|
|
|
pkt->makeTimingResponse();
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s attempting to send a response\n", cachePort->name());
|
2006-10-11 05:53:10 +02:00
|
|
|
if (!cachePort->drainList.empty() || cachePort->waitingOnRetry) {
|
2006-10-10 23:10:56 +02:00
|
|
|
//Already have a list, just append
|
2006-10-07 18:20:29 +02:00
|
|
|
cachePort->drainList.push_back(pkt);
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s appending response onto drain list\n", cachePort->name());
|
2006-10-07 18:20:29 +02:00
|
|
|
}
|
|
|
|
else if (!cachePort->sendTiming(pkt)) {
|
2006-10-07 18:02:59 +02:00
|
|
|
//It failed, save it to list of drain events
|
2006-10-11 04:50:36 +02:00
|
|
|
DPRINTF(CachePort, "%s now waiting for a retry\n", cachePort->name());
|
2006-10-07 18:02:59 +02:00
|
|
|
cachePort->drainList.push_back(pkt);
|
2006-10-10 23:10:56 +02:00
|
|
|
cachePort->waitingOnRetry = true;
|
2006-10-07 18:02:59 +02:00
|
|
|
}
|
2006-07-06 21:15:37 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
const char *
|
|
|
|
BaseCache::CacheEvent::description()
|
|
|
|
{
|
|
|
|
return "timing event\n";
|
|
|
|
}
|
|
|
|
|
2006-06-28 17:02:14 +02:00
|
|
|
Port*
|
2006-06-30 17:34:27 +02:00
|
|
|
BaseCache::getPort(const std::string &if_name, int idx)
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2006-06-30 22:25:35 +02:00
|
|
|
if (if_name == "")
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2006-06-30 22:25:35 +02:00
|
|
|
if(cpuSidePort == NULL)
|
|
|
|
cpuSidePort = new CachePort(name() + "-cpu_side_port", this, true);
|
2006-06-28 17:02:14 +02:00
|
|
|
return cpuSidePort;
|
|
|
|
}
|
2006-07-07 21:15:11 +02:00
|
|
|
else if (if_name == "functional")
|
|
|
|
{
|
|
|
|
if(cpuSidePort == NULL)
|
|
|
|
cpuSidePort = new CachePort(name() + "-cpu_side_port", this, true);
|
|
|
|
return cpuSidePort;
|
|
|
|
}
|
|
|
|
else if (if_name == "cpu_side")
|
2006-06-28 17:02:14 +02:00
|
|
|
{
|
2006-06-30 22:25:35 +02:00
|
|
|
if(cpuSidePort == NULL)
|
|
|
|
cpuSidePort = new CachePort(name() + "-cpu_side_port", this, true);
|
|
|
|
return cpuSidePort;
|
|
|
|
}
|
|
|
|
else if (if_name == "mem_side")
|
|
|
|
{
|
|
|
|
if (memSidePort != NULL)
|
2006-06-28 17:02:14 +02:00
|
|
|
panic("Already have a mem side for this cache\n");
|
|
|
|
memSidePort = new CachePort(name() + "-mem_side_port", this, false);
|
|
|
|
return memSidePort;
|
|
|
|
}
|
|
|
|
else panic("Port name %s unrecognized\n", if_name);
|
|
|
|
}
|
|
|
|
|
2006-07-07 22:02:22 +02:00
|
|
|
void
|
|
|
|
BaseCache::init()
|
|
|
|
{
|
|
|
|
if (!cpuSidePort || !memSidePort)
|
|
|
|
panic("Cache not hooked up on both sides\n");
|
|
|
|
cpuSidePort->sendStatusChange(Port::RangeChange);
|
|
|
|
}
|
|
|
|
|
2006-06-28 17:02:14 +02:00
|
|
|
void
|
|
|
|
BaseCache::regStats()
|
|
|
|
{
|
2006-06-30 22:25:35 +02:00
|
|
|
Request temp_req((Addr) NULL, 4, 0);
|
2006-06-28 20:35:00 +02:00
|
|
|
Packet::Command temp_cmd = Packet::ReadReq;
|
|
|
|
Packet temp_pkt(&temp_req, temp_cmd, 0); //@todo FIx command strings so this isn't neccessary
|
2006-06-30 22:25:35 +02:00
|
|
|
temp_pkt.allocate(); //Temp allocate, all need data
|
2006-06-28 20:35:00 +02:00
|
|
|
|
2006-06-28 17:02:14 +02:00
|
|
|
using namespace Stats;
|
|
|
|
|
|
|
|
// Hit statistics
|
|
|
|
for (int access_idx = 0; access_idx < NUM_MEM_CMDS; ++access_idx) {
|
2006-06-28 20:35:00 +02:00
|
|
|
Packet::Command cmd = (Packet::Command)access_idx;
|
|
|
|
const string &cstr = temp_pkt.cmdIdxToString(cmd);
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
hits[access_idx]
|
|
|
|
.init(maxThreadsPerCPU)
|
|
|
|
.name(name() + "." + cstr + "_hits")
|
|
|
|
.desc("number of " + cstr + " hits")
|
|
|
|
.flags(total | nozero | nonan)
|
|
|
|
;
|
|
|
|
}
|
|
|
|
|
|
|
|
demandHits
|
|
|
|
.name(name() + ".demand_hits")
|
|
|
|
.desc("number of demand (read+write) hits")
|
|
|
|
.flags(total)
|
|
|
|
;
|
2006-06-28 20:35:00 +02:00
|
|
|
demandHits = hits[Packet::ReadReq] + hits[Packet::WriteReq];
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
overallHits
|
|
|
|
.name(name() + ".overall_hits")
|
|
|
|
.desc("number of overall hits")
|
|
|
|
.flags(total)
|
|
|
|
;
|
2006-06-28 20:35:00 +02:00
|
|
|
overallHits = demandHits + hits[Packet::SoftPFReq] + hits[Packet::HardPFReq]
|
|
|
|
+ hits[Packet::Writeback];
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
// Miss statistics
|
|
|
|
for (int access_idx = 0; access_idx < NUM_MEM_CMDS; ++access_idx) {
|
2006-06-28 20:35:00 +02:00
|
|
|
Packet::Command cmd = (Packet::Command)access_idx;
|
|
|
|
const string &cstr = temp_pkt.cmdIdxToString(cmd);
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
misses[access_idx]
|
|
|
|
.init(maxThreadsPerCPU)
|
|
|
|
.name(name() + "." + cstr + "_misses")
|
|
|
|
.desc("number of " + cstr + " misses")
|
|
|
|
.flags(total | nozero | nonan)
|
|
|
|
;
|
|
|
|
}
|
|
|
|
|
|
|
|
demandMisses
|
|
|
|
.name(name() + ".demand_misses")
|
|
|
|
.desc("number of demand (read+write) misses")
|
|
|
|
.flags(total)
|
|
|
|
;
|
2006-06-28 20:35:00 +02:00
|
|
|
demandMisses = misses[Packet::ReadReq] + misses[Packet::WriteReq];
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
overallMisses
|
|
|
|
.name(name() + ".overall_misses")
|
|
|
|
.desc("number of overall misses")
|
|
|
|
.flags(total)
|
|
|
|
;
|
2006-06-28 20:35:00 +02:00
|
|
|
overallMisses = demandMisses + misses[Packet::SoftPFReq] +
|
|
|
|
misses[Packet::HardPFReq] + misses[Packet::Writeback];
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
// Miss latency statistics
|
|
|
|
for (int access_idx = 0; access_idx < NUM_MEM_CMDS; ++access_idx) {
|
2006-06-28 20:35:00 +02:00
|
|
|
Packet::Command cmd = (Packet::Command)access_idx;
|
|
|
|
const string &cstr = temp_pkt.cmdIdxToString(cmd);
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
missLatency[access_idx]
|
|
|
|
.init(maxThreadsPerCPU)
|
|
|
|
.name(name() + "." + cstr + "_miss_latency")
|
|
|
|
.desc("number of " + cstr + " miss cycles")
|
|
|
|
.flags(total | nozero | nonan)
|
|
|
|
;
|
|
|
|
}
|
|
|
|
|
|
|
|
demandMissLatency
|
|
|
|
.name(name() + ".demand_miss_latency")
|
|
|
|
.desc("number of demand (read+write) miss cycles")
|
|
|
|
.flags(total)
|
|
|
|
;
|
2006-06-28 20:35:00 +02:00
|
|
|
demandMissLatency = missLatency[Packet::ReadReq] + missLatency[Packet::WriteReq];
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
overallMissLatency
|
|
|
|
.name(name() + ".overall_miss_latency")
|
|
|
|
.desc("number of overall miss cycles")
|
|
|
|
.flags(total)
|
|
|
|
;
|
2006-06-28 20:35:00 +02:00
|
|
|
overallMissLatency = demandMissLatency + missLatency[Packet::SoftPFReq] +
|
|
|
|
missLatency[Packet::HardPFReq];
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
// access formulas
|
|
|
|
for (int access_idx = 0; access_idx < NUM_MEM_CMDS; ++access_idx) {
|
2006-06-28 20:35:00 +02:00
|
|
|
Packet::Command cmd = (Packet::Command)access_idx;
|
|
|
|
const string &cstr = temp_pkt.cmdIdxToString(cmd);
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
accesses[access_idx]
|
|
|
|
.name(name() + "." + cstr + "_accesses")
|
|
|
|
.desc("number of " + cstr + " accesses(hits+misses)")
|
|
|
|
.flags(total | nozero | nonan)
|
|
|
|
;
|
|
|
|
|
|
|
|
accesses[access_idx] = hits[access_idx] + misses[access_idx];
|
|
|
|
}
|
|
|
|
|
|
|
|
demandAccesses
|
|
|
|
.name(name() + ".demand_accesses")
|
|
|
|
.desc("number of demand (read+write) accesses")
|
|
|
|
.flags(total)
|
|
|
|
;
|
|
|
|
demandAccesses = demandHits + demandMisses;
|
|
|
|
|
|
|
|
overallAccesses
|
|
|
|
.name(name() + ".overall_accesses")
|
|
|
|
.desc("number of overall (read+write) accesses")
|
|
|
|
.flags(total)
|
|
|
|
;
|
|
|
|
overallAccesses = overallHits + overallMisses;
|
|
|
|
|
|
|
|
// miss rate formulas
|
|
|
|
for (int access_idx = 0; access_idx < NUM_MEM_CMDS; ++access_idx) {
|
2006-06-28 20:35:00 +02:00
|
|
|
Packet::Command cmd = (Packet::Command)access_idx;
|
|
|
|
const string &cstr = temp_pkt.cmdIdxToString(cmd);
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
missRate[access_idx]
|
|
|
|
.name(name() + "." + cstr + "_miss_rate")
|
|
|
|
.desc("miss rate for " + cstr + " accesses")
|
|
|
|
.flags(total | nozero | nonan)
|
|
|
|
;
|
|
|
|
|
|
|
|
missRate[access_idx] = misses[access_idx] / accesses[access_idx];
|
|
|
|
}
|
|
|
|
|
|
|
|
demandMissRate
|
|
|
|
.name(name() + ".demand_miss_rate")
|
|
|
|
.desc("miss rate for demand accesses")
|
|
|
|
.flags(total)
|
|
|
|
;
|
|
|
|
demandMissRate = demandMisses / demandAccesses;
|
|
|
|
|
|
|
|
overallMissRate
|
|
|
|
.name(name() + ".overall_miss_rate")
|
|
|
|
.desc("miss rate for overall accesses")
|
|
|
|
.flags(total)
|
|
|
|
;
|
|
|
|
overallMissRate = overallMisses / overallAccesses;
|
|
|
|
|
|
|
|
// miss latency formulas
|
|
|
|
for (int access_idx = 0; access_idx < NUM_MEM_CMDS; ++access_idx) {
|
2006-06-28 20:35:00 +02:00
|
|
|
Packet::Command cmd = (Packet::Command)access_idx;
|
|
|
|
const string &cstr = temp_pkt.cmdIdxToString(cmd);
|
2006-06-28 17:02:14 +02:00
|
|
|
|
|
|
|
avgMissLatency[access_idx]
|
|
|
|
.name(name() + "." + cstr + "_avg_miss_latency")
|
|
|
|
.desc("average " + cstr + " miss latency")
|
|
|
|
.flags(total | nozero | nonan)
|
|
|
|
;
|
|
|
|
|
|
|
|
avgMissLatency[access_idx] =
|
|
|
|
missLatency[access_idx] / misses[access_idx];
|
|
|
|
}
|
|
|
|
|
|
|
|
demandAvgMissLatency
|
|
|
|
.name(name() + ".demand_avg_miss_latency")
|
|
|
|
.desc("average overall miss latency")
|
|
|
|
.flags(total)
|
|
|
|
;
|
|
|
|
demandAvgMissLatency = demandMissLatency / demandMisses;
|
|
|
|
|
|
|
|
overallAvgMissLatency
|
|
|
|
.name(name() + ".overall_avg_miss_latency")
|
|
|
|
.desc("average overall miss latency")
|
|
|
|
.flags(total)
|
|
|
|
;
|
|
|
|
overallAvgMissLatency = overallMissLatency / overallMisses;
|
|
|
|
|
|
|
|
blocked_cycles.init(NUM_BLOCKED_CAUSES);
|
|
|
|
blocked_cycles
|
|
|
|
.name(name() + ".blocked_cycles")
|
|
|
|
.desc("number of cycles access was blocked")
|
|
|
|
.subname(Blocked_NoMSHRs, "no_mshrs")
|
|
|
|
.subname(Blocked_NoTargets, "no_targets")
|
|
|
|
;
|
|
|
|
|
|
|
|
|
|
|
|
blocked_causes.init(NUM_BLOCKED_CAUSES);
|
|
|
|
blocked_causes
|
|
|
|
.name(name() + ".blocked")
|
|
|
|
.desc("number of cycles access was blocked")
|
|
|
|
.subname(Blocked_NoMSHRs, "no_mshrs")
|
|
|
|
.subname(Blocked_NoTargets, "no_targets")
|
|
|
|
;
|
|
|
|
|
|
|
|
avg_blocked
|
|
|
|
.name(name() + ".avg_blocked_cycles")
|
|
|
|
.desc("average number of cycles each access was blocked")
|
|
|
|
.subname(Blocked_NoMSHRs, "no_mshrs")
|
|
|
|
.subname(Blocked_NoTargets, "no_targets")
|
|
|
|
;
|
|
|
|
|
|
|
|
avg_blocked = blocked_cycles / blocked_causes;
|
|
|
|
|
|
|
|
fastWrites
|
|
|
|
.name(name() + ".fast_writes")
|
|
|
|
.desc("number of fast writes performed")
|
|
|
|
;
|
|
|
|
|
|
|
|
cacheCopies
|
|
|
|
.name(name() + ".cache_copies")
|
|
|
|
.desc("number of cache copies performed")
|
|
|
|
;
|
2006-06-30 22:25:35 +02:00
|
|
|
|
2006-06-28 17:02:14 +02:00
|
|
|
}
|