32eae8094d
This patch adds a creation-time check to the CPU to ensure that the interrupt controller is created for the cases where it is needed, i.e. if the CPU is not being switched in later and not a checker CPU. The patch also adds the "createInterruptController" call to a number of the regression scripts.
1800 lines
50 KiB
C++
1800 lines
50 KiB
C++
/*
|
|
* Copyright (c) 2012 ARM Limited
|
|
* All rights reserved
|
|
*
|
|
* The license below extends only to copyright in the software and shall
|
|
* not be construed as granting a license to any other intellectual
|
|
* property including but not limited to intellectual property relating
|
|
* to a hardware implementation of the functionality of the software
|
|
* licensed hereunder. You may use the software subject to the license
|
|
* terms below provided that you ensure that this notice is replicated
|
|
* unmodified and in its entirety in all distributions of the software,
|
|
* modified or unmodified, in source code or in binary form.
|
|
*
|
|
* Copyright (c) 2007 MIPS Technologies, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are
|
|
* met: redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer;
|
|
* redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution;
|
|
* neither the name of the copyright holders nor the names of its
|
|
* contributors may be used to endorse or promote products derived from
|
|
* this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* Authors: Korey Sewell
|
|
*
|
|
*/
|
|
|
|
#include <algorithm>
|
|
|
|
#include "arch/utility.hh"
|
|
#include "base/bigint.hh"
|
|
#include "config/the_isa.hh"
|
|
#include "cpu/inorder/resources/cache_unit.hh"
|
|
#include "cpu/inorder/resources/resource_list.hh"
|
|
#include "cpu/inorder/cpu.hh"
|
|
#include "cpu/inorder/first_stage.hh"
|
|
#include "cpu/inorder/inorder_dyn_inst.hh"
|
|
#include "cpu/inorder/pipeline_traits.hh"
|
|
#include "cpu/inorder/resource_pool.hh"
|
|
#include "cpu/inorder/thread_context.hh"
|
|
#include "cpu/inorder/thread_state.hh"
|
|
#include "cpu/activity.hh"
|
|
#include "cpu/base.hh"
|
|
#include "cpu/exetrace.hh"
|
|
#include "cpu/quiesce_event.hh"
|
|
#include "cpu/simple_thread.hh"
|
|
#include "cpu/thread_context.hh"
|
|
#include "debug/Activity.hh"
|
|
#include "debug/InOrderCPU.hh"
|
|
#include "debug/InOrderCachePort.hh"
|
|
#include "debug/Interrupt.hh"
|
|
#include "debug/Quiesce.hh"
|
|
#include "debug/RefCount.hh"
|
|
#include "debug/SkedCache.hh"
|
|
#include "params/InOrderCPU.hh"
|
|
#include "sim/full_system.hh"
|
|
#include "sim/process.hh"
|
|
#include "sim/stat_control.hh"
|
|
#include "sim/system.hh"
|
|
|
|
#if THE_ISA == ALPHA_ISA
|
|
#include "arch/alpha/osfpal.hh"
|
|
#endif
|
|
|
|
using namespace std;
|
|
using namespace TheISA;
|
|
using namespace ThePipeline;
|
|
|
|
InOrderCPU::CachePort::CachePort(CacheUnit *_cacheUnit) :
|
|
CpuPort(_cacheUnit->name() + "-cache-port", _cacheUnit->cpu),
|
|
cacheUnit(_cacheUnit)
|
|
{ }
|
|
|
|
bool
|
|
InOrderCPU::CachePort::recvTiming(Packet *pkt)
|
|
{
|
|
if (pkt->isError())
|
|
DPRINTF(InOrderCachePort, "Got error packet back for address: %x\n",
|
|
pkt->getAddr());
|
|
else if (pkt->isResponse())
|
|
cacheUnit->processCacheCompletion(pkt);
|
|
else {
|
|
//@note: depending on consistency model, update here
|
|
DPRINTF(InOrderCachePort, "Received snoop pkt %x,Ignoring\n",
|
|
pkt->getAddr());
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::CachePort::recvRetry()
|
|
{
|
|
cacheUnit->recvRetry();
|
|
}
|
|
|
|
InOrderCPU::TickEvent::TickEvent(InOrderCPU *c)
|
|
: Event(CPU_Tick_Pri), cpu(c)
|
|
{ }
|
|
|
|
|
|
void
|
|
InOrderCPU::TickEvent::process()
|
|
{
|
|
cpu->tick();
|
|
}
|
|
|
|
|
|
const char *
|
|
InOrderCPU::TickEvent::description() const
|
|
{
|
|
return "InOrderCPU tick event";
|
|
}
|
|
|
|
InOrderCPU::CPUEvent::CPUEvent(InOrderCPU *_cpu, CPUEventType e_type,
|
|
Fault fault, ThreadID _tid, DynInstPtr inst,
|
|
CPUEventPri event_pri)
|
|
: Event(event_pri), cpu(_cpu)
|
|
{
|
|
setEvent(e_type, fault, _tid, inst);
|
|
}
|
|
|
|
|
|
std::string InOrderCPU::eventNames[NumCPUEvents] =
|
|
{
|
|
"ActivateThread",
|
|
"ActivateNextReadyThread",
|
|
"DeactivateThread",
|
|
"HaltThread",
|
|
"SuspendThread",
|
|
"Trap",
|
|
"Syscall",
|
|
"SquashFromMemStall",
|
|
"UpdatePCs"
|
|
};
|
|
|
|
void
|
|
InOrderCPU::CPUEvent::process()
|
|
{
|
|
switch (cpuEventType)
|
|
{
|
|
case ActivateThread:
|
|
cpu->activateThread(tid);
|
|
cpu->resPool->activateThread(tid);
|
|
break;
|
|
|
|
case ActivateNextReadyThread:
|
|
cpu->activateNextReadyThread();
|
|
break;
|
|
|
|
case DeactivateThread:
|
|
cpu->deactivateThread(tid);
|
|
cpu->resPool->deactivateThread(tid);
|
|
break;
|
|
|
|
case HaltThread:
|
|
cpu->haltThread(tid);
|
|
cpu->resPool->deactivateThread(tid);
|
|
break;
|
|
|
|
case SuspendThread:
|
|
cpu->suspendThread(tid);
|
|
cpu->resPool->suspendThread(tid);
|
|
break;
|
|
|
|
case SquashFromMemStall:
|
|
cpu->squashDueToMemStall(inst->squashingStage, inst->seqNum, tid);
|
|
cpu->resPool->squashDueToMemStall(inst, inst->squashingStage,
|
|
inst->seqNum, tid);
|
|
break;
|
|
|
|
case Trap:
|
|
DPRINTF(InOrderCPU, "Trapping CPU\n");
|
|
cpu->trap(fault, tid, inst);
|
|
cpu->resPool->trap(fault, tid, inst);
|
|
cpu->trapPending[tid] = false;
|
|
break;
|
|
|
|
case Syscall:
|
|
cpu->syscall(inst->syscallNum, tid);
|
|
cpu->resPool->trap(fault, tid, inst);
|
|
break;
|
|
|
|
default:
|
|
fatal("Unrecognized Event Type %s", eventNames[cpuEventType]);
|
|
}
|
|
|
|
cpu->cpuEventRemoveList.push(this);
|
|
}
|
|
|
|
|
|
|
|
const char *
|
|
InOrderCPU::CPUEvent::description() const
|
|
{
|
|
return "InOrderCPU event";
|
|
}
|
|
|
|
void
|
|
InOrderCPU::CPUEvent::scheduleEvent(int delay)
|
|
{
|
|
assert(!scheduled() || squashed());
|
|
cpu->reschedule(this, cpu->nextCycle(curTick() + cpu->ticks(delay)), true);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::CPUEvent::unscheduleEvent()
|
|
{
|
|
if (scheduled())
|
|
squash();
|
|
}
|
|
|
|
InOrderCPU::InOrderCPU(Params *params)
|
|
: BaseCPU(params),
|
|
cpu_id(params->cpu_id),
|
|
coreType("default"),
|
|
_status(Idle),
|
|
tickEvent(this),
|
|
stageWidth(params->stageWidth),
|
|
resPool(new ResourcePool(this, params)),
|
|
timeBuffer(2 , 2),
|
|
dataPort(resPool->getDataUnit()),
|
|
instPort(resPool->getInstUnit()),
|
|
removeInstsThisCycle(false),
|
|
activityRec(params->name, NumStages, 10, params->activity),
|
|
system(params->system),
|
|
#ifdef DEBUG
|
|
cpuEventNum(0),
|
|
resReqCount(0),
|
|
#endif // DEBUG
|
|
drainCount(0),
|
|
deferRegistration(false/*params->deferRegistration*/),
|
|
stageTracing(params->stageTracing),
|
|
lastRunningCycle(0),
|
|
instsPerSwitch(0)
|
|
{
|
|
cpu_params = params;
|
|
|
|
// Resize for Multithreading CPUs
|
|
thread.resize(numThreads);
|
|
|
|
ThreadID active_threads = params->workload.size();
|
|
if (FullSystem) {
|
|
active_threads = 1;
|
|
} else {
|
|
active_threads = params->workload.size();
|
|
|
|
if (active_threads > MaxThreads) {
|
|
panic("Workload Size too large. Increase the 'MaxThreads'"
|
|
"in your InOrder implementation or "
|
|
"edit your workload size.");
|
|
}
|
|
|
|
|
|
if (active_threads > 1) {
|
|
threadModel = (InOrderCPU::ThreadModel) params->threadModel;
|
|
|
|
if (threadModel == SMT) {
|
|
DPRINTF(InOrderCPU, "Setting Thread Model to SMT.\n");
|
|
} else if (threadModel == SwitchOnCacheMiss) {
|
|
DPRINTF(InOrderCPU, "Setting Thread Model to "
|
|
"Switch On Cache Miss\n");
|
|
}
|
|
|
|
} else {
|
|
threadModel = Single;
|
|
}
|
|
}
|
|
|
|
for (ThreadID tid = 0; tid < numThreads; ++tid) {
|
|
pc[tid].set(0);
|
|
lastCommittedPC[tid].set(0);
|
|
|
|
if (FullSystem) {
|
|
// SMT is not supported in FS mode yet.
|
|
assert(numThreads == 1);
|
|
thread[tid] = new Thread(this, 0, NULL);
|
|
} else {
|
|
if (tid < (ThreadID)params->workload.size()) {
|
|
DPRINTF(InOrderCPU, "Workload[%i] process is %#x\n",
|
|
tid, params->workload[tid]->prog_fname);
|
|
thread[tid] =
|
|
new Thread(this, tid, params->workload[tid]);
|
|
} else {
|
|
//Allocate Empty thread so M5 can use later
|
|
//when scheduling threads to CPU
|
|
Process* dummy_proc = params->workload[0];
|
|
thread[tid] = new Thread(this, tid, dummy_proc);
|
|
}
|
|
|
|
// Eventually set this with parameters...
|
|
asid[tid] = tid;
|
|
}
|
|
|
|
// Setup the TC that will serve as the interface to the threads/CPU.
|
|
InOrderThreadContext *tc = new InOrderThreadContext;
|
|
tc->cpu = this;
|
|
tc->thread = thread[tid];
|
|
|
|
// Setup quiesce event.
|
|
this->thread[tid]->quiesceEvent = new EndQuiesceEvent(tc);
|
|
|
|
// Give the thread the TC.
|
|
thread[tid]->tc = tc;
|
|
thread[tid]->setFuncExeInst(0);
|
|
globalSeqNum[tid] = 1;
|
|
|
|
// Add the TC to the CPU's list of TC's.
|
|
this->threadContexts.push_back(tc);
|
|
}
|
|
|
|
// Initialize TimeBuffer Stage Queues
|
|
for (int stNum=0; stNum < NumStages - 1; stNum++) {
|
|
stageQueue[stNum] = new StageQueue(NumStages, NumStages);
|
|
stageQueue[stNum]->id(stNum);
|
|
}
|
|
|
|
|
|
// Set Up Pipeline Stages
|
|
for (int stNum=0; stNum < NumStages; stNum++) {
|
|
if (stNum == 0)
|
|
pipelineStage[stNum] = new FirstStage(params, stNum);
|
|
else
|
|
pipelineStage[stNum] = new PipelineStage(params, stNum);
|
|
|
|
pipelineStage[stNum]->setCPU(this);
|
|
pipelineStage[stNum]->setActiveThreads(&activeThreads);
|
|
pipelineStage[stNum]->setTimeBuffer(&timeBuffer);
|
|
|
|
// Take Care of 1st/Nth stages
|
|
if (stNum > 0)
|
|
pipelineStage[stNum]->setPrevStageQueue(stageQueue[stNum - 1]);
|
|
if (stNum < NumStages - 1)
|
|
pipelineStage[stNum]->setNextStageQueue(stageQueue[stNum]);
|
|
}
|
|
|
|
// Initialize thread specific variables
|
|
for (ThreadID tid = 0; tid < numThreads; tid++) {
|
|
archRegDepMap[tid].setCPU(this);
|
|
|
|
nonSpecInstActive[tid] = false;
|
|
nonSpecSeqNum[tid] = 0;
|
|
|
|
squashSeqNum[tid] = MaxAddr;
|
|
lastSquashCycle[tid] = 0;
|
|
|
|
memset(intRegs[tid], 0, sizeof(intRegs[tid]));
|
|
memset(floatRegs.i[tid], 0, sizeof(floatRegs.i[tid]));
|
|
isa[tid].clear();
|
|
|
|
// Define dummy instructions and resource requests to be used.
|
|
dummyInst[tid] = new InOrderDynInst(this,
|
|
thread[tid],
|
|
0,
|
|
tid,
|
|
asid[tid]);
|
|
|
|
dummyReq[tid] = new ResourceRequest(resPool->getResource(0));
|
|
|
|
|
|
if (FullSystem) {
|
|
// Use this dummy inst to force squashing behind every instruction
|
|
// in pipeline
|
|
dummyTrapInst[tid] = new InOrderDynInst(this, NULL, 0, 0, 0);
|
|
dummyTrapInst[tid]->seqNum = 0;
|
|
dummyTrapInst[tid]->squashSeqNum = 0;
|
|
dummyTrapInst[tid]->setTid(tid);
|
|
}
|
|
|
|
trapPending[tid] = false;
|
|
|
|
}
|
|
|
|
// InOrderCPU always requires an interrupt controller.
|
|
if (!params->defer_registration && !interrupts) {
|
|
fatal("InOrderCPU %s has no interrupt controller.\n"
|
|
"Ensure createInterruptController() is called.\n", name());
|
|
}
|
|
|
|
dummyReqInst = new InOrderDynInst(this, NULL, 0, 0, 0);
|
|
dummyReqInst->setSquashed();
|
|
dummyReqInst->resetInstCount();
|
|
|
|
dummyBufferInst = new InOrderDynInst(this, NULL, 0, 0, 0);
|
|
dummyBufferInst->setSquashed();
|
|
dummyBufferInst->resetInstCount();
|
|
|
|
endOfSkedIt = skedCache.end();
|
|
frontEndSked = createFrontEndSked();
|
|
faultSked = createFaultSked();
|
|
|
|
lastRunningCycle = curTick();
|
|
|
|
lockAddr = 0;
|
|
lockFlag = false;
|
|
|
|
// Schedule First Tick Event, CPU will reschedule itself from here on out.
|
|
scheduleTickEvent(0);
|
|
}
|
|
|
|
InOrderCPU::~InOrderCPU()
|
|
{
|
|
delete resPool;
|
|
|
|
SkedCacheIt sked_it = skedCache.begin();
|
|
SkedCacheIt sked_end = skedCache.end();
|
|
|
|
while (sked_it != sked_end) {
|
|
delete (*sked_it).second;
|
|
sked_it++;
|
|
}
|
|
skedCache.clear();
|
|
}
|
|
|
|
m5::hash_map<InOrderCPU::SkedID, ThePipeline::RSkedPtr> InOrderCPU::skedCache;
|
|
|
|
RSkedPtr
|
|
InOrderCPU::createFrontEndSked()
|
|
{
|
|
RSkedPtr res_sked = new ResourceSked();
|
|
int stage_num = 0;
|
|
StageScheduler F(res_sked, stage_num++);
|
|
StageScheduler D(res_sked, stage_num++);
|
|
|
|
// FETCH
|
|
F.needs(FetchSeq, FetchSeqUnit::AssignNextPC);
|
|
F.needs(ICache, FetchUnit::InitiateFetch);
|
|
|
|
// DECODE
|
|
D.needs(ICache, FetchUnit::CompleteFetch);
|
|
D.needs(Decode, DecodeUnit::DecodeInst);
|
|
D.needs(BPred, BranchPredictor::PredictBranch);
|
|
D.needs(FetchSeq, FetchSeqUnit::UpdateTargetPC);
|
|
|
|
|
|
DPRINTF(SkedCache, "Resource Sked created for instruction Front End\n");
|
|
|
|
return res_sked;
|
|
}
|
|
|
|
RSkedPtr
|
|
InOrderCPU::createFaultSked()
|
|
{
|
|
RSkedPtr res_sked = new ResourceSked();
|
|
StageScheduler W(res_sked, NumStages - 1);
|
|
W.needs(Grad, GraduationUnit::CheckFault);
|
|
DPRINTF(SkedCache, "Resource Sked created for instruction Faults\n");
|
|
return res_sked;
|
|
}
|
|
|
|
RSkedPtr
|
|
InOrderCPU::createBackEndSked(DynInstPtr inst)
|
|
{
|
|
RSkedPtr res_sked = lookupSked(inst);
|
|
if (res_sked != NULL) {
|
|
DPRINTF(SkedCache, "Found %s in sked cache.\n",
|
|
inst->instName());
|
|
return res_sked;
|
|
} else {
|
|
res_sked = new ResourceSked();
|
|
}
|
|
|
|
int stage_num = ThePipeline::BackEndStartStage;
|
|
StageScheduler X(res_sked, stage_num++);
|
|
StageScheduler M(res_sked, stage_num++);
|
|
StageScheduler W(res_sked, stage_num++);
|
|
|
|
if (!inst->staticInst) {
|
|
warn_once("Static Instruction Object Not Set. Can't Create"
|
|
" Back End Schedule");
|
|
return NULL;
|
|
}
|
|
|
|
// EXECUTE
|
|
X.needs(RegManager, UseDefUnit::MarkDestRegs);
|
|
for (int idx=0; idx < inst->numSrcRegs(); idx++) {
|
|
if (!idx || !inst->isStore()) {
|
|
X.needs(RegManager, UseDefUnit::ReadSrcReg, idx);
|
|
}
|
|
}
|
|
|
|
//@todo: schedule non-spec insts to operate on this cycle
|
|
// as long as all previous insts are done
|
|
if ( inst->isNonSpeculative() ) {
|
|
// skip execution of non speculative insts until later
|
|
} else if ( inst->isMemRef() ) {
|
|
if ( inst->isLoad() ) {
|
|
X.needs(AGEN, AGENUnit::GenerateAddr);
|
|
}
|
|
} else if (inst->opClass() == IntMultOp || inst->opClass() == IntDivOp) {
|
|
X.needs(MDU, MultDivUnit::StartMultDiv);
|
|
} else {
|
|
X.needs(ExecUnit, ExecutionUnit::ExecuteInst);
|
|
}
|
|
|
|
// MEMORY
|
|
if (!inst->isNonSpeculative()) {
|
|
if (inst->opClass() == IntMultOp || inst->opClass() == IntDivOp) {
|
|
M.needs(MDU, MultDivUnit::EndMultDiv);
|
|
}
|
|
|
|
if ( inst->isLoad() ) {
|
|
M.needs(DCache, CacheUnit::InitiateReadData);
|
|
if (inst->splitInst)
|
|
M.needs(DCache, CacheUnit::InitSecondSplitRead);
|
|
} else if ( inst->isStore() ) {
|
|
for (int i = 1; i < inst->numSrcRegs(); i++ ) {
|
|
M.needs(RegManager, UseDefUnit::ReadSrcReg, i);
|
|
}
|
|
M.needs(AGEN, AGENUnit::GenerateAddr);
|
|
M.needs(DCache, CacheUnit::InitiateWriteData);
|
|
if (inst->splitInst)
|
|
M.needs(DCache, CacheUnit::InitSecondSplitWrite);
|
|
}
|
|
}
|
|
|
|
// WRITEBACK
|
|
if (!inst->isNonSpeculative()) {
|
|
if ( inst->isLoad() ) {
|
|
W.needs(DCache, CacheUnit::CompleteReadData);
|
|
if (inst->splitInst)
|
|
W.needs(DCache, CacheUnit::CompleteSecondSplitRead);
|
|
} else if ( inst->isStore() ) {
|
|
W.needs(DCache, CacheUnit::CompleteWriteData);
|
|
if (inst->splitInst)
|
|
W.needs(DCache, CacheUnit::CompleteSecondSplitWrite);
|
|
}
|
|
} else {
|
|
// Finally, Execute Speculative Data
|
|
if (inst->isMemRef()) {
|
|
if (inst->isLoad()) {
|
|
W.needs(AGEN, AGENUnit::GenerateAddr);
|
|
W.needs(DCache, CacheUnit::InitiateReadData);
|
|
if (inst->splitInst)
|
|
W.needs(DCache, CacheUnit::InitSecondSplitRead);
|
|
W.needs(DCache, CacheUnit::CompleteReadData);
|
|
if (inst->splitInst)
|
|
W.needs(DCache, CacheUnit::CompleteSecondSplitRead);
|
|
} else if (inst->isStore()) {
|
|
if ( inst->numSrcRegs() >= 2 ) {
|
|
W.needs(RegManager, UseDefUnit::ReadSrcReg, 1);
|
|
}
|
|
W.needs(AGEN, AGENUnit::GenerateAddr);
|
|
W.needs(DCache, CacheUnit::InitiateWriteData);
|
|
if (inst->splitInst)
|
|
W.needs(DCache, CacheUnit::InitSecondSplitWrite);
|
|
W.needs(DCache, CacheUnit::CompleteWriteData);
|
|
if (inst->splitInst)
|
|
W.needs(DCache, CacheUnit::CompleteSecondSplitWrite);
|
|
}
|
|
} else {
|
|
W.needs(ExecUnit, ExecutionUnit::ExecuteInst);
|
|
}
|
|
}
|
|
|
|
W.needs(Grad, GraduationUnit::CheckFault);
|
|
|
|
for (int idx=0; idx < inst->numDestRegs(); idx++) {
|
|
W.needs(RegManager, UseDefUnit::WriteDestReg, idx);
|
|
}
|
|
|
|
if (inst->isControl())
|
|
W.needs(BPred, BranchPredictor::UpdatePredictor);
|
|
|
|
W.needs(Grad, GraduationUnit::GraduateInst);
|
|
|
|
// Insert Back Schedule into our cache of
|
|
// resource schedules
|
|
addToSkedCache(inst, res_sked);
|
|
|
|
DPRINTF(SkedCache, "Back End Sked Created for instruction: %s (%08p)\n",
|
|
inst->instName(), inst->getMachInst());
|
|
res_sked->print();
|
|
|
|
return res_sked;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::regStats()
|
|
{
|
|
/* Register the Resource Pool's stats here.*/
|
|
resPool->regStats();
|
|
|
|
/* Register for each Pipeline Stage */
|
|
for (int stage_num=0; stage_num < ThePipeline::NumStages; stage_num++) {
|
|
pipelineStage[stage_num]->regStats();
|
|
}
|
|
|
|
/* Register any of the InOrderCPU's stats here.*/
|
|
instsPerCtxtSwitch
|
|
.name(name() + ".instsPerContextSwitch")
|
|
.desc("Instructions Committed Per Context Switch")
|
|
.prereq(instsPerCtxtSwitch);
|
|
|
|
numCtxtSwitches
|
|
.name(name() + ".contextSwitches")
|
|
.desc("Number of context switches");
|
|
|
|
comLoads
|
|
.name(name() + ".comLoads")
|
|
.desc("Number of Load instructions committed");
|
|
|
|
comStores
|
|
.name(name() + ".comStores")
|
|
.desc("Number of Store instructions committed");
|
|
|
|
comBranches
|
|
.name(name() + ".comBranches")
|
|
.desc("Number of Branches instructions committed");
|
|
|
|
comNops
|
|
.name(name() + ".comNops")
|
|
.desc("Number of Nop instructions committed");
|
|
|
|
comNonSpec
|
|
.name(name() + ".comNonSpec")
|
|
.desc("Number of Non-Speculative instructions committed");
|
|
|
|
comInts
|
|
.name(name() + ".comInts")
|
|
.desc("Number of Integer instructions committed");
|
|
|
|
comFloats
|
|
.name(name() + ".comFloats")
|
|
.desc("Number of Floating Point instructions committed");
|
|
|
|
timesIdled
|
|
.name(name() + ".timesIdled")
|
|
.desc("Number of times that the entire CPU went into an idle state and"
|
|
" unscheduled itself")
|
|
.prereq(timesIdled);
|
|
|
|
idleCycles
|
|
.name(name() + ".idleCycles")
|
|
.desc("Number of cycles cpu's stages were not processed");
|
|
|
|
runCycles
|
|
.name(name() + ".runCycles")
|
|
.desc("Number of cycles cpu stages are processed.");
|
|
|
|
activity
|
|
.name(name() + ".activity")
|
|
.desc("Percentage of cycles cpu is active")
|
|
.precision(6);
|
|
activity = (runCycles / numCycles) * 100;
|
|
|
|
threadCycles
|
|
.init(numThreads)
|
|
.name(name() + ".threadCycles")
|
|
.desc("Total Number of Cycles A Thread Was Active in CPU (Per-Thread)");
|
|
|
|
smtCycles
|
|
.name(name() + ".smtCycles")
|
|
.desc("Total number of cycles that the CPU was in SMT-mode");
|
|
|
|
committedInsts
|
|
.init(numThreads)
|
|
.name(name() + ".committedInsts")
|
|
.desc("Number of Instructions committed (Per-Thread)");
|
|
|
|
committedOps
|
|
.init(numThreads)
|
|
.name(name() + ".committedOps")
|
|
.desc("Number of Ops committed (Per-Thread)");
|
|
|
|
smtCommittedInsts
|
|
.init(numThreads)
|
|
.name(name() + ".smtCommittedInsts")
|
|
.desc("Number of SMT Instructions committed (Per-Thread)");
|
|
|
|
totalCommittedInsts
|
|
.name(name() + ".committedInsts_total")
|
|
.desc("Number of Instructions committed (Total)");
|
|
|
|
cpi
|
|
.name(name() + ".cpi")
|
|
.desc("CPI: Cycles Per Instruction (Per-Thread)")
|
|
.precision(6);
|
|
cpi = numCycles / committedInsts;
|
|
|
|
smtCpi
|
|
.name(name() + ".smt_cpi")
|
|
.desc("CPI: Total SMT-CPI")
|
|
.precision(6);
|
|
smtCpi = smtCycles / smtCommittedInsts;
|
|
|
|
totalCpi
|
|
.name(name() + ".cpi_total")
|
|
.desc("CPI: Total CPI of All Threads")
|
|
.precision(6);
|
|
totalCpi = numCycles / totalCommittedInsts;
|
|
|
|
ipc
|
|
.name(name() + ".ipc")
|
|
.desc("IPC: Instructions Per Cycle (Per-Thread)")
|
|
.precision(6);
|
|
ipc = committedInsts / numCycles;
|
|
|
|
smtIpc
|
|
.name(name() + ".smt_ipc")
|
|
.desc("IPC: Total SMT-IPC")
|
|
.precision(6);
|
|
smtIpc = smtCommittedInsts / smtCycles;
|
|
|
|
totalIpc
|
|
.name(name() + ".ipc_total")
|
|
.desc("IPC: Total IPC of All Threads")
|
|
.precision(6);
|
|
totalIpc = totalCommittedInsts / numCycles;
|
|
|
|
BaseCPU::regStats();
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::tick()
|
|
{
|
|
DPRINTF(InOrderCPU, "\n\nInOrderCPU: Ticking main, InOrderCPU.\n");
|
|
|
|
++numCycles;
|
|
|
|
checkForInterrupts();
|
|
|
|
bool pipes_idle = true;
|
|
//Tick each of the stages
|
|
for (int stNum=NumStages - 1; stNum >= 0 ; stNum--) {
|
|
pipelineStage[stNum]->tick();
|
|
|
|
pipes_idle = pipes_idle && pipelineStage[stNum]->idle;
|
|
}
|
|
|
|
if (pipes_idle)
|
|
idleCycles++;
|
|
else
|
|
runCycles++;
|
|
|
|
// Now advance the time buffers one tick
|
|
timeBuffer.advance();
|
|
for (int sqNum=0; sqNum < NumStages - 1; sqNum++) {
|
|
stageQueue[sqNum]->advance();
|
|
}
|
|
activityRec.advance();
|
|
|
|
// Any squashed events, or insts then remove them now
|
|
cleanUpRemovedEvents();
|
|
cleanUpRemovedInsts();
|
|
|
|
// Re-schedule CPU for this cycle
|
|
if (!tickEvent.scheduled()) {
|
|
if (_status == SwitchedOut) {
|
|
// increment stat
|
|
lastRunningCycle = curTick();
|
|
} else if (!activityRec.active()) {
|
|
DPRINTF(InOrderCPU, "sleeping CPU.\n");
|
|
lastRunningCycle = curTick();
|
|
timesIdled++;
|
|
} else {
|
|
//Tick next_tick = curTick() + cycles(1);
|
|
//tickEvent.schedule(next_tick);
|
|
schedule(&tickEvent, nextCycle(curTick() + 1));
|
|
DPRINTF(InOrderCPU, "Scheduled CPU for next tick @ %i.\n",
|
|
nextCycle(curTick() + 1));
|
|
}
|
|
}
|
|
|
|
tickThreadStats();
|
|
updateThreadPriority();
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::init()
|
|
{
|
|
if (!deferRegistration) {
|
|
registerThreadContexts();
|
|
}
|
|
|
|
// Set inSyscall so that the CPU doesn't squash when initially
|
|
// setting up registers.
|
|
for (ThreadID tid = 0; tid < numThreads; ++tid)
|
|
thread[tid]->inSyscall = true;
|
|
|
|
if (FullSystem) {
|
|
for (ThreadID tid = 0; tid < numThreads; tid++) {
|
|
ThreadContext *src_tc = threadContexts[tid];
|
|
TheISA::initCPU(src_tc, src_tc->contextId());
|
|
// Initialise the ThreadContext's memory proxies
|
|
thread[tid]->initMemProxies(thread[tid]->getTC());
|
|
}
|
|
}
|
|
|
|
// Clear inSyscall.
|
|
for (ThreadID tid = 0; tid < numThreads; ++tid)
|
|
thread[tid]->inSyscall = false;
|
|
|
|
// Call Initializiation Routine for Resource Pool
|
|
resPool->init();
|
|
}
|
|
|
|
Fault
|
|
InOrderCPU::hwrei(ThreadID tid)
|
|
{
|
|
#if THE_ISA == ALPHA_ISA
|
|
// Need to clear the lock flag upon returning from an interrupt.
|
|
setMiscRegNoEffect(AlphaISA::MISCREG_LOCKFLAG, false, tid);
|
|
|
|
thread[tid]->kernelStats->hwrei();
|
|
// FIXME: XXX check for interrupts? XXX
|
|
#endif
|
|
|
|
return NoFault;
|
|
}
|
|
|
|
|
|
bool
|
|
InOrderCPU::simPalCheck(int palFunc, ThreadID tid)
|
|
{
|
|
#if THE_ISA == ALPHA_ISA
|
|
if (this->thread[tid]->kernelStats)
|
|
this->thread[tid]->kernelStats->callpal(palFunc,
|
|
this->threadContexts[tid]);
|
|
|
|
switch (palFunc) {
|
|
case PAL::halt:
|
|
halt();
|
|
if (--System::numSystemsRunning == 0)
|
|
exitSimLoop("all cpus halted");
|
|
break;
|
|
|
|
case PAL::bpt:
|
|
case PAL::bugchk:
|
|
if (this->system->breakpoint())
|
|
return false;
|
|
break;
|
|
}
|
|
#endif
|
|
return true;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::checkForInterrupts()
|
|
{
|
|
for (int i = 0; i < threadContexts.size(); i++) {
|
|
ThreadContext *tc = threadContexts[i];
|
|
|
|
if (interrupts->checkInterrupts(tc)) {
|
|
Fault interrupt = interrupts->getInterrupt(tc);
|
|
|
|
if (interrupt != NoFault) {
|
|
DPRINTF(Interrupt, "Processing Intterupt for [tid:%i].\n",
|
|
tc->threadId());
|
|
|
|
ThreadID tid = tc->threadId();
|
|
interrupts->updateIntrInfo(tc);
|
|
|
|
// Squash from Last Stage in Pipeline
|
|
unsigned last_stage = NumStages - 1;
|
|
dummyTrapInst[tid]->squashingStage = last_stage;
|
|
pipelineStage[last_stage]->setupSquash(dummyTrapInst[tid],
|
|
tid);
|
|
|
|
// By default, setupSquash will always squash from stage + 1
|
|
pipelineStage[BackEndStartStage - 1]->setupSquash(dummyTrapInst[tid],
|
|
tid);
|
|
|
|
// Schedule Squash Through-out Resource Pool
|
|
resPool->scheduleEvent(
|
|
(InOrderCPU::CPUEventType)ResourcePool::SquashAll,
|
|
dummyTrapInst[tid], 0);
|
|
|
|
// Finally, Setup Trap to happen at end of cycle
|
|
trapContext(interrupt, tid, dummyTrapInst[tid]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
Fault
|
|
InOrderCPU::getInterrupts()
|
|
{
|
|
// Check if there are any outstanding interrupts
|
|
return interrupts->getInterrupt(threadContexts[0]);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::processInterrupts(Fault interrupt)
|
|
{
|
|
// Check for interrupts here. For now can copy the code that
|
|
// exists within isa_fullsys_traits.hh. Also assume that thread 0
|
|
// is the one that handles the interrupts.
|
|
// @todo: Possibly consolidate the interrupt checking code.
|
|
// @todo: Allow other threads to handle interrupts.
|
|
|
|
assert(interrupt != NoFault);
|
|
interrupts->updateIntrInfo(threadContexts[0]);
|
|
|
|
DPRINTF(InOrderCPU, "Interrupt %s being handled\n", interrupt->name());
|
|
|
|
// Note: Context ID ok here? Impl. of FS mode needs to revisit this
|
|
trap(interrupt, threadContexts[0]->contextId(), dummyBufferInst);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::trapContext(Fault fault, ThreadID tid, DynInstPtr inst, int delay)
|
|
{
|
|
scheduleCpuEvent(Trap, fault, tid, inst, delay);
|
|
trapPending[tid] = true;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::trap(Fault fault, ThreadID tid, DynInstPtr inst)
|
|
{
|
|
fault->invoke(tcBase(tid), inst->staticInst);
|
|
removePipelineStalls(tid);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::squashFromMemStall(DynInstPtr inst, ThreadID tid, int delay)
|
|
{
|
|
scheduleCpuEvent(SquashFromMemStall, NoFault, tid, inst, delay);
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::squashDueToMemStall(int stage_num, InstSeqNum seq_num,
|
|
ThreadID tid)
|
|
{
|
|
DPRINTF(InOrderCPU, "Squashing Pipeline Stages Due to Memory Stall...\n");
|
|
|
|
// Squash all instructions in each stage including
|
|
// instruction that caused the squash (seq_num - 1)
|
|
// NOTE: The stage bandwidth needs to be cleared so thats why
|
|
// the stalling instruction is squashed as well. The stalled
|
|
// instruction is previously placed in another intermediate buffer
|
|
// while it's stall is being handled.
|
|
InstSeqNum squash_seq_num = seq_num - 1;
|
|
|
|
for (int stNum=stage_num; stNum >= 0 ; stNum--) {
|
|
pipelineStage[stNum]->squashDueToMemStall(squash_seq_num, tid);
|
|
}
|
|
}
|
|
|
|
void
|
|
InOrderCPU::scheduleCpuEvent(CPUEventType c_event, Fault fault,
|
|
ThreadID tid, DynInstPtr inst,
|
|
unsigned delay, CPUEventPri event_pri)
|
|
{
|
|
CPUEvent *cpu_event = new CPUEvent(this, c_event, fault, tid, inst,
|
|
event_pri);
|
|
|
|
Tick sked_tick = nextCycle(curTick() + ticks(delay));
|
|
if (delay >= 0) {
|
|
DPRINTF(InOrderCPU, "Scheduling CPU Event (%s) for cycle %i, [tid:%i].\n",
|
|
eventNames[c_event], curTick() + delay, tid);
|
|
schedule(cpu_event, sked_tick);
|
|
} else {
|
|
cpu_event->process();
|
|
cpuEventRemoveList.push(cpu_event);
|
|
}
|
|
|
|
// Broadcast event to the Resource Pool
|
|
// Need to reset tid just in case this is a dummy instruction
|
|
inst->setTid(tid);
|
|
resPool->scheduleEvent(c_event, inst, 0, 0, tid);
|
|
}
|
|
|
|
bool
|
|
InOrderCPU::isThreadActive(ThreadID tid)
|
|
{
|
|
list<ThreadID>::iterator isActive =
|
|
std::find(activeThreads.begin(), activeThreads.end(), tid);
|
|
|
|
return (isActive != activeThreads.end());
|
|
}
|
|
|
|
bool
|
|
InOrderCPU::isThreadReady(ThreadID tid)
|
|
{
|
|
list<ThreadID>::iterator isReady =
|
|
std::find(readyThreads.begin(), readyThreads.end(), tid);
|
|
|
|
return (isReady != readyThreads.end());
|
|
}
|
|
|
|
bool
|
|
InOrderCPU::isThreadSuspended(ThreadID tid)
|
|
{
|
|
list<ThreadID>::iterator isSuspended =
|
|
std::find(suspendedThreads.begin(), suspendedThreads.end(), tid);
|
|
|
|
return (isSuspended != suspendedThreads.end());
|
|
}
|
|
|
|
void
|
|
InOrderCPU::activateNextReadyThread()
|
|
{
|
|
if (readyThreads.size() >= 1) {
|
|
ThreadID ready_tid = readyThreads.front();
|
|
|
|
// Activate in Pipeline
|
|
activateThread(ready_tid);
|
|
|
|
// Activate in Resource Pool
|
|
resPool->activateThread(ready_tid);
|
|
|
|
list<ThreadID>::iterator ready_it =
|
|
std::find(readyThreads.begin(), readyThreads.end(), ready_tid);
|
|
readyThreads.erase(ready_it);
|
|
} else {
|
|
DPRINTF(InOrderCPU,
|
|
"Attempting to activate new thread, but No Ready Threads to"
|
|
"activate.\n");
|
|
DPRINTF(InOrderCPU,
|
|
"Unable to switch to next active thread.\n");
|
|
}
|
|
}
|
|
|
|
void
|
|
InOrderCPU::activateThread(ThreadID tid)
|
|
{
|
|
if (isThreadSuspended(tid)) {
|
|
DPRINTF(InOrderCPU,
|
|
"Removing [tid:%i] from suspended threads list.\n", tid);
|
|
|
|
list<ThreadID>::iterator susp_it =
|
|
std::find(suspendedThreads.begin(), suspendedThreads.end(),
|
|
tid);
|
|
suspendedThreads.erase(susp_it);
|
|
}
|
|
|
|
if (threadModel == SwitchOnCacheMiss &&
|
|
numActiveThreads() == 1) {
|
|
DPRINTF(InOrderCPU,
|
|
"Ignoring activation of [tid:%i], since [tid:%i] is "
|
|
"already running.\n", tid, activeThreadId());
|
|
|
|
DPRINTF(InOrderCPU,"Placing [tid:%i] on ready threads list\n",
|
|
tid);
|
|
|
|
readyThreads.push_back(tid);
|
|
|
|
} else if (!isThreadActive(tid)) {
|
|
DPRINTF(InOrderCPU,
|
|
"Adding [tid:%i] to active threads list.\n", tid);
|
|
activeThreads.push_back(tid);
|
|
|
|
activateThreadInPipeline(tid);
|
|
|
|
thread[tid]->lastActivate = curTick();
|
|
|
|
tcBase(tid)->setStatus(ThreadContext::Active);
|
|
|
|
wakeCPU();
|
|
|
|
numCtxtSwitches++;
|
|
}
|
|
}
|
|
|
|
void
|
|
InOrderCPU::activateThreadInPipeline(ThreadID tid)
|
|
{
|
|
for (int stNum=0; stNum < NumStages; stNum++) {
|
|
pipelineStage[stNum]->activateThread(tid);
|
|
}
|
|
}
|
|
|
|
void
|
|
InOrderCPU::deactivateContext(ThreadID tid, int delay)
|
|
{
|
|
DPRINTF(InOrderCPU,"[tid:%i]: Deactivating ...\n", tid);
|
|
|
|
scheduleCpuEvent(DeactivateThread, NoFault, tid, dummyInst[tid], delay);
|
|
|
|
// Be sure to signal that there's some activity so the CPU doesn't
|
|
// deschedule itself.
|
|
activityRec.activity();
|
|
|
|
_status = Running;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::deactivateThread(ThreadID tid)
|
|
{
|
|
DPRINTF(InOrderCPU, "[tid:%i]: Calling deactivate thread.\n", tid);
|
|
|
|
if (isThreadActive(tid)) {
|
|
DPRINTF(InOrderCPU,"[tid:%i]: Removing from active threads list\n",
|
|
tid);
|
|
list<ThreadID>::iterator thread_it =
|
|
std::find(activeThreads.begin(), activeThreads.end(), tid);
|
|
|
|
removePipelineStalls(*thread_it);
|
|
|
|
activeThreads.erase(thread_it);
|
|
|
|
// Ideally, this should be triggered from the
|
|
// suspendContext/Thread functions
|
|
tcBase(tid)->setStatus(ThreadContext::Suspended);
|
|
}
|
|
|
|
assert(!isThreadActive(tid));
|
|
}
|
|
|
|
void
|
|
InOrderCPU::removePipelineStalls(ThreadID tid)
|
|
{
|
|
DPRINTF(InOrderCPU,"[tid:%i]: Removing all pipeline stalls\n",
|
|
tid);
|
|
|
|
for (int stNum = 0; stNum < NumStages ; stNum++) {
|
|
pipelineStage[stNum]->removeStalls(tid);
|
|
}
|
|
|
|
}
|
|
|
|
void
|
|
InOrderCPU::updateThreadPriority()
|
|
{
|
|
if (activeThreads.size() > 1)
|
|
{
|
|
//DEFAULT TO ROUND ROBIN SCHEME
|
|
//e.g. Move highest priority to end of thread list
|
|
list<ThreadID>::iterator list_begin = activeThreads.begin();
|
|
|
|
unsigned high_thread = *list_begin;
|
|
|
|
activeThreads.erase(list_begin);
|
|
|
|
activeThreads.push_back(high_thread);
|
|
}
|
|
}
|
|
|
|
inline void
|
|
InOrderCPU::tickThreadStats()
|
|
{
|
|
/** Keep track of cycles that each thread is active */
|
|
list<ThreadID>::iterator thread_it = activeThreads.begin();
|
|
while (thread_it != activeThreads.end()) {
|
|
threadCycles[*thread_it]++;
|
|
thread_it++;
|
|
}
|
|
|
|
// Keep track of cycles where SMT is active
|
|
if (activeThreads.size() > 1) {
|
|
smtCycles++;
|
|
}
|
|
}
|
|
|
|
void
|
|
InOrderCPU::activateContext(ThreadID tid, int delay)
|
|
{
|
|
DPRINTF(InOrderCPU,"[tid:%i]: Activating ...\n", tid);
|
|
|
|
|
|
scheduleCpuEvent(ActivateThread, NoFault, tid, dummyInst[tid], delay);
|
|
|
|
// Be sure to signal that there's some activity so the CPU doesn't
|
|
// deschedule itself.
|
|
activityRec.activity();
|
|
|
|
_status = Running;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::activateNextReadyContext(int delay)
|
|
{
|
|
DPRINTF(InOrderCPU,"Activating next ready thread\n");
|
|
|
|
scheduleCpuEvent(ActivateNextReadyThread, NoFault, 0/*tid*/, dummyInst[0],
|
|
delay, ActivateNextReadyThread_Pri);
|
|
|
|
// Be sure to signal that there's some activity so the CPU doesn't
|
|
// deschedule itself.
|
|
activityRec.activity();
|
|
|
|
_status = Running;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::haltContext(ThreadID tid)
|
|
{
|
|
DPRINTF(InOrderCPU, "[tid:%i]: Calling Halt Context...\n", tid);
|
|
|
|
scheduleCpuEvent(HaltThread, NoFault, tid, dummyInst[tid]);
|
|
|
|
activityRec.activity();
|
|
}
|
|
|
|
void
|
|
InOrderCPU::haltThread(ThreadID tid)
|
|
{
|
|
DPRINTF(InOrderCPU, "[tid:%i]: Placing on Halted Threads List...\n", tid);
|
|
deactivateThread(tid);
|
|
squashThreadInPipeline(tid);
|
|
haltedThreads.push_back(tid);
|
|
|
|
tcBase(tid)->setStatus(ThreadContext::Halted);
|
|
|
|
if (threadModel == SwitchOnCacheMiss) {
|
|
activateNextReadyContext();
|
|
}
|
|
}
|
|
|
|
void
|
|
InOrderCPU::suspendContext(ThreadID tid)
|
|
{
|
|
scheduleCpuEvent(SuspendThread, NoFault, tid, dummyInst[tid]);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::suspendThread(ThreadID tid)
|
|
{
|
|
DPRINTF(InOrderCPU, "[tid:%i]: Placing on Suspended Threads List...\n",
|
|
tid);
|
|
deactivateThread(tid);
|
|
suspendedThreads.push_back(tid);
|
|
thread[tid]->lastSuspend = curTick();
|
|
|
|
tcBase(tid)->setStatus(ThreadContext::Suspended);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::squashThreadInPipeline(ThreadID tid)
|
|
{
|
|
//Squash all instructions in each stage
|
|
for (int stNum=NumStages - 1; stNum >= 0 ; stNum--) {
|
|
pipelineStage[stNum]->squash(0 /*seq_num*/, tid);
|
|
}
|
|
}
|
|
|
|
PipelineStage*
|
|
InOrderCPU::getPipeStage(int stage_num)
|
|
{
|
|
return pipelineStage[stage_num];
|
|
}
|
|
|
|
|
|
RegIndex
|
|
InOrderCPU::flattenRegIdx(RegIndex reg_idx, RegType ®_type, ThreadID tid)
|
|
{
|
|
if (reg_idx < FP_Base_DepTag) {
|
|
reg_type = IntType;
|
|
return isa[tid].flattenIntIndex(reg_idx);
|
|
} else if (reg_idx < Ctrl_Base_DepTag) {
|
|
reg_type = FloatType;
|
|
reg_idx -= FP_Base_DepTag;
|
|
return isa[tid].flattenFloatIndex(reg_idx);
|
|
} else {
|
|
reg_type = MiscType;
|
|
return reg_idx - TheISA::Ctrl_Base_DepTag;
|
|
}
|
|
}
|
|
|
|
uint64_t
|
|
InOrderCPU::readIntReg(RegIndex reg_idx, ThreadID tid)
|
|
{
|
|
DPRINTF(IntRegs, "[tid:%i]: Reading Int. Reg %i as %x\n",
|
|
tid, reg_idx, intRegs[tid][reg_idx]);
|
|
|
|
return intRegs[tid][reg_idx];
|
|
}
|
|
|
|
FloatReg
|
|
InOrderCPU::readFloatReg(RegIndex reg_idx, ThreadID tid)
|
|
{
|
|
DPRINTF(FloatRegs, "[tid:%i]: Reading Float Reg %i as %x, %08f\n",
|
|
tid, reg_idx, floatRegs.i[tid][reg_idx], floatRegs.f[tid][reg_idx]);
|
|
|
|
return floatRegs.f[tid][reg_idx];
|
|
}
|
|
|
|
FloatRegBits
|
|
InOrderCPU::readFloatRegBits(RegIndex reg_idx, ThreadID tid)
|
|
{
|
|
DPRINTF(FloatRegs, "[tid:%i]: Reading Float Reg %i as %x, %08f\n",
|
|
tid, reg_idx, floatRegs.i[tid][reg_idx], floatRegs.f[tid][reg_idx]);
|
|
|
|
return floatRegs.i[tid][reg_idx];
|
|
}
|
|
|
|
void
|
|
InOrderCPU::setIntReg(RegIndex reg_idx, uint64_t val, ThreadID tid)
|
|
{
|
|
if (reg_idx == TheISA::ZeroReg) {
|
|
DPRINTF(IntRegs, "[tid:%i]: Ignoring Setting of ISA-ZeroReg "
|
|
"(Int. Reg %i) to %x\n", tid, reg_idx, val);
|
|
return;
|
|
} else {
|
|
DPRINTF(IntRegs, "[tid:%i]: Setting Int. Reg %i to %x\n",
|
|
tid, reg_idx, val);
|
|
|
|
intRegs[tid][reg_idx] = val;
|
|
}
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::setFloatReg(RegIndex reg_idx, FloatReg val, ThreadID tid)
|
|
{
|
|
floatRegs.f[tid][reg_idx] = val;
|
|
DPRINTF(FloatRegs, "[tid:%i]: Setting Float. Reg %i bits to "
|
|
"%x, %08f\n",
|
|
tid, reg_idx,
|
|
floatRegs.i[tid][reg_idx],
|
|
floatRegs.f[tid][reg_idx]);
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::setFloatRegBits(RegIndex reg_idx, FloatRegBits val, ThreadID tid)
|
|
{
|
|
floatRegs.i[tid][reg_idx] = val;
|
|
DPRINTF(FloatRegs, "[tid:%i]: Setting Float. Reg %i bits to "
|
|
"%x, %08f\n",
|
|
tid, reg_idx,
|
|
floatRegs.i[tid][reg_idx],
|
|
floatRegs.f[tid][reg_idx]);
|
|
}
|
|
|
|
uint64_t
|
|
InOrderCPU::readRegOtherThread(unsigned reg_idx, ThreadID tid)
|
|
{
|
|
// If Default value is set, then retrieve target thread
|
|
if (tid == InvalidThreadID) {
|
|
tid = TheISA::getTargetThread(tcBase(tid));
|
|
}
|
|
|
|
if (reg_idx < FP_Base_DepTag) {
|
|
// Integer Register File
|
|
return readIntReg(reg_idx, tid);
|
|
} else if (reg_idx < Ctrl_Base_DepTag) {
|
|
// Float Register File
|
|
reg_idx -= FP_Base_DepTag;
|
|
return readFloatRegBits(reg_idx, tid);
|
|
} else {
|
|
reg_idx -= Ctrl_Base_DepTag;
|
|
return readMiscReg(reg_idx, tid); // Misc. Register File
|
|
}
|
|
}
|
|
void
|
|
InOrderCPU::setRegOtherThread(unsigned reg_idx, const MiscReg &val,
|
|
ThreadID tid)
|
|
{
|
|
// If Default value is set, then retrieve target thread
|
|
if (tid == InvalidThreadID) {
|
|
tid = TheISA::getTargetThread(tcBase(tid));
|
|
}
|
|
|
|
if (reg_idx < FP_Base_DepTag) { // Integer Register File
|
|
setIntReg(reg_idx, val, tid);
|
|
} else if (reg_idx < Ctrl_Base_DepTag) { // Float Register File
|
|
reg_idx -= FP_Base_DepTag;
|
|
setFloatRegBits(reg_idx, val, tid);
|
|
} else {
|
|
reg_idx -= Ctrl_Base_DepTag;
|
|
setMiscReg(reg_idx, val, tid); // Misc. Register File
|
|
}
|
|
}
|
|
|
|
MiscReg
|
|
InOrderCPU::readMiscRegNoEffect(int misc_reg, ThreadID tid)
|
|
{
|
|
return isa[tid].readMiscRegNoEffect(misc_reg);
|
|
}
|
|
|
|
MiscReg
|
|
InOrderCPU::readMiscReg(int misc_reg, ThreadID tid)
|
|
{
|
|
return isa[tid].readMiscReg(misc_reg, tcBase(tid));
|
|
}
|
|
|
|
void
|
|
InOrderCPU::setMiscRegNoEffect(int misc_reg, const MiscReg &val, ThreadID tid)
|
|
{
|
|
isa[tid].setMiscRegNoEffect(misc_reg, val);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::setMiscReg(int misc_reg, const MiscReg &val, ThreadID tid)
|
|
{
|
|
isa[tid].setMiscReg(misc_reg, val, tcBase(tid));
|
|
}
|
|
|
|
|
|
InOrderCPU::ListIt
|
|
InOrderCPU::addInst(DynInstPtr inst)
|
|
{
|
|
ThreadID tid = inst->readTid();
|
|
|
|
instList[tid].push_back(inst);
|
|
|
|
return --(instList[tid].end());
|
|
}
|
|
|
|
InOrderCPU::ListIt
|
|
InOrderCPU::findInst(InstSeqNum seq_num, ThreadID tid)
|
|
{
|
|
ListIt it = instList[tid].begin();
|
|
ListIt end = instList[tid].end();
|
|
|
|
while (it != end) {
|
|
if ((*it)->seqNum == seq_num)
|
|
return it;
|
|
else if ((*it)->seqNum > seq_num)
|
|
break;
|
|
|
|
it++;
|
|
}
|
|
|
|
return instList[tid].end();
|
|
}
|
|
|
|
void
|
|
InOrderCPU::updateContextSwitchStats()
|
|
{
|
|
// Set Average Stat Here, then reset to 0
|
|
instsPerCtxtSwitch = instsPerSwitch;
|
|
instsPerSwitch = 0;
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::instDone(DynInstPtr inst, ThreadID tid)
|
|
{
|
|
// Set the nextPC to be fetched if this is the last instruction
|
|
// committed
|
|
// ========
|
|
// This contributes to the precise state of the CPU
|
|
// which can be used when restoring a thread to the CPU after after any
|
|
// type of context switching activity (fork, exception, etc.)
|
|
TheISA::PCState comm_pc = inst->pcState();
|
|
lastCommittedPC[tid] = comm_pc;
|
|
TheISA::advancePC(comm_pc, inst->staticInst);
|
|
pcState(comm_pc, tid);
|
|
|
|
//@todo: may be unnecessary with new-ISA-specific branch handling code
|
|
if (inst->isControl()) {
|
|
thread[tid]->lastGradIsBranch = true;
|
|
thread[tid]->lastBranchPC = inst->pcState();
|
|
TheISA::advancePC(thread[tid]->lastBranchPC, inst->staticInst);
|
|
} else {
|
|
thread[tid]->lastGradIsBranch = false;
|
|
}
|
|
|
|
|
|
// Finalize Trace Data For Instruction
|
|
if (inst->traceData) {
|
|
//inst->traceData->setCycle(curTick());
|
|
inst->traceData->setFetchSeq(inst->seqNum);
|
|
//inst->traceData->setCPSeq(cpu->tcBase(tid)->numInst);
|
|
inst->traceData->dump();
|
|
delete inst->traceData;
|
|
inst->traceData = NULL;
|
|
}
|
|
|
|
// Increment active thread's instruction count
|
|
instsPerSwitch++;
|
|
|
|
// Increment thread-state's instruction count
|
|
thread[tid]->numInst++;
|
|
thread[tid]->numOp++;
|
|
|
|
// Increment thread-state's instruction stats
|
|
thread[tid]->numInsts++;
|
|
thread[tid]->numOps++;
|
|
|
|
// Count committed insts per thread stats
|
|
if (!inst->isMicroop() || inst->isLastMicroop()) {
|
|
committedInsts[tid]++;
|
|
|
|
// Count total insts committed stat
|
|
totalCommittedInsts++;
|
|
}
|
|
|
|
committedOps[tid]++;
|
|
|
|
// Count SMT-committed insts per thread stat
|
|
if (numActiveThreads() > 1) {
|
|
if (!inst->isMicroop() || inst->isLastMicroop())
|
|
smtCommittedInsts[tid]++;
|
|
}
|
|
|
|
// Instruction-Mix Stats
|
|
if (inst->isLoad()) {
|
|
comLoads++;
|
|
} else if (inst->isStore()) {
|
|
comStores++;
|
|
} else if (inst->isControl()) {
|
|
comBranches++;
|
|
} else if (inst->isNop()) {
|
|
comNops++;
|
|
} else if (inst->isNonSpeculative()) {
|
|
comNonSpec++;
|
|
} else if (inst->isInteger()) {
|
|
comInts++;
|
|
} else if (inst->isFloating()) {
|
|
comFloats++;
|
|
}
|
|
|
|
// Check for instruction-count-based events.
|
|
comInstEventQueue[tid]->serviceEvents(thread[tid]->numOp);
|
|
|
|
// Finally, remove instruction from CPU
|
|
removeInst(inst);
|
|
}
|
|
|
|
// currently unused function, but substitute repetitive code w/this function
|
|
// call
|
|
void
|
|
InOrderCPU::addToRemoveList(DynInstPtr inst)
|
|
{
|
|
removeInstsThisCycle = true;
|
|
if (!inst->isRemoveList()) {
|
|
DPRINTF(InOrderCPU, "Pushing instruction [tid:%i] PC %s "
|
|
"[sn:%lli] to remove list\n",
|
|
inst->threadNumber, inst->pcState(), inst->seqNum);
|
|
inst->setRemoveList();
|
|
removeList.push(inst->getInstListIt());
|
|
} else {
|
|
DPRINTF(InOrderCPU, "Ignoring instruction removal for [tid:%i] PC %s "
|
|
"[sn:%lli], already remove list\n",
|
|
inst->threadNumber, inst->pcState(), inst->seqNum);
|
|
}
|
|
|
|
}
|
|
|
|
void
|
|
InOrderCPU::removeInst(DynInstPtr inst)
|
|
{
|
|
DPRINTF(InOrderCPU, "Removing graduated instruction [tid:%i] PC %s "
|
|
"[sn:%lli]\n",
|
|
inst->threadNumber, inst->pcState(), inst->seqNum);
|
|
|
|
removeInstsThisCycle = true;
|
|
|
|
// Remove the instruction.
|
|
if (!inst->isRemoveList()) {
|
|
DPRINTF(InOrderCPU, "Pushing instruction [tid:%i] PC %s "
|
|
"[sn:%lli] to remove list\n",
|
|
inst->threadNumber, inst->pcState(), inst->seqNum);
|
|
inst->setRemoveList();
|
|
removeList.push(inst->getInstListIt());
|
|
} else {
|
|
DPRINTF(InOrderCPU, "Ignoring instruction removal for [tid:%i] PC %s "
|
|
"[sn:%lli], already on remove list\n",
|
|
inst->threadNumber, inst->pcState(), inst->seqNum);
|
|
}
|
|
|
|
}
|
|
|
|
void
|
|
InOrderCPU::removeInstsUntil(const InstSeqNum &seq_num, ThreadID tid)
|
|
{
|
|
//assert(!instList[tid].empty());
|
|
|
|
removeInstsThisCycle = true;
|
|
|
|
ListIt inst_iter = instList[tid].end();
|
|
|
|
inst_iter--;
|
|
|
|
DPRINTF(InOrderCPU, "Squashing instructions from CPU instruction "
|
|
"list that are from [tid:%i] and above [sn:%lli] (end=%lli).\n",
|
|
tid, seq_num, (*inst_iter)->seqNum);
|
|
|
|
while ((*inst_iter)->seqNum > seq_num) {
|
|
|
|
bool break_loop = (inst_iter == instList[tid].begin());
|
|
|
|
squashInstIt(inst_iter, tid);
|
|
|
|
inst_iter--;
|
|
|
|
if (break_loop)
|
|
break;
|
|
}
|
|
}
|
|
|
|
|
|
inline void
|
|
InOrderCPU::squashInstIt(const ListIt inst_it, ThreadID tid)
|
|
{
|
|
DynInstPtr inst = (*inst_it);
|
|
if (inst->threadNumber == tid) {
|
|
DPRINTF(InOrderCPU, "Squashing instruction, "
|
|
"[tid:%i] [sn:%lli] PC %s\n",
|
|
inst->threadNumber,
|
|
inst->seqNum,
|
|
inst->pcState());
|
|
|
|
inst->setSquashed();
|
|
archRegDepMap[tid].remove(inst);
|
|
|
|
if (!inst->isRemoveList()) {
|
|
DPRINTF(InOrderCPU, "Pushing instruction [tid:%i] PC %s "
|
|
"[sn:%lli] to remove list\n",
|
|
inst->threadNumber, inst->pcState(),
|
|
inst->seqNum);
|
|
inst->setRemoveList();
|
|
removeList.push(inst_it);
|
|
} else {
|
|
DPRINTF(InOrderCPU, "Ignoring instruction removal for [tid:%i]"
|
|
" PC %s [sn:%lli], already on remove list\n",
|
|
inst->threadNumber, inst->pcState(),
|
|
inst->seqNum);
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::cleanUpRemovedInsts()
|
|
{
|
|
while (!removeList.empty()) {
|
|
DPRINTF(InOrderCPU, "Removing instruction, "
|
|
"[tid:%i] [sn:%lli] PC %s\n",
|
|
(*removeList.front())->threadNumber,
|
|
(*removeList.front())->seqNum,
|
|
(*removeList.front())->pcState());
|
|
|
|
DynInstPtr inst = *removeList.front();
|
|
ThreadID tid = inst->threadNumber;
|
|
|
|
// Remove From Register Dependency Map, If Necessary
|
|
// archRegDepMap[tid].remove(inst);
|
|
|
|
// Clear if Non-Speculative
|
|
if (inst->staticInst &&
|
|
inst->seqNum == nonSpecSeqNum[tid] &&
|
|
nonSpecInstActive[tid] == true) {
|
|
nonSpecInstActive[tid] = false;
|
|
}
|
|
|
|
inst->onInstList = false;
|
|
|
|
instList[tid].erase(removeList.front());
|
|
|
|
removeList.pop();
|
|
}
|
|
|
|
removeInstsThisCycle = false;
|
|
}
|
|
|
|
void
|
|
InOrderCPU::cleanUpRemovedEvents()
|
|
{
|
|
while (!cpuEventRemoveList.empty()) {
|
|
Event *cpu_event = cpuEventRemoveList.front();
|
|
cpuEventRemoveList.pop();
|
|
delete cpu_event;
|
|
}
|
|
}
|
|
|
|
|
|
void
|
|
InOrderCPU::dumpInsts()
|
|
{
|
|
int num = 0;
|
|
|
|
ListIt inst_list_it = instList[0].begin();
|
|
|
|
cprintf("Dumping Instruction List\n");
|
|
|
|
while (inst_list_it != instList[0].end()) {
|
|
cprintf("Instruction:%i\nPC:%s\n[tid:%i]\n[sn:%lli]\nIssued:%i\n"
|
|
"Squashed:%i\n\n",
|
|
num, (*inst_list_it)->pcState(),
|
|
(*inst_list_it)->threadNumber,
|
|
(*inst_list_it)->seqNum, (*inst_list_it)->isIssued(),
|
|
(*inst_list_it)->isSquashed());
|
|
inst_list_it++;
|
|
++num;
|
|
}
|
|
}
|
|
|
|
void
|
|
InOrderCPU::wakeCPU()
|
|
{
|
|
if (/*activityRec.active() || */tickEvent.scheduled()) {
|
|
DPRINTF(Activity, "CPU already running.\n");
|
|
return;
|
|
}
|
|
|
|
DPRINTF(Activity, "Waking up CPU\n");
|
|
|
|
Tick extra_cycles = tickToCycles((curTick() - 1) - lastRunningCycle);
|
|
|
|
idleCycles += extra_cycles;
|
|
for (int stage_num = 0; stage_num < NumStages; stage_num++) {
|
|
pipelineStage[stage_num]->idleCycles += extra_cycles;
|
|
}
|
|
|
|
numCycles += extra_cycles;
|
|
|
|
schedule(&tickEvent, nextCycle(curTick()));
|
|
}
|
|
|
|
// Lots of copied full system code...place into BaseCPU class?
|
|
void
|
|
InOrderCPU::wakeup()
|
|
{
|
|
if (thread[0]->status() != ThreadContext::Suspended)
|
|
return;
|
|
|
|
wakeCPU();
|
|
|
|
DPRINTF(Quiesce, "Suspended Processor woken\n");
|
|
threadContexts[0]->activate();
|
|
}
|
|
|
|
void
|
|
InOrderCPU::syscallContext(Fault fault, ThreadID tid, DynInstPtr inst, int delay)
|
|
{
|
|
// Syscall must be non-speculative, so squash from last stage
|
|
unsigned squash_stage = NumStages - 1;
|
|
inst->setSquashInfo(squash_stage);
|
|
|
|
// Squash In Pipeline Stage
|
|
pipelineStage[squash_stage]->setupSquash(inst, tid);
|
|
|
|
// Schedule Squash Through-out Resource Pool
|
|
resPool->scheduleEvent(
|
|
(InOrderCPU::CPUEventType)ResourcePool::SquashAll, inst, 0);
|
|
scheduleCpuEvent(Syscall, fault, tid, inst, delay, Syscall_Pri);
|
|
}
|
|
|
|
void
|
|
InOrderCPU::syscall(int64_t callnum, ThreadID tid)
|
|
{
|
|
DPRINTF(InOrderCPU, "[tid:%i] Executing syscall().\n\n", tid);
|
|
|
|
DPRINTF(Activity,"Activity: syscall() called.\n");
|
|
|
|
// Temporarily increase this by one to account for the syscall
|
|
// instruction.
|
|
++(this->thread[tid]->funcExeInst);
|
|
|
|
// Execute the actual syscall.
|
|
this->thread[tid]->syscall(callnum);
|
|
|
|
// Decrease funcExeInst by one as the normal commit will handle
|
|
// incrementing it.
|
|
--(this->thread[tid]->funcExeInst);
|
|
|
|
// Clear Non-Speculative Block Variable
|
|
nonSpecInstActive[tid] = false;
|
|
}
|
|
|
|
TheISA::TLB*
|
|
InOrderCPU::getITBPtr()
|
|
{
|
|
CacheUnit *itb_res = resPool->getInstUnit();
|
|
return itb_res->tlb();
|
|
}
|
|
|
|
|
|
TheISA::TLB*
|
|
InOrderCPU::getDTBPtr()
|
|
{
|
|
return resPool->getDataUnit()->tlb();
|
|
}
|
|
|
|
Decoder *
|
|
InOrderCPU::getDecoderPtr()
|
|
{
|
|
return &resPool->getInstUnit()->decoder;
|
|
}
|
|
|
|
Fault
|
|
InOrderCPU::read(DynInstPtr inst, Addr addr,
|
|
uint8_t *data, unsigned size, unsigned flags)
|
|
{
|
|
return resPool->getDataUnit()->read(inst, addr, data, size, flags);
|
|
}
|
|
|
|
Fault
|
|
InOrderCPU::write(DynInstPtr inst, uint8_t *data, unsigned size,
|
|
Addr addr, unsigned flags, uint64_t *write_res)
|
|
{
|
|
return resPool->getDataUnit()->write(inst, data, size, addr, flags,
|
|
write_res);
|
|
}
|