2009-05-11 19:38:43 +02:00
|
|
|
/*
|
2012-01-11 20:48:48 +01:00
|
|
|
* Copyright (c) 1999-2011 Mark D. Hill and David A. Wood
|
2009-05-11 19:38:43 +02:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions are
|
|
|
|
* met: redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer;
|
|
|
|
* redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution;
|
|
|
|
* neither the name of the copyright holders nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived from
|
|
|
|
* this software without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
2012-01-11 20:48:48 +01:00
|
|
|
#include <fcntl.h>
|
|
|
|
#include <zlib.h>
|
|
|
|
|
|
|
|
#include <cstdio>
|
|
|
|
|
2010-04-02 20:20:32 +02:00
|
|
|
#include "base/intmath.hh"
|
2012-11-02 18:18:25 +01:00
|
|
|
#include "base/statistics.hh"
|
2012-04-06 22:47:07 +02:00
|
|
|
#include "debug/RubyCacheTrace.hh"
|
2012-10-02 21:35:42 +02:00
|
|
|
#include "debug/RubySystem.hh"
|
2009-07-07 00:49:47 +02:00
|
|
|
#include "mem/ruby/common/Address.hh"
|
2009-05-11 19:38:45 +02:00
|
|
|
#include "mem/ruby/network/Network.hh"
|
2010-03-23 02:43:53 +01:00
|
|
|
#include "mem/ruby/profiler/Profiler.hh"
|
|
|
|
#include "mem/ruby/system/System.hh"
|
2012-04-06 22:47:07 +02:00
|
|
|
#include "sim/eventq.hh"
|
2012-01-11 20:48:48 +01:00
|
|
|
#include "sim/simulate.hh"
|
2009-07-07 00:49:47 +02:00
|
|
|
|
2010-04-02 20:20:32 +02:00
|
|
|
using namespace std;
|
|
|
|
|
2009-07-07 00:49:47 +02:00
|
|
|
int RubySystem::m_random_seed;
|
|
|
|
bool RubySystem::m_randomization;
|
|
|
|
int RubySystem::m_block_size_bytes;
|
|
|
|
int RubySystem::m_block_size_bits;
|
|
|
|
uint64 RubySystem::m_memory_size_bytes;
|
|
|
|
int RubySystem::m_memory_size_bits;
|
|
|
|
|
2010-01-30 05:29:17 +01:00
|
|
|
RubySystem::RubySystem(const Params *p)
|
2012-09-10 19:21:01 +02:00
|
|
|
: ClockedObject(p)
|
2009-05-11 19:38:46 +02:00
|
|
|
{
|
2010-01-30 05:29:17 +01:00
|
|
|
if (g_system_ptr != NULL)
|
|
|
|
fatal("Only one RubySystem object currently allowed.\n");
|
|
|
|
|
|
|
|
m_random_seed = p->random_seed;
|
|
|
|
srandom(m_random_seed);
|
|
|
|
m_randomization = p->randomization;
|
2010-01-30 05:29:18 +01:00
|
|
|
|
2010-01-30 05:29:17 +01:00
|
|
|
m_block_size_bytes = p->block_size_bytes;
|
2010-04-02 20:20:32 +02:00
|
|
|
assert(isPowerOf2(m_block_size_bytes));
|
|
|
|
m_block_size_bits = floorLog2(m_block_size_bytes);
|
2010-01-30 05:29:18 +01:00
|
|
|
|
2010-01-30 05:29:23 +01:00
|
|
|
m_memory_size_bytes = p->mem_size;
|
2010-03-22 05:22:21 +01:00
|
|
|
if (m_memory_size_bytes == 0) {
|
|
|
|
m_memory_size_bits = 0;
|
|
|
|
} else {
|
2013-01-31 16:44:20 +01:00
|
|
|
m_memory_size_bits = ceilLog2(m_memory_size_bytes);
|
2010-03-22 05:22:21 +01:00
|
|
|
}
|
2010-01-30 05:29:18 +01:00
|
|
|
|
2010-03-22 05:22:21 +01:00
|
|
|
if (p->no_mem_vec) {
|
|
|
|
m_mem_vec_ptr = NULL;
|
|
|
|
} else {
|
|
|
|
m_mem_vec_ptr = new MemoryVector;
|
2010-06-11 08:17:07 +02:00
|
|
|
m_mem_vec_ptr->resize(m_memory_size_bytes);
|
2010-03-22 05:22:21 +01:00
|
|
|
}
|
2010-01-30 05:29:21 +01:00
|
|
|
|
2012-11-02 18:18:25 +01:00
|
|
|
// Print ruby configuration and stats at exit and when asked for
|
|
|
|
Stats::registerDumpCallback(new RubyDumpStatsCallback(p->stats_filename,
|
|
|
|
this));
|
2012-10-02 21:35:45 +02:00
|
|
|
|
2012-01-11 20:48:48 +01:00
|
|
|
m_warmup_enabled = false;
|
|
|
|
m_cooldown_enabled = false;
|
2013-02-11 04:26:22 +01:00
|
|
|
|
|
|
|
// Setup the global variables used in Ruby
|
|
|
|
g_system_ptr = this;
|
|
|
|
|
|
|
|
// Resize to the size of different machine types
|
|
|
|
g_abs_controls.resize(MachineType_NUM);
|
2009-05-11 19:38:46 +02:00
|
|
|
}
|
|
|
|
|
2010-03-23 02:43:53 +01:00
|
|
|
void
|
|
|
|
RubySystem::init()
|
2009-05-11 19:38:43 +02:00
|
|
|
{
|
2010-03-23 02:43:53 +01:00
|
|
|
m_profiler_ptr->clearStats();
|
2012-11-02 18:18:25 +01:00
|
|
|
m_network_ptr->clearStats();
|
2009-05-11 19:38:46 +02:00
|
|
|
}
|
2009-05-11 19:38:43 +02:00
|
|
|
|
2011-07-01 02:49:26 +02:00
|
|
|
void
|
|
|
|
RubySystem::registerNetwork(Network* network_ptr)
|
|
|
|
{
|
|
|
|
m_network_ptr = network_ptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
RubySystem::registerProfiler(Profiler* profiler_ptr)
|
|
|
|
{
|
|
|
|
m_profiler_ptr = profiler_ptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
RubySystem::registerAbstractController(AbstractController* cntrl)
|
|
|
|
{
|
|
|
|
m_abs_cntrl_vec.push_back(cntrl);
|
2013-02-11 04:26:22 +01:00
|
|
|
|
|
|
|
MachineID id = cntrl->getMachineID();
|
|
|
|
g_abs_controls[id.getType()][id.getNum()] = cntrl;
|
2011-07-01 02:49:26 +02:00
|
|
|
}
|
|
|
|
|
2012-01-11 20:48:48 +01:00
|
|
|
void
|
|
|
|
RubySystem::registerSparseMemory(SparseMemory* s)
|
|
|
|
{
|
|
|
|
m_sparse_memory_vector.push_back(s);
|
|
|
|
}
|
|
|
|
|
2012-07-11 07:51:54 +02:00
|
|
|
void
|
|
|
|
RubySystem::registerMemController(MemoryControl *mc) {
|
2012-10-16 00:27:17 +02:00
|
|
|
m_memory_controller_vec.push_back(mc);
|
2012-07-11 07:51:54 +02:00
|
|
|
}
|
|
|
|
|
2009-07-07 00:49:47 +02:00
|
|
|
RubySystem::~RubySystem()
|
2009-05-11 19:38:46 +02:00
|
|
|
{
|
2010-03-23 02:43:53 +01:00
|
|
|
delete m_network_ptr;
|
|
|
|
delete m_profiler_ptr;
|
|
|
|
if (m_mem_vec_ptr)
|
|
|
|
delete m_mem_vec_ptr;
|
2009-07-07 00:49:47 +02:00
|
|
|
}
|
|
|
|
|
2010-03-23 02:43:53 +01:00
|
|
|
void
|
|
|
|
RubySystem::printStats(ostream& out)
|
2009-05-11 19:38:43 +02:00
|
|
|
{
|
2010-03-23 02:43:53 +01:00
|
|
|
const time_t T = time(NULL);
|
|
|
|
tm *localTime = localtime(&T);
|
|
|
|
char buf[100];
|
|
|
|
strftime(buf, 100, "%b/%d/%Y %H:%M:%S", localTime);
|
2009-07-07 00:49:47 +02:00
|
|
|
|
2010-03-23 02:43:53 +01:00
|
|
|
out << "Real time: " << buf << endl;
|
2009-05-11 19:38:43 +02:00
|
|
|
|
2010-03-23 02:43:53 +01:00
|
|
|
m_profiler_ptr->printStats(out);
|
|
|
|
m_network_ptr->printStats(out);
|
2013-02-11 04:26:22 +01:00
|
|
|
|
|
|
|
for (uint32_t i = 0;i < g_abs_controls.size(); ++i) {
|
|
|
|
for (map<uint32_t, AbstractController *>::iterator it =
|
|
|
|
g_abs_controls[i].begin();
|
|
|
|
it != g_abs_controls[i].end(); ++it) {
|
|
|
|
|
|
|
|
((*it).second)->printStats(out);
|
|
|
|
}
|
|
|
|
}
|
2009-05-11 19:38:43 +02:00
|
|
|
}
|
|
|
|
|
2012-01-11 20:48:48 +01:00
|
|
|
void
|
2012-09-11 16:23:56 +02:00
|
|
|
RubySystem::writeCompressedTrace(uint8_t *raw_data, string filename,
|
2012-01-11 20:48:48 +01:00
|
|
|
uint64 uncompressed_trace_size)
|
|
|
|
{
|
|
|
|
// Create the checkpoint file for the memory
|
|
|
|
string thefile = Checkpoint::dir() + "/" + filename.c_str();
|
|
|
|
|
|
|
|
int fd = creat(thefile.c_str(), 0664);
|
|
|
|
if (fd < 0) {
|
|
|
|
perror("creat");
|
|
|
|
fatal("Can't open memory trace file '%s'\n", filename);
|
|
|
|
}
|
|
|
|
|
|
|
|
gzFile compressedMemory = gzdopen(fd, "wb");
|
|
|
|
if (compressedMemory == NULL)
|
|
|
|
fatal("Insufficient memory to allocate compression state for %s\n",
|
|
|
|
filename);
|
|
|
|
|
|
|
|
if (gzwrite(compressedMemory, raw_data, uncompressed_trace_size) !=
|
|
|
|
uncompressed_trace_size) {
|
|
|
|
fatal("Write failed on memory trace file '%s'\n", filename);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (gzclose(compressedMemory)) {
|
|
|
|
fatal("Close failed on memory trace file '%s'\n", filename);
|
|
|
|
}
|
|
|
|
delete raw_data;
|
|
|
|
}
|
|
|
|
|
2010-08-20 20:46:13 +02:00
|
|
|
void
|
|
|
|
RubySystem::serialize(std::ostream &os)
|
|
|
|
{
|
2012-01-11 20:48:48 +01:00
|
|
|
m_cooldown_enabled = true;
|
|
|
|
|
|
|
|
vector<Sequencer*> sequencer_map;
|
|
|
|
Sequencer* sequencer_ptr = NULL;
|
|
|
|
int cntrl_id = -1;
|
|
|
|
|
|
|
|
|
|
|
|
for (int cntrl = 0; cntrl < m_abs_cntrl_vec.size(); cntrl++) {
|
|
|
|
sequencer_map.push_back(m_abs_cntrl_vec[cntrl]->getSequencer());
|
|
|
|
if (sequencer_ptr == NULL) {
|
|
|
|
sequencer_ptr = sequencer_map[cntrl];
|
|
|
|
cntrl_id = cntrl;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(sequencer_ptr != NULL);
|
|
|
|
|
|
|
|
for (int cntrl = 0; cntrl < m_abs_cntrl_vec.size(); cntrl++) {
|
|
|
|
if (sequencer_map[cntrl] == NULL) {
|
|
|
|
sequencer_map[cntrl] = sequencer_ptr;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-04-06 22:47:07 +02:00
|
|
|
DPRINTF(RubyCacheTrace, "Recording Cache Trace\n");
|
2012-01-11 20:48:48 +01:00
|
|
|
// Create the CacheRecorder and record the cache trace
|
|
|
|
m_cache_recorder = new CacheRecorder(NULL, 0, sequencer_map);
|
|
|
|
|
|
|
|
for (int cntrl = 0; cntrl < m_abs_cntrl_vec.size(); cntrl++) {
|
|
|
|
m_abs_cntrl_vec[cntrl]->recordCacheTrace(cntrl, m_cache_recorder);
|
|
|
|
}
|
|
|
|
|
2012-04-06 22:47:07 +02:00
|
|
|
DPRINTF(RubyCacheTrace, "Cache Trace Complete\n");
|
2012-01-11 20:48:48 +01:00
|
|
|
// save the current tick value
|
|
|
|
Tick curtick_original = curTick();
|
|
|
|
// save the event queue head
|
|
|
|
Event* eventq_head = eventq->replaceHead(NULL);
|
2012-04-06 22:47:07 +02:00
|
|
|
DPRINTF(RubyCacheTrace, "Recording current tick %ld and event queue\n",
|
|
|
|
curtick_original);
|
2012-01-11 20:48:48 +01:00
|
|
|
|
|
|
|
// Schedule an event to start cache cooldown
|
2012-04-06 22:47:07 +02:00
|
|
|
DPRINTF(RubyCacheTrace, "Starting cache flush\n");
|
|
|
|
enqueueRubyEvent(curTick());
|
2012-01-11 20:48:48 +01:00
|
|
|
simulate();
|
2012-04-06 22:47:07 +02:00
|
|
|
DPRINTF(RubyCacheTrace, "Cache flush complete\n");
|
2012-01-11 20:48:48 +01:00
|
|
|
|
|
|
|
// Restore eventq head
|
|
|
|
eventq_head = eventq->replaceHead(eventq_head);
|
|
|
|
// Restore curTick
|
2012-11-16 17:27:47 +01:00
|
|
|
setCurTick(curtick_original);
|
2010-08-20 20:46:13 +02:00
|
|
|
|
2012-09-11 16:23:56 +02:00
|
|
|
uint8_t *raw_data = NULL;
|
2012-01-11 20:48:48 +01:00
|
|
|
|
|
|
|
if (m_mem_vec_ptr != NULL) {
|
|
|
|
uint64 memory_trace_size = m_mem_vec_ptr->collatePages(raw_data);
|
|
|
|
|
|
|
|
string memory_trace_file = name() + ".memory.gz";
|
|
|
|
writeCompressedTrace(raw_data, memory_trace_file,
|
|
|
|
memory_trace_size);
|
|
|
|
|
|
|
|
SERIALIZE_SCALAR(memory_trace_file);
|
|
|
|
SERIALIZE_SCALAR(memory_trace_size);
|
|
|
|
|
|
|
|
} else {
|
|
|
|
for (int i = 0; i < m_sparse_memory_vector.size(); ++i) {
|
|
|
|
m_sparse_memory_vector[i]->recordBlocks(cntrl_id,
|
|
|
|
m_cache_recorder);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Aggergate the trace entries together into a single array
|
|
|
|
raw_data = new uint8_t[4096];
|
|
|
|
uint64 cache_trace_size = m_cache_recorder->aggregateRecords(&raw_data,
|
|
|
|
4096);
|
|
|
|
string cache_trace_file = name() + ".cache.gz";
|
|
|
|
writeCompressedTrace(raw_data, cache_trace_file, cache_trace_size);
|
|
|
|
|
|
|
|
SERIALIZE_SCALAR(cache_trace_file);
|
|
|
|
SERIALIZE_SCALAR(cache_trace_size);
|
|
|
|
|
|
|
|
m_cooldown_enabled = false;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2012-09-11 16:23:56 +02:00
|
|
|
RubySystem::readCompressedTrace(string filename, uint8_t *&raw_data,
|
2012-01-11 20:48:48 +01:00
|
|
|
uint64& uncompressed_trace_size)
|
|
|
|
{
|
|
|
|
// Read the trace file
|
|
|
|
gzFile compressedTrace;
|
|
|
|
|
|
|
|
// trace file
|
|
|
|
int fd = open(filename.c_str(), O_RDONLY);
|
|
|
|
if (fd < 0) {
|
|
|
|
perror("open");
|
|
|
|
fatal("Unable to open trace file %s", filename);
|
|
|
|
}
|
|
|
|
|
|
|
|
compressedTrace = gzdopen(fd, "rb");
|
|
|
|
if (compressedTrace == NULL) {
|
|
|
|
fatal("Insufficient memory to allocate compression state for %s\n",
|
|
|
|
filename);
|
|
|
|
}
|
|
|
|
|
|
|
|
raw_data = new uint8_t[uncompressed_trace_size];
|
|
|
|
if (gzread(compressedTrace, raw_data, uncompressed_trace_size) <
|
|
|
|
uncompressed_trace_size) {
|
|
|
|
fatal("Unable to read complete trace from file %s\n", filename);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (gzclose(compressedTrace)) {
|
|
|
|
fatal("Failed to close cache trace file '%s'\n", filename);
|
|
|
|
}
|
2010-08-20 20:46:13 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
RubySystem::unserialize(Checkpoint *cp, const string §ion)
|
|
|
|
{
|
|
|
|
//
|
|
|
|
// The main purpose for clearing stats in the unserialize process is so
|
|
|
|
// that the profiler can correctly set its start time to the unserialized
|
2011-01-08 06:50:29 +01:00
|
|
|
// value of curTick()
|
2010-08-20 20:46:13 +02:00
|
|
|
//
|
2012-11-02 18:18:25 +01:00
|
|
|
resetStats();
|
2012-09-11 16:23:56 +02:00
|
|
|
uint8_t *uncompressed_trace = NULL;
|
2012-01-11 20:48:48 +01:00
|
|
|
|
|
|
|
if (m_mem_vec_ptr != NULL) {
|
|
|
|
string memory_trace_file;
|
|
|
|
uint64 memory_trace_size = 0;
|
|
|
|
|
|
|
|
UNSERIALIZE_SCALAR(memory_trace_file);
|
|
|
|
UNSERIALIZE_SCALAR(memory_trace_size);
|
|
|
|
memory_trace_file = cp->cptDir + "/" + memory_trace_file;
|
|
|
|
|
|
|
|
readCompressedTrace(memory_trace_file, uncompressed_trace,
|
|
|
|
memory_trace_size);
|
|
|
|
m_mem_vec_ptr->populatePages(uncompressed_trace);
|
|
|
|
|
|
|
|
delete uncompressed_trace;
|
|
|
|
uncompressed_trace = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
string cache_trace_file;
|
|
|
|
uint64 cache_trace_size = 0;
|
|
|
|
|
|
|
|
UNSERIALIZE_SCALAR(cache_trace_file);
|
|
|
|
UNSERIALIZE_SCALAR(cache_trace_size);
|
|
|
|
cache_trace_file = cp->cptDir + "/" + cache_trace_file;
|
|
|
|
|
|
|
|
readCompressedTrace(cache_trace_file, uncompressed_trace,
|
|
|
|
cache_trace_size);
|
|
|
|
m_warmup_enabled = true;
|
|
|
|
|
|
|
|
vector<Sequencer*> sequencer_map;
|
|
|
|
Sequencer* t = NULL;
|
|
|
|
for (int cntrl = 0; cntrl < m_abs_cntrl_vec.size(); cntrl++) {
|
|
|
|
sequencer_map.push_back(m_abs_cntrl_vec[cntrl]->getSequencer());
|
2012-10-02 21:35:42 +02:00
|
|
|
if (t == NULL) t = sequencer_map[cntrl];
|
2012-01-11 20:48:48 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
assert(t != NULL);
|
|
|
|
|
|
|
|
for (int cntrl = 0; cntrl < m_abs_cntrl_vec.size(); cntrl++) {
|
|
|
|
if (sequencer_map[cntrl] == NULL) {
|
|
|
|
sequencer_map[cntrl] = t;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
m_cache_recorder = new CacheRecorder(uncompressed_trace, cache_trace_size,
|
|
|
|
sequencer_map);
|
2010-08-20 20:46:13 +02:00
|
|
|
}
|
|
|
|
|
2010-03-23 02:43:53 +01:00
|
|
|
void
|
2012-01-11 20:48:48 +01:00
|
|
|
RubySystem::startup()
|
2009-05-11 19:38:43 +02:00
|
|
|
{
|
2012-01-11 20:48:48 +01:00
|
|
|
if (m_warmup_enabled) {
|
|
|
|
// save the current tick value
|
|
|
|
Tick curtick_original = curTick();
|
|
|
|
// save the event queue head
|
|
|
|
Event* eventq_head = eventq->replaceHead(NULL);
|
2012-10-16 00:27:15 +02:00
|
|
|
// set curTick to 0 and reset Ruby System's clock
|
2012-11-16 17:27:47 +01:00
|
|
|
setCurTick(0);
|
2012-10-16 00:27:15 +02:00
|
|
|
resetClock();
|
2012-01-11 20:48:48 +01:00
|
|
|
|
|
|
|
// Schedule an event to start cache warmup
|
2012-04-06 22:47:07 +02:00
|
|
|
enqueueRubyEvent(curTick());
|
2012-01-11 20:48:48 +01:00
|
|
|
simulate();
|
|
|
|
|
|
|
|
delete m_cache_recorder;
|
|
|
|
m_cache_recorder = NULL;
|
|
|
|
m_warmup_enabled = false;
|
2012-10-16 00:27:17 +02:00
|
|
|
|
2012-07-11 18:44:34 +02:00
|
|
|
// reset DRAM so that it's not waiting for events on the old event
|
|
|
|
// queue
|
2012-10-16 00:27:17 +02:00
|
|
|
for (int i = 0; i < m_memory_controller_vec.size(); ++i) {
|
|
|
|
m_memory_controller_vec[i]->reset();
|
|
|
|
}
|
|
|
|
|
2012-01-11 20:48:48 +01:00
|
|
|
// Restore eventq head
|
|
|
|
eventq_head = eventq->replaceHead(eventq_head);
|
2012-10-16 00:27:15 +02:00
|
|
|
// Restore curTick and Ruby System's clock
|
2012-11-16 17:27:47 +01:00
|
|
|
setCurTick(curtick_original);
|
2012-10-16 00:27:15 +02:00
|
|
|
resetClock();
|
2012-01-11 20:48:48 +01:00
|
|
|
}
|
2009-05-11 19:38:43 +02:00
|
|
|
}
|
|
|
|
|
2010-03-23 02:43:53 +01:00
|
|
|
void
|
2012-01-11 20:48:48 +01:00
|
|
|
RubySystem::RubyEvent::process()
|
2009-05-11 19:38:43 +02:00
|
|
|
{
|
2012-01-11 20:48:48 +01:00
|
|
|
if (ruby_system->m_warmup_enabled) {
|
|
|
|
ruby_system->m_cache_recorder->enqueueNextFetchRequest();
|
|
|
|
} else if (ruby_system->m_cooldown_enabled) {
|
|
|
|
ruby_system->m_cache_recorder->enqueueNextFlushRequest();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2012-11-02 18:18:25 +01:00
|
|
|
RubySystem::resetStats()
|
2012-01-11 20:48:48 +01:00
|
|
|
{
|
|
|
|
m_profiler_ptr->clearStats();
|
|
|
|
m_network_ptr->clearStats();
|
2013-02-11 04:26:22 +01:00
|
|
|
for (uint32_t cntrl = 0; cntrl < m_abs_cntrl_vec.size(); cntrl++) {
|
|
|
|
m_abs_cntrl_vec[cntrl]->clearStats();
|
|
|
|
}
|
2009-05-11 19:38:43 +02:00
|
|
|
}
|
|
|
|
|
2012-10-02 21:35:42 +02:00
|
|
|
bool
|
|
|
|
RubySystem::functionalRead(PacketPtr pkt)
|
|
|
|
{
|
|
|
|
Address address(pkt->getAddr());
|
|
|
|
Address line_address(address);
|
|
|
|
line_address.makeLineAddress();
|
|
|
|
|
|
|
|
AccessPermission access_perm = AccessPermission_NotPresent;
|
|
|
|
int num_controllers = m_abs_cntrl_vec.size();
|
|
|
|
|
|
|
|
DPRINTF(RubySystem, "Functional Read request for %s\n",address);
|
|
|
|
|
|
|
|
unsigned int num_ro = 0;
|
|
|
|
unsigned int num_rw = 0;
|
|
|
|
unsigned int num_busy = 0;
|
|
|
|
unsigned int num_backing_store = 0;
|
|
|
|
unsigned int num_invalid = 0;
|
|
|
|
|
|
|
|
// In this loop we count the number of controllers that have the given
|
|
|
|
// address in read only, read write and busy states.
|
2012-10-16 00:51:57 +02:00
|
|
|
for (unsigned int i = 0; i < num_controllers; ++i) {
|
2012-10-02 21:35:42 +02:00
|
|
|
access_perm = m_abs_cntrl_vec[i]-> getAccessPermission(line_address);
|
|
|
|
if (access_perm == AccessPermission_Read_Only)
|
|
|
|
num_ro++;
|
|
|
|
else if (access_perm == AccessPermission_Read_Write)
|
|
|
|
num_rw++;
|
|
|
|
else if (access_perm == AccessPermission_Busy)
|
|
|
|
num_busy++;
|
|
|
|
else if (access_perm == AccessPermission_Backing_Store)
|
|
|
|
// See RubySlicc_Exports.sm for details, but Backing_Store is meant
|
|
|
|
// to represent blocks in memory *for Broadcast/Snooping protocols*,
|
|
|
|
// where memory has no idea whether it has an exclusive copy of data
|
|
|
|
// or not.
|
|
|
|
num_backing_store++;
|
|
|
|
else if (access_perm == AccessPermission_Invalid ||
|
|
|
|
access_perm == AccessPermission_NotPresent)
|
|
|
|
num_invalid++;
|
|
|
|
}
|
|
|
|
assert(num_rw <= 1);
|
|
|
|
|
|
|
|
uint8_t *data = pkt->getPtr<uint8_t>(true);
|
|
|
|
unsigned int size_in_bytes = pkt->getSize();
|
|
|
|
unsigned startByte = address.getAddress() - line_address.getAddress();
|
|
|
|
|
|
|
|
// This if case is meant to capture what happens in a Broadcast/Snoop
|
|
|
|
// protocol where the block does not exist in the cache hierarchy. You
|
|
|
|
// only want to read from the Backing_Store memory if there is no copy in
|
|
|
|
// the cache hierarchy, otherwise you want to try to read the RO or RW
|
|
|
|
// copies existing in the cache hierarchy (covered by the else statement).
|
|
|
|
// The reason is because the Backing_Store memory could easily be stale, if
|
|
|
|
// there are copies floating around the cache hierarchy, so you want to read
|
|
|
|
// it only if it's not in the cache hierarchy at all.
|
|
|
|
if (num_invalid == (num_controllers - 1) &&
|
|
|
|
num_backing_store == 1) {
|
|
|
|
DPRINTF(RubySystem, "only copy in Backing_Store memory, read from it\n");
|
2012-10-16 00:51:57 +02:00
|
|
|
for (unsigned int i = 0; i < num_controllers; ++i) {
|
2012-10-02 21:35:42 +02:00
|
|
|
access_perm = m_abs_cntrl_vec[i]->getAccessPermission(line_address);
|
|
|
|
if (access_perm == AccessPermission_Backing_Store) {
|
|
|
|
DataBlock& block = m_abs_cntrl_vec[i]->
|
|
|
|
getDataBlock(line_address);
|
|
|
|
|
|
|
|
DPRINTF(RubySystem, "reading from %s block %s\n",
|
|
|
|
m_abs_cntrl_vec[i]->name(), block);
|
|
|
|
for (unsigned i = 0; i < size_in_bytes; ++i) {
|
|
|
|
data[i] = block.getByte(i + startByte);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
2012-10-16 00:51:57 +02:00
|
|
|
} else if (num_ro > 0 || num_rw == 1) {
|
2012-10-02 21:35:42 +02:00
|
|
|
// In Broadcast/Snoop protocols, this covers if you know the block
|
|
|
|
// exists somewhere in the caching hierarchy, then you want to read any
|
|
|
|
// valid RO or RW block. In directory protocols, same thing, you want
|
|
|
|
// to read any valid readable copy of the block.
|
|
|
|
DPRINTF(RubySystem, "num_busy = %d, num_ro = %d, num_rw = %d\n",
|
|
|
|
num_busy, num_ro, num_rw);
|
|
|
|
// In this loop, we try to figure which controller has a read only or
|
|
|
|
// a read write copy of the given address. Any valid copy would suffice
|
|
|
|
// for a functional read.
|
2012-10-16 00:51:57 +02:00
|
|
|
for (unsigned int i = 0;i < num_controllers;++i) {
|
2012-10-02 21:35:42 +02:00
|
|
|
access_perm = m_abs_cntrl_vec[i]->getAccessPermission(line_address);
|
|
|
|
if (access_perm == AccessPermission_Read_Only ||
|
|
|
|
access_perm == AccessPermission_Read_Write) {
|
|
|
|
DataBlock& block = m_abs_cntrl_vec[i]->
|
|
|
|
getDataBlock(line_address);
|
|
|
|
|
|
|
|
DPRINTF(RubySystem, "reading from %s block %s\n",
|
|
|
|
m_abs_cntrl_vec[i]->name(), block);
|
|
|
|
for (unsigned i = 0; i < size_in_bytes; ++i) {
|
|
|
|
data[i] = block.getByte(i + startByte);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2012-10-16 00:51:57 +02:00
|
|
|
|
2012-10-02 21:35:42 +02:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2012-10-16 00:51:57 +02:00
|
|
|
// The function searches through all the buffers that exist in different
|
|
|
|
// cache, directory and memory controllers, and in the network components
|
|
|
|
// and writes the data portion of those that hold the address specified
|
|
|
|
// in the packet.
|
2012-10-02 21:35:42 +02:00
|
|
|
bool
|
|
|
|
RubySystem::functionalWrite(PacketPtr pkt)
|
|
|
|
{
|
|
|
|
Address addr(pkt->getAddr());
|
|
|
|
Address line_addr = line_address(addr);
|
|
|
|
AccessPermission access_perm = AccessPermission_NotPresent;
|
|
|
|
int num_controllers = m_abs_cntrl_vec.size();
|
|
|
|
|
|
|
|
DPRINTF(RubySystem, "Functional Write request for %s\n",addr);
|
|
|
|
|
|
|
|
uint8_t *data = pkt->getPtr<uint8_t>(true);
|
|
|
|
unsigned int size_in_bytes = pkt->getSize();
|
|
|
|
unsigned startByte = addr.getAddress() - line_addr.getAddress();
|
|
|
|
|
2012-10-16 00:51:57 +02:00
|
|
|
for (unsigned int i = 0; i < num_controllers;++i) {
|
|
|
|
m_abs_cntrl_vec[i]->functionalWriteBuffers(pkt);
|
2012-10-02 21:35:42 +02:00
|
|
|
|
2012-10-16 00:51:57 +02:00
|
|
|
access_perm = m_abs_cntrl_vec[i]->getAccessPermission(line_addr);
|
|
|
|
if (access_perm != AccessPermission_Invalid &&
|
|
|
|
access_perm != AccessPermission_NotPresent) {
|
|
|
|
|
|
|
|
DataBlock& block = m_abs_cntrl_vec[i]->getDataBlock(line_addr);
|
|
|
|
DPRINTF(RubySystem, "%s\n",block);
|
|
|
|
for (unsigned i = 0; i < size_in_bytes; ++i) {
|
|
|
|
block.setByte(i + startByte, data[i]);
|
2012-10-02 21:35:42 +02:00
|
|
|
}
|
2012-10-16 00:51:57 +02:00
|
|
|
DPRINTF(RubySystem, "%s\n",block);
|
2012-10-02 21:35:42 +02:00
|
|
|
}
|
|
|
|
}
|
2012-10-16 00:51:57 +02:00
|
|
|
|
|
|
|
uint32_t M5_VAR_USED num_functional_writes = 0;
|
|
|
|
for (unsigned int i = 0; i < m_memory_controller_vec.size() ;++i) {
|
|
|
|
num_functional_writes +=
|
|
|
|
m_memory_controller_vec[i]->functionalWriteBuffers(pkt);
|
|
|
|
}
|
|
|
|
|
|
|
|
num_functional_writes += m_network_ptr->functionalWrite(pkt);
|
|
|
|
DPRINTF(RubySystem, "Messages written = %u\n", num_functional_writes);
|
|
|
|
|
|
|
|
return true;
|
2012-10-02 21:35:42 +02:00
|
|
|
}
|
|
|
|
|
2009-05-11 19:38:43 +02:00
|
|
|
#ifdef CHECK_COHERENCE
|
|
|
|
// This code will check for cases if the given cache block is exclusive in
|
|
|
|
// one node and shared in another-- a coherence violation
|
|
|
|
//
|
|
|
|
// To use, the SLICC specification must call sequencer.checkCoherence(address)
|
|
|
|
// when the controller changes to a state with new permissions. Do this
|
|
|
|
// in setState. The SLICC spec must also define methods "isBlockShared"
|
|
|
|
// and "isBlockExclusive" that are specific to that protocol
|
|
|
|
//
|
2010-03-23 02:43:53 +01:00
|
|
|
void
|
|
|
|
RubySystem::checkGlobalCoherenceInvariant(const Address& addr)
|
|
|
|
{
|
|
|
|
#if 0
|
|
|
|
NodeID exclusive = -1;
|
|
|
|
bool sharedDetected = false;
|
|
|
|
NodeID lastShared = -1;
|
|
|
|
|
|
|
|
for (int i = 0; i < m_chip_vector.size(); i++) {
|
|
|
|
if (m_chip_vector[i]->isBlockExclusive(addr)) {
|
|
|
|
if (exclusive != -1) {
|
|
|
|
// coherence violation
|
|
|
|
WARN_EXPR(exclusive);
|
|
|
|
WARN_EXPR(m_chip_vector[i]->getID());
|
|
|
|
WARN_EXPR(addr);
|
2012-08-27 08:00:55 +02:00
|
|
|
WARN_EXPR(getTime());
|
2010-03-23 02:43:53 +01:00
|
|
|
ERROR_MSG("Coherence Violation Detected -- 2 exclusive chips");
|
|
|
|
} else if (sharedDetected) {
|
|
|
|
WARN_EXPR(lastShared);
|
|
|
|
WARN_EXPR(m_chip_vector[i]->getID());
|
|
|
|
WARN_EXPR(addr);
|
2012-08-27 08:00:55 +02:00
|
|
|
WARN_EXPR(getTime());
|
2010-03-23 02:43:53 +01:00
|
|
|
ERROR_MSG("Coherence Violation Detected -- exclusive chip with >=1 shared");
|
|
|
|
} else {
|
|
|
|
exclusive = m_chip_vector[i]->getID();
|
|
|
|
}
|
|
|
|
} else if (m_chip_vector[i]->isBlockShared(addr)) {
|
|
|
|
sharedDetected = true;
|
|
|
|
lastShared = m_chip_vector[i]->getID();
|
|
|
|
|
|
|
|
if (exclusive != -1) {
|
|
|
|
WARN_EXPR(lastShared);
|
|
|
|
WARN_EXPR(exclusive);
|
|
|
|
WARN_EXPR(addr);
|
2012-08-27 08:00:55 +02:00
|
|
|
WARN_EXPR(getTime());
|
2010-03-23 02:43:53 +01:00
|
|
|
ERROR_MSG("Coherence Violation Detected -- exclusive chip with >=1 shared");
|
|
|
|
}
|
|
|
|
}
|
2009-05-11 19:38:43 +02:00
|
|
|
}
|
2010-03-23 02:43:53 +01:00
|
|
|
#endif
|
2009-05-11 19:38:43 +02:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-01-30 05:29:17 +01:00
|
|
|
RubySystem *
|
|
|
|
RubySystemParams::create()
|
|
|
|
{
|
|
|
|
return new RubySystem(this);
|
|
|
|
}
|
2010-01-30 05:29:21 +01:00
|
|
|
|
|
|
|
/**
|
|
|
|
* virtual process function that is invoked when the callback
|
|
|
|
* queue is executed.
|
|
|
|
*/
|
2010-03-23 02:43:53 +01:00
|
|
|
void
|
2012-11-02 18:18:25 +01:00
|
|
|
RubyDumpStatsCallback::process()
|
2010-01-30 05:29:21 +01:00
|
|
|
{
|
2012-10-02 21:35:45 +02:00
|
|
|
ruby_system->printStats(*os);
|
2010-01-30 05:29:21 +01:00
|
|
|
}
|