summaryrefslogtreecommitdiff
path: root/configs/ruby/MOESI_AMD_Base.py
diff options
context:
space:
mode:
Diffstat (limited to 'configs/ruby/MOESI_AMD_Base.py')
-rw-r--r--configs/ruby/MOESI_AMD_Base.py326
1 files changed, 326 insertions, 0 deletions
diff --git a/configs/ruby/MOESI_AMD_Base.py b/configs/ruby/MOESI_AMD_Base.py
new file mode 100644
index 000000000..4c8ad28b0
--- /dev/null
+++ b/configs/ruby/MOESI_AMD_Base.py
@@ -0,0 +1,326 @@
+#
+# Copyright (c) 2010-2015 Advanced Micro Devices, Inc.
+# All rights reserved.
+#
+# For use for simulation and test purposes only
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions are met:
+#
+# 1. Redistributions of source code must retain the above copyright notice,
+# this list of conditions and the following disclaimer.
+#
+# 2. Redistributions in binary form must reproduce the above copyright notice,
+# this list of conditions and the following disclaimer in the documentation
+# and/or other materials provided with the distribution.
+#
+# 3. Neither the name of the copyright holder nor the names of its contributors
+# may be used to endorse or promote products derived from this software
+# without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
+# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+# POSSIBILITY OF SUCH DAMAGE.
+#
+# Author: Lisa Hsu
+#
+
+import math
+import m5
+from m5.objects import *
+from m5.defines import buildEnv
+from Ruby import create_topology
+from Ruby import send_evicts
+
+from Cluster import Cluster
+from Crossbar import Crossbar
+
+class CntrlBase:
+ _seqs = 0
+ @classmethod
+ def seqCount(cls):
+ # Use SeqCount not class since we need global count
+ CntrlBase._seqs += 1
+ return CntrlBase._seqs - 1
+
+ _cntrls = 0
+ @classmethod
+ def cntrlCount(cls):
+ # Use CntlCount not class since we need global count
+ CntrlBase._cntrls += 1
+ return CntrlBase._cntrls - 1
+
+ _version = 0
+ @classmethod
+ def versionCount(cls):
+ cls._version += 1 # Use count for this particular type
+ return cls._version - 1
+
+class L1DCache(RubyCache):
+ resourceStalls = False
+ def create(self, options):
+ self.size = MemorySize(options.l1d_size)
+ self.assoc = options.l1d_assoc
+ self.replacement_policy = PseudoLRUReplacementPolicy()
+
+class L1ICache(RubyCache):
+ resourceStalls = False
+ def create(self, options):
+ self.size = MemorySize(options.l1i_size)
+ self.assoc = options.l1i_assoc
+ self.replacement_policy = PseudoLRUReplacementPolicy()
+
+class L2Cache(RubyCache):
+ resourceStalls = False
+ def create(self, options):
+ self.size = MemorySize(options.l2_size)
+ self.assoc = options.l2_assoc
+ self.replacement_policy = PseudoLRUReplacementPolicy()
+
+class CPCntrl(CorePair_Controller, CntrlBase):
+
+ def create(self, options, ruby_system, system):
+ self.version = self.versionCount()
+
+ self.L1Icache = L1ICache()
+ self.L1Icache.create(options)
+ self.L1D0cache = L1DCache()
+ self.L1D0cache.create(options)
+ self.L1D1cache = L1DCache()
+ self.L1D1cache.create(options)
+ self.L2cache = L2Cache()
+ self.L2cache.create(options)
+
+ self.sequencer = RubySequencer()
+ self.sequencer.icache_hit_latency = 2
+ self.sequencer.dcache_hit_latency = 2
+ self.sequencer.version = self.seqCount()
+ self.sequencer.icache = self.L1Icache
+ self.sequencer.dcache = self.L1D0cache
+ self.sequencer.ruby_system = ruby_system
+ self.sequencer.coreid = 0
+ self.sequencer.is_cpu_sequencer = True
+
+ self.sequencer1 = RubySequencer()
+ self.sequencer1.version = self.seqCount()
+ self.sequencer1.icache = self.L1Icache
+ self.sequencer1.dcache = self.L1D1cache
+ self.sequencer1.icache_hit_latency = 2
+ self.sequencer1.dcache_hit_latency = 2
+ self.sequencer1.ruby_system = ruby_system
+ self.sequencer1.coreid = 1
+ self.sequencer1.is_cpu_sequencer = True
+
+ self.issue_latency = options.cpu_to_dir_latency
+ self.send_evictions = send_evicts(options)
+
+ self.ruby_system = ruby_system
+
+ if options.recycle_latency:
+ self.recycle_latency = options.recycle_latency
+
+class L3Cache(RubyCache):
+ assoc = 8
+ dataArrayBanks = 256
+ tagArrayBanks = 256
+
+ def create(self, options, ruby_system, system):
+ self.size = MemorySize(options.l3_size)
+ self.size.value /= options.num_dirs
+ self.dataArrayBanks /= options.num_dirs
+ self.tagArrayBanks /= options.num_dirs
+ self.dataArrayBanks /= options.num_dirs
+ self.tagArrayBanks /= options.num_dirs
+ self.dataAccessLatency = options.l3_data_latency
+ self.tagAccessLatency = options.l3_tag_latency
+ self.resourceStalls = options.no_resource_stalls
+ self.replacement_policy = PseudoLRUReplacementPolicy()
+
+class L3Cntrl(L3Cache_Controller, CntrlBase):
+ def create(self, options, ruby_system, system):
+ self.version = self.versionCount()
+ self.L3cache = L3Cache()
+ self.L3cache.create(options, ruby_system, system)
+
+ self.l3_response_latency = max(self.L3cache.dataAccessLatency,
+ self.L3cache.tagAccessLatency)
+ self.ruby_system = ruby_system
+
+ if options.recycle_latency:
+ self.recycle_latency = options.recycle_latency
+
+ def connectWireBuffers(self, req_to_dir, resp_to_dir, l3_unblock_to_dir,
+ req_to_l3, probe_to_l3, resp_to_l3):
+ self.reqToDir = req_to_dir
+ self.respToDir = resp_to_dir
+ self.l3UnblockToDir = l3_unblock_to_dir
+ self.reqToL3 = req_to_l3
+ self.probeToL3 = probe_to_l3
+ self.respToL3 = resp_to_l3
+
+class DirMem(RubyDirectoryMemory, CntrlBase):
+ def create(self, options, ruby_system, system):
+ self.version = self.versionCount()
+
+ phys_mem_size = AddrRange(options.mem_size).size()
+ mem_module_size = phys_mem_size / options.num_dirs
+ dir_size = MemorySize('0B')
+ dir_size.value = mem_module_size
+ self.size = dir_size
+
+class DirCntrl(Directory_Controller, CntrlBase):
+ def create(self, options, ruby_system, system):
+ self.version = self.versionCount()
+
+ self.response_latency = 30
+
+ self.directory = DirMem()
+ self.directory.create(options, ruby_system, system)
+
+ self.L3CacheMemory = L3Cache()
+ self.L3CacheMemory.create(options, ruby_system, system)
+
+ self.l3_hit_latency = max(self.L3CacheMemory.dataAccessLatency,
+ self.L3CacheMemory.tagAccessLatency)
+
+ self.number_of_TBEs = options.num_tbes
+
+ self.ruby_system = ruby_system
+
+ if options.recycle_latency:
+ self.recycle_latency = options.recycle_latency
+
+ self.CPUonly = True
+
+ def connectWireBuffers(self, req_to_dir, resp_to_dir, l3_unblock_to_dir,
+ req_to_l3, probe_to_l3, resp_to_l3):
+ self.reqToDir = req_to_dir
+ self.respToDir = resp_to_dir
+ self.l3UnblockToDir = l3_unblock_to_dir
+ self.reqToL3 = req_to_l3
+ self.probeToL3 = probe_to_l3
+ self.respToL3 = resp_to_l3
+
+def define_options(parser):
+ parser.add_option("--num-subcaches", type="int", default=4)
+ parser.add_option("--l3-data-latency", type="int", default=20)
+ parser.add_option("--l3-tag-latency", type="int", default=15)
+ parser.add_option("--cpu-to-dir-latency", type="int", default=15)
+ parser.add_option("--no-resource-stalls", action="store_false",
+ default=True)
+ parser.add_option("--num-tbes", type="int", default=256)
+ parser.add_option("--l2-latency", type="int", default=50) # load to use
+
+def create_system(options, full_system, system, dma_devices, ruby_system):
+ if buildEnv['PROTOCOL'] != 'MOESI_AMD_Base':
+ panic("This script requires the MOESI_AMD_Base protocol.")
+
+ cpu_sequencers = []
+
+ #
+ # The ruby network creation expects the list of nodes in the system to
+ # be consistent with the NetDest list. Therefore the l1 controller
+ # nodes must be listed before the directory nodes and directory nodes
+ # before dma nodes, etc.
+ #
+ l1_cntrl_nodes = []
+ l3_cntrl_nodes = []
+ dir_cntrl_nodes = []
+
+ control_count = 0
+
+ #
+ # Must create the individual controllers before the network to ensure
+ # the controller constructors are called before the network constructor
+ #
+
+ # This is the base crossbar that connects the L3s, Dirs, and cpu
+ # Cluster
+ mainCluster = Cluster(extBW = 512, intBW = 512) # 1 TB/s
+ for i in xrange(options.num_dirs):
+
+ dir_cntrl = DirCntrl(TCC_select_num_bits = 0)
+ dir_cntrl.create(options, ruby_system, system)
+
+ # Connect the Directory controller to the ruby network
+ dir_cntrl.requestFromCores = MessageBuffer(ordered = True)
+ dir_cntrl.requestFromCores.slave = ruby_system.network.master
+
+ dir_cntrl.responseFromCores = MessageBuffer()
+ dir_cntrl.responseFromCores.slave = ruby_system.network.master
+
+ dir_cntrl.unblockFromCores = MessageBuffer()
+ dir_cntrl.unblockFromCores.slave = ruby_system.network.master
+
+ dir_cntrl.probeToCore = MessageBuffer()
+ dir_cntrl.probeToCore.master = ruby_system.network.slave
+
+ dir_cntrl.responseToCore = MessageBuffer()
+ dir_cntrl.responseToCore.master = ruby_system.network.slave
+
+ dir_cntrl.triggerQueue = MessageBuffer(ordered = True)
+ dir_cntrl.L3triggerQueue = MessageBuffer(ordered = True)
+ dir_cntrl.responseFromMemory = MessageBuffer()
+
+ exec("system.dir_cntrl%d = dir_cntrl" % i)
+ dir_cntrl_nodes.append(dir_cntrl)
+
+ mainCluster.add(dir_cntrl)
+
+ # Technically this config can support an odd number of cpus, but the top
+ # level config files, such as the ruby_random_tester, will get confused if
+ # the number of cpus does not equal the number of sequencers. Thus make
+ # sure that an even number of cpus is specified.
+ assert((options.num_cpus % 2) == 0)
+
+ # For an odd number of CPUs, still create the right number of controllers
+ cpuCluster = Cluster(extBW = 512, intBW = 512) # 1 TB/s
+ for i in xrange((options.num_cpus + 1) / 2):
+
+ cp_cntrl = CPCntrl()
+ cp_cntrl.create(options, ruby_system, system)
+
+ exec("system.cp_cntrl%d = cp_cntrl" % i)
+ #
+ # Add controllers and sequencers to the appropriate lists
+ #
+ cpu_sequencers.extend([cp_cntrl.sequencer, cp_cntrl.sequencer1])
+
+ # Connect the CP controllers and the network
+ cp_cntrl.requestFromCore = MessageBuffer()
+ cp_cntrl.requestFromCore.master = ruby_system.network.slave
+
+ cp_cntrl.responseFromCore = MessageBuffer()
+ cp_cntrl.responseFromCore.master = ruby_system.network.slave
+
+ cp_cntrl.unblockFromCore = MessageBuffer()
+ cp_cntrl.unblockFromCore.master = ruby_system.network.slave
+
+ cp_cntrl.probeToCore = MessageBuffer()
+ cp_cntrl.probeToCore.slave = ruby_system.network.master
+
+ cp_cntrl.responseToCore = MessageBuffer()
+ cp_cntrl.responseToCore.slave = ruby_system.network.master
+
+ cp_cntrl.mandatoryQueue = MessageBuffer()
+ cp_cntrl.triggerQueue = MessageBuffer(ordered = True)
+
+ cpuCluster.add(cp_cntrl)
+
+ # Assuming no DMA devices
+ assert(len(dma_devices) == 0)
+
+ # Add cpu/gpu clusters to main cluster
+ mainCluster.add(cpuCluster)
+
+ ruby_system.network.number_of_virtual_networks = 10
+
+ return (cpu_sequencers, dir_cntrl_nodes, mainCluster)