summaryrefslogtreecommitdiff
path: root/src/mem/cache
diff options
context:
space:
mode:
authorAndreas Hansson <andreas.hansson@arm.com>2014-12-02 06:07:54 -0500
committerAndreas Hansson <andreas.hansson@arm.com>2014-12-02 06:07:54 -0500
commitea5ccc70417db08379027ca7344e50cba53063dd (patch)
treedb523845787bb9104a8619ececb814156b0054a8 /src/mem/cache
parentf012166bb600ebaeefa48e74f7dd7fdfc9742506 (diff)
downloadgem5-ea5ccc70417db08379027ca7344e50cba53063dd.tar.xz
mem: Clean up packet data allocation
This patch attempts to make the rules for data allocation in the packet explicit, understandable, and easy to verify. The constructor that copies a packet is extended with an additional flag "alloc_data" to enable the call site to explicitly say whether the newly created packet is short-lived (a zero-time snoop), or has an unknown life-time and therefore should allocate its own data (or copy a static pointer in the case of static data). The tricky case is the static data. In essence this is a copy-avoidance scheme where the original source of the request (DMA, CPU etc) does not ask the memory system to return data as part of the packet, but instead provides a pointer, and then the memory system carries this pointer around, and copies the appropriate data to the location itself. Thus any derived packet actually never copies any data. As the original source does not copy any data from the response packet when arriving back at the source, we must maintain the copy of the original pointer to not break the system. We might want to revisit this one day and pay the price for a few extra memcpy invocations. All in all this patch should make it easier to grok what is going on in the memory system and how data is actually copied (or not).
Diffstat (limited to 'src/mem/cache')
-rw-r--r--src/mem/cache/cache_impl.hh38
-rw-r--r--src/mem/cache/mshr.cc6
2 files changed, 31 insertions, 13 deletions
diff --git a/src/mem/cache/cache_impl.hh b/src/mem/cache/cache_impl.hh
index 296f31ebd..a161f8085 100644
--- a/src/mem/cache/cache_impl.hh
+++ b/src/mem/cache/cache_impl.hh
@@ -487,17 +487,24 @@ Cache<TagStore>::recvTimingReq(PacketPtr pkt)
// that have shared copies. Not necessary if we know that
// supplier had exclusive copy to begin with.
if (pkt->needsExclusive() && !pkt->isSupplyExclusive()) {
- Packet *snoopPkt = new Packet(pkt, true); // clear flags
+ // create a downstream express snoop with cleared packet
+ // flags, there is no need to allocate any data as the
+ // packet is merely used to co-ordinate state transitions
+ Packet *snoop_pkt = new Packet(pkt, true, false);
+
// also reset the bus time that the original packet has
// not yet paid for
- snoopPkt->firstWordDelay = snoopPkt->lastWordDelay = 0;
- snoopPkt->setExpressSnoop();
- snoopPkt->assertMemInhibit();
- bool M5_VAR_USED success = memSidePort->sendTimingReq(snoopPkt);
- // the packet is marked inhibited and will thus bypass any
- // flow control
+ snoop_pkt->firstWordDelay = snoop_pkt->lastWordDelay = 0;
+
+ // make this an instantaneous express snoop, and let the
+ // other caches in the system know that the packet is
+ // inhibited
+ snoop_pkt->setExpressSnoop();
+ snoop_pkt->assertMemInhibit();
+ bool M5_VAR_USED success = memSidePort->sendTimingReq(snoop_pkt);
+ // express snoops always succeed
assert(success);
- // main memory will delete snoopPkt
+ // main memory will delete the packet
}
// since we're the official target but we aren't responding,
// delete the packet now.
@@ -1570,12 +1577,19 @@ doTimingSupplyResponse(PacketPtr req_pkt, const uint8_t *blk_data,
{
// sanity check
assert(req_pkt->isRequest());
+ assert(req_pkt->needsResponse());
DPRINTF(Cache, "%s for %s address %x size %d\n", __func__,
req_pkt->cmdString(), req_pkt->getAddr(), req_pkt->getSize());
// timing-mode snoop responses require a new packet, unless we
// already made a copy...
- PacketPtr pkt = already_copied ? req_pkt : new Packet(req_pkt);
+ PacketPtr pkt = req_pkt;
+ if (!already_copied)
+ // do not clear flags, and allocate space for data if the
+ // packet needs it (the only packets that carry data are read
+ // responses)
+ pkt = new Packet(req_pkt, false, req_pkt->isRead());
+
assert(req_pkt->isInvalidate() || pkt->sharedAsserted());
pkt->makeTimingResponse();
// @todo Make someone pay for this
@@ -1624,7 +1638,7 @@ Cache<TagStore>::handleSnoop(PacketPtr pkt, BlkType *blk,
// rewritten to be relative to cpu-side bus (if any)
bool alreadyResponded = pkt->memInhibitAsserted();
if (is_timing) {
- Packet snoopPkt(pkt, true); // clear flags
+ Packet snoopPkt(pkt, true, false); // clear flags, no allocation
snoopPkt.setExpressSnoop();
snoopPkt.pushSenderState(new ForwardResponseRecord(pkt->getSrc()));
// the snoop packet does not need to wait any additional
@@ -1996,7 +2010,7 @@ Cache<TagStore>::getTimingPacket()
// at the moment. Without this check we could get a stale
// copy from memory that might get used in place of the
// dirty one.
- Packet snoop_pkt(tgt_pkt, true);
+ Packet snoop_pkt(tgt_pkt, true, false);
snoop_pkt.setExpressSnoop();
snoop_pkt.senderState = mshr;
cpuSidePort->sendTimingSnoopReq(&snoop_pkt);
@@ -2032,7 +2046,7 @@ Cache<TagStore>::getTimingPacket()
// not a cache block request, but a response is expected
// make copy of current packet to forward, keep current
// copy for response handling
- pkt = new Packet(tgt_pkt);
+ pkt = new Packet(tgt_pkt, false, true);
if (pkt->isWrite()) {
pkt->setData(tgt_pkt->getConstPtr<uint8_t>());
}
diff --git a/src/mem/cache/mshr.cc b/src/mem/cache/mshr.cc
index bc46ed267..e4b62e230 100644
--- a/src/mem/cache/mshr.cc
+++ b/src/mem/cache/mshr.cc
@@ -366,7 +366,11 @@ MSHR::handleSnoop(PacketPtr pkt, Counter _order)
// Actual target device (typ. a memory) will delete the
// packet on reception, so we need to save a copy here.
- PacketPtr cp_pkt = new Packet(pkt, true);
+
+ // Clear flags and also allocate new data as the original
+ // packet data storage may have been deleted by the time we
+ // get to send this packet.
+ PacketPtr cp_pkt = new Packet(pkt, true, true);
targets.add(cp_pkt, curTick(), _order, Target::FromSnoop,
downstreamPending && targets.needsExclusive);