summaryrefslogtreecommitdiff
path: root/src/mem/cache
diff options
context:
space:
mode:
authorCurtis Dunham <Curtis.Dunham@arm.com>2014-12-02 06:08:17 -0500
committerCurtis Dunham <Curtis.Dunham@arm.com>2014-12-02 06:08:17 -0500
commit7ca27dd3ccc2bcd3b77480179030d07f50c3d2d9 (patch)
treed663a1fc2e9797d90672656a9fd110df6d777ba6 /src/mem/cache
parentdf37cad0fdf262ffbfd1b680e7fb8ef7689885ad (diff)
downloadgem5-7ca27dd3ccc2bcd3b77480179030d07f50c3d2d9.tar.xz
mem: Remove WriteInvalidate support
Prepare for a different implementation following in the next patch
Diffstat (limited to 'src/mem/cache')
-rw-r--r--src/mem/cache/base.hh1
-rw-r--r--src/mem/cache/blk.hh3
-rw-r--r--src/mem/cache/cache_impl.hh189
-rw-r--r--src/mem/cache/mshr.cc25
-rw-r--r--src/mem/cache/mshr.hh18
5 files changed, 25 insertions, 211 deletions
diff --git a/src/mem/cache/base.hh b/src/mem/cache/base.hh
index 297b80180..845a689a4 100644
--- a/src/mem/cache/base.hh
+++ b/src/mem/cache/base.hh
@@ -94,7 +94,6 @@ class BaseCache : public MemObject
Blocked_NoMSHRs = MSHRQueue_MSHRs,
Blocked_NoWBBuffers = MSHRQueue_WriteBuffer,
Blocked_NoTargets,
- Blocked_PendingWriteInvalidate,
NUM_BLOCKED_CAUSES
};
diff --git a/src/mem/cache/blk.hh b/src/mem/cache/blk.hh
index ff09b42c4..6c72817c0 100644
--- a/src/mem/cache/blk.hh
+++ b/src/mem/cache/blk.hh
@@ -73,9 +73,6 @@ enum CacheBlkStatusBits {
BlkHWPrefetched = 0x20,
/** block holds data from the secure memory space */
BlkSecure = 0x40,
- /** can the block transition to E? (hasn't been shared with another cache)
- * used to close a timing gap when handling WriteInvalidate packets */
- BlkCanGoExclusive = 0x80
};
/**
diff --git a/src/mem/cache/cache_impl.hh b/src/mem/cache/cache_impl.hh
index a161f8085..b46717f14 100644
--- a/src/mem/cache/cache_impl.hh
+++ b/src/mem/cache/cache_impl.hh
@@ -70,7 +70,7 @@ Cache<TagStore>::Cache(const Params *p)
: BaseCache(p),
tags(dynamic_cast<TagStore*>(p->tags)),
prefetcher(p->prefetcher),
- doFastWrites(true),
+ doFastWrites(false),
prefetchOnAccess(p->prefetch_on_access)
{
tempBlock = new BlkType();
@@ -321,20 +321,15 @@ Cache<TagStore>::access(PacketPtr pkt, BlkType *&blk,
pkt->getAddr(), pkt->isSecure() ? "s" : "ns",
blk ? "hit" : "miss", blk ? blk->print() : "");
- // Writeback handling is special case. We can write the block
- // into the cache without having a writeable copy (or any copy at
- // all). Like writebacks, we write into the cache upon initial
- // receipt of a write-invalidate packets as well.
- if ((pkt->cmd == MemCmd::Writeback) ||
- ((pkt->cmd == MemCmd::WriteInvalidateReq) && isTopLevel)) {
+ // Writeback handling is special case. We can write the block into
+ // the cache without having a writeable copy (or any copy at all).
+ if (pkt->cmd == MemCmd::Writeback) {
assert(blkSize == pkt->getSize());
if (blk == NULL) {
// need to do a replacement
blk = allocateBlock(pkt->getAddr(), pkt->isSecure(), writebacks);
if (blk == NULL) {
- // no replaceable block available, give up.
- // Writeback will be forwarded to next level,
- // WriteInvalidate will be retried.
+ // no replaceable block available: give up, fwd to next level.
incMissCount(pkt);
return false;
}
@@ -345,29 +340,16 @@ Cache<TagStore>::access(PacketPtr pkt, BlkType *&blk,
blk->status |= BlkSecure;
}
}
- if (pkt->cmd == MemCmd::Writeback) {
- blk->status |= BlkDirty;
- if (pkt->isSupplyExclusive()) {
- blk->status |= BlkWritable;
- }
- // nothing else to do; writeback doesn't expect response
- assert(!pkt->needsResponse());
- } else if (pkt->cmd == MemCmd::WriteInvalidateReq) {
- blk->status |= (BlkReadable | BlkDirty | BlkCanGoExclusive);
- blk->status &= ~BlkWritable;
- ++fastWrites;
+ blk->status |= BlkDirty;
+ if (pkt->isSupplyExclusive()) {
+ blk->status |= BlkWritable;
}
+ // nothing else to do; writeback doesn't expect response
+ assert(!pkt->needsResponse());
std::memcpy(blk->data, pkt->getConstPtr<uint8_t>(), blkSize);
DPRINTF(Cache, "%s new state is %s\n", __func__, blk->print());
incHitCount(pkt);
return true;
- } else if ((pkt->cmd == MemCmd::WriteInvalidateReq) && !isTopLevel) {
- if (blk != NULL) {
- assert(blk != tempBlock);
- tags->invalidate(blk);
- blk->invalidate();
- }
- return true;
} else if ((blk != NULL) &&
(pkt->needsExclusive() ? blk->isWritable()
: blk->isReadable())) {
@@ -550,18 +532,6 @@ Cache<TagStore>::recvTimingReq(PacketPtr pkt)
bool needsResponse = pkt->needsResponse();
- if (pkt->cmd == MemCmd::WriteInvalidateReq) {
- if (!satisfied && isTopLevel) {
- // access() tried to allocate a block but it could not; abort.
- setBlocked(Blocked_PendingWriteInvalidate);
- return false;
- }
- satisfied = false;
- // we need to take the miss path (allocate MSHR, etc.) for
- // WriteInvalidates because they always need to propagate
- // throughout the memory system
- }
-
if (satisfied) {
// hit (for all other request types)
@@ -591,16 +561,6 @@ Cache<TagStore>::recvTimingReq(PacketPtr pkt)
// @todo: Make someone pay for this
pkt->firstWordDelay = pkt->lastWordDelay = 0;
- if (blk && blk->isValid() && (blk->status & BlkCanGoExclusive) &&
- pkt->isWrite() && (pkt->cmd != MemCmd::WriteInvalidateReq)) {
- // Packet is a Write (needs exclusive) should be delayed because
- // a WriteInvalidate is pending. Instead of going the MSHR route,
- // the Packet should be replayed, since if the block transitions
- // to Exclusive the write can complete immediately.
- setBlocked(Blocked_PendingWriteInvalidate);
- return false;
- }
-
Addr blk_addr = blockAlign(pkt->getAddr());
MSHR *mshr = mshrQueue.findMatch(blk_addr, pkt->isSecure());
@@ -647,20 +607,7 @@ Cache<TagStore>::recvTimingReq(PacketPtr pkt)
pkt = pf;
}
- if (pkt && (pkt->cmd == MemCmd::WriteInvalidateReq)) {
- // WriteInvalidates cannot coalesce with other requests, so
- // we cannot use an existing MSHR. If one exists, we mark it
- // as 'obsolete' so they don't modify the cache.
- if (mshr) {
- // Everything up to this point is obsolete, meaning
- // they should not modify the cache.
- DPRINTF(Cache, "%s: marking MSHR obsolete in %s of %x\n",
- __func__, pkt->cmdString(), pkt->getAddr());
-
- mshr->markObsolete();
- }
- allocateMissBuffer(pkt, time, true);
- } else if (mshr) {
+ if (mshr) {
/// MSHR hit
/// @note writebacks will be checked in getNextMSHR()
/// for any conflicting requests to the same block
@@ -707,10 +654,7 @@ Cache<TagStore>::recvTimingReq(PacketPtr pkt)
if (pkt->cmd == MemCmd::Writeback) {
allocateWriteBuffer(pkt, time, true);
} else {
- if (pkt->cmd == MemCmd::WriteInvalidateReq) {
- // a WriteInvalidate is not a normal write miss;
- // the assertions below are not applicable.
- } else if (blk && blk->isValid()) {
+ if (blk && blk->isValid()) {
// If we have a write miss to a valid block, we
// need to mark the block non-readable. Otherwise
// if we allow reads while there's an outstanding
@@ -769,12 +713,6 @@ Cache<TagStore>::getBusPacket(PacketPtr cpu_pkt, BlkType *blk,
return NULL;
}
- // WriteInvalidates for cache line clearing instructions don't
- // require a read; just send directly to the bus.
- if (cpu_pkt->cmd == MemCmd::WriteInvalidateReq) {
- return NULL;
- }
-
if (!blkValid &&
(cpu_pkt->cmd == MemCmd::Writeback || cpu_pkt->isUpgrade())) {
// Writebacks that weren't allocated in access() and upgrades
@@ -870,9 +808,6 @@ Cache<TagStore>::recvAtomic(PacketPtr pkt)
if (!access(pkt, blk, lat, writebacks)) {
// MISS
- // WriteInvalidates should never fail an access() in Atomic mode
- assert(pkt->cmd != MemCmd::WriteInvalidateReq);
-
PacketPtr bus_pkt = getBusPacket(pkt, blk, pkt->needsExclusive());
bool is_forward = (bus_pkt == NULL);
@@ -943,29 +878,6 @@ Cache<TagStore>::recvAtomic(PacketPtr pkt)
delete wbPkt;
}
- // We now have the block one way or another (hit or completed miss),
- // except for Request types that perform an invalidate, where the point
- // is to make sure there is no block.
-
- if (pkt->cmd == MemCmd::WriteInvalidateReq) {
- memSidePort->sendAtomic(pkt); // complete writeback
- if (isTopLevel) {
- // @todo Static analysis suggests this can actually happen
- assert(blk);
-
- // top level caches allocate and write the data
- assert(blk->isDirty());
- assert(!blk->isWritable());
- assert(blk->status & BlkCanGoExclusive);
- blk->status &= ~(BlkDirty | BlkCanGoExclusive); // and mark clean
- blk->status |= BlkWritable; // i.e. O(+cgE) -> E
- } else {
- // other caches invalidate.
- // if the block was found, it was invalidated.
- assert(!blk || !blk->isValid());
- }
- }
-
if (pkt->needsResponse()) {
pkt->makeAtomicResponse();
}
@@ -1101,10 +1013,7 @@ Cache<TagStore>::recvTimingResp(PacketPtr pkt)
bool is_fill = !mshr->isForward &&
(pkt->isRead() || pkt->cmd == MemCmd::UpgradeResp);
- if (mshr->isObsolete()) {
- DPRINTF(Cache, "%s: skipping cache fills; data for %s of %x "
- "is obsolete\n", __func__, pkt->cmdString(), pkt->getAddr());
- } else if (is_fill && !is_error) {
+ if (is_fill && !is_error) {
DPRINTF(Cache, "Block for addr %x being updated in Cache\n",
pkt->getAddr());
@@ -1140,19 +1049,9 @@ Cache<TagStore>::recvTimingResp(PacketPtr pkt)
}
if (is_fill) {
- // Presently the only situation leading to 'obsolete'
- // data is when a WriteInvalidate blows away an already
- // pending/in-progress read. We don't want to overwrite
- // cache data in that case.
- if (mshr->isObsolete()) {
- DPRINTF(Cache, "%s: skipping satisfyCpuSideRequest; "
- "data for %s of %x is obsolete\n",
- __func__, target->pkt->cmdString(),
- target->pkt->getAddr());
- } else {
- satisfyCpuSideRequest(target->pkt, blk,
- true, mshr->hasPostDowngrade());
- }
+ satisfyCpuSideRequest(target->pkt, blk,
+ true, mshr->hasPostDowngrade());
+
// How many bytes past the first request is this one
int transfer_offset =
target->pkt->getOffset(blkSize) - initial_offset;
@@ -1184,38 +1083,6 @@ Cache<TagStore>::recvTimingResp(PacketPtr pkt)
completion_time = clockEdge(responseLatency) +
pkt->lastWordDelay;
target->pkt->req->setExtraData(0);
- } else if (pkt->cmd == MemCmd::WriteInvalidateResp) {
- if (blk) {
- assert(blk->isDirty() && !blk->isWritable());
- // The block, having been written back, is no longer dirty,
- // nor do we have any reason to see if it was snooped in the
- // meantime (which CanGoExclusive tracks). If it can go
- // exclusive, we put it in that state, and otherwise S.
- // In short: O(+cgE) -> E, O(-cgE) -> S
- if (blk->status & BlkCanGoExclusive) {
- blk->status |= BlkWritable;
- }
- blk->status &= ~(BlkDirty | BlkCanGoExclusive);
- }
- if (isTopLevel) {
- // makeTimingResponse() will turn it into a WriteResp
- target->pkt->cmd = MemCmd::WriteReq;
- // Writes may have been blocked - quite rare case, but
- // it does happen. Prevent deadlock by telling the core
- if (isBlocked()) { // to retry.
- clearBlocked(Blocked_PendingWriteInvalidate);
- }
- }
- // If the block managed to get evicted before its own
- // writeback (e.g. by a Read/Upgrade (from O(-cgE)/S to
- // I/E) or ReadExclusive (direct to I/E); either way a
- // cache-to-cache ownership transfer) completed, that's
- // OK, we just ignore this response. If the new owner
- // doesn't actually modify it, a superfluous writeback
- // will occur for its impatience (since it will think it
- // has dirty data), but it really can't be helped.
- completion_time = clockEdge(responseLatency) +
- pkt->lastWordDelay;
} else {
// not a cache fill, just forwarding response
// responseLatency is the latency of the return path
@@ -1443,8 +1310,7 @@ Cache<TagStore>::allocateBlock(Addr addr, bool is_secure,
Addr repl_addr = tags->regenerateBlkAddr(blk->tag, blk->set);
MSHR *repl_mshr = mshrQueue.findMatch(repl_addr, blk->isSecure());
if (repl_mshr) {
- // must be an outstanding upgrade request (common case)
- // or WriteInvalidate pending writeback (very uncommon case)
+ // must be an outstanding upgrade request
// on a block we're about to replace...
assert(!blk->isWritable() || blk->isDirty());
assert(repl_mshr->needsExclusive());
@@ -1532,20 +1398,10 @@ Cache<TagStore>::handleFill(PacketPtr pkt, BlkType *blk,
// there are cases (such as failed store conditionals or
// compare-and-swaps) where we'll demand an exclusive copy but
// end up not writing it.
- // Caveat: if a Read takes a value from a WriteInvalidate MSHR,
- // it will get marked Dirty even though it is Clean (once the
- // WriteInvalidate completes). This is due to insufficient meta-
- // data and overly presumptive interpretation of the inhibit flag.
- // The result is an unnecessary extra writeback.
if (pkt->memInhibitAsserted())
blk->status |= BlkDirty;
}
- if (pkt->cmd == MemCmd::WriteInvalidateReq) {
- // a block written immediately, all at once, pre-writeback is dirty
- blk->status |= BlkDirty;
- }
-
DPRINTF(Cache, "Block addr %x (%s) moving from state %x to %s\n",
addr, is_secure ? "s" : "ns", old_state, blk->print());
@@ -1683,13 +1539,8 @@ Cache<TagStore>::handleSnoop(PacketPtr pkt, BlkType *blk,
// we may end up modifying both the block state and the packet (if
// we respond in atomic mode), so just figure out what to do now
- // and then do it later. If we find dirty data while snooping for a
- // WriteInvalidate, we don't care, since no merging needs to take place.
- // We need the eviction to happen as normal, but the data needn't be
- // sent anywhere, nor should the writeback be inhibited at the memory
- // controller for any reason.
- bool respond = blk->isDirty() && pkt->needsResponse()
- && (pkt->cmd != MemCmd::WriteInvalidateReq);
+ // and then do it later.
+ bool respond = blk->isDirty() && pkt->needsResponse();
bool have_exclusive = blk->isWritable();
// Invalidate any prefetch's from below that would strip write permissions
@@ -1706,7 +1557,7 @@ Cache<TagStore>::handleSnoop(PacketPtr pkt, BlkType *blk,
if (pkt->isRead() && !invalidate) {
assert(!needs_exclusive);
pkt->assertShared();
- int bits_to_clear = BlkWritable | BlkCanGoExclusive;
+ int bits_to_clear = BlkWritable;
const bool haveOwnershipState = true; // for now
if (!haveOwnershipState) {
// if we don't support pure ownership (dirty && !writable),
diff --git a/src/mem/cache/mshr.cc b/src/mem/cache/mshr.cc
index e4b62e230..793db02c2 100644
--- a/src/mem/cache/mshr.cc
+++ b/src/mem/cache/mshr.cc
@@ -62,9 +62,9 @@
using namespace std;
MSHR::MSHR() : readyTime(0), _isUncacheable(false), downstreamPending(false),
- pendingDirty(false), pendingClean(false),
+ pendingDirty(false),
postInvalidate(false), postDowngrade(false),
- _isObsolete(false), queue(NULL), order(0), addr(0),
+ queue(NULL), order(0), addr(0),
size(0), isSecure(false), inService(false),
isForward(false), threadNum(InvalidThreadID), data(NULL)
{
@@ -214,9 +214,7 @@ MSHR::allocate(Addr _addr, int _size, PacketPtr target, Tick whenReady,
isForward = false;
_isUncacheable = target->req->isUncacheable();
inService = false;
- pendingClean = (target->cmd == MemCmd::WriteInvalidateReq);
downstreamPending = false;
- _isObsolete = false;
threadNum = 0;
assert(targets.isReset());
// Don't know of a case where we would allocate a new MSHR for a
@@ -253,9 +251,8 @@ MSHR::markInService(PacketPtr pkt)
assert(pkt != NULL);
inService = true;
- pendingDirty = ((targets.needsExclusive &&
- (pkt->cmd != MemCmd::WriteInvalidateReq)) ||
- (!pkt->sharedAsserted() && pkt->memInhibitAsserted()));
+ pendingDirty = targets.needsExclusive ||
+ (!pkt->sharedAsserted() && pkt->memInhibitAsserted());
postInvalidate = postDowngrade = false;
if (!downstreamPending) {
@@ -374,12 +371,7 @@ MSHR::handleSnoop(PacketPtr pkt, Counter _order)
targets.add(cp_pkt, curTick(), _order, Target::FromSnoop,
downstreamPending && targets.needsExclusive);
- // WriteInvalidates must writeback and should not be inhibited on
- // account of its snoops discovering MSHRs wanting exclusive access
- // to what it wrote. We don't want to push this check higher,
- // however, because we want to be sure to add an invalidating
- // Target::FromSnoop, above.
- if (isPendingDirty() && (pkt->cmd != MemCmd::WriteInvalidateReq)) {
+ if (isPendingDirty()) {
pkt->assertMemInhibit();
pkt->setSupplyExclusive();
}
@@ -387,13 +379,6 @@ MSHR::handleSnoop(PacketPtr pkt, Counter _order)
if (pkt->needsExclusive()) {
// This transaction will take away our pending copy
postInvalidate = true;
-
- // Do not defer (i.e. return true) the snoop if the block is
- // going to be clean once the MSHR completes, as the data is
- // ready now.
- if (isPendingClean()) {
- return false;
- }
}
}
diff --git a/src/mem/cache/mshr.hh b/src/mem/cache/mshr.hh
index e5a15b61d..65357b9e6 100644
--- a/src/mem/cache/mshr.hh
+++ b/src/mem/cache/mshr.hh
@@ -83,18 +83,12 @@ class MSHR : public Packet::SenderState, public Printable
/** Will we have a dirty copy after this request? */
bool pendingDirty;
- /** Will we have a clean copy after this request? (i.e. is writeback) */
- bool pendingClean;
-
/** Did we snoop an invalidate while waiting for data? */
bool postInvalidate;
/** Did we snoop a read while waiting for data? */
bool postDowngrade;
- /** Did we get WriteInvalidate'd (and therefore obsoleted)? */
- bool _isObsolete;
-
public:
class Target {
@@ -182,10 +176,6 @@ class MSHR : public Packet::SenderState, public Printable
assert(inService); return pendingDirty;
}
- bool isPendingClean() const {
- return pendingClean;
- }
-
bool hasPostInvalidate() const {
assert(inService); return postInvalidate;
}
@@ -224,8 +214,6 @@ class MSHR : public Packet::SenderState, public Printable
bool isUncacheable() const { return _isUncacheable; }
- bool isObsolete() const { return _isObsolete; }
-
/**
* Allocate a miss to this MSHR.
* @param cmd The requesting command.
@@ -301,12 +289,6 @@ class MSHR : public Packet::SenderState, public Printable
bool checkFunctional(PacketPtr pkt);
- /** Mark this MSHR as tracking a transaction with obsoleted data. It still
- * needs to complete its lifecycle, but should not modify the cache. */
- void markObsolete() {
- _isObsolete = true;
- }
-
/**
* Prints the contents of this MSHR for debugging.
*/