cachepc-qemu

Fork of AMDESE/qemu with changes for cachepc side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-qemu
Log | Files | Refs | Submodules | LICENSE | sfeed.txt

memory_ldst_cached.h.inc (3902B)


      1/*
      2 *  Memory access templates for MemoryRegionCache
      3 *
      4 *  Copyright (c) 2018 Red Hat, Inc.
      5 *
      6 * This library is free software; you can redistribute it and/or
      7 * modify it under the terms of the GNU Lesser General Public
      8 * License as published by the Free Software Foundation; either
      9 * version 2.1 of the License, or (at your option) any later version.
     10 *
     11 * This library is distributed in the hope that it will be useful,
     12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
     13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
     14 * Lesser General Public License for more details.
     15 *
     16 * You should have received a copy of the GNU Lesser General Public
     17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
     18 */
     19
     20#define ADDRESS_SPACE_LD_CACHED(size) \
     21    glue(glue(address_space_ld, size), glue(ENDIANNESS, _cached))
     22#define ADDRESS_SPACE_LD_CACHED_SLOW(size) \
     23    glue(glue(address_space_ld, size), glue(ENDIANNESS, _cached_slow))
     24#define LD_P(size) \
     25    glue(glue(ld, size), glue(ENDIANNESS, _p))
     26
     27static inline uint16_t ADDRESS_SPACE_LD_CACHED(uw)(MemoryRegionCache *cache,
     28    hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
     29{
     30    assert(addr < cache->len && 2 <= cache->len - addr);
     31    fuzz_dma_read_cb(cache->xlat + addr, 2, cache->mrs.mr);
     32    if (likely(cache->ptr)) {
     33        return LD_P(uw)(cache->ptr + addr);
     34    } else {
     35        return ADDRESS_SPACE_LD_CACHED_SLOW(uw)(cache, addr, attrs, result);
     36    }
     37}
     38
     39static inline uint32_t ADDRESS_SPACE_LD_CACHED(l)(MemoryRegionCache *cache,
     40    hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
     41{
     42    assert(addr < cache->len && 4 <= cache->len - addr);
     43    fuzz_dma_read_cb(cache->xlat + addr, 4, cache->mrs.mr);
     44    if (likely(cache->ptr)) {
     45        return LD_P(l)(cache->ptr + addr);
     46    } else {
     47        return ADDRESS_SPACE_LD_CACHED_SLOW(l)(cache, addr, attrs, result);
     48    }
     49}
     50
     51static inline uint64_t ADDRESS_SPACE_LD_CACHED(q)(MemoryRegionCache *cache,
     52    hwaddr addr, MemTxAttrs attrs, MemTxResult *result)
     53{
     54    assert(addr < cache->len && 8 <= cache->len - addr);
     55    fuzz_dma_read_cb(cache->xlat + addr, 8, cache->mrs.mr);
     56    if (likely(cache->ptr)) {
     57        return LD_P(q)(cache->ptr + addr);
     58    } else {
     59        return ADDRESS_SPACE_LD_CACHED_SLOW(q)(cache, addr, attrs, result);
     60    }
     61}
     62
     63#undef ADDRESS_SPACE_LD_CACHED
     64#undef ADDRESS_SPACE_LD_CACHED_SLOW
     65#undef LD_P
     66
     67#define ADDRESS_SPACE_ST_CACHED(size) \
     68    glue(glue(address_space_st, size), glue(ENDIANNESS, _cached))
     69#define ADDRESS_SPACE_ST_CACHED_SLOW(size) \
     70    glue(glue(address_space_st, size), glue(ENDIANNESS, _cached_slow))
     71#define ST_P(size) \
     72    glue(glue(st, size), glue(ENDIANNESS, _p))
     73
     74static inline void ADDRESS_SPACE_ST_CACHED(w)(MemoryRegionCache *cache,
     75    hwaddr addr, uint16_t val, MemTxAttrs attrs, MemTxResult *result)
     76{
     77    assert(addr < cache->len && 2 <= cache->len - addr);
     78    if (likely(cache->ptr)) {
     79        ST_P(w)(cache->ptr + addr, val);
     80    } else {
     81        ADDRESS_SPACE_ST_CACHED_SLOW(w)(cache, addr, val, attrs, result);
     82    }
     83}
     84
     85static inline void ADDRESS_SPACE_ST_CACHED(l)(MemoryRegionCache *cache,
     86    hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result)
     87{
     88    assert(addr < cache->len && 4 <= cache->len - addr);
     89    if (likely(cache->ptr)) {
     90        ST_P(l)(cache->ptr + addr, val);
     91    } else {
     92        ADDRESS_SPACE_ST_CACHED_SLOW(l)(cache, addr, val, attrs, result);
     93    }
     94}
     95
     96static inline void ADDRESS_SPACE_ST_CACHED(q)(MemoryRegionCache *cache,
     97    hwaddr addr, uint64_t val, MemTxAttrs attrs, MemTxResult *result)
     98{
     99    assert(addr < cache->len && 8 <= cache->len - addr);
    100    if (likely(cache->ptr)) {
    101        ST_P(q)(cache->ptr + addr, val);
    102    } else {
    103        ADDRESS_SPACE_ST_CACHED_SLOW(q)(cache, addr, val, attrs, result);
    104    }
    105}
    106
    107#undef ADDRESS_SPACE_ST_CACHED
    108#undef ADDRESS_SPACE_ST_CACHED_SLOW
    109#undef ST_P
    110
    111#undef ENDIANNESS