Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
accel/tcg/cputlb.c | 151 ++++++++++++++++++++++-----------------------
1 file changed, 74 insertions(+), 77 deletions(-)
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
index a33bebf55a..8f459be5a8 100644
--- a/accel/tcg/cputlb.c
+++ b/accel/tcg/cputlb.c
@@ -1684,10 +1684,7 @@ bool tlb_plugin_lookup(CPUState *cpu, vaddr addr, int
mmu_idx,
*/
typedef struct MMULookupPageData {
- CPUTLBEntryFull *full;
- void *haddr;
vaddr addr;
- int flags;
int size;
TLBLookupOutput o;
} MMULookupPageData;
@@ -1724,10 +1721,6 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData
*data, MemOp memop,
};
tlb_lookup_nofail(cpu, &data->o, &i);
-
- data->full = &data->o.full;
- data->flags = data->o.flags;
- data->haddr = data->o.haddr;
}
/**
@@ -1743,24 +1736,22 @@ static void mmu_lookup1(CPUState *cpu,
MMULookupPageData *data, MemOp memop,
static void mmu_watch_or_dirty(CPUState *cpu, MMULookupPageData *data,
MMUAccessType access_type, uintptr_t ra)
{
- CPUTLBEntryFull *full = data->full;
- vaddr addr = data->addr;
- int flags = data->flags;
- int size = data->size;
+ int flags = data->o.flags;
/* On watchpoint hit, this will longjmp out. */
if (flags & TLB_WATCHPOINT) {
int wp = access_type == MMU_DATA_STORE ? BP_MEM_WRITE : BP_MEM_READ;
- cpu_check_watchpoint(cpu, addr, size, full->attrs, wp, ra);
+ cpu_check_watchpoint(cpu, data->addr, data->size,
+ data->o.full.attrs, wp, ra);
flags &= ~TLB_WATCHPOINT;
}
/* Note that notdirty is only set for writes. */
if (flags & TLB_NOTDIRTY) {
- notdirty_write(cpu, addr, size, full, ra);
+ notdirty_write(cpu, data->addr, data->size, &data->o.full, ra);
flags &= ~TLB_NOTDIRTY;
}
- data->flags = flags;
+ data->o.flags = flags;
}
/**
@@ -1795,7 +1786,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr,
MemOpIdx oi,
if (likely(!crosspage)) {
mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
- flags = l->page[0].flags;
+ flags = l->page[0].o.flags;
if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) {
mmu_watch_or_dirty(cpu, &l->page[0], type, ra);
}
@@ -1812,7 +1803,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr,
MemOpIdx oi,
mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra);
- flags = l->page[0].flags | l->page[1].flags;
+ flags = l->page[0].o.flags | l->page[1].o.flags;
if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) {
mmu_watch_or_dirty(cpu, &l->page[0], type, ra);
mmu_watch_or_dirty(cpu, &l->page[1], type, ra);
@@ -2029,7 +2020,7 @@ static Int128 do_ld16_mmio_beN(CPUState *cpu,
CPUTLBEntryFull *full,
*/
static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be)
{
- uint8_t *haddr = p->haddr;
+ uint8_t *haddr = p->o.haddr;
int i, size = p->size;
for (i = 0; i < size; i++) {
@@ -2047,7 +2038,7 @@ static uint64_t do_ld_bytes_beN(MMULookupPageData *p,
uint64_t ret_be)
*/
static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be)
{
- void *haddr = p->haddr;
+ void *haddr = p->o.haddr;
int size = p->size;
do {
@@ -2097,7 +2088,7 @@ static uint64_t do_ld_parts_beN(MMULookupPageData *p,
uint64_t ret_be)
static uint64_t do_ld_whole_be4(MMULookupPageData *p, uint64_t ret_be)
{
int o = p->addr & 3;
- uint32_t x = load_atomic4(p->haddr - o);
+ uint32_t x = load_atomic4(p->o.haddr - o);
x = cpu_to_be32(x);
x <<= o * 8;
@@ -2117,7 +2108,7 @@ static uint64_t do_ld_whole_be8(CPUState *cpu, uintptr_t
ra,
MMULookupPageData *p, uint64_t ret_be)
{
int o = p->addr & 7;
- uint64_t x = load_atomic8_or_exit(cpu, ra, p->haddr - o);
+ uint64_t x = load_atomic8_or_exit(cpu, ra, p->o.haddr - o);
x = cpu_to_be64(x);
x <<= o * 8;
@@ -2137,7 +2128,7 @@ static Int128 do_ld_whole_be16(CPUState *cpu, uintptr_t
ra,
MMULookupPageData *p, uint64_t ret_be)
{
int o = p->addr & 15;
- Int128 x, y = load_atomic16_or_exit(cpu, ra, p->haddr - o);
+ Int128 x, y = load_atomic16_or_exit(cpu, ra, p->o.haddr - o);
int size = p->size;
if (!HOST_BIG_ENDIAN) {
@@ -2160,8 +2151,8 @@ static uint64_t do_ld_beN(CPUState *cpu,
MMULookupPageData *p,
MemOp atom;
unsigned tmp, half_size;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_ld_mmio_beN(cpu, p->full, ret_be, p->addr, p->size,
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_ld_mmio_beN(cpu, &p->o.full, ret_be, p->addr, p->size,
mmu_idx, type, ra);
}
@@ -2210,8 +2201,9 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p,
uint64_t b;
MemOp atom;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_ld16_mmio_beN(cpu, p->full, a, p->addr, size, mmu_idx, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_ld16_mmio_beN(cpu, &p->o.full, a, p->addr,
+ size, mmu_idx, ra);
}
/*
@@ -2223,7 +2215,7 @@ static Int128 do_ld16_beN(CPUState *cpu,
MMULookupPageData *p,
case MO_ATOM_SUBALIGN:
p->size = size - 8;
a = do_ld_parts_beN(p, a);
- p->haddr += size - 8;
+ p->o.haddr += size - 8;
p->size = 8;
b = do_ld_parts_beN(p, 0);
break;
@@ -2242,7 +2234,7 @@ static Int128 do_ld16_beN(CPUState *cpu,
MMULookupPageData *p,
case MO_ATOM_NONE:
p->size = size - 8;
a = do_ld_bytes_beN(p, a);
- b = ldq_be_p(p->haddr + size - 8);
+ b = ldq_be_p(p->o.haddr + size - 8);
break;
default:
@@ -2255,10 +2247,11 @@ static Int128 do_ld16_beN(CPUState *cpu,
MMULookupPageData *p,
static uint8_t do_ld_1(CPUState *cpu, MMULookupPageData *p, int mmu_idx,
MMUAccessType type, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_ld_mmio_beN(cpu, p->full, 0, p->addr, 1, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 1,
+ mmu_idx, type, ra);
} else {
- return *(uint8_t *)p->haddr;
+ return *(uint8_t *)p->o.haddr;
}
}
@@ -2267,14 +2260,15 @@ static uint16_t do_ld_2(CPUState *cpu, MMULookupPageData *p, int mmu_idx,
{
uint16_t ret;
- if (unlikely(p->flags & TLB_MMIO)) {
- ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 2, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 2,
+ mmu_idx, type, ra);
if ((memop & MO_BSWAP) == MO_LE) {
ret = bswap16(ret);
}
} else {
/* Perform the load host endian, then swap if necessary. */
- ret = load_atom_2(cpu, ra, p->haddr, memop);
+ ret = load_atom_2(cpu, ra, p->o.haddr, memop);
if (memop & MO_BSWAP) {
ret = bswap16(ret);
}
@@ -2287,14 +2281,15 @@ static uint32_t do_ld_4(CPUState *cpu,
MMULookupPageData *p, int mmu_idx,
{
uint32_t ret;
- if (unlikely(p->flags & TLB_MMIO)) {
- ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 4, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 4,
+ mmu_idx, type, ra);
if ((memop & MO_BSWAP) == MO_LE) {
ret = bswap32(ret);
}
} else {
/* Perform the load host endian. */
- ret = load_atom_4(cpu, ra, p->haddr, memop);
+ ret = load_atom_4(cpu, ra, p->o.haddr, memop);
if (memop & MO_BSWAP) {
ret = bswap32(ret);
}
@@ -2307,14 +2302,15 @@ static uint64_t do_ld_8(CPUState *cpu,
MMULookupPageData *p, int mmu_idx,
{
uint64_t ret;
- if (unlikely(p->flags & TLB_MMIO)) {
- ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 8, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 8,
+ mmu_idx, type, ra);
if ((memop & MO_BSWAP) == MO_LE) {
ret = bswap64(ret);
}
} else {
/* Perform the load host endian. */
- ret = load_atom_8(cpu, ra, p->haddr, memop);
+ ret = load_atom_8(cpu, ra, p->o.haddr, memop);
if (memop & MO_BSWAP) {
ret = bswap64(ret);
}
@@ -2414,15 +2410,15 @@ static Int128 do_ld16_mmu(CPUState *cpu, vaddr addr,
cpu_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD);
crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_LOAD, &l);
if (likely(!crosspage)) {
- if (unlikely(l.page[0].flags & TLB_MMIO)) {
- ret = do_ld16_mmio_beN(cpu, l.page[0].full, 0, addr, 16,
+ if (unlikely(l.page[0].o.flags & TLB_MMIO)) {
+ ret = do_ld16_mmio_beN(cpu, &l.page[0].o.full, 0, addr, 16,
l.mmu_idx, ra);
if ((l.memop & MO_BSWAP) == MO_LE) {
ret = bswap128(ret);
}
} else {
/* Perform the load host endian. */
- ret = load_atom_16(cpu, ra, l.page[0].haddr, l.memop);
+ ret = load_atom_16(cpu, ra, l.page[0].o.haddr, l.memop);
if (l.memop & MO_BSWAP) {
ret = bswap128(ret);
}
@@ -2568,10 +2564,10 @@ static uint64_t do_st_leN(CPUState *cpu,
MMULookupPageData *p,
MemOp atom;
unsigned tmp, half_size;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_st_mmio_leN(cpu, p->full, val_le, p->addr,
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_st_mmio_leN(cpu, &p->o.full, val_le, p->addr,
p->size, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
return val_le >> (p->size * 8);
}
@@ -2582,7 +2578,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p,
atom = mop & MO_ATOM_MASK;
switch (atom) {
case MO_ATOM_SUBALIGN:
- return store_parts_leN(p->haddr, p->size, val_le);
+ return store_parts_leN(p->o.haddr, p->size, val_le);
case MO_ATOM_IFALIGN_PAIR:
case MO_ATOM_WITHIN16_PAIR:
@@ -2593,9 +2589,9 @@ static uint64_t do_st_leN(CPUState *cpu,
MMULookupPageData *p,
? p->size == half_size
: p->size >= half_size) {
if (!HAVE_al8_fast && p->size <= 4) {
- return store_whole_le4(p->haddr, p->size, val_le);
+ return store_whole_le4(p->o.haddr, p->size, val_le);
} else if (HAVE_al8) {
- return store_whole_le8(p->haddr, p->size, val_le);
+ return store_whole_le8(p->o.haddr, p->size, val_le);
} else {
cpu_loop_exit_atomic(cpu, ra);
}
@@ -2605,7 +2601,7 @@ static uint64_t do_st_leN(CPUState *cpu,
MMULookupPageData *p,
case MO_ATOM_IFALIGN:
case MO_ATOM_WITHIN16:
case MO_ATOM_NONE:
- return store_bytes_leN(p->haddr, p->size, val_le);
+ return store_bytes_leN(p->o.haddr, p->size, val_le);
default:
g_assert_not_reached();
@@ -2622,10 +2618,10 @@ static uint64_t do_st16_leN(CPUState *cpu,
MMULookupPageData *p,
int size = p->size;
MemOp atom;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_st16_mmio_leN(cpu, p->full, val_le, p->addr,
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_st16_mmio_leN(cpu, &p->o.full, val_le, p->addr,
size, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
return int128_gethi(val_le) >> ((size - 8) * 8);
}
@@ -2636,8 +2632,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p,
atom = mop & MO_ATOM_MASK;
switch (atom) {
case MO_ATOM_SUBALIGN:
- store_parts_leN(p->haddr, 8, int128_getlo(val_le));
- return store_parts_leN(p->haddr + 8, p->size - 8,
+ store_parts_leN(p->o.haddr, 8, int128_getlo(val_le));
+ return store_parts_leN(p->o.haddr + 8, p->size - 8,
int128_gethi(val_le));
case MO_ATOM_WITHIN16_PAIR:
@@ -2645,7 +2641,7 @@ static uint64_t do_st16_leN(CPUState *cpu,
MMULookupPageData *p,
if (!HAVE_CMPXCHG128) {
cpu_loop_exit_atomic(cpu, ra);
}
- return store_whole_le16(p->haddr, p->size, val_le);
+ return store_whole_le16(p->o.haddr, p->size, val_le);
case MO_ATOM_IFALIGN_PAIR:
/*
@@ -2655,8 +2651,8 @@ static uint64_t do_st16_leN(CPUState *cpu,
MMULookupPageData *p,
case MO_ATOM_IFALIGN:
case MO_ATOM_WITHIN16:
case MO_ATOM_NONE:
- stq_le_p(p->haddr, int128_getlo(val_le));
- return store_bytes_leN(p->haddr + 8, p->size - 8,
+ stq_le_p(p->o.haddr, int128_getlo(val_le));
+ return store_bytes_leN(p->o.haddr + 8, p->size - 8,
int128_gethi(val_le));
default:
@@ -2667,69 +2663,69 @@ static uint64_t do_st16_leN(CPUState *cpu,
MMULookupPageData *p,
static void do_st_1(CPUState *cpu, MMULookupPageData *p, uint8_t val,
int mmu_idx, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
- do_st_mmio_leN(cpu, p->full, val, p->addr, 1, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 1, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
- *(uint8_t *)p->haddr = val;
+ *(uint8_t *)p->o.haddr = val;
}
}
static void do_st_2(CPUState *cpu, MMULookupPageData *p, uint16_t val,
int mmu_idx, MemOp memop, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
if ((memop & MO_BSWAP) != MO_LE) {
val = bswap16(val);
}
- do_st_mmio_leN(cpu, p->full, val, p->addr, 2, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 2, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (memop & MO_BSWAP) {
val = bswap16(val);
}
- store_atom_2(cpu, ra, p->haddr, memop, val);
+ store_atom_2(cpu, ra, p->o.haddr, memop, val);
}
}
static void do_st_4(CPUState *cpu, MMULookupPageData *p, uint32_t val,
int mmu_idx, MemOp memop, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
if ((memop & MO_BSWAP) != MO_LE) {
val = bswap32(val);
}
- do_st_mmio_leN(cpu, p->full, val, p->addr, 4, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 4, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (memop & MO_BSWAP) {
val = bswap32(val);
}
- store_atom_4(cpu, ra, p->haddr, memop, val);
+ store_atom_4(cpu, ra, p->o.haddr, memop, val);
}
}
static void do_st_8(CPUState *cpu, MMULookupPageData *p, uint64_t val,
int mmu_idx, MemOp memop, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
if ((memop & MO_BSWAP) != MO_LE) {
val = bswap64(val);
}
- do_st_mmio_leN(cpu, p->full, val, p->addr, 8, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 8, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (memop & MO_BSWAP) {
val = bswap64(val);
}
- store_atom_8(cpu, ra, p->haddr, memop, val);
+ store_atom_8(cpu, ra, p->o.haddr, memop, val);
}
}
@@ -2822,19 +2818,20 @@ static void do_st16_mmu(CPUState *cpu, vaddr addr, Int128 val,
cpu_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST);
crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_STORE, &l);
if (likely(!crosspage)) {
- if (unlikely(l.page[0].flags & TLB_MMIO)) {
+ if (unlikely(l.page[0].o.flags & TLB_MMIO)) {
if ((l.memop & MO_BSWAP) != MO_LE) {
val = bswap128(val);
}
- do_st16_mmio_leN(cpu, l.page[0].full, val, addr, 16, l.mmu_idx,
ra);
- } else if (unlikely(l.page[0].flags & TLB_DISCARD_WRITE)) {
+ do_st16_mmio_leN(cpu, &l.page[0].o.full, val, addr,
+ 16, l.mmu_idx, ra);
+ } else if (unlikely(l.page[0].o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (l.memop & MO_BSWAP) {
val = bswap128(val);
}
- store_atom_16(cpu, ra, l.page[0].haddr, l.memop, val);
+ store_atom_16(cpu, ra, l.page[0].o.haddr, l.memop, val);
}
return;
}