[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [PATCH v2 08/10] target/ppc: Introduce ppc_xlate
From: |
David Gibson |
Subject: |
Re: [PATCH v2 08/10] target/ppc: Introduce ppc_xlate |
Date: |
Thu, 24 Jun 2021 16:34:24 +1000 |
On Mon, Jun 21, 2021 at 09:51:13AM -0300, Bruno Larsen (billionai) wrote:
> From: Richard Henderson <richard.henderson@linaro.org>
>
> Create one common dispatch for all of the ppc_*_xlate functions.
> Use ppc64_v3_radix to directly dispatch between ppc_radix64_xlate
> and ppc_hash64_xlate.
>
> Remove the separate *_handle_mmu_fault and *_get_phys_page_debug
> functions, using common code for ppc_cpu_tlb_fill and
> ppc_cpu_get_phys_page_debug.
>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Applied to ppc-for-6.1, thanks.
> ---
> target/ppc/mmu-book3s-v3.c | 19 -------
> target/ppc/mmu-book3s-v3.h | 5 --
> target/ppc/mmu-hash32.c | 38 ++------------
> target/ppc/mmu-hash32.h | 6 +--
> target/ppc/mmu-hash64.c | 37 ++------------
> target/ppc/mmu-hash64.h | 6 +--
> target/ppc/mmu-radix64.c | 38 ++------------
> target/ppc/mmu-radix64.h | 6 +--
> target/ppc/mmu_helper.c | 100 ++++++++++++++-----------------------
> 9 files changed, 55 insertions(+), 200 deletions(-)
>
> diff --git a/target/ppc/mmu-book3s-v3.c b/target/ppc/mmu-book3s-v3.c
> index c78fd8dc0e..f4985bae78 100644
> --- a/target/ppc/mmu-book3s-v3.c
> +++ b/target/ppc/mmu-book3s-v3.c
> @@ -23,25 +23,6 @@
> #include "mmu-book3s-v3.h"
> #include "mmu-radix64.h"
>
> -int ppc64_v3_handle_mmu_fault(PowerPCCPU *cpu, vaddr eaddr, int rwx,
> - int mmu_idx)
> -{
> - if (ppc64_v3_radix(cpu)) { /* Guest uses radix */
> - return ppc_radix64_handle_mmu_fault(cpu, eaddr, rwx, mmu_idx);
> - } else { /* Guest uses hash */
> - return ppc_hash64_handle_mmu_fault(cpu, eaddr, rwx, mmu_idx);
> - }
> -}
> -
> -hwaddr ppc64_v3_get_phys_page_debug(PowerPCCPU *cpu, vaddr eaddr)
> -{
> - if (ppc64_v3_radix(cpu)) {
> - return ppc_radix64_get_phys_page_debug(cpu, eaddr);
> - } else {
> - return ppc_hash64_get_phys_page_debug(cpu, eaddr);
> - }
> -}
> -
> bool ppc64_v3_get_pate(PowerPCCPU *cpu, target_ulong lpid, ppc_v3_pate_t
> *entry)
> {
> uint64_t patb = cpu->env.spr[SPR_PTCR] & PTCR_PATB;
> diff --git a/target/ppc/mmu-book3s-v3.h b/target/ppc/mmu-book3s-v3.h
> index 7b89be54b8..a1326df969 100644
> --- a/target/ppc/mmu-book3s-v3.h
> +++ b/target/ppc/mmu-book3s-v3.h
> @@ -67,11 +67,6 @@ static inline bool ppc64_v3_radix(PowerPCCPU *cpu)
> return !!(cpu->env.spr[SPR_LPCR] & LPCR_HR);
> }
>
> -hwaddr ppc64_v3_get_phys_page_debug(PowerPCCPU *cpu, vaddr eaddr);
> -
> -int ppc64_v3_handle_mmu_fault(PowerPCCPU *cpu, vaddr eaddr, int rwx,
> - int mmu_idx);
> -
> static inline hwaddr ppc_hash64_hpt_base(PowerPCCPU *cpu)
> {
> uint64_t base;
> diff --git a/target/ppc/mmu-hash32.c b/target/ppc/mmu-hash32.c
> index ad22372c07..6a07c345e4 100644
> --- a/target/ppc/mmu-hash32.c
> +++ b/target/ppc/mmu-hash32.c
> @@ -424,10 +424,9 @@ static hwaddr ppc_hash32_pte_raddr(target_ulong sr,
> ppc_hash_pte32_t pte,
> return (rpn & ~mask) | (eaddr & mask);
> }
>
> -static bool ppc_hash32_xlate(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type,
> - hwaddr *raddrp, int *psizep, int *protp,
> - bool guest_visible)
> +bool ppc_hash32_xlate(PowerPCCPU *cpu, vaddr eaddr, MMUAccessType
> access_type,
> + hwaddr *raddrp, int *psizep, int *protp,
> + bool guest_visible)
> {
> CPUState *cs = CPU(cpu);
> CPUPPCState *env = &cpu->env;
> @@ -569,34 +568,3 @@ static bool ppc_hash32_xlate(PowerPCCPU *cpu, vaddr
> eaddr,
> *protp = prot;
> return true;
> }
> -
> -int ppc_hash32_handle_mmu_fault(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type, int mmu_idx)
> -{
> - CPUState *cs = CPU(cpu);
> - int page_size, prot;
> - hwaddr raddr;
> -
> - /* Translate eaddr to raddr (where raddr is addr qemu needs for access)
> */
> - if (!ppc_hash32_xlate(cpu, eaddr, access_type, &raddr,
> - &page_size, &prot, true)) {
> - return 1;
> - }
> -
> - tlb_set_page(cs, eaddr & TARGET_PAGE_MASK, raddr & TARGET_PAGE_MASK,
> - prot, mmu_idx, 1UL << page_size);
> - return 0;
> -}
> -
> -hwaddr ppc_hash32_get_phys_page_debug(PowerPCCPU *cpu, target_ulong eaddr)
> -{
> - int psize, prot;
> - hwaddr raddr;
> -
> - if (!ppc_hash32_xlate(cpu, eaddr, MMU_DATA_LOAD, &raddr,
> - &psize, &prot, false)) {
> - return -1;
> - }
> -
> - return raddr & TARGET_PAGE_MASK;
> -}
> diff --git a/target/ppc/mmu-hash32.h b/target/ppc/mmu-hash32.h
> index 30e35718a7..8694eccabd 100644
> --- a/target/ppc/mmu-hash32.h
> +++ b/target/ppc/mmu-hash32.h
> @@ -4,9 +4,9 @@
> #ifndef CONFIG_USER_ONLY
>
> hwaddr get_pteg_offset32(PowerPCCPU *cpu, hwaddr hash);
> -hwaddr ppc_hash32_get_phys_page_debug(PowerPCCPU *cpu, target_ulong addr);
> -int ppc_hash32_handle_mmu_fault(PowerPCCPU *cpu, vaddr address,
> - MMUAccessType access_type, int mmu_idx);
> +bool ppc_hash32_xlate(PowerPCCPU *cpu, vaddr eaddr, MMUAccessType
> access_type,
> + hwaddr *raddrp, int *psizep, int *protp,
> + bool guest_visible);
>
> /*
> * Segment register definitions
> diff --git a/target/ppc/mmu-hash64.c b/target/ppc/mmu-hash64.c
> index c6b167b4dc..c1b98a97e9 100644
> --- a/target/ppc/mmu-hash64.c
> +++ b/target/ppc/mmu-hash64.c
> @@ -873,10 +873,9 @@ static int build_vrma_slbe(PowerPCCPU *cpu, ppc_slb_t
> *slb)
> return -1;
> }
>
> -static bool ppc_hash64_xlate(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type,
> - hwaddr *raddrp, int *psizep, int *protp,
> - bool guest_visible)
> +bool ppc_hash64_xlate(PowerPCCPU *cpu, vaddr eaddr, MMUAccessType
> access_type,
> + hwaddr *raddrp, int *psizep, int *protp,
> + bool guest_visible)
> {
> CPUState *cs = CPU(cpu);
> CPUPPCState *env = &cpu->env;
> @@ -1094,36 +1093,6 @@ static bool ppc_hash64_xlate(PowerPCCPU *cpu, vaddr
> eaddr,
> return true;
> }
>
> -int ppc_hash64_handle_mmu_fault(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type, int mmu_idx)
> -{
> - CPUState *cs = CPU(cpu);
> - int page_size, prot;
> - hwaddr raddr;
> -
> - if (!ppc_hash64_xlate(cpu, eaddr, access_type, &raddr,
> - &page_size, &prot, true)) {
> - return 1;
> - }
> -
> - tlb_set_page(cs, eaddr & TARGET_PAGE_MASK, raddr & TARGET_PAGE_MASK,
> - prot, mmu_idx, 1UL << page_size);
> - return 0;
> -}
> -
> -hwaddr ppc_hash64_get_phys_page_debug(PowerPCCPU *cpu, target_ulong eaddr)
> -{
> - int psize, prot;
> - hwaddr raddr;
> -
> - if (!ppc_hash64_xlate(cpu, eaddr, MMU_DATA_LOAD, &raddr,
> - &psize, &prot, false)) {
> - return -1;
> - }
> -
> - return raddr & TARGET_PAGE_MASK;
> -}
> -
> void ppc_hash64_tlb_flush_hpte(PowerPCCPU *cpu, target_ulong ptex,
> target_ulong pte0, target_ulong pte1)
> {
> diff --git a/target/ppc/mmu-hash64.h b/target/ppc/mmu-hash64.h
> index 3e8a8eec1f..9f338e1fe9 100644
> --- a/target/ppc/mmu-hash64.h
> +++ b/target/ppc/mmu-hash64.h
> @@ -7,9 +7,9 @@
> void dump_slb(PowerPCCPU *cpu);
> int ppc_store_slb(PowerPCCPU *cpu, target_ulong slot,
> target_ulong esid, target_ulong vsid);
> -hwaddr ppc_hash64_get_phys_page_debug(PowerPCCPU *cpu, target_ulong addr);
> -int ppc_hash64_handle_mmu_fault(PowerPCCPU *cpu, vaddr address,
> - MMUAccessType access_type, int mmu_idx);
> +bool ppc_hash64_xlate(PowerPCCPU *cpu, vaddr eaddr, MMUAccessType
> access_type,
> + hwaddr *raddrp, int *psizep, int *protp,
> + bool guest_visible);
> void ppc_hash64_tlb_flush_hpte(PowerPCCPU *cpu,
> target_ulong pte_index,
> target_ulong pte0, target_ulong pte1);
> diff --git a/target/ppc/mmu-radix64.c b/target/ppc/mmu-radix64.c
> index 2d5f0850c9..cbd404bfa4 100644
> --- a/target/ppc/mmu-radix64.c
> +++ b/target/ppc/mmu-radix64.c
> @@ -463,10 +463,9 @@ static int ppc_radix64_process_scoped_xlate(PowerPCCPU
> *cpu,
> * | = On | Process Scoped | Scoped |
> * +-------------+----------------+---------------+
> */
> -static bool ppc_radix64_xlate(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type,
> - hwaddr *raddr, int *psizep, int *protp,
> - bool guest_visible)
> +bool ppc_radix64_xlate(PowerPCCPU *cpu, vaddr eaddr, MMUAccessType
> access_type,
> + hwaddr *raddr, int *psizep, int *protp,
> + bool guest_visible)
> {
> CPUPPCState *env = &cpu->env;
> uint64_t lpid, pid;
> @@ -584,34 +583,3 @@ static bool ppc_radix64_xlate(PowerPCCPU *cpu, vaddr
> eaddr,
>
> return true;
> }
> -
> -int ppc_radix64_handle_mmu_fault(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type, int mmu_idx)
> -{
> - CPUState *cs = CPU(cpu);
> - int page_size, prot;
> - hwaddr raddr;
> -
> - /* Translate eaddr to raddr (where raddr is addr qemu needs for access)
> */
> - if (!ppc_radix64_xlate(cpu, eaddr, access_type, &raddr,
> - &page_size, &prot, true)) {
> - return 1;
> - }
> -
> - tlb_set_page(cs, eaddr & TARGET_PAGE_MASK, raddr & TARGET_PAGE_MASK,
> - prot, mmu_idx, 1UL << page_size);
> - return 0;
> -}
> -
> -hwaddr ppc_radix64_get_phys_page_debug(PowerPCCPU *cpu, target_ulong eaddr)
> -{
> - int psize, prot;
> - hwaddr raddr;
> -
> - if (!ppc_radix64_xlate(cpu, eaddr, MMU_DATA_LOAD, &raddr,
> - &psize, &prot, false)) {
> - return -1;
> - }
> -
> - return raddr & TARGET_PAGE_MASK;
> -}
> diff --git a/target/ppc/mmu-radix64.h b/target/ppc/mmu-radix64.h
> index 94bd72cb38..6b13b89b64 100644
> --- a/target/ppc/mmu-radix64.h
> +++ b/target/ppc/mmu-radix64.h
> @@ -44,9 +44,9 @@
>
> #ifdef TARGET_PPC64
>
> -int ppc_radix64_handle_mmu_fault(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type, int mmu_idx);
> -hwaddr ppc_radix64_get_phys_page_debug(PowerPCCPU *cpu, target_ulong addr);
> +bool ppc_radix64_xlate(PowerPCCPU *cpu, vaddr eaddr, MMUAccessType
> access_type,
> + hwaddr *raddr, int *psizep, int *protp,
> + bool guest_visible);
>
> static inline int ppc_radix64_get_prot_eaa(uint64_t pte)
> {
> diff --git a/target/ppc/mmu_helper.c b/target/ppc/mmu_helper.c
> index 2e92deb105..a0e4e027d3 100644
> --- a/target/ppc/mmu_helper.c
> +++ b/target/ppc/mmu_helper.c
> @@ -2899,98 +2899,72 @@ void helper_check_tlb_flush_global(CPUPPCState *env)
>
>
> /*****************************************************************************/
>
> -static int cpu_ppc_handle_mmu_fault(PowerPCCPU *cpu, vaddr eaddr,
> - MMUAccessType access_type, int mmu_idx)
> +static bool ppc_xlate(PowerPCCPU *cpu, vaddr eaddr, MMUAccessType
> access_type,
> + hwaddr *raddrp, int *psizep, int *protp,
> + int mmu_idx, bool guest_visible)
> {
> - CPUState *cs = CPU(cpu);
> - int page_size, prot;
> - hwaddr raddr;
> -
> - if (!ppc_jumbo_xlate(cpu, eaddr, access_type, &raddr,
> - &page_size, &prot, mmu_idx, true)) {
> - return 1;
> - }
> -
> - tlb_set_page(cs, eaddr & TARGET_PAGE_MASK, raddr & TARGET_PAGE_MASK,
> - prot, mmu_idx, 1UL << page_size);
> - return 0;
> -}
> -
> -hwaddr ppc_cpu_get_phys_page_debug(CPUState *cs, vaddr addr)
> -{
> - PowerPCCPU *cpu = POWERPC_CPU(cs);
> - CPUPPCState *env = &cpu->env;
> - hwaddr raddr;
> - int s, p;
> -
> - switch (env->mmu_model) {
> + switch (cpu->env.mmu_model) {
> #if defined(TARGET_PPC64)
> + case POWERPC_MMU_3_00:
> + if (ppc64_v3_radix(cpu)) {
> + return ppc_radix64_xlate(cpu, eaddr, access_type,
> + raddrp, psizep, protp, guest_visible);
> + }
> + /* fall through */
> case POWERPC_MMU_64B:
> case POWERPC_MMU_2_03:
> case POWERPC_MMU_2_06:
> case POWERPC_MMU_2_07:
> - return ppc_hash64_get_phys_page_debug(cpu, addr);
> - case POWERPC_MMU_3_00:
> - return ppc64_v3_get_phys_page_debug(cpu, addr);
> + return ppc_hash64_xlate(cpu, eaddr, access_type,
> + raddrp, psizep, protp, guest_visible);
> #endif
>
> case POWERPC_MMU_32B:
> case POWERPC_MMU_601:
> - return ppc_hash32_get_phys_page_debug(cpu, addr);
> + return ppc_hash32_xlate(cpu, eaddr, access_type,
> + raddrp, psizep, protp, guest_visible);
>
> default:
> - ;
> + return ppc_jumbo_xlate(cpu, eaddr, access_type, raddrp,
> + psizep, protp, mmu_idx, guest_visible);
> }
> +}
> +
> +hwaddr ppc_cpu_get_phys_page_debug(CPUState *cs, vaddr addr)
> +{
> + PowerPCCPU *cpu = POWERPC_CPU(cs);
> + hwaddr raddr;
> + int s, p;
>
> /*
> * Some MMUs have separate TLBs for code and data. If we only
> * try an MMU_DATA_LOAD, we may not be able to read instructions
> * mapped by code TLBs, so we also try a MMU_INST_FETCH.
> */
> - if (ppc_jumbo_xlate(cpu, addr, MMU_DATA_LOAD, &raddr, &s, &p, 0, false)
> ||
> - ppc_jumbo_xlate(cpu, addr, MMU_INST_FETCH, &raddr, &s, &p, 0,
> false)) {
> + if (ppc_xlate(cpu, addr, MMU_DATA_LOAD, &raddr, &s, &p, 0, false) ||
> + ppc_xlate(cpu, addr, MMU_INST_FETCH, &raddr, &s, &p, 0, false)) {
> return raddr & TARGET_PAGE_MASK;
> }
> return -1;
> }
>
> -
> -bool ppc_cpu_tlb_fill(CPUState *cs, vaddr addr, int size,
> +bool ppc_cpu_tlb_fill(CPUState *cs, vaddr eaddr, int size,
> MMUAccessType access_type, int mmu_idx,
> bool probe, uintptr_t retaddr)
> {
> PowerPCCPU *cpu = POWERPC_CPU(cs);
> - CPUPPCState *env = &cpu->env;
> - int ret;
> -
> - switch (env->mmu_model) {
> -#if defined(TARGET_PPC64)
> - case POWERPC_MMU_64B:
> - case POWERPC_MMU_2_03:
> - case POWERPC_MMU_2_06:
> - case POWERPC_MMU_2_07:
> - ret = ppc_hash64_handle_mmu_fault(cpu, addr, access_type, mmu_idx);
> - break;
> - case POWERPC_MMU_3_00:
> - ret = ppc64_v3_handle_mmu_fault(cpu, addr, access_type, mmu_idx);
> - break;
> -#endif
> -
> - case POWERPC_MMU_32B:
> - case POWERPC_MMU_601:
> - ret = ppc_hash32_handle_mmu_fault(cpu, addr, access_type, mmu_idx);
> - break;
> + hwaddr raddr;
> + int page_size, prot;
>
> - default:
> - ret = cpu_ppc_handle_mmu_fault(cpu, addr, access_type, mmu_idx);
> - break;
> + if (ppc_xlate(cpu, eaddr, access_type, &raddr,
> + &page_size, &prot, mmu_idx, !probe)) {
> + tlb_set_page(cs, eaddr & TARGET_PAGE_MASK, raddr & TARGET_PAGE_MASK,
> + prot, mmu_idx, 1UL << page_size);
> + return true;
> }
> - if (unlikely(ret != 0)) {
> - if (probe) {
> - return false;
> - }
> - raise_exception_err_ra(env, cs->exception_index, env->error_code,
> - retaddr);
> + if (probe) {
> + return false;
> }
> - return true;
> + raise_exception_err_ra(&cpu->env, cs->exception_index,
> + cpu->env.error_code, retaddr);
> }
--
David Gibson | I'll have my music baroque, and my code
david AT gibson.dropbear.id.au | minimalist, thank you. NOT _the_ _other_
| _way_ _around_!
http://www.ozlabs.org/~dgibson
signature.asc
Description: PGP signature
- [PATCH v2 03/10] target/ppc: Push real-mode handling into ppc_radix64_xlate, (continued)
- [PATCH v2 03/10] target/ppc: Push real-mode handling into ppc_radix64_xlate, Bruno Larsen (billionai), 2021/06/21
- [PATCH v2 04/10] target/ppc: Use bool success for ppc_radix64_xlate, Bruno Larsen (billionai), 2021/06/21
- [PATCH v2 05/10] target/ppc: Split out ppc_hash64_xlate, Bruno Larsen (billionai), 2021/06/21
- [PATCH v2 06/10] target/ppc: Split out ppc_hash32_xlate, Bruno Larsen (billionai), 2021/06/21
- [PATCH v2 07/10] target/ppc: Split out ppc_jumbo_xlate, Bruno Larsen (billionai), 2021/06/21
- [PATCH v2 08/10] target/ppc: Introduce ppc_xlate, Bruno Larsen (billionai), 2021/06/21
- Re: [PATCH v2 08/10] target/ppc: Introduce ppc_xlate,
David Gibson <=
- [PATCH v2 10/10] target/ppc: fix address translation bug for radix mmus, Bruno Larsen (billionai), 2021/06/21
- [PATCH v2 09/10] target/ppc: Restrict ppc_cpu_tlb_fill to TCG, Bruno Larsen (billionai), 2021/06/21