|  | /* | 
|  | *  Software MMU support | 
|  | * | 
|  | * This library is free software; you can redistribute it and/or | 
|  | * modify it under the terms of the GNU Lesser General Public | 
|  | * License as published by the Free Software Foundation; either | 
|  | * version 2 of the License, or (at your option) any later version. | 
|  | * | 
|  | * This library is distributed in the hope that it will be useful, | 
|  | * but WITHOUT ANY WARRANTY; without even the implied warranty of | 
|  | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU | 
|  | * Lesser General Public License for more details. | 
|  | * | 
|  | * You should have received a copy of the GNU Lesser General Public | 
|  | * License along with this library; if not, see <http://www.gnu.org/licenses/>. | 
|  | * | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * Generate inline load/store functions for all MMU modes (typically | 
|  | * at least _user and _kernel) as well as _data versions, for all data | 
|  | * sizes. | 
|  | * | 
|  | * Used by target op helpers. | 
|  | * | 
|  | * MMU mode suffixes are defined in target cpu.h. | 
|  | */ | 
|  | #ifndef CPU_LDST_H | 
|  | #define CPU_LDST_H | 
|  |  | 
|  | #if defined(CONFIG_USER_ONLY) | 
|  | /* All direct uses of g2h and h2g need to go away for usermode softmmu.  */ | 
|  | #define g2h(x) ((void *)((unsigned long)(target_ulong)(x) + GUEST_BASE)) | 
|  |  | 
|  | #if HOST_LONG_BITS <= TARGET_VIRT_ADDR_SPACE_BITS | 
|  | #define h2g_valid(x) 1 | 
|  | #else | 
|  | #define h2g_valid(x) ({ \ | 
|  | unsigned long __guest = (unsigned long)(x) - GUEST_BASE; \ | 
|  | (__guest < (1ul << TARGET_VIRT_ADDR_SPACE_BITS)) && \ | 
|  | (!RESERVED_VA || (__guest < RESERVED_VA)); \ | 
|  | }) | 
|  | #endif | 
|  |  | 
|  | #define h2g_nocheck(x) ({ \ | 
|  | unsigned long __ret = (unsigned long)(x) - GUEST_BASE; \ | 
|  | (abi_ulong)__ret; \ | 
|  | }) | 
|  |  | 
|  | #define h2g(x) ({ \ | 
|  | /* Check if given address fits target address space */ \ | 
|  | assert(h2g_valid(x)); \ | 
|  | h2g_nocheck(x); \ | 
|  | }) | 
|  |  | 
|  | #define saddr(x) g2h(x) | 
|  | #define laddr(x) g2h(x) | 
|  |  | 
|  | #else /* !CONFIG_USER_ONLY */ | 
|  | /* NOTE: we use double casts if pointers and target_ulong have | 
|  | different sizes */ | 
|  | #define saddr(x) (uint8_t *)(intptr_t)(x) | 
|  | #define laddr(x) (uint8_t *)(intptr_t)(x) | 
|  | #endif | 
|  |  | 
|  | #define ldub_raw(p) ldub_p(laddr((p))) | 
|  | #define ldsb_raw(p) ldsb_p(laddr((p))) | 
|  | #define lduw_raw(p) lduw_p(laddr((p))) | 
|  | #define ldsw_raw(p) ldsw_p(laddr((p))) | 
|  | #define ldl_raw(p) ldl_p(laddr((p))) | 
|  | #define ldq_raw(p) ldq_p(laddr((p))) | 
|  | #define ldfl_raw(p) ldfl_p(laddr((p))) | 
|  | #define ldfq_raw(p) ldfq_p(laddr((p))) | 
|  | #define stb_raw(p, v) stb_p(saddr((p)), v) | 
|  | #define stw_raw(p, v) stw_p(saddr((p)), v) | 
|  | #define stl_raw(p, v) stl_p(saddr((p)), v) | 
|  | #define stq_raw(p, v) stq_p(saddr((p)), v) | 
|  | #define stfl_raw(p, v) stfl_p(saddr((p)), v) | 
|  | #define stfq_raw(p, v) stfq_p(saddr((p)), v) | 
|  |  | 
|  |  | 
|  | #if defined(CONFIG_USER_ONLY) | 
|  |  | 
|  | /* if user mode, no other memory access functions */ | 
|  | #define ldub(p) ldub_raw(p) | 
|  | #define ldsb(p) ldsb_raw(p) | 
|  | #define lduw(p) lduw_raw(p) | 
|  | #define ldsw(p) ldsw_raw(p) | 
|  | #define ldl(p) ldl_raw(p) | 
|  | #define ldq(p) ldq_raw(p) | 
|  | #define ldfl(p) ldfl_raw(p) | 
|  | #define ldfq(p) ldfq_raw(p) | 
|  | #define stb(p, v) stb_raw(p, v) | 
|  | #define stw(p, v) stw_raw(p, v) | 
|  | #define stl(p, v) stl_raw(p, v) | 
|  | #define stq(p, v) stq_raw(p, v) | 
|  | #define stfl(p, v) stfl_raw(p, v) | 
|  | #define stfq(p, v) stfq_raw(p, v) | 
|  |  | 
|  | #define cpu_ldub_code(env1, p) ldub_raw(p) | 
|  | #define cpu_ldsb_code(env1, p) ldsb_raw(p) | 
|  | #define cpu_lduw_code(env1, p) lduw_raw(p) | 
|  | #define cpu_ldsw_code(env1, p) ldsw_raw(p) | 
|  | #define cpu_ldl_code(env1, p) ldl_raw(p) | 
|  | #define cpu_ldq_code(env1, p) ldq_raw(p) | 
|  |  | 
|  | #define cpu_ldub_data(env, addr) ldub_raw(addr) | 
|  | #define cpu_lduw_data(env, addr) lduw_raw(addr) | 
|  | #define cpu_ldsw_data(env, addr) ldsw_raw(addr) | 
|  | #define cpu_ldl_data(env, addr) ldl_raw(addr) | 
|  | #define cpu_ldq_data(env, addr) ldq_raw(addr) | 
|  |  | 
|  | #define cpu_stb_data(env, addr, data) stb_raw(addr, data) | 
|  | #define cpu_stw_data(env, addr, data) stw_raw(addr, data) | 
|  | #define cpu_stl_data(env, addr, data) stl_raw(addr, data) | 
|  | #define cpu_stq_data(env, addr, data) stq_raw(addr, data) | 
|  |  | 
|  | #define cpu_ldub_kernel(env, addr) ldub_raw(addr) | 
|  | #define cpu_lduw_kernel(env, addr) lduw_raw(addr) | 
|  | #define cpu_ldsw_kernel(env, addr) ldsw_raw(addr) | 
|  | #define cpu_ldl_kernel(env, addr) ldl_raw(addr) | 
|  | #define cpu_ldq_kernel(env, addr) ldq_raw(addr) | 
|  |  | 
|  | #define cpu_stb_kernel(env, addr, data) stb_raw(addr, data) | 
|  | #define cpu_stw_kernel(env, addr, data) stw_raw(addr, data) | 
|  | #define cpu_stl_kernel(env, addr, data) stl_raw(addr, data) | 
|  | #define cpu_stq_kernel(env, addr, data) stq_raw(addr, data) | 
|  |  | 
|  | #define ldub_kernel(p) ldub_raw(p) | 
|  | #define ldsb_kernel(p) ldsb_raw(p) | 
|  | #define lduw_kernel(p) lduw_raw(p) | 
|  | #define ldsw_kernel(p) ldsw_raw(p) | 
|  | #define ldl_kernel(p) ldl_raw(p) | 
|  | #define ldq_kernel(p) ldq_raw(p) | 
|  | #define ldfl_kernel(p) ldfl_raw(p) | 
|  | #define ldfq_kernel(p) ldfq_raw(p) | 
|  | #define stb_kernel(p, v) stb_raw(p, v) | 
|  | #define stw_kernel(p, v) stw_raw(p, v) | 
|  | #define stl_kernel(p, v) stl_raw(p, v) | 
|  | #define stq_kernel(p, v) stq_raw(p, v) | 
|  | #define stfl_kernel(p, v) stfl_raw(p, v) | 
|  | #define stfq_kernel(p, vt) stfq_raw(p, v) | 
|  |  | 
|  | #define cpu_ldub_data(env, addr) ldub_raw(addr) | 
|  | #define cpu_lduw_data(env, addr) lduw_raw(addr) | 
|  | #define cpu_ldl_data(env, addr) ldl_raw(addr) | 
|  |  | 
|  | #define cpu_stb_data(env, addr, data) stb_raw(addr, data) | 
|  | #define cpu_stw_data(env, addr, data) stw_raw(addr, data) | 
|  | #define cpu_stl_data(env, addr, data) stl_raw(addr, data) | 
|  |  | 
|  | #else | 
|  |  | 
|  | /* XXX: find something cleaner. | 
|  | * Furthermore, this is false for 64 bits targets | 
|  | */ | 
|  | #define ldul_user       ldl_user | 
|  | #define ldul_kernel     ldl_kernel | 
|  | #define ldul_hypv       ldl_hypv | 
|  | #define ldul_executive  ldl_executive | 
|  | #define ldul_supervisor ldl_supervisor | 
|  |  | 
|  | /* The memory helpers for tcg-generated code need tcg_target_long etc.  */ | 
|  | #include "tcg.h" | 
|  |  | 
|  | uint8_t helper_ldb_mmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  | uint16_t helper_ldw_mmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  | uint32_t helper_ldl_mmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  | uint64_t helper_ldq_mmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  |  | 
|  | void helper_stb_mmu(CPUArchState *env, target_ulong addr, | 
|  | uint8_t val, int mmu_idx); | 
|  | void helper_stw_mmu(CPUArchState *env, target_ulong addr, | 
|  | uint16_t val, int mmu_idx); | 
|  | void helper_stl_mmu(CPUArchState *env, target_ulong addr, | 
|  | uint32_t val, int mmu_idx); | 
|  | void helper_stq_mmu(CPUArchState *env, target_ulong addr, | 
|  | uint64_t val, int mmu_idx); | 
|  |  | 
|  | uint8_t helper_ldb_cmmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  | uint16_t helper_ldw_cmmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  | uint32_t helper_ldl_cmmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  | uint64_t helper_ldq_cmmu(CPUArchState *env, target_ulong addr, int mmu_idx); | 
|  |  | 
|  | #define CPU_MMU_INDEX 0 | 
|  | #define MEMSUFFIX MMU_MODE0_SUFFIX | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  |  | 
|  | #define CPU_MMU_INDEX 1 | 
|  | #define MEMSUFFIX MMU_MODE1_SUFFIX | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  |  | 
|  | #if (NB_MMU_MODES >= 3) | 
|  |  | 
|  | #define CPU_MMU_INDEX 2 | 
|  | #define MEMSUFFIX MMU_MODE2_SUFFIX | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  | #endif /* (NB_MMU_MODES >= 3) */ | 
|  |  | 
|  | #if (NB_MMU_MODES >= 4) | 
|  |  | 
|  | #define CPU_MMU_INDEX 3 | 
|  | #define MEMSUFFIX MMU_MODE3_SUFFIX | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  | #endif /* (NB_MMU_MODES >= 4) */ | 
|  |  | 
|  | #if (NB_MMU_MODES >= 5) | 
|  |  | 
|  | #define CPU_MMU_INDEX 4 | 
|  | #define MEMSUFFIX MMU_MODE4_SUFFIX | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  | #endif /* (NB_MMU_MODES >= 5) */ | 
|  |  | 
|  | #if (NB_MMU_MODES >= 6) | 
|  |  | 
|  | #define CPU_MMU_INDEX 5 | 
|  | #define MEMSUFFIX MMU_MODE5_SUFFIX | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  | #endif /* (NB_MMU_MODES >= 6) */ | 
|  |  | 
|  | #if (NB_MMU_MODES > 6) | 
|  | #error "NB_MMU_MODES > 6 is not supported for now" | 
|  | #endif /* (NB_MMU_MODES > 6) */ | 
|  |  | 
|  | /* these access are slower, they must be as rare as possible */ | 
|  | #define CPU_MMU_INDEX (cpu_mmu_index(env)) | 
|  | #define MEMSUFFIX _data | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  |  | 
|  | #define ldub(p) ldub_data(p) | 
|  | #define ldsb(p) ldsb_data(p) | 
|  | #define lduw(p) lduw_data(p) | 
|  | #define ldsw(p) ldsw_data(p) | 
|  | #define ldl(p) ldl_data(p) | 
|  | #define ldq(p) ldq_data(p) | 
|  |  | 
|  | #define stb(p, v) stb_data(p, v) | 
|  | #define stw(p, v) stw_data(p, v) | 
|  | #define stl(p, v) stl_data(p, v) | 
|  | #define stq(p, v) stq_data(p, v) | 
|  |  | 
|  | #define CPU_MMU_INDEX (cpu_mmu_index(env)) | 
|  | #define MEMSUFFIX _code | 
|  | #define SOFTMMU_CODE_ACCESS | 
|  |  | 
|  | #define DATA_SIZE 1 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 2 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 4 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #define DATA_SIZE 8 | 
|  | #include "exec/cpu_ldst_template.h" | 
|  |  | 
|  | #undef CPU_MMU_INDEX | 
|  | #undef MEMSUFFIX | 
|  | #undef SOFTMMU_CODE_ACCESS | 
|  |  | 
|  | /** | 
|  | * tlb_vaddr_to_host: | 
|  | * @env: CPUArchState | 
|  | * @addr: guest virtual address to look up | 
|  | * @access_type: 0 for read, 1 for write, 2 for execute | 
|  | * @mmu_idx: MMU index to use for lookup | 
|  | * | 
|  | * Look up the specified guest virtual index in the TCG softmmu TLB. | 
|  | * If the TLB contains a host virtual address suitable for direct RAM | 
|  | * access, then return it. Otherwise (TLB miss, TLB entry is for an | 
|  | * I/O access, etc) return NULL. | 
|  | * | 
|  | * This is the equivalent of the initial fast-path code used by | 
|  | * TCG backends for guest load and store accesses. | 
|  | */ | 
|  | static inline void *tlb_vaddr_to_host(CPUArchState *env, target_ulong addr, | 
|  | int access_type, int mmu_idx) | 
|  | { | 
|  | int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | 
|  | CPUTLBEntry *tlbentry = &env->tlb_table[mmu_idx][index]; | 
|  | target_ulong tlb_addr; | 
|  | uintptr_t haddr; | 
|  |  | 
|  | switch (access_type) { | 
|  | case 0: | 
|  | tlb_addr = tlbentry->addr_read; | 
|  | break; | 
|  | case 1: | 
|  | tlb_addr = tlbentry->addr_write; | 
|  | break; | 
|  | case 2: | 
|  | tlb_addr = tlbentry->addr_code; | 
|  | break; | 
|  | default: | 
|  | g_assert_not_reached(); | 
|  | } | 
|  |  | 
|  | if ((addr & TARGET_PAGE_MASK) | 
|  | != (tlb_addr & (TARGET_PAGE_MASK | TLB_INVALID_MASK))) { | 
|  | /* TLB entry is for a different page */ | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | if (tlb_addr & ~TARGET_PAGE_MASK) { | 
|  | /* IO access */ | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | haddr = addr + env->tlb_table[mmu_idx][index].addend; | 
|  | return (void *)haddr; | 
|  | } | 
|  |  | 
|  | #endif /* defined(CONFIG_USER_ONLY) */ | 
|  |  | 
|  | #endif /* CPU_LDST_H */ |