mirror of
https://github.com/xemu-project/xemu.git
synced 2024-11-23 11:39:53 +00:00
cfa29dd506
The value passed is always true. Reviewed-by: Claudio Fontana <cfontana@suse.de> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
123 lines
3.7 KiB
C
123 lines
3.7 KiB
C
/*
|
|
* Internal execution defines for qemu
|
|
*
|
|
* Copyright (c) 2003 Fabrice Bellard
|
|
*
|
|
* SPDX-License-Identifier: LGPL-2.1-or-later
|
|
*/
|
|
|
|
#ifndef ACCEL_TCG_INTERNAL_H
|
|
#define ACCEL_TCG_INTERNAL_H
|
|
|
|
#include "exec/exec-all.h"
|
|
|
|
/*
|
|
* Access to the various translations structures need to be serialised
|
|
* via locks for consistency. In user-mode emulation access to the
|
|
* memory related structures are protected with mmap_lock.
|
|
* In !user-mode we use per-page locks.
|
|
*/
|
|
#ifdef CONFIG_SOFTMMU
|
|
#define assert_memory_lock()
|
|
#else
|
|
#define assert_memory_lock() tcg_debug_assert(have_mmap_lock())
|
|
#endif
|
|
|
|
typedef struct PageDesc {
|
|
/* list of TBs intersecting this ram page */
|
|
uintptr_t first_tb;
|
|
#ifdef CONFIG_USER_ONLY
|
|
unsigned long flags;
|
|
void *target_data;
|
|
#endif
|
|
#ifdef CONFIG_SOFTMMU
|
|
QemuSpin lock;
|
|
#endif
|
|
} PageDesc;
|
|
|
|
/* Size of the L2 (and L3, etc) page tables. */
|
|
#define V_L2_BITS 10
|
|
#define V_L2_SIZE (1 << V_L2_BITS)
|
|
|
|
/*
|
|
* L1 Mapping properties
|
|
*/
|
|
extern int v_l1_size;
|
|
extern int v_l1_shift;
|
|
extern int v_l2_levels;
|
|
|
|
/*
|
|
* The bottom level has pointers to PageDesc, and is indexed by
|
|
* anything from 4 to (V_L2_BITS + 3) bits, depending on target page size.
|
|
*/
|
|
#define V_L1_MIN_BITS 4
|
|
#define V_L1_MAX_BITS (V_L2_BITS + 3)
|
|
#define V_L1_MAX_SIZE (1 << V_L1_MAX_BITS)
|
|
|
|
extern void *l1_map[V_L1_MAX_SIZE];
|
|
|
|
PageDesc *page_find_alloc(tb_page_addr_t index, bool alloc);
|
|
|
|
static inline PageDesc *page_find(tb_page_addr_t index)
|
|
{
|
|
return page_find_alloc(index, false);
|
|
}
|
|
|
|
/* list iterators for lists of tagged pointers in TranslationBlock */
|
|
#define TB_FOR_EACH_TAGGED(head, tb, n, field) \
|
|
for (n = (head) & 1, tb = (TranslationBlock *)((head) & ~1); \
|
|
tb; tb = (TranslationBlock *)tb->field[n], n = (uintptr_t)tb & 1, \
|
|
tb = (TranslationBlock *)((uintptr_t)tb & ~1))
|
|
|
|
#define PAGE_FOR_EACH_TB(pagedesc, tb, n) \
|
|
TB_FOR_EACH_TAGGED((pagedesc)->first_tb, tb, n, page_next)
|
|
|
|
#define TB_FOR_EACH_JMP(head_tb, tb, n) \
|
|
TB_FOR_EACH_TAGGED((head_tb)->jmp_list_head, tb, n, jmp_list_next)
|
|
|
|
/* In user-mode page locks aren't used; mmap_lock is enough */
|
|
#ifdef CONFIG_USER_ONLY
|
|
#define assert_page_locked(pd) tcg_debug_assert(have_mmap_lock())
|
|
static inline void page_lock(PageDesc *pd) { }
|
|
static inline void page_unlock(PageDesc *pd) { }
|
|
#else
|
|
#ifdef CONFIG_DEBUG_TCG
|
|
void do_assert_page_locked(const PageDesc *pd, const char *file, int line);
|
|
#define assert_page_locked(pd) do_assert_page_locked(pd, __FILE__, __LINE__)
|
|
#else
|
|
#define assert_page_locked(pd)
|
|
#endif
|
|
void page_lock(PageDesc *pd);
|
|
void page_unlock(PageDesc *pd);
|
|
#endif
|
|
#if !defined(CONFIG_USER_ONLY) && defined(CONFIG_DEBUG_TCG)
|
|
void assert_no_pages_locked(void);
|
|
#else
|
|
static inline void assert_no_pages_locked(void) { }
|
|
#endif
|
|
|
|
TranslationBlock *tb_gen_code(CPUState *cpu, target_ulong pc,
|
|
target_ulong cs_base, uint32_t flags,
|
|
int cflags);
|
|
G_NORETURN void cpu_io_recompile(CPUState *cpu, uintptr_t retaddr);
|
|
void page_init(void);
|
|
void tb_htable_init(void);
|
|
void tb_reset_jump(TranslationBlock *tb, int n);
|
|
TranslationBlock *tb_link_page(TranslationBlock *tb, tb_page_addr_t phys_pc,
|
|
tb_page_addr_t phys_page2);
|
|
bool tb_invalidate_phys_page_unwind(tb_page_addr_t addr, uintptr_t pc);
|
|
void cpu_restore_state_from_tb(CPUState *cpu, TranslationBlock *tb,
|
|
uintptr_t host_pc);
|
|
|
|
/* Return the current PC from CPU, which may be cached in TB. */
|
|
static inline target_ulong log_pc(CPUState *cpu, const TranslationBlock *tb)
|
|
{
|
|
#if TARGET_TB_PCREL
|
|
return cpu->cc->get_pc(cpu);
|
|
#else
|
|
return tb_pc(tb);
|
|
#endif
|
|
}
|
|
|
|
#endif /* ACCEL_TCG_INTERNAL_H */
|