|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#ifdef HAVE_CONFIG_H |
|
# include <config.h> |
|
#endif |
|
|
|
|
|
#if ENABLE_JIT |
|
|
|
#include <stdio.h> |
|
#include <lightening.h> |
|
|
|
#include "frames.h" |
|
#include "gsubr.h" |
|
#include "gc-inline.h" |
|
#include "instructions.h" |
|
#include "intrinsics.h" |
|
#include "simpos.h" |
|
#include "threads.h" |
|
#include "vm-builtins.h" |
|
#include "vm-operations.h" |
|
|
|
#ifdef __MINGW32__ |
|
#ifndef WIN32_LEAN_AND_MEAN |
|
#define WIN32_LEAN_AND_MEAN |
|
#endif |
|
#include <windows.h> |
|
#else |
|
#include <sys/mman.h> |
|
#endif |
|
|
|
#if defined __APPLE__ && HAVE_PTHREAD_JIT_WRITE_PROTECT_NP |
|
#include <libkern/OSCacheControl.h> |
|
#endif |
|
|
|
#include "jit.h" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static const uint32_t default_jit_threshold = 1000; |
|
|
|
|
|
|
|
uint32_t scm_jit_counter_threshold = -1; |
|
|
|
|
|
|
|
static int jit_stop_after = -1; |
|
|
|
|
|
|
|
static int jit_pause_when_stopping = 0; |
|
|
|
|
|
static int jit_log_level = 0; |
|
|
|
|
|
|
|
static void (*enter_mcode) (scm_thread *thread, const uint8_t *mcode); |
|
|
|
|
|
static void *exit_mcode; |
|
|
|
|
|
|
|
static void *handle_interrupts_trampoline; |
|
|
|
|
|
|
|
void *scm_jit_return_to_interpreter_trampoline; |
|
|
|
|
|
struct code_arena |
|
{ |
|
#ifdef __MINGW32__ |
|
HANDLE handle; |
|
#endif |
|
uint8_t *base; |
|
size_t used; |
|
size_t size; |
|
struct code_arena *prev; |
|
}; |
|
|
|
|
|
struct pending_reloc |
|
{ |
|
jit_reloc_t reloc; |
|
|
|
|
|
|
|
|
|
|
|
ptrdiff_t target_label_offset; |
|
}; |
|
|
|
|
|
struct scm_jit_state { |
|
jit_state_t *jit; |
|
scm_thread *thread; |
|
const uint32_t *start; |
|
uint32_t *ip; |
|
uint32_t *next_ip; |
|
const uint32_t *end; |
|
uint32_t *entry; |
|
uint8_t *op_attrs; |
|
struct pending_reloc *relocs; |
|
size_t reloc_idx; |
|
size_t reloc_count; |
|
void **labels; |
|
int32_t frame_size_min; |
|
int32_t frame_size_max; |
|
uint32_t register_state; |
|
jit_gpr_t sp_cache_gpr; |
|
jit_fpr_t sp_cache_fpr; |
|
uint32_t sp_cache_gpr_idx; |
|
uint32_t sp_cache_fpr_idx; |
|
struct code_arena *code_arena; |
|
}; |
|
|
|
typedef struct scm_jit_state scm_jit_state; |
|
|
|
static const uint32_t program_word_offset_free_variable = 2; |
|
|
|
static const uint32_t frame_offset_mra = 0 * sizeof(union scm_vm_stack_element); |
|
static const uint32_t frame_offset_vra = 1 * sizeof(union scm_vm_stack_element); |
|
static const uint32_t frame_offset_prev = 2 * sizeof(union scm_vm_stack_element); |
|
static const uint32_t frame_overhead_slots = 3; |
|
|
|
#define DEFINE_THREAD_OFFSET(f) \ |
|
static const uint32_t thread_offset_##f = \ |
|
offsetof (struct scm_thread, f) |
|
|
|
DEFINE_THREAD_OFFSET (handle); |
|
DEFINE_THREAD_OFFSET (pending_asyncs); |
|
DEFINE_THREAD_OFFSET (block_asyncs); |
|
|
|
#define DEFINE_THREAD_VP_OFFSET(f) \ |
|
static const uint32_t thread_offset_##f = \ |
|
offsetof (struct scm_thread, vm) + offsetof (struct scm_vm, f) |
|
|
|
DEFINE_THREAD_VP_OFFSET (fp); |
|
DEFINE_THREAD_VP_OFFSET (sp); |
|
DEFINE_THREAD_VP_OFFSET (ip); |
|
DEFINE_THREAD_VP_OFFSET (stack_limit); |
|
|
|
|
|
static const jit_gpr_t THREAD = JIT_V0; |
|
|
|
|
|
|
|
|
|
static const jit_gpr_t SP = JIT_R0; |
|
|
|
|
|
|
|
|
|
|
|
static const jit_gpr_t FP = JIT_R1; |
|
|
|
|
|
|
|
|
|
static const jit_gpr_t OLD_FP_FOR_RETURN_TRAMPOLINE = JIT_V1; |
|
|
|
|
|
static const jit_gpr_t T0 = JIT_V1; |
|
static const jit_gpr_t T1 = JIT_V2; |
|
static const jit_gpr_t T2 = JIT_R2; |
|
SCM_UNUSED static const jit_gpr_t T3_OR_FP = JIT_R1; |
|
SCM_UNUSED static const jit_gpr_t T4_OR_SP = JIT_R0; |
|
|
|
|
|
|
|
SCM_UNUSED static const jit_gpr_t T0_PRESERVED = JIT_V1; |
|
static const jit_gpr_t T1_PRESERVED = JIT_V2; |
|
|
|
static const uint32_t SP_IN_REGISTER = 0x1; |
|
static const uint32_t FP_IN_REGISTER = 0x2; |
|
static const uint32_t UNREACHABLE = 0x4; |
|
static const uint32_t SP_CACHE_GPR = 0x8; |
|
static const uint32_t SP_CACHE_FPR = 0x10; |
|
|
|
static const uint8_t OP_ATTR_BLOCK = 0x1; |
|
static const uint8_t OP_ATTR_ENTRY = 0x2; |
|
|
|
#ifdef WORDS_BIGENDIAN |
|
#define JIT_BIGENDIAN 1 |
|
#else |
|
#define JIT_BIGENDIAN 0 |
|
#endif |
|
|
|
#if SCM_SIZEOF_UINTPTR_T == 4 |
|
static const uint32_t log2_sizeof_uintptr_t = 2; |
|
#elif SCM_SIZEOF_UINTPTR_T == 8 |
|
static const uint32_t log2_sizeof_uintptr_t = 3; |
|
#else |
|
#error unhandled uintptr_t size |
|
#endif |
|
|
|
#define LENGTH_NOP 0 |
|
#define LENGTH_OP1(a) 1 |
|
#define LENGTH_OP2(a,b) 2 |
|
#define LENGTH_OP3(a,b,c) 3 |
|
#define LENGTH_OP4(a,b,c,d) 4 |
|
#define LENGTH_DOP1(a) 1 |
|
#define LENGTH_DOP2(a,b) 2 |
|
#define LENGTH_DOP3(a,b,c) 3 |
|
#define LENGTH_DOP4(a,b,c,d) 4 |
|
|
|
static const uint8_t op_lengths[256] = { |
|
#define OP_LENGTH(code, cname, name, arity) LENGTH_##arity, |
|
FOR_EACH_VM_OPERATION(OP_LENGTH) |
|
#undef OP_LENGTH |
|
}; |
|
|
|
static void die (int line, const char *msg) SCM_NORETURN; |
|
static void |
|
die (int line, const char *msg) |
|
{ |
|
fprintf (stderr, "jit.c:%d: fatal: %s\n", line, msg); |
|
abort (); |
|
} |
|
|
|
#define DIE(msg) die(__LINE__, msg) |
|
|
|
#define ASSERT(x) \ |
|
do { if (SCM_UNLIKELY (!(x))) DIE ("assertion failed"); } while (0) |
|
|
|
#define UNREACHABLE() \ |
|
DIE ("unreachable") |
|
|
|
#define _LOG(level, ...) \ |
|
do { \ |
|
if (SCM_UNLIKELY (jit_log_level >= level)) \ |
|
fprintf (stderr, "jit: " __VA_ARGS__); \ |
|
} while (0) |
|
|
|
enum { |
|
LOG_LEVEL_NONE, |
|
LOG_LEVEL_INFO, |
|
LOG_LEVEL_DEBUG, |
|
LOG_LEVEL_LOG |
|
}; |
|
|
|
#define INFO(...) _LOG(LOG_LEVEL_INFO, __VA_ARGS__) |
|
#define DEBUG(...) _LOG(LOG_LEVEL_DEBUG, __VA_ARGS__) |
|
#define LOG(...) _LOG(LOG_LEVEL_LOG, __VA_ARGS__) |
|
|
|
static void |
|
reset_register_state (scm_jit_state *j, uint32_t state) |
|
{ |
|
j->register_state = state; |
|
} |
|
|
|
static void |
|
clear_register_state (scm_jit_state *j, uint32_t state) |
|
{ |
|
j->register_state &= ~state; |
|
} |
|
|
|
static void |
|
clear_scratch_register_state (scm_jit_state *j) |
|
{ |
|
reset_register_state (j, 0); |
|
} |
|
|
|
static void |
|
set_register_state (scm_jit_state *j, uint32_t state) |
|
{ |
|
j->register_state |= state; |
|
} |
|
|
|
static uint32_t |
|
unreachable (scm_jit_state *j) |
|
{ |
|
return j->register_state & UNREACHABLE; |
|
} |
|
|
|
static uint32_t |
|
has_register_state (scm_jit_state *j, uint32_t state) |
|
{ |
|
return (j->register_state & state) == state; |
|
} |
|
|
|
#define ASSERT_HAS_REGISTER_STATE(state) \ |
|
ASSERT (unreachable (j) || has_register_state (j, state)); |
|
|
|
static void |
|
record_gpr_clobber (scm_jit_state *j, jit_gpr_t r) |
|
{ |
|
if (jit_same_gprs (j->sp_cache_gpr, r)) |
|
clear_register_state (j, SP_CACHE_GPR); |
|
|
|
if (jit_same_gprs (r, SP)) |
|
clear_register_state (j, SP_IN_REGISTER); |
|
else if (jit_same_gprs (r, FP)) |
|
clear_register_state (j, FP_IN_REGISTER); |
|
} |
|
|
|
static void |
|
record_fpr_clobber (scm_jit_state *j, jit_fpr_t r) |
|
{ |
|
if (jit_same_fprs (j->sp_cache_fpr, r)) |
|
clear_register_state (j, SP_CACHE_FPR); |
|
} |
|
|
|
static void |
|
set_sp_cache_gpr (scm_jit_state *j, uint32_t idx, jit_gpr_t r) |
|
{ |
|
set_register_state (j, SP_CACHE_GPR); |
|
j->sp_cache_gpr_idx = idx; |
|
if (j->sp_cache_fpr_idx == idx) |
|
clear_register_state (j, SP_CACHE_FPR); |
|
} |
|
|
|
static void |
|
set_sp_cache_fpr (scm_jit_state *j, uint32_t idx, jit_fpr_t r) |
|
{ |
|
set_register_state (j, SP_CACHE_FPR); |
|
j->sp_cache_fpr_idx = idx; |
|
if (j->sp_cache_gpr_idx == idx) |
|
clear_register_state (j, SP_CACHE_GPR); |
|
} |
|
|
|
static inline ptrdiff_t |
|
inline_label_offset (uint32_t vcode_offset) |
|
{ |
|
return vcode_offset * 2; |
|
} |
|
|
|
static inline ptrdiff_t |
|
slow_label_offset (uint32_t vcode_offset) |
|
{ |
|
return vcode_offset * 2 + 1; |
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static void |
|
emit_retval (scm_jit_state *j, jit_gpr_t r) |
|
{ |
|
jit_retval (j->jit, r); |
|
record_gpr_clobber (j, r); |
|
} |
|
|
|
static void |
|
emit_retval_d (scm_jit_state *j, jit_fpr_t r) |
|
{ |
|
jit_retval_d (j->jit, r); |
|
record_fpr_clobber (j, r); |
|
} |
|
|
|
static void |
|
emit_movi (scm_jit_state *j, jit_gpr_t r, jit_word_t i) |
|
{ |
|
jit_movi (j->jit, r, i); |
|
record_gpr_clobber (j, r); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_mov_addr (scm_jit_state *j, jit_gpr_t r) |
|
{ |
|
record_gpr_clobber (j, r); |
|
return jit_mov_addr (j->jit, r); |
|
} |
|
|
|
static void |
|
emit_ldxi (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t src, jit_word_t offset) |
|
{ |
|
if (offset == 0) |
|
jit_ldr (j->jit, dst, src); |
|
else |
|
jit_ldxi (j->jit, dst, src, offset); |
|
record_gpr_clobber (j, dst); |
|
} |
|
|
|
#define DEFINE_CLOBBER_RECORDING_EMITTER_R(stem, typ) \ |
|
static void \ |
|
emit_##stem (scm_jit_state *j, jit_##typ##_t dst, jit_##typ##_t a) \ |
|
{ \ |
|
jit_##stem (j->jit, dst, a); \ |
|
record_##typ##_clobber (j, dst); \ |
|
} |
|
|
|
#define DEFINE_CLOBBER_RECORDING_EMITTER_P(stem, typ) \ |
|
static void \ |
|
emit_##stem (scm_jit_state *j, jit_##typ##_t dst, jit_pointer_t a) \ |
|
{ \ |
|
jit_##stem (j->jit, dst, a); \ |
|
record_##typ##_clobber (j, dst); \ |
|
} |
|
|
|
#define DEFINE_CLOBBER_RECORDING_EMITTER_R_I(stem, typ) \ |
|
static void \ |
|
emit_##stem (scm_jit_state *j, jit_##typ##_t dst, \ |
|
jit_##typ##_t a, jit_word_t b) \ |
|
{ \ |
|
jit_##stem (j->jit, dst, a, b); \ |
|
record_##typ##_clobber (j, dst); \ |
|
} |
|
|
|
#define DEFINE_CLOBBER_RECORDING_EMITTER_R_R(stem, typ) \ |
|
static void \ |
|
emit_##stem (scm_jit_state *j, jit_##typ##_t dst, \ |
|
jit_##typ##_t a, jit_##typ##_t b) \ |
|
{ \ |
|
jit_##stem (j->jit, dst, a, b); \ |
|
record_##typ##_clobber (j, dst); \ |
|
} |
|
|
|
#define DEFINE_CLOBBER_RECORDING_EMITTER_R_R_2(stem, typ) \ |
|
static void \ |
|
emit_##stem (scm_jit_state *j, \ |
|
jit_##typ##_t dst1, jit_##typ##_t dst2, \ |
|
jit_##typ##_t a, jit_##typ##_t b) \ |
|
{ \ |
|
jit_##stem (j->jit, dst1, dst2, a, b); \ |
|
record_##typ##_clobber (j, dst1); \ |
|
record_##typ##_clobber (j, dst2); \ |
|
} |
|
|
|
DEFINE_CLOBBER_RECORDING_EMITTER_R(ldr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_P(ldi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R(comr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(ldxr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(addi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(addr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(addr_d, fpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(subi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(subr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(subr_d, fpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(muli, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(mulr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(mulr_d, fpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(divr_d, fpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R(absr_d, fpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R(sqrtr_d, fpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(andi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(andr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(orr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(xorr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(rshi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(rshi_u, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(rshr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(rshr_u, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(lshi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(lshr, gpr) |
|
|
|
#if SIZEOF_UINTPTR_T < 8 |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R(movr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R(negr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(addci, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(addcr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(addxi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(addxr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(subci, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(subcr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_I(subxi, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R(subxr, gpr) |
|
DEFINE_CLOBBER_RECORDING_EMITTER_R_R_2(qmulr_u, gpr) |
|
#endif |
|
|
|
static void |
|
emit_reload_sp (scm_jit_state *j) |
|
{ |
|
emit_ldxi (j, SP, THREAD, thread_offset_sp); |
|
set_register_state (j, SP_IN_REGISTER); |
|
} |
|
|
|
static void |
|
emit_store_sp (scm_jit_state *j) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
jit_stxi (j->jit, thread_offset_sp, THREAD, SP); |
|
} |
|
|
|
static void |
|
emit_reload_fp (scm_jit_state *j) |
|
{ |
|
emit_ldxi (j, FP, THREAD, thread_offset_fp); |
|
set_register_state (j, FP_IN_REGISTER); |
|
} |
|
|
|
static void |
|
emit_store_fp (scm_jit_state *j) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
jit_stxi (j->jit, thread_offset_fp, THREAD, FP); |
|
} |
|
|
|
static uint32_t |
|
save_reloadable_register_state (scm_jit_state *j) |
|
{ |
|
return j->register_state & (SP_IN_REGISTER | FP_IN_REGISTER); |
|
} |
|
|
|
static void |
|
restore_reloadable_register_state (scm_jit_state *j, uint32_t state) |
|
{ |
|
if ((state & SP_IN_REGISTER) && !has_register_state (j, SP_IN_REGISTER)) |
|
emit_reload_sp (j); |
|
if ((state & FP_IN_REGISTER) && !has_register_state (j, FP_IN_REGISTER)) |
|
emit_reload_fp (j); |
|
} |
|
|
|
static void |
|
emit_subtract_stack_slots (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t src, |
|
uint32_t n) |
|
{ |
|
emit_subi (j, dst, src, n * sizeof (union scm_vm_stack_element)); |
|
} |
|
|
|
static void |
|
emit_load_mra (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t fp) |
|
{ |
|
emit_ldxi (j, dst, fp, frame_offset_mra); |
|
} |
|
|
|
static void |
|
emit_store_mra (scm_jit_state *j, jit_gpr_t fp, jit_gpr_t mra) |
|
{ |
|
ASSERT (frame_offset_mra == 0); |
|
jit_str (j->jit, fp, mra); |
|
} |
|
|
|
static void |
|
emit_load_vra (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t fp) |
|
{ |
|
emit_ldxi (j, dst, fp, frame_offset_vra); |
|
} |
|
|
|
static void |
|
emit_store_vra (scm_jit_state *j, jit_gpr_t fp, jit_gpr_t t, const uint32_t *vra) |
|
{ |
|
emit_movi (j, t, (intptr_t) vra); |
|
jit_stxi (j->jit, frame_offset_vra, fp, t); |
|
} |
|
|
|
static void |
|
emit_load_prev_fp_offset (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t fp) |
|
{ |
|
emit_ldxi (j, dst, fp, frame_offset_prev); |
|
} |
|
|
|
static void |
|
emit_store_prev_fp_offset (scm_jit_state *j, jit_gpr_t fp, jit_gpr_t t, |
|
uint32_t n) |
|
{ |
|
emit_movi (j, t, n); |
|
jit_stxi (j->jit, frame_offset_prev, fp, t); |
|
} |
|
|
|
static void |
|
emit_store_ip (scm_jit_state *j, jit_gpr_t ip) |
|
{ |
|
jit_stxi (j->jit, thread_offset_ip, THREAD, ip); |
|
} |
|
|
|
static void |
|
emit_store_current_ip (scm_jit_state *j, jit_gpr_t t) |
|
{ |
|
emit_movi (j, t, (intptr_t) j->ip); |
|
emit_store_ip (j, t); |
|
} |
|
|
|
static void |
|
emit_pop_fp (scm_jit_state *j, jit_gpr_t old_fp) |
|
{ |
|
emit_ldxi (j, old_fp, THREAD, thread_offset_fp); |
|
emit_load_prev_fp_offset (j, FP, old_fp); |
|
emit_lshi (j, FP, FP, 3); |
|
emit_addr (j, FP, old_fp, FP); |
|
set_register_state (j, FP_IN_REGISTER); |
|
emit_store_fp (j); |
|
} |
|
|
|
static void |
|
emit_reset_frame (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
emit_subtract_stack_slots (j, SP, FP, nlocals); |
|
set_register_state (j, SP_IN_REGISTER); |
|
emit_store_sp (j); |
|
clear_register_state (j, SP_CACHE_GPR | SP_CACHE_FPR); |
|
} |
|
|
|
static jit_operand_t |
|
thread_operand (void) |
|
{ |
|
return jit_operand_gpr (JIT_OPERAND_ABI_POINTER, THREAD); |
|
} |
|
|
|
static void |
|
emit_call_0 (scm_jit_state *j, void *f) |
|
{ |
|
jit_calli_0 (j->jit, f); |
|
clear_scratch_register_state (j); |
|
} |
|
|
|
static void |
|
emit_call_1 (scm_jit_state *j, void *f, jit_operand_t a) |
|
{ |
|
jit_calli_1 (j->jit, f, a); |
|
clear_scratch_register_state (j); |
|
} |
|
|
|
static void |
|
emit_call_2 (scm_jit_state *j, void *f, jit_operand_t a, jit_operand_t b) |
|
{ |
|
jit_calli_2 (j->jit, f, a, b); |
|
clear_scratch_register_state (j); |
|
} |
|
|
|
static void |
|
emit_call_3 (scm_jit_state *j, void *f, jit_operand_t a, jit_operand_t b, |
|
jit_operand_t c) |
|
{ |
|
jit_calli_3 (j->jit, f, a, b, c); |
|
clear_scratch_register_state (j); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_alloc_frame_for_sp_fast (scm_jit_state *j, jit_gpr_t t) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
emit_ldxi (j, t, THREAD, thread_offset_stack_limit); |
|
jit_reloc_t slow = jit_bltr (j->jit, SP, t); |
|
emit_store_sp (j); |
|
clear_register_state (j, SP_CACHE_GPR | SP_CACHE_FPR); |
|
|
|
return slow; |
|
} |
|
|
|
static void |
|
emit_alloc_frame_for_sp_slow (scm_jit_state *j, jit_gpr_t t) |
|
{ |
|
|
|
emit_store_current_ip (j, t); |
|
emit_call_2 (j, scm_vm_intrinsics.expand_stack, thread_operand (), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, SP)); |
|
restore_reloadable_register_state (j, SP_IN_REGISTER | FP_IN_REGISTER); |
|
} |
|
|
|
static void |
|
emit_alloc_frame (scm_jit_state *j, jit_gpr_t t, uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
emit_subtract_stack_slots (j, SP, FP, nlocals); |
|
set_register_state (j, SP_IN_REGISTER); |
|
jit_reloc_t slow = emit_alloc_frame_for_sp_fast (j, t); |
|
jit_reloc_t k = jit_jmp (j->jit); |
|
jit_patch_here (j->jit, slow); |
|
emit_alloc_frame_for_sp_slow (j, t); |
|
jit_patch_here (j->jit, k); |
|
} |
|
|
|
static void |
|
emit_get_callee_vcode (scm_jit_state *j, jit_gpr_t dst) |
|
{ |
|
emit_call_1 (j, scm_vm_intrinsics.get_callee_vcode, thread_operand ()); |
|
emit_retval (j, dst); |
|
emit_reload_sp (j); |
|
emit_reload_fp (j); |
|
} |
|
|
|
static void |
|
emit_get_ip_relative_addr (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t ip, |
|
uint32_t offset) |
|
{ |
|
uint32_t byte_offset = offset * sizeof (uint32_t); |
|
jit_ldxi_i (j->jit, dst, ip, byte_offset); |
|
record_gpr_clobber (j, dst); |
|
emit_lshi (j, dst, dst, 2); |
|
emit_addr (j, dst, dst, ip); |
|
} |
|
|
|
static void |
|
emit_exit (scm_jit_state *j) |
|
{ |
|
jit_jmpi (j->jit, exit_mcode); |
|
} |
|
|
|
static void |
|
emit_push_frame (scm_jit_state *j, uint32_t proc_slot, uint32_t nlocals, |
|
const uint32_t *vra) |
|
{ |
|
jit_gpr_t t = T0; |
|
|
|
emit_reload_fp (j); |
|
emit_subtract_stack_slots (j, FP, FP, proc_slot); |
|
set_register_state (j, FP_IN_REGISTER); |
|
emit_store_vra (j, FP, t, vra); |
|
emit_store_prev_fp_offset (j, FP, t, proc_slot); |
|
emit_store_fp (j); |
|
emit_reset_frame (j, nlocals); |
|
} |
|
|
|
static void |
|
emit_indirect_tail_call (scm_jit_state *j) |
|
{ |
|
emit_get_callee_vcode (j, T0); |
|
emit_get_ip_relative_addr (j, T1, T0, 1); |
|
emit_ldxi (j, T1, T1, 0); |
|
jit_reloc_t no_mcode = jit_beqi (j->jit, T1, 0); |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER | SP_IN_REGISTER); |
|
jit_jmpr (j->jit, T1); |
|
|
|
jit_patch_here (j->jit, no_mcode); |
|
|
|
emit_store_ip (j, T0); |
|
emit_exit (j); |
|
} |
|
|
|
static void |
|
emit_direct_tail_call (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER | SP_IN_REGISTER); |
|
|
|
ASSERT ((vcode[0] & 0xff) == scm_op_instrument_entry); |
|
|
|
if (vcode == j->start) |
|
{ |
|
uint8_t *mcode = j->labels[inline_label_offset (0)]; |
|
ASSERT (mcode); |
|
jit_jmpi (j->jit, mcode); |
|
} |
|
else |
|
{ |
|
struct scm_jit_function_data *data; |
|
data = (struct scm_jit_function_data *) (vcode + (int32_t)(vcode[1])); |
|
|
|
if (data->mcode) |
|
{ |
|
|
|
|
|
jit_jmpi (j->jit, data->mcode); |
|
} |
|
else |
|
{ |
|
jit_reloc_t no_mcode; |
|
|
|
|
|
jit_ldi (j->jit, T0, &data->mcode); |
|
no_mcode = jit_beqi (j->jit, T0, 0); |
|
jit_jmpr (j->jit, T0); |
|
jit_patch_here (j->jit, no_mcode); |
|
jit_movi (j->jit, T0, (intptr_t) vcode); |
|
emit_store_ip (j, T0); |
|
emit_exit (j); |
|
} |
|
} |
|
} |
|
|
|
static jit_operand_t |
|
fp_scm_operand (scm_jit_state *j, uint32_t slot) SCM_UNUSED; |
|
static jit_operand_t |
|
fp_scm_operand (scm_jit_state *j, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
|
|
return jit_operand_mem (JIT_OPERAND_ABI_POINTER, FP, |
|
-8 * ((ptrdiff_t) slot + 1)); |
|
} |
|
|
|
static void |
|
emit_fp_ref_scm (scm_jit_state *j, jit_gpr_t dst, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
|
|
emit_ldxi (j, dst, FP, -8 * ((ptrdiff_t) slot + 1)); |
|
} |
|
|
|
static void |
|
emit_fp_set_scm (scm_jit_state *j, uint32_t slot, jit_gpr_t val) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
|
|
jit_stxi (j->jit, -8 * ((ptrdiff_t) slot + 1), FP, val); |
|
clear_register_state (j, SP_CACHE_GPR); |
|
} |
|
|
|
static jit_operand_t |
|
sp_slot_operand (scm_jit_state *j, uint32_t slot) SCM_UNUSED; |
|
static jit_operand_t |
|
sp_slot_operand (scm_jit_state *j, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
return jit_operand_addi (jit_operand_gpr (JIT_OPERAND_ABI_POINTER, SP), |
|
8 * slot); |
|
} |
|
|
|
static jit_operand_t |
|
sp_scm_operand (scm_jit_state *j, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
return jit_operand_mem (JIT_OPERAND_ABI_POINTER, SP, 8 * slot); |
|
} |
|
|
|
static void |
|
emit_sp_ref_scm (scm_jit_state *j, jit_gpr_t dst, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
emit_ldxi (j, dst, SP, 8 * slot); |
|
} |
|
|
|
static void |
|
emit_sp_set_scm (scm_jit_state *j, uint32_t slot, jit_gpr_t val) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
if (slot == 0) |
|
jit_str (j->jit, SP, val); |
|
else |
|
jit_stxi (j->jit, 8 * slot, SP, val); |
|
|
|
set_sp_cache_gpr (j, slot, val); |
|
} |
|
|
|
|
|
|
|
static jit_operand_t |
|
sp_sz_operand (scm_jit_state *j, uint32_t src) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
enum jit_operand_abi abi = |
|
sizeof (size_t) == 4 ? JIT_OPERAND_ABI_UINT32 : JIT_OPERAND_ABI_UINT64; |
|
|
|
if (JIT_BIGENDIAN && sizeof (size_t) == 4) |
|
return jit_operand_mem (abi, SP, src * 8 + 4); |
|
else |
|
return jit_operand_mem (abi, SP, src * 8); |
|
} |
|
|
|
static void |
|
emit_sp_ref_sz (scm_jit_state *j, jit_gpr_t dst, uint32_t src) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
if (JIT_BIGENDIAN && sizeof (size_t) == 4) |
|
emit_ldxi (j, dst, SP, src * 8 + 4); |
|
else |
|
emit_ldxi (j, dst, SP, src * 8); |
|
} |
|
|
|
static void |
|
emit_sp_set_sz (scm_jit_state *j, uint32_t dst, jit_gpr_t src) |
|
{ |
|
size_t offset = dst * 8; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
if (sizeof (size_t) == 4) |
|
{ |
|
size_t lo, hi; |
|
if (JIT_BIGENDIAN) |
|
lo = offset + 4, hi = offset; |
|
else |
|
lo = offset, hi = offset + 4; |
|
|
|
jit_stxi (j->jit, lo, SP, src); |
|
|
|
emit_xorr (j, src, src, src); |
|
jit_stxi (j->jit, hi, SP, src); |
|
} |
|
else |
|
{ |
|
jit_stxi (j->jit, offset, SP, src); |
|
set_sp_cache_gpr (j, dst, src); |
|
} |
|
} |
|
|
|
static jit_operand_t |
|
sp_u64_operand (scm_jit_state *j, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
return jit_operand_mem (JIT_OPERAND_ABI_UINT64, SP, 8 * slot); |
|
} |
|
|
|
#if SIZEOF_UINTPTR_T >= 8 |
|
static void |
|
emit_sp_ref_u64 (scm_jit_state *j, jit_gpr_t dst, uint32_t src) |
|
{ |
|
size_t offset = src * 8; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
emit_ldxi (j, dst, SP, offset); |
|
} |
|
|
|
static void |
|
emit_sp_set_u64 (scm_jit_state *j, uint32_t dst, jit_gpr_t src) |
|
{ |
|
size_t offset = dst * 8; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
if (dst == 0) |
|
jit_str (j->jit, SP, src); |
|
else |
|
jit_stxi (j->jit, offset, SP, src); |
|
|
|
set_sp_cache_gpr (j, dst, src); |
|
} |
|
|
|
static void |
|
emit_sp_ref_s64 (scm_jit_state *j, jit_gpr_t dst, uint32_t src) |
|
{ |
|
emit_sp_ref_u64 (j, dst, src); |
|
} |
|
|
|
static void |
|
emit_sp_set_s64 (scm_jit_state *j, uint32_t dst, jit_gpr_t src) |
|
{ |
|
emit_sp_set_u64 (j, dst, src); |
|
} |
|
|
|
static void |
|
emit_sp_ref_ptr (scm_jit_state *j, jit_gpr_t dst, uint32_t src) |
|
{ |
|
emit_sp_ref_u64 (j, dst, src); |
|
} |
|
|
|
#else |
|
|
|
static jit_operand_t |
|
sp_s32_operand (scm_jit_state *j, uint32_t src) |
|
{ |
|
return sp_sz_operand (j, src); |
|
} |
|
|
|
static void |
|
emit_sp_ref_s32 (scm_jit_state *j, jit_gpr_t dst, uint32_t src) |
|
{ |
|
emit_sp_ref_sz (j, dst, src); |
|
} |
|
|
|
static void |
|
emit_sp_ref_u64 (scm_jit_state *j, jit_gpr_t dst_lo, jit_gpr_t dst_hi, |
|
uint32_t src) |
|
{ |
|
size_t offset = src * 8; |
|
jit_gpr_t first, second; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
#if JIT_BIGENDIAN |
|
first = dst_hi, second = dst_lo; |
|
#else |
|
first = dst_lo, second = dst_hi; |
|
#endif |
|
|
|
emit_ldxi (j, first, SP, offset); |
|
emit_ldxi (j, second, SP, offset + 4); |
|
} |
|
|
|
static void |
|
emit_sp_set_u64 (scm_jit_state *j, uint32_t dst, jit_gpr_t lo, jit_gpr_t hi) |
|
{ |
|
size_t offset = dst * 8; |
|
jit_gpr_t first, second; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
#if JIT_BIGENDIAN |
|
first = hi, second = lo; |
|
#else |
|
first = lo, second = hi; |
|
#endif |
|
|
|
if (offset == 0) |
|
jit_str (j->jit, SP, first); |
|
else |
|
jit_stxi (j->jit, offset, SP, first); |
|
jit_stxi (j->jit, offset + 4, SP, second); |
|
|
|
clear_register_state (j, SP_CACHE_GPR); |
|
} |
|
|
|
static void |
|
emit_sp_ref_s64 (scm_jit_state *j, jit_gpr_t dst_lo, jit_gpr_t dst_hi, |
|
uint32_t src) |
|
{ |
|
emit_sp_ref_u64 (j, dst_lo, dst_hi, src); |
|
} |
|
|
|
static void |
|
emit_sp_set_s64 (scm_jit_state *j, uint32_t dst, jit_gpr_t lo, jit_gpr_t hi) |
|
{ |
|
emit_sp_set_u64 (j, dst, lo, hi); |
|
} |
|
|
|
static void |
|
emit_sp_ref_u64_lower_half (scm_jit_state *j, jit_gpr_t dst, uint32_t src) |
|
{ |
|
size_t offset = src * 8; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
emit_ldxi (j, dst, SP, offset); |
|
} |
|
|
|
static void |
|
emit_sp_ref_ptr (scm_jit_state *j, jit_gpr_t dst, uint32_t src) |
|
{ |
|
emit_sp_ref_u64_lower_half (j, dst, src); |
|
} |
|
#endif |
|
|
|
static jit_operand_t |
|
sp_f64_operand (scm_jit_state *j, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
return jit_operand_mem (JIT_OPERAND_ABI_DOUBLE, SP, 8 * slot); |
|
} |
|
|
|
static void |
|
emit_sp_ref_f64 (scm_jit_state *j, jit_fpr_t dst, uint32_t src) |
|
{ |
|
size_t offset = src * 8; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
if (offset == 0) |
|
jit_ldr_d (j->jit, dst, SP); |
|
else |
|
jit_ldxi_d (j->jit, dst, SP, offset); |
|
|
|
record_fpr_clobber (j, dst); |
|
} |
|
|
|
static void |
|
emit_sp_set_f64 (scm_jit_state *j, uint32_t dst, jit_fpr_t src) |
|
{ |
|
size_t offset = dst * 8; |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
|
|
if (offset == 0) |
|
jit_str_d (j->jit, SP, src); |
|
else |
|
jit_stxi_d (j->jit, offset, SP, src); |
|
|
|
set_sp_cache_fpr (j, dst, src); |
|
} |
|
|
|
static void |
|
emit_mov (scm_jit_state *j, uint32_t dst, uint32_t src, jit_gpr_t t) |
|
{ |
|
emit_sp_ref_scm (j, t, src); |
|
emit_sp_set_scm (j, dst, t); |
|
|
|
|
|
|
|
|
|
|
|
if (sizeof (void*) < sizeof (union scm_vm_stack_element)) |
|
{ |
|
|
|
uintptr_t src_offset = src * sizeof (union scm_vm_stack_element); |
|
uintptr_t dst_offset = dst * sizeof (union scm_vm_stack_element); |
|
|
|
jit_ldxi (j->jit, t, SP, src_offset + sizeof (void*)); |
|
jit_stxi (j->jit, dst_offset + sizeof (void*), SP, t); |
|
|
|
clear_register_state (j, SP_CACHE_GPR | SP_CACHE_FPR); |
|
} |
|
else |
|
|
|
|
|
clear_register_state (j, SP_CACHE_FPR); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_frame_locals_count_less_than (scm_jit_state *j, jit_gpr_t t, |
|
uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER | FP_IN_REGISTER); |
|
|
|
emit_subr (j, t, FP, SP); |
|
return jit_blti (j->jit, t, nlocals * sizeof (union scm_vm_stack_element)); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_frame_locals_count_eq (scm_jit_state *j, jit_gpr_t t, |
|
uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER | FP_IN_REGISTER); |
|
|
|
emit_subr (j, t, FP, SP); |
|
return jit_beqi (j->jit, t, nlocals * sizeof (union scm_vm_stack_element)); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_frame_locals_count_not_eq (scm_jit_state *j, jit_gpr_t t, |
|
uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER | FP_IN_REGISTER); |
|
|
|
emit_subr (j, t, FP, SP); |
|
return jit_bnei (j->jit, t, nlocals * sizeof (union scm_vm_stack_element)); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_frame_locals_count_greater_than (scm_jit_state *j, jit_gpr_t t, |
|
uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER | FP_IN_REGISTER); |
|
|
|
emit_subr (j, t, FP, SP); |
|
return jit_bgti (j->jit, t, nlocals * sizeof (union scm_vm_stack_element)); |
|
} |
|
|
|
static void |
|
emit_load_fp_slot (scm_jit_state *j, jit_gpr_t dst, uint32_t slot) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
|
|
emit_subi (j, dst, FP, (slot + 1) * sizeof (union scm_vm_stack_element)); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_immediate (scm_jit_state *j, jit_gpr_t r) |
|
{ |
|
return jit_bmsi (j->jit, r, 6); |
|
} |
|
|
|
static void |
|
emit_load_heap_object_word (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t r, |
|
uint32_t word) |
|
{ |
|
emit_ldxi (j, dst, r, word * sizeof(SCM)); |
|
} |
|
|
|
static void |
|
emit_load_heap_object_tc (scm_jit_state *j, jit_gpr_t dst, jit_gpr_t r, |
|
scm_t_bits mask) |
|
{ |
|
emit_load_heap_object_word (j, dst, r, 0); |
|
emit_andi (j, dst, dst, mask); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_heap_object_has_tc (scm_jit_state *j, jit_gpr_t r, jit_gpr_t t, |
|
scm_t_bits mask, scm_t_bits tc) |
|
{ |
|
emit_load_heap_object_tc (j, t, r, mask); |
|
return jit_beqi (j->jit, t, tc); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_heap_object_not_tc (scm_jit_state *j, jit_gpr_t r, jit_gpr_t t, |
|
scm_t_bits mask, scm_t_bits tc) |
|
{ |
|
emit_load_heap_object_tc (j, t, r, mask); |
|
return jit_bnei (j->jit, t, tc); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_heap_object_has_tc7 (scm_jit_state *j, jit_gpr_t r, jit_gpr_t t, |
|
scm_t_bits tc7) |
|
{ |
|
return emit_branch_if_heap_object_has_tc (j, r, t, 0x7f, tc7); |
|
} |
|
|
|
static jit_reloc_t |
|
emit_branch_if_heap_object_not_tc7 (scm_jit_state *j, jit_gpr_t r, jit_gpr_t t, |
|
scm_t_bits tc7) |
|
{ |
|
return emit_branch_if_heap_object_not_tc (j, r, t, 0x7f, tc7); |
|
} |
|
|
|
static void |
|
emit_entry_trampoline (scm_jit_state *j) |
|
{ |
|
size_t align = jit_enter_jit_abi(j->jit, 3, 0, 0); |
|
|
|
|
|
|
|
jit_load_args_2 (j->jit, thread_operand (), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, T0)); |
|
emit_reload_sp (j); |
|
|
|
|
|
emit_reload_fp (j); |
|
|
|
|
|
jit_jmpr (j->jit, T0); |
|
|
|
|
|
exit_mcode = jit_address (j->jit); |
|
|
|
jit_leave_jit_abi(j->jit, 3, 0, align); |
|
|
|
|
|
jit_ret (j->jit); |
|
} |
|
|
|
static void |
|
emit_handle_interrupts_trampoline (scm_jit_state *j) |
|
{ |
|
|
|
jit_pop_link_register (j->jit); |
|
emit_call_2 (j, scm_vm_intrinsics.push_interrupt_frame, |
|
thread_operand (), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, JIT_LR)); |
|
emit_reload_sp (j); |
|
emit_reload_fp (j); |
|
emit_direct_tail_call (j, scm_vm_intrinsics.handle_interrupt_code); |
|
} |
|
|
|
|
|
|
|
|
|
|
|
static const size_t default_code_arena_size = 0x40000; |
|
|
|
static struct code_arena * |
|
allocate_code_arena (size_t size, struct code_arena *prev) |
|
{ |
|
struct code_arena *ret = malloc (sizeof (struct code_arena)); |
|
|
|
if (!ret) return NULL; |
|
|
|
memset (ret, 0, sizeof (*ret)); |
|
ret->used = 0; |
|
ret->size = size; |
|
ret->prev = prev; |
|
#ifndef __MINGW32__ |
|
int flags = MAP_PRIVATE | MAP_ANONYMOUS; |
|
#if defined __APPLE__ && HAVE_PTHREAD_JIT_WRITE_PROTECT_NP |
|
flags |= MAP_JIT; |
|
#endif |
|
ret->base = mmap (NULL, ret->size, |
|
PROT_EXEC | PROT_READ | PROT_WRITE, |
|
flags, -1, 0); |
|
if (ret->base == MAP_FAILED) |
|
{ |
|
perror ("allocating JIT code buffer failed"); |
|
free (ret); |
|
return NULL; |
|
} |
|
#else |
|
ret->handle = CreateFileMappingA(INVALID_HANDLE_VALUE, NULL, |
|
PAGE_EXECUTE_READWRITE, |
|
size >> 32, size & 0xffffffff, NULL); |
|
if (ret->handle == NULL) |
|
{ |
|
fprintf (stderr, "allocating JIT code buffer failed: %lu\n", |
|
GetLastError()); |
|
free (ret); |
|
return NULL; |
|
} |
|
ret->base = MapViewOfFile (ret->handle, |
|
FILE_MAP_WRITE | FILE_MAP_EXECUTE | FILE_MAP_COPY, |
|
0, 0, size); |
|
if (ret->base == NULL) |
|
{ |
|
CloseHandle (ret->handle); |
|
fprintf (stderr, "memory mapping JIT code buffer failed: %lu\n", |
|
GetLastError()); |
|
free (ret); |
|
return NULL; |
|
} |
|
#endif |
|
|
|
|
|
INFO ("allocated code arena, %p-%p\n", ret->base, ret->base + ret->size); |
|
|
|
return ret; |
|
} |
|
|
|
static void * |
|
emit_code (scm_jit_state *j, void (*emit) (scm_jit_state *)) |
|
{ |
|
if (!j->code_arena) |
|
j->code_arena = allocate_code_arena (default_code_arena_size, NULL); |
|
|
|
if (!j->code_arena) |
|
|
|
return NULL; |
|
|
|
while (1) |
|
{ |
|
struct code_arena *arena = j->code_arena; |
|
|
|
jit_begin(j->jit, arena->base + arena->used, arena->size - arena->used); |
|
|
|
uint8_t *ret = jit_address (j->jit); |
|
|
|
#if defined __APPLE__ && HAVE_PTHREAD_JIT_WRITE_PROTECT_NP |
|
pthread_jit_write_protect_np(0); |
|
#endif |
|
|
|
emit (j); |
|
|
|
size_t size; |
|
if (!jit_has_overflow (j->jit) && jit_end (j->jit, &size)) |
|
{ |
|
#if defined __APPLE__ && HAVE_PTHREAD_JIT_WRITE_PROTECT_NP |
|
|
|
|
|
pthread_jit_write_protect_np(1); |
|
sys_icache_invalidate(arena->base, arena->size); |
|
#endif |
|
ASSERT (size <= (arena->size - arena->used)); |
|
DEBUG ("mcode: %p,+%zu\n", ret, size); |
|
arena->used += size; |
|
|
|
|
|
arena->used = (arena->used + 15) & ~15; |
|
|
|
|
|
ASSERT (arena->used <= arena->size); |
|
return ret; |
|
} |
|
else |
|
{ |
|
#if defined __APPLE__ && HAVE_PTHREAD_JIT_WRITE_PROTECT_NP |
|
|
|
pthread_jit_write_protect_np(1); |
|
sys_icache_invalidate(arena->base, arena->size); |
|
#endif |
|
jit_reset (j->jit); |
|
if (arena->used == 0) |
|
{ |
|
|
|
|
|
INFO ("code didn't fit in empty arena of size %zu\n", arena->size); |
|
arena = allocate_code_arena (arena->size * 2, arena->prev); |
|
if (!arena) |
|
return NULL; |
|
#ifndef __MINGW32__ |
|
munmap (j->code_arena->base, j->code_arena->size); |
|
#else |
|
UnmapViewOfFile (j->code_arena->base); |
|
CloseHandle (j->code_arena->handle); |
|
#endif |
|
free (j->code_arena); |
|
j->code_arena = arena; |
|
} |
|
else |
|
{ |
|
|
|
|
|
|
|
|
|
INFO ("code didn't fit in arena tail %zu\n", |
|
arena->size - arena->used); |
|
arena = allocate_code_arena (arena->size, arena); |
|
if (!arena) |
|
return NULL; |
|
j->code_arena = arena; |
|
} |
|
} |
|
} |
|
} |
|
|
|
static jit_operand_t |
|
free_variable_operand (scm_jit_state *j, jit_gpr_t src, size_t n) |
|
{ |
|
ptrdiff_t offset = (n + program_word_offset_free_variable) * sizeof(SCM); |
|
return jit_operand_mem (JIT_OPERAND_ABI_POINTER, src, offset); |
|
} |
|
|
|
static void |
|
add_pending_reloc (scm_jit_state *j, jit_reloc_t reloc, ptrdiff_t offset) |
|
{ |
|
if (j->reloc_idx >= j->reloc_count) |
|
{ |
|
size_t count = j->reloc_count * 2; |
|
if (!count) count = 10; |
|
size_t size = sizeof(*j->relocs) * count; |
|
ASSERT(size / sizeof(*j->relocs) == count); |
|
struct pending_reloc *relocs = realloc (j->relocs, size); |
|
if (relocs) |
|
{ |
|
j->reloc_count = count; |
|
j->relocs = relocs; |
|
} |
|
} |
|
|
|
ASSERT (j->reloc_idx < j->reloc_count); |
|
ASSERT (0 <= offset && offset < (j->end - j->start) * 2); |
|
j->relocs[j->reloc_idx].reloc = reloc; |
|
j->relocs[j->reloc_idx].target_label_offset = offset; |
|
j->reloc_idx++; |
|
} |
|
|
|
static void |
|
add_inter_instruction_patch (scm_jit_state *j, jit_reloc_t reloc, |
|
const uint32_t *target) |
|
{ |
|
ASSERT (j->start <= target && target < j->end); |
|
ptrdiff_t offset = inline_label_offset (target - j->start); |
|
|
|
if (j->labels[offset]) |
|
{ |
|
jit_patch_there (j->jit, reloc, j->labels[offset]); |
|
return; |
|
} |
|
|
|
add_pending_reloc (j, reloc, offset); |
|
} |
|
|
|
static void |
|
add_slow_path_patch (scm_jit_state *j, jit_reloc_t reloc) |
|
{ |
|
ASSERT (j->start <= j->ip && j->ip < j->end); |
|
ptrdiff_t offset = slow_label_offset (j->ip - j->start); |
|
add_pending_reloc (j, reloc, offset); |
|
|
|
} |
|
|
|
static void |
|
continue_after_slow_path (scm_jit_state *j, const uint32_t *target) |
|
{ |
|
void *label = j->labels[inline_label_offset (target - j->start)]; |
|
ASSERT (label); |
|
restore_reloadable_register_state (j, SP_IN_REGISTER | FP_IN_REGISTER); |
|
jit_jmpi (j->jit, label); |
|
} |
|
|
|
|
|
|
|
static void |
|
bad_instruction (scm_jit_state *j) |
|
{ |
|
ASSERT (0); |
|
} |
|
|
|
static void |
|
compile_halt (scm_jit_state *j) |
|
{ |
|
bad_instruction (j); |
|
} |
|
static void |
|
compile_halt_slow (scm_jit_state *j) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call (scm_jit_state *j, uint32_t proc, uint32_t nlocals) |
|
{ |
|
jit_reloc_t push_frame = jit_jmp (j->jit); |
|
|
|
void *trampoline = jit_address (j->jit); |
|
reset_register_state (j, FP_IN_REGISTER | SP_IN_REGISTER); |
|
jit_pop_link_register (j->jit); |
|
emit_store_mra (j, FP, JIT_LR); |
|
emit_indirect_tail_call (j); |
|
|
|
jit_patch_here (j->jit, push_frame); |
|
|
|
emit_push_frame (j, proc, nlocals, j->ip + 2); |
|
jit_jmpi_with_link (j->jit, trampoline); |
|
|
|
reset_register_state (j, FP_IN_REGISTER | SP_IN_REGISTER); |
|
j->frame_size_min = proc; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_call_slow (scm_jit_state *j, uint32_t proc, uint32_t nlocals) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_label (scm_jit_state *j, uint32_t proc, uint32_t nlocals, const uint32_t *vcode) |
|
{ |
|
jit_reloc_t push_frame = jit_jmp (j->jit); |
|
|
|
void *trampoline = jit_address (j->jit); |
|
reset_register_state (j, FP_IN_REGISTER | SP_IN_REGISTER); |
|
jit_pop_link_register (j->jit); |
|
emit_store_mra (j, FP, JIT_LR); |
|
emit_direct_tail_call (j, vcode); |
|
|
|
jit_patch_here (j->jit, push_frame); |
|
|
|
emit_push_frame (j, proc, nlocals, j->ip + 3); |
|
jit_jmpi_with_link (j->jit, trampoline); |
|
|
|
reset_register_state (j, FP_IN_REGISTER | SP_IN_REGISTER); |
|
j->frame_size_min = proc; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_call_label_slow (scm_jit_state *j, uint32_t proc, uint32_t nlocals, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_tail_call (scm_jit_state *j) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
restore_reloadable_register_state (j, FP_IN_REGISTER); |
|
|
|
emit_indirect_tail_call (j); |
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_tail_call_slow (scm_jit_state *j) |
|
{ |
|
} |
|
|
|
static void |
|
compile_tail_call_label (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER); |
|
restore_reloadable_register_state (j, FP_IN_REGISTER); |
|
|
|
emit_direct_tail_call (j, vcode); |
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_tail_call_label_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_instrument_entry (scm_jit_state *j, void *data) |
|
{ |
|
} |
|
static void |
|
compile_instrument_entry_slow (scm_jit_state *j, void *data) |
|
{ |
|
} |
|
|
|
static void |
|
compile_instrument_loop (scm_jit_state *j, void *data) |
|
{ |
|
|
|
} |
|
static void |
|
compile_instrument_loop_slow (scm_jit_state *j, void *data) |
|
{ |
|
} |
|
|
|
static void |
|
compile_receive (scm_jit_state *j, uint32_t dst, uint32_t proc, uint32_t nlocals) |
|
{ |
|
jit_gpr_t t = T0; |
|
|
|
add_slow_path_patch |
|
(j, emit_branch_if_frame_locals_count_less_than (j, t, proc + 1)); |
|
emit_fp_ref_scm (j, t, proc); |
|
emit_fp_set_scm (j, dst, t); |
|
emit_reset_frame (j, nlocals); |
|
|
|
j->frame_size_min = j->frame_size_max = nlocals; |
|
} |
|
static void |
|
compile_receive_slow (scm_jit_state *j, uint32_t dst, uint32_t proc, uint32_t nlocals) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_0 (j, scm_vm_intrinsics.error_no_values); |
|
} |
|
|
|
static void |
|
compile_receive_values (scm_jit_state *j, uint32_t proc, uint8_t allow_extra, |
|
uint32_t nvalues) |
|
{ |
|
jit_gpr_t t = T0; |
|
|
|
|
|
|
|
|
|
|
|
restore_reloadable_register_state (j, FP_IN_REGISTER); |
|
|
|
if (allow_extra) |
|
add_slow_path_patch |
|
(j, emit_branch_if_frame_locals_count_less_than (j, t, proc + nvalues)); |
|
else |
|
add_slow_path_patch |
|
(j, emit_branch_if_frame_locals_count_not_eq (j, t, proc + nvalues)); |
|
|
|
j->frame_size_min = proc + nvalues; |
|
j->frame_size_max = allow_extra ? INT32_MAX : j->frame_size_min; |
|
clear_register_state (j, SP_CACHE_GPR | SP_CACHE_FPR); |
|
} |
|
static void |
|
compile_receive_values_slow (scm_jit_state *j, uint32_t proc, uint8_t allow_extra, |
|
uint32_t nvalues) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
if (allow_extra) |
|
emit_call_0 (j, scm_vm_intrinsics.error_not_enough_values); |
|
else |
|
emit_call_1 (j, scm_vm_intrinsics.error_wrong_number_of_values, |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT32, nvalues)); |
|
} |
|
|
|
static void |
|
compile_shuffle_down (scm_jit_state *j, uint32_t from, uint32_t to) |
|
{ |
|
jit_gpr_t walk = T0, t = T1; |
|
size_t offset = (from - to) * sizeof (union scm_vm_stack_element); |
|
|
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER | FP_IN_REGISTER); |
|
|
|
emit_load_fp_slot (j, walk, from); |
|
jit_reloc_t done = jit_bltr (j->jit, walk, SP); |
|
void *head = jit_address (j->jit); |
|
jit_ldr (j->jit, t, walk); |
|
jit_stxi (j->jit, offset, walk, t); |
|
jit_subi (j->jit, walk, walk, sizeof (union scm_vm_stack_element)); |
|
jit_patch_there (j->jit, jit_bger (j->jit, walk, SP), head); |
|
jit_patch_here (j->jit, done); |
|
jit_addi (j->jit, SP, SP, offset); |
|
emit_store_sp (j); |
|
|
|
clear_register_state (j, SP_CACHE_GPR | SP_CACHE_FPR); |
|
|
|
j->frame_size_min -= (from - to); |
|
if (j->frame_size_max != INT32_MAX) |
|
j->frame_size_max -= (from - to); |
|
} |
|
static void |
|
compile_shuffle_down_slow (scm_jit_state *j, uint32_t from, uint32_t to) |
|
{ |
|
} |
|
|
|
static void |
|
compile_return_values (scm_jit_state *j) |
|
{ |
|
emit_pop_fp (j, OLD_FP_FOR_RETURN_TRAMPOLINE); |
|
emit_load_mra (j, JIT_LR, OLD_FP_FOR_RETURN_TRAMPOLINE); |
|
jit_push_link_register (j->jit); |
|
jit_ret (j->jit); |
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_return_values_slow (scm_jit_state *j) |
|
{ |
|
} |
|
|
|
static void |
|
emit_return_to_interpreter_trampoline (scm_jit_state *j) |
|
{ |
|
jit_gpr_t ra = T1; |
|
|
|
emit_load_vra (j, ra, OLD_FP_FOR_RETURN_TRAMPOLINE); |
|
emit_store_ip (j, ra); |
|
emit_exit (j); |
|
} |
|
|
|
static void |
|
compile_subr_call (scm_jit_state *j, uint32_t idx) |
|
{ |
|
jit_gpr_t t = T0, ret = T1; |
|
void *subr; |
|
jit_reloc_t immediate; |
|
jit_operand_t args[SCM_GSUBR_MAX]; |
|
|
|
ASSERT (j->frame_size_min == j->frame_size_max); |
|
size_t argc = j->frame_size_max - 1; |
|
ASSERT (argc <= SCM_GSUBR_MAX); |
|
|
|
subr = scm_subr_function_by_index (idx); |
|
emit_store_current_ip (j, t); |
|
for (size_t i = 2; i <= j->frame_size_max; i++) |
|
args[i - 2] = sp_scm_operand (j, (j->frame_size_max - i)); |
|
jit_calli (j->jit, subr, argc, args); |
|
clear_scratch_register_state (j); |
|
jit_retval (j->jit, ret); |
|
|
|
immediate = emit_branch_if_immediate (j, ret); |
|
add_slow_path_patch |
|
(j, emit_branch_if_heap_object_has_tc7 (j, ret, t, scm_tc7_values)); |
|
|
|
jit_patch_here (j->jit, immediate); |
|
emit_reload_fp (j); |
|
emit_subtract_stack_slots (j, SP, FP, 1); |
|
set_register_state (j, SP_IN_REGISTER); |
|
emit_store_sp (j); |
|
jit_str (j->jit, SP, ret); |
|
|
|
clear_register_state (j, SP_CACHE_GPR | SP_CACHE_FPR); |
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_subr_call_slow (scm_jit_state *j, uint32_t idx) |
|
{ |
|
jit_gpr_t ret = T1; |
|
emit_call_2 (j, scm_vm_intrinsics.unpack_values_object, thread_operand (), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, ret)); |
|
continue_after_slow_path (j, j->next_ip); |
|
} |
|
|
|
static void |
|
compile_foreign_call (scm_jit_state *j, uint32_t cif_idx, uint32_t ptr_idx) |
|
{ |
|
uint32_t saved_state; |
|
|
|
ASSERT (j->frame_size_min == j->frame_size_max); |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_sp_ref_scm (j, T0, j->frame_size_min - 1); |
|
|
|
|
|
saved_state = save_reloadable_register_state (j); |
|
emit_call_3 (j, scm_vm_intrinsics.foreign_call, thread_operand (), |
|
free_variable_operand (j, T0, cif_idx), |
|
free_variable_operand (j, T0, ptr_idx)); |
|
restore_reloadable_register_state (j, saved_state); |
|
|
|
j->frame_size_min = j->frame_size_max = 2; |
|
} |
|
static void |
|
compile_foreign_call_slow (scm_jit_state *j, uint32_t cif_idx, uint32_t ptr_idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_continuation_call (scm_jit_state *j, uint32_t contregs_idx) |
|
{ |
|
emit_reload_fp (j); |
|
emit_store_current_ip (j, T0); |
|
emit_fp_ref_scm (j, T0, 0); |
|
emit_call_2 (j, scm_vm_intrinsics.reinstate_continuation_x, |
|
thread_operand (), free_variable_operand (j, T0, contregs_idx)); |
|
|
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_continuation_call_slow (scm_jit_state *j, uint32_t contregs_idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_compose_continuation (scm_jit_state *j, uint32_t cont_idx) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (SP_IN_REGISTER | FP_IN_REGISTER); |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_fp_ref_scm (j, T0, 0); |
|
emit_call_2 (j, scm_vm_intrinsics.compose_continuation, |
|
thread_operand (), free_variable_operand (j, T0, cont_idx)); |
|
jit_retval (j->jit, T0); |
|
add_slow_path_patch (j, jit_beqi (j->jit, T0, 0)); |
|
emit_reload_sp (j); |
|
emit_reload_fp (j); |
|
jit_jmpr (j->jit, T0); |
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_compose_continuation_slow (scm_jit_state *j, uint32_t cont_idx) |
|
{ |
|
emit_exit (j); |
|
} |
|
|
|
static void |
|
compile_capture_continuation (scm_jit_state *j, uint32_t dst) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, scm_vm_intrinsics.capture_continuation, thread_operand ()); |
|
jit_retval (j->jit, T0); |
|
emit_reload_sp (j); |
|
emit_reload_fp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_capture_continuation_slow (scm_jit_state *j, uint32_t dst) |
|
{ |
|
} |
|
|
|
static void |
|
compile_abort (scm_jit_state *j) |
|
{ |
|
jit_movi (j->jit, T0, (intptr_t) (j->ip + 1)); |
|
emit_store_ip (j, T0); |
|
jit_reloc_t k = jit_mov_addr (j->jit, T0); |
|
emit_call_2 (j, scm_vm_intrinsics.abort_to_prompt, thread_operand (), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, T0)); |
|
jit_retval (j->jit, T1_PRESERVED); |
|
|
|
add_slow_path_patch(j, jit_beqi (j->jit, T1_PRESERVED, 0)); |
|
emit_reload_sp (j); |
|
emit_reload_fp (j); |
|
jit_jmpr (j->jit, T1_PRESERVED); |
|
|
|
jit_patch_here (j->jit, k); |
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_abort_slow (scm_jit_state *j) |
|
{ |
|
emit_exit (j); |
|
} |
|
|
|
static void |
|
compile_builtin_ref (scm_jit_state *j, uint32_t dst, uint16_t idx) |
|
{ |
|
SCM builtin = scm_vm_builtin_ref (idx); |
|
|
|
emit_movi (j, T0, SCM_UNPACK (builtin)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_builtin_ref_slow (scm_jit_state *j, uint32_t dst, uint16_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_throw (scm_jit_state *j, uint32_t key, uint32_t args) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_2 (j, scm_vm_intrinsics.throw_, sp_scm_operand (j, key), |
|
sp_scm_operand (j, args)); |
|
|
|
set_register_state (j, UNREACHABLE); |
|
} |
|
static void |
|
compile_throw_slow (scm_jit_state *j, uint32_t key, uint32_t args) |
|
{ |
|
} |
|
|
|
static void |
|
compile_throw_value (scm_jit_state *j, uint32_t val, |
|
const void *key_subr_and_message) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_2 (j, scm_vm_intrinsics.throw_with_value, sp_scm_operand (j, val), |
|
jit_operand_imm (JIT_OPERAND_ABI_POINTER, |
|
(intptr_t) key_subr_and_message)); |
|
|
|
set_register_state (j, UNREACHABLE); |
|
} |
|
static void |
|
compile_throw_value_slow (scm_jit_state *j, uint32_t val, |
|
const void *key_subr_and_message) |
|
{ |
|
} |
|
|
|
static void |
|
compile_throw_value_and_data (scm_jit_state *j, uint32_t val, |
|
const void *key_subr_and_message) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_2 (j, scm_vm_intrinsics.throw_with_value_and_data, |
|
sp_scm_operand (j, val), |
|
jit_operand_imm (JIT_OPERAND_ABI_POINTER, |
|
(intptr_t) key_subr_and_message)); |
|
|
|
set_register_state (j, UNREACHABLE); |
|
} |
|
static void |
|
compile_throw_value_and_data_slow (scm_jit_state *j, uint32_t val, |
|
const void *key_subr_and_message) |
|
{ |
|
} |
|
|
|
static void |
|
compile_unreachable (scm_jit_state *j) |
|
{ |
|
jit_breakpoint (j->jit); |
|
set_register_state (j, UNREACHABLE); |
|
} |
|
static void |
|
compile_unreachable_slow (scm_jit_state *j) |
|
{ |
|
} |
|
|
|
static void |
|
compile_assert_nargs_ee (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
add_slow_path_patch |
|
(j, emit_branch_if_frame_locals_count_not_eq (j, T0, nlocals)); |
|
|
|
j->frame_size_min = j->frame_size_max = nlocals; |
|
} |
|
static void |
|
compile_assert_nargs_ee_slow (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, scm_vm_intrinsics.error_wrong_num_args, |
|
thread_operand ()); |
|
} |
|
|
|
static void |
|
compile_assert_nargs_ge (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
if (nlocals > 0) |
|
add_slow_path_patch |
|
(j, emit_branch_if_frame_locals_count_less_than (j, T0, nlocals)); |
|
|
|
j->frame_size_min = nlocals; |
|
} |
|
static void |
|
compile_assert_nargs_ge_slow (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, scm_vm_intrinsics.error_wrong_num_args, |
|
thread_operand ()); |
|
} |
|
|
|
static void |
|
compile_assert_nargs_le (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
add_slow_path_patch |
|
(j, emit_branch_if_frame_locals_count_greater_than (j, T0, nlocals)); |
|
|
|
j->frame_size_max = nlocals; |
|
} |
|
static void |
|
compile_assert_nargs_le_slow (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, scm_vm_intrinsics.error_wrong_num_args, |
|
thread_operand ()); |
|
} |
|
|
|
static void |
|
compile_alloc_frame (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER); |
|
emit_subtract_stack_slots (j, SP, FP, nlocals); |
|
set_register_state (j, SP_IN_REGISTER); |
|
add_slow_path_patch (j, emit_alloc_frame_for_sp_fast (j, T0)); |
|
|
|
j->frame_size_min = j->frame_size_max = nlocals; |
|
} |
|
static void |
|
compile_alloc_frame_slow (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
emit_alloc_frame_for_sp_slow (j, T0); |
|
continue_after_slow_path (j, j->next_ip); |
|
} |
|
|
|
static void |
|
compile_reset_frame (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
restore_reloadable_register_state (j, FP_IN_REGISTER); |
|
emit_reset_frame (j, nlocals); |
|
|
|
j->frame_size_min = j->frame_size_max = nlocals; |
|
} |
|
static void |
|
compile_reset_frame_slow (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
} |
|
|
|
static void |
|
compile_push (scm_jit_state *j, uint32_t src) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_push_slow (scm_jit_state *j, uint32_t src) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
|
|
static void |
|
compile_pop (scm_jit_state *j, uint32_t dst) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_pop_slow (scm_jit_state *j, uint32_t dst) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
|
|
static void |
|
compile_drop (scm_jit_state *j, uint32_t nvalues) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_drop_slow (scm_jit_state *j, uint32_t nvalues) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
|
|
static void |
|
compile_assert_nargs_ee_locals (scm_jit_state *j, uint32_t expected, |
|
uint32_t nlocals) |
|
{ |
|
jit_gpr_t t = T0; |
|
|
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER | SP_IN_REGISTER); |
|
if (nlocals) |
|
{ |
|
emit_subtract_stack_slots (j, SP, SP, nlocals); |
|
set_register_state (j, SP_IN_REGISTER); |
|
} |
|
add_slow_path_patch |
|
(j, emit_branch_if_frame_locals_count_not_eq (j, t, expected + nlocals)); |
|
|
|
if (nlocals) |
|
add_slow_path_patch (j, emit_alloc_frame_for_sp_fast (j, t)); |
|
|
|
j->frame_size_min = j->frame_size_max = expected + nlocals; |
|
} |
|
static void |
|
compile_assert_nargs_ee_locals_slow (scm_jit_state *j, uint32_t expected, |
|
uint32_t nlocals) |
|
{ |
|
jit_gpr_t t = T0; |
|
|
|
reset_register_state (j, SP_IN_REGISTER | FP_IN_REGISTER); |
|
jit_reloc_t args_ok = |
|
emit_branch_if_frame_locals_count_eq (j, t, expected + nlocals); |
|
emit_store_current_ip (j, t); |
|
emit_call_1 (j, scm_vm_intrinsics.error_wrong_num_args, |
|
thread_operand ()); |
|
jit_patch_here (j->jit, args_ok); |
|
|
|
if (nlocals) |
|
emit_alloc_frame_for_sp_slow (j, t); |
|
|
|
continue_after_slow_path (j, j->next_ip); |
|
} |
|
|
|
static void |
|
compile_expand_apply_argument (scm_jit_state *j) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, scm_vm_intrinsics.expand_apply_argument, thread_operand ()); |
|
emit_reload_sp (j); |
|
emit_reload_fp (j); |
|
|
|
j->frame_size_min--; |
|
j->frame_size_max = INT32_MAX; |
|
} |
|
static void |
|
compile_expand_apply_argument_slow (scm_jit_state *j) |
|
{ |
|
} |
|
|
|
static void |
|
compile_bind_kwargs (scm_jit_state *j, uint32_t nreq, uint8_t flags, |
|
uint32_t nreq_and_opt, uint32_t ntotal, const void *kw) |
|
{ |
|
uint8_t allow_other_keys = flags & 0x1, has_rest = flags & 0x2; |
|
jit_gpr_t t = T0, npositional = T1; |
|
|
|
emit_store_current_ip (j, t); |
|
|
|
emit_call_3 (j, scm_vm_intrinsics.compute_kwargs_npositional, |
|
thread_operand (), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT32, nreq), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT32, nreq_and_opt - nreq)); |
|
jit_retval_i (j->jit, npositional); |
|
|
|
jit_operand_t args[] = |
|
{ jit_operand_gpr (JIT_OPERAND_ABI_POINTER, THREAD), |
|
jit_operand_gpr (JIT_OPERAND_ABI_UINT32, npositional), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT32, ntotal), |
|
jit_operand_imm (JIT_OPERAND_ABI_POINTER, (intptr_t)kw), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT8, !has_rest), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT8, allow_other_keys) }; |
|
jit_calli (j->jit, scm_vm_intrinsics.bind_kwargs, 6, args); |
|
clear_scratch_register_state (j); |
|
|
|
if (has_rest) |
|
{ |
|
emit_call_2 (j, scm_vm_intrinsics.cons_rest, thread_operand (), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT32, ntotal)); |
|
jit_retval (j->jit, t); |
|
emit_reload_fp (j); |
|
emit_fp_set_scm (j, nreq_and_opt, t); |
|
} |
|
else |
|
emit_reload_fp (j); |
|
|
|
emit_reset_frame (j, ntotal); |
|
j->frame_size_min = j->frame_size_max = ntotal; |
|
} |
|
static void |
|
compile_bind_kwargs_slow (scm_jit_state *j, uint32_t nreq, uint8_t flags, |
|
uint32_t nreq_and_opt, uint32_t ntotal, const void *kw) |
|
{ |
|
} |
|
|
|
static void |
|
compile_bind_rest (scm_jit_state *j, uint32_t dst) |
|
{ |
|
jit_reloc_t k, cons; |
|
jit_gpr_t t = T1; |
|
|
|
|
|
|
|
|
|
|
|
restore_reloadable_register_state (j, FP_IN_REGISTER); |
|
|
|
cons = emit_branch_if_frame_locals_count_greater_than (j, t, dst); |
|
|
|
emit_alloc_frame (j, t, dst + 1); |
|
emit_movi (j, t, SCM_UNPACK (SCM_EOL)); |
|
emit_sp_set_scm (j, 0, t); |
|
k = jit_jmp (j->jit); |
|
|
|
jit_patch_here (j->jit, cons); |
|
emit_store_current_ip (j, t); |
|
emit_call_2 (j, scm_vm_intrinsics.cons_rest, thread_operand (), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT32, dst)); |
|
emit_retval (j, t); |
|
compile_reset_frame (j, dst + 1); |
|
emit_sp_set_scm (j, 0, t); |
|
|
|
jit_patch_here (j->jit, k); |
|
|
|
j->frame_size_min = dst + 1; |
|
} |
|
static void |
|
compile_bind_rest_slow (scm_jit_state *j, uint32_t dst) |
|
{ |
|
} |
|
|
|
static void |
|
compile_bind_optionals (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER | SP_IN_REGISTER); |
|
ASSERT(j->frame_size_min < nlocals); |
|
ASSERT(j->frame_size_min < j->frame_size_max); |
|
|
|
jit_gpr_t saved_frame_size = T1_PRESERVED; |
|
jit_subr (j->jit, saved_frame_size, FP, SP); |
|
|
|
jit_reloc_t no_optionals = jit_bgei |
|
(j->jit, saved_frame_size, nlocals * sizeof (union scm_vm_stack_element)); |
|
|
|
emit_alloc_frame (j, T0, nlocals); |
|
j->frame_size_min = nlocals; |
|
|
|
jit_gpr_t walk = saved_frame_size; |
|
jit_subr (j->jit, walk, FP, saved_frame_size); |
|
|
|
jit_reloc_t done = jit_bler (j->jit, walk, SP); |
|
jit_movi (j->jit, T0, SCM_UNPACK (SCM_UNDEFINED)); |
|
|
|
void *head = jit_address (j->jit); |
|
jit_subi (j->jit, walk, walk, sizeof (union scm_vm_stack_element)); |
|
jit_str (j->jit, walk, T0); |
|
jit_patch_there (j->jit, jit_bner (j->jit, walk, SP), head); |
|
|
|
jit_patch_here (j->jit, done); |
|
jit_patch_here (j->jit, no_optionals); |
|
} |
|
static void |
|
compile_bind_optionals_slow (scm_jit_state *j, uint32_t nlocals) |
|
{ |
|
} |
|
|
|
static void |
|
compile_allocate_words (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
jit_gpr_t t = T0; |
|
|
|
emit_store_current_ip (j, t); |
|
emit_call_2 (j, scm_vm_intrinsics.allocate_words, thread_operand (), |
|
sp_sz_operand (j, nwords)); |
|
emit_retval (j, t); |
|
record_gpr_clobber (j, t); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, t); |
|
} |
|
static void |
|
compile_allocate_words_slow (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
} |
|
|
|
static void |
|
compile_allocate_words_immediate (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
size_t bytes = nwords * sizeof(SCM); |
|
size_t idx = scm_inline_gc_bytes_to_freelist_index (bytes); |
|
|
|
if (SCM_UNLIKELY (idx >= SCM_INLINE_GC_FREELIST_COUNT)) |
|
{ |
|
jit_gpr_t t = T0; |
|
emit_store_current_ip (j, t); |
|
emit_call_1 (j, GC_malloc, jit_operand_imm (JIT_OPERAND_ABI_WORD, bytes)); |
|
emit_retval (j, t); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, t); |
|
} |
|
else |
|
{ |
|
jit_gpr_t res = T0; |
|
ptrdiff_t offset = offsetof(struct scm_thread, freelists); |
|
offset += idx * sizeof(void*); |
|
emit_ldxi (j, res, THREAD, offset); |
|
add_slow_path_patch (j, jit_beqi (j->jit, res, 0)); |
|
jit_gpr_t new_freelist = T1; |
|
emit_ldr (j, new_freelist, res); |
|
jit_stxi (j->jit, offset, THREAD, new_freelist); |
|
emit_sp_set_scm (j, dst, res); |
|
} |
|
} |
|
static void |
|
compile_allocate_words_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
size_t bytes = nwords * sizeof(SCM); |
|
size_t idx = scm_inline_gc_bytes_to_freelist_index (bytes); |
|
|
|
if (SCM_UNLIKELY (idx >= SCM_INLINE_GC_FREELIST_COUNT)) |
|
{ |
|
} |
|
else |
|
{ |
|
jit_gpr_t res = T0; |
|
emit_store_current_ip (j, res); |
|
emit_call_2 (j, scm_vm_intrinsics.allocate_words_with_freelist, |
|
thread_operand (), |
|
jit_operand_imm (JIT_OPERAND_ABI_WORD, idx)); |
|
emit_retval (j, res); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, res); |
|
continue_after_slow_path (j, j->next_ip); |
|
} |
|
} |
|
|
|
static void |
|
compile_allocate_pointerless_words (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
jit_gpr_t t = T0; |
|
|
|
emit_store_current_ip (j, t); |
|
emit_call_2 (j, scm_vm_intrinsics.allocate_pointerless_words, thread_operand (), |
|
sp_sz_operand (j, nwords)); |
|
emit_retval (j, t); |
|
record_gpr_clobber (j, t); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, t); |
|
} |
|
static void |
|
compile_allocate_pointerless_words_slow (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
} |
|
|
|
static void |
|
compile_allocate_pointerless_words_immediate (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
size_t bytes = nwords * sizeof(SCM); |
|
size_t idx = scm_inline_gc_bytes_to_freelist_index (bytes); |
|
|
|
if (SCM_UNLIKELY (idx >= SCM_INLINE_GC_FREELIST_COUNT)) |
|
{ |
|
jit_gpr_t t = T0; |
|
emit_store_current_ip (j, t); |
|
emit_call_1 (j, GC_malloc_atomic, jit_operand_imm (JIT_OPERAND_ABI_WORD, bytes)); |
|
emit_retval (j, t); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, t); |
|
} |
|
else |
|
{ |
|
jit_gpr_t res = T0; |
|
ptrdiff_t offset = offsetof(struct scm_thread, pointerless_freelists); |
|
offset += idx * sizeof(void*); |
|
emit_ldxi (j, res, THREAD, offset); |
|
add_slow_path_patch (j, jit_beqi (j->jit, res, 0)); |
|
jit_gpr_t new_freelist = T1; |
|
emit_ldr (j, new_freelist, res); |
|
jit_stxi (j->jit, offset, THREAD, new_freelist); |
|
emit_sp_set_scm (j, dst, res); |
|
} |
|
} |
|
static void |
|
compile_allocate_pointerless_words_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t nwords) |
|
{ |
|
size_t bytes = nwords * sizeof(SCM); |
|
size_t idx = scm_inline_gc_bytes_to_freelist_index (bytes); |
|
|
|
if (SCM_UNLIKELY (idx >= SCM_INLINE_GC_FREELIST_COUNT)) |
|
{ |
|
} |
|
else |
|
{ |
|
jit_gpr_t res = T0; |
|
emit_store_current_ip (j, res); |
|
emit_call_2 (j, scm_vm_intrinsics.allocate_pointerless_words_with_freelist, |
|
thread_operand (), |
|
jit_operand_imm (JIT_OPERAND_ABI_WORD, idx)); |
|
emit_retval (j, res); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, res); |
|
continue_after_slow_path (j, j->next_ip); |
|
} |
|
} |
|
|
|
static void |
|
compile_scm_ref (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_sz (j, T1, idx); |
|
emit_lshi (j, T1, T1, log2_sizeof_uintptr_t); |
|
emit_ldxr (j, T0, T0, T1); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_scm_ref_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_scm_set (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_sz (j, T1, idx); |
|
emit_sp_ref_scm (j, T2, val); |
|
emit_lshi (j, T1, T1, log2_sizeof_uintptr_t); |
|
jit_stxr (j->jit, T0, T1, T2); |
|
} |
|
static void |
|
compile_scm_set_slow (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_scm_ref_tag (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t tag) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_ldr (j, T0, T0); |
|
emit_subi (j, T0, T0, tag); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_scm_ref_tag_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t tag) |
|
{ |
|
} |
|
|
|
static void |
|
compile_scm_set_tag (scm_jit_state *j, uint32_t obj, uint32_t tag, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_scm (j, T1, val); |
|
emit_addi (j, T1, T1, tag); |
|
jit_str (j->jit, T0, T1); |
|
} |
|
static void |
|
compile_scm_set_tag_slow (scm_jit_state *j, uint32_t obj, uint32_t tag, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_scm_ref_immediate (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_ldxi (j, T0, T0, idx * sizeof (SCM)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_scm_ref_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_scm_set_immediate (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_scm (j, T1, val); |
|
jit_stxi (j->jit, idx * sizeof (SCM), T0, T1); |
|
} |
|
static void |
|
compile_scm_set_immediate_slow (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_word_ref (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_sz (j, T1, idx); |
|
emit_lshi (j, T1, T1, log2_sizeof_uintptr_t); |
|
emit_ldxr (j, T0, T0, T1); |
|
emit_sp_set_sz (j, dst, T0); |
|
} |
|
static void |
|
compile_word_ref_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_word_set (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_sz (j, T1, idx); |
|
emit_sp_ref_sz (j, T2, val); |
|
emit_lshi (j, T1, T1, log2_sizeof_uintptr_t); |
|
jit_stxr (j->jit, T0, T1, T2); |
|
} |
|
static void |
|
compile_word_set_slow (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_word_ref_immediate (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_ldxi (j, T0, T0, idx * sizeof (SCM)); |
|
emit_sp_set_sz (j, dst, T0); |
|
} |
|
static void |
|
compile_word_ref_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_word_set_immediate (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_sz (j, T1, val); |
|
jit_stxi (j->jit, idx * sizeof (SCM), T0, T1); |
|
} |
|
static void |
|
compile_word_set_immediate_slow (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_pointer_ref_immediate (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_ldxi (j, T0, T0, idx * sizeof (SCM)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_pointer_ref_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_pointer_set_immediate (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_scm (j, T1, val); |
|
jit_stxi (j->jit, idx * sizeof (SCM), T0, T1); |
|
} |
|
static void |
|
compile_pointer_set_immediate_slow (scm_jit_state *j, uint32_t obj, uint32_t idx, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_tail_pointer_ref_immediate (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_addi (j, T0, T0, idx * sizeof (SCM)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_tail_pointer_ref_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_mov (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
emit_mov (j, dst, src, T0); |
|
} |
|
static void |
|
compile_mov_slow (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
} |
|
|
|
static void |
|
compile_long_mov (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
emit_mov (j, dst, src, T0); |
|
} |
|
static void |
|
compile_long_mov_slow (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
} |
|
|
|
static void |
|
compile_long_fmov (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
jit_gpr_t t = T0; |
|
restore_reloadable_register_state (j, FP_IN_REGISTER); |
|
emit_fp_ref_scm (j, t, src); |
|
emit_fp_set_scm (j, dst, t); |
|
} |
|
static void |
|
compile_long_fmov_slow (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_scm_scm (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
switch ((enum scm_vm_intrinsic) idx) |
|
{ |
|
case SCM_VM_INTRINSIC_ADD: |
|
{ |
|
emit_sp_ref_scm (j, T0, a); |
|
emit_sp_ref_scm (j, T1, b); |
|
add_slow_path_patch (j, jit_bmci (j->jit, T0, scm_tc2_int)); |
|
add_slow_path_patch (j, jit_bmci (j->jit, T1, scm_tc2_int)); |
|
jit_subi (j->jit, T0, T0, scm_tc2_int); |
|
add_slow_path_patch (j, jit_boaddr (j->jit, T0, T1)); |
|
break; |
|
} |
|
case SCM_VM_INTRINSIC_SUB: |
|
{ |
|
emit_sp_ref_scm (j, T0, a); |
|
emit_sp_ref_scm (j, T1, b); |
|
add_slow_path_patch (j, jit_bmci (j->jit, T0, scm_tc2_int)); |
|
add_slow_path_patch (j, jit_bmci (j->jit, T1, scm_tc2_int)); |
|
jit_subi (j->jit, T1, T1, scm_tc2_int); |
|
add_slow_path_patch (j, jit_bosubr (j->jit, T0, T1)); |
|
break; |
|
} |
|
default: |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
jit_operand_t op_a = sp_scm_operand (j, a); |
|
jit_operand_t op_b = sp_scm_operand (j, b); |
|
emit_store_current_ip (j, T2); |
|
emit_call_2 (j, intrinsic, op_a, op_b); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
} |
|
} |
|
|
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_scm_scm_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
switch ((enum scm_vm_intrinsic) idx) |
|
{ |
|
case SCM_VM_INTRINSIC_ADD: |
|
case SCM_VM_INTRINSIC_SUB: |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
jit_operand_t op_a = sp_scm_operand (j, a); |
|
jit_operand_t op_b = sp_scm_operand (j, b); |
|
emit_store_current_ip (j, T1); |
|
emit_call_2 (j, intrinsic, op_a, op_b); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
continue_after_slow_path (j, j->next_ip); |
|
break; |
|
} |
|
default: |
|
break; |
|
} |
|
} |
|
|
|
static void |
|
compile_call_scm_from_scm_uimm (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
switch ((enum scm_vm_intrinsic) idx) |
|
{ |
|
case SCM_VM_INTRINSIC_ADD_IMMEDIATE: |
|
{ |
|
emit_sp_ref_scm (j, T0, a); |
|
scm_t_bits addend = b << 2; |
|
add_slow_path_patch (j, jit_bmci (j->jit, T0, 2)); |
|
add_slow_path_patch (j, jit_boaddi (j->jit, T0, addend)); |
|
break; |
|
} |
|
case SCM_VM_INTRINSIC_SUB_IMMEDIATE: |
|
{ |
|
emit_sp_ref_scm (j, T0, a); |
|
scm_t_bits subtrahend = b << 2; |
|
add_slow_path_patch (j, jit_bmci (j->jit, T0, 2)); |
|
add_slow_path_patch (j, jit_bosubi (j->jit, T0, subtrahend)); |
|
break; |
|
} |
|
default: |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
jit_operand_t op_a = sp_scm_operand (j, a); |
|
jit_operand_t op_b = jit_operand_imm (JIT_OPERAND_ABI_UINT8, b); |
|
|
|
emit_store_current_ip (j, T1); |
|
emit_call_2 (j, intrinsic, op_a, op_b); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
break; |
|
} |
|
} |
|
|
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_scm_uimm_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
switch ((enum scm_vm_intrinsic) idx) |
|
{ |
|
case SCM_VM_INTRINSIC_ADD_IMMEDIATE: |
|
case SCM_VM_INTRINSIC_SUB_IMMEDIATE: |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
jit_operand_t op_a = sp_scm_operand (j, a); |
|
jit_operand_t op_b = jit_operand_imm (JIT_OPERAND_ABI_UINT8, b); |
|
emit_store_current_ip (j, T1); |
|
emit_call_2 (j, intrinsic, op_a, op_b); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
continue_after_slow_path (j, j->next_ip); |
|
break; |
|
} |
|
default: |
|
break; |
|
} |
|
} |
|
|
|
static void |
|
compile_call_scm_sz_u32 (scm_jit_state *j, uint32_t a, uint32_t b, uint32_t c, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_3 (j, intrinsic, sp_scm_operand (j, a), sp_sz_operand (j, b), |
|
sp_sz_operand (j, c)); |
|
emit_reload_sp (j); |
|
} |
|
static void |
|
compile_call_scm_sz_u32_slow (scm_jit_state *j, uint32_t a, uint32_t b, uint32_t c, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_scm (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, intrinsic, sp_scm_operand (j, a)); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_scm_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_f64_from_scm (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, intrinsic, sp_scm_operand (j, a)); |
|
emit_retval_d (j, JIT_F0); |
|
emit_reload_sp (j); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_call_f64_from_scm_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_f64_from_f64 (scm_jit_state *j, uint32_t dst, uint32_t src, uint32_t idx) |
|
{ |
|
switch ((enum scm_vm_intrinsic) idx) |
|
{ |
|
case SCM_VM_INTRINSIC_FABS: |
|
{ |
|
emit_sp_ref_f64 (j, JIT_F0, src); |
|
emit_absr_d (j, JIT_F0, JIT_F0); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
break; |
|
} |
|
case SCM_VM_INTRINSIC_FSQRT: |
|
{ |
|
emit_sp_ref_f64 (j, JIT_F0, src); |
|
emit_sqrtr_d (j, JIT_F0, JIT_F0); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
break; |
|
} |
|
default: |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
emit_call_1 (j, intrinsic, sp_f64_operand (j, src)); |
|
emit_retval_d (j, JIT_F0); |
|
emit_reload_sp (j); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
break; |
|
} |
|
} |
|
} |
|
static void |
|
compile_call_f64_from_f64_slow (scm_jit_state *j, uint32_t dst, uint32_t src, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_f64_from_f64_f64 (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
emit_call_2 (j, intrinsic, sp_f64_operand (j, a), sp_f64_operand (j, b)); |
|
emit_retval_d (j, JIT_F0); |
|
emit_reload_sp (j); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_call_f64_from_f64_f64_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_u64_from_scm (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
#if INDIRECT_INT64_INTRINSICS |
|
emit_call_2 (j, intrinsic, sp_slot_operand (j, dst), sp_scm_operand (j, a)); |
|
emit_reload_sp (j); |
|
#else |
|
emit_call_1 (j, intrinsic, sp_scm_operand (j, a)); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#endif |
|
} |
|
static void |
|
compile_call_u64_from_scm_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_make_immediate (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
emit_movi (j, T0, SCM_UNPACK (a)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_make_immediate_slow (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_make_short_immediate (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
emit_movi (j, T0, SCM_UNPACK (a)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_make_short_immediate_slow (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_make_long_immediate (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
emit_movi (j, T0, SCM_UNPACK (a)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_make_long_immediate_slow (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_make_long_long_immediate (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
emit_movi (j, T0, SCM_UNPACK (a)); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_make_long_long_immediate_slow (scm_jit_state *j, uint32_t dst, SCM a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_make_non_immediate (scm_jit_state *j, uint32_t dst, const void *data) |
|
{ |
|
emit_movi (j, T0, (uintptr_t)data); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_make_non_immediate_slow (scm_jit_state *j, uint32_t dst, const void *data) |
|
{ |
|
} |
|
|
|
static void |
|
compile_static_ref (scm_jit_state *j, uint32_t dst, void *loc) |
|
{ |
|
emit_ldi (j, T0, loc); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_static_ref_slow (scm_jit_state *j, uint32_t dst, void *loc) |
|
{ |
|
} |
|
|
|
static void |
|
compile_static_set (scm_jit_state *j, uint32_t obj, void *loc) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
jit_sti (j->jit, loc, T0); |
|
} |
|
static void |
|
compile_static_set_slow (scm_jit_state *j, uint32_t obj, void *loc) |
|
{ |
|
} |
|
|
|
static void |
|
compile_static_patch (scm_jit_state *j, void *dst, const void *src) |
|
{ |
|
emit_movi (j, T0, (uintptr_t) src); |
|
jit_sti (j->jit, dst, T0); |
|
} |
|
static void |
|
compile_static_patch_slow (scm_jit_state *j, void *dst, const void *src) |
|
{ |
|
} |
|
|
|
static void |
|
compile_prompt (scm_jit_state *j, uint32_t tag, uint8_t escape_only_p, |
|
uint32_t proc_slot, const uint32_t *vcode) |
|
{ |
|
emit_store_current_ip (j, T0); |
|
|
|
emit_reload_fp (j); |
|
jit_subi (j->jit, FP, FP, proc_slot * sizeof (union scm_vm_stack_element)); |
|
jit_reloc_t mra = emit_mov_addr (j, T2); |
|
|
|
jit_operand_t args[] = |
|
{ thread_operand (), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT8, escape_only_p), |
|
sp_scm_operand (j, tag), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, FP), |
|
jit_operand_imm (JIT_OPERAND_ABI_POINTER, (uintptr_t)vcode), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, T2) }; |
|
jit_calli (j->jit, scm_vm_intrinsics.push_prompt, 6, args); |
|
clear_scratch_register_state (j); |
|
emit_reload_sp (j); |
|
emit_reload_fp (j); |
|
add_inter_instruction_patch (j, mra, vcode); |
|
} |
|
static void |
|
compile_prompt_slow (scm_jit_state *j, uint32_t tag, uint8_t escape_only_p, |
|
uint32_t proc_slot, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_load_label (scm_jit_state *j, uint32_t dst, const uint32_t *vcode) |
|
{ |
|
emit_movi (j, T0, (uintptr_t) vcode); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_movi (j, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_load_label_slow (scm_jit_state *j, uint32_t dst, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_s64_from_scm (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
compile_call_u64_from_scm (j, dst, a, idx); |
|
} |
|
static void |
|
compile_call_s64_from_scm_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_u64 (scm_jit_state *j, uint32_t dst, uint32_t src, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
#if INDIRECT_INT64_INTRINSICS |
|
emit_call_1 (j, intrinsic, sp_slot_operand (j, src)); |
|
#else |
|
emit_call_1 (j, intrinsic, sp_u64_operand (j, src)); |
|
#endif |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_u64_slow (scm_jit_state *j, uint32_t dst, uint32_t src, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_s64 (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
compile_call_scm_from_u64 (j, dst, a, b); |
|
} |
|
static void |
|
compile_call_scm_from_s64_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_tag_char (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, src); |
|
#else |
|
emit_sp_ref_u64_lower_half (j, T0, src); |
|
#endif |
|
emit_lshi (j, T0, T0, 8); |
|
emit_addi (j, T0, T0, scm_tc8_char); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_tag_char_slow (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
} |
|
|
|
static void |
|
compile_untag_char (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
emit_sp_ref_scm (j, T0, src); |
|
emit_rshi (j, T0, T0, 8); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_movi (j, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_untag_char_slow (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
} |
|
|
|
static void |
|
compile_atomic_scm_ref_immediate (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t offset) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_addi (j, T0, T0, offset * sizeof (SCM)); |
|
jit_ldr_atomic (j->jit, T0, T0); |
|
record_gpr_clobber (j, T0); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_atomic_scm_ref_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t offset) |
|
{ |
|
} |
|
|
|
static void |
|
compile_atomic_scm_set_immediate (scm_jit_state *j, uint32_t obj, uint32_t offset, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_scm (j, T1, val); |
|
emit_addi (j, T0, T0, offset * sizeof (SCM)); |
|
jit_str_atomic (j->jit, T0, T1); |
|
} |
|
static void |
|
compile_atomic_scm_set_immediate_slow (scm_jit_state *j, uint32_t obj, uint32_t offset, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_atomic_scm_swap_immediate (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t offset, uint32_t val) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_scm (j, T1, val); |
|
emit_addi (j, T0, T0, offset * sizeof (SCM)); |
|
jit_swap_atomic (j->jit, T1, T0, T1); |
|
record_gpr_clobber (j, T1); |
|
emit_sp_set_scm (j, dst, T1); |
|
} |
|
static void |
|
compile_atomic_scm_swap_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t obj, uint32_t offset, uint32_t val) |
|
{ |
|
} |
|
|
|
static void |
|
compile_atomic_scm_compare_and_swap_immediate (scm_jit_state *j, uint32_t dst, |
|
uint32_t obj, uint32_t offset, |
|
uint32_t expected, uint32_t desired) |
|
{ |
|
emit_sp_ref_scm (j, T0, obj); |
|
emit_sp_ref_scm (j, T1, expected); |
|
emit_sp_ref_scm (j, T2, desired); |
|
emit_addi (j, T0, T0, offset * sizeof (SCM)); |
|
jit_cas_atomic (j->jit, T1, T0, T1, T2); |
|
record_gpr_clobber (j, T1); |
|
emit_sp_set_scm (j, dst, T1); |
|
} |
|
static void |
|
compile_atomic_scm_compare_and_swap_immediate_slow (scm_jit_state *j, uint32_t dst, |
|
uint32_t obj, uint32_t offset, |
|
uint32_t expected, uint32_t desired) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_thread_scm_scm (scm_jit_state *j, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_3 (j, intrinsic, thread_operand (), sp_scm_operand (j, a), |
|
sp_scm_operand (j, b)); |
|
emit_reload_sp (j); |
|
} |
|
static void |
|
compile_call_thread_scm_scm_slow (scm_jit_state *j, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_thread (scm_jit_state *j, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, intrinsic, thread_operand ()); |
|
emit_reload_sp (j); |
|
} |
|
static void |
|
compile_call_thread_slow (scm_jit_state *j, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_thread_scm (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_2 (j, intrinsic, thread_operand (), sp_scm_operand (j, a)); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_thread_scm_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_thread_scm (scm_jit_state *j, uint32_t a, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_2 (j, intrinsic, thread_operand (), sp_scm_operand (j, a)); |
|
emit_reload_sp (j); |
|
} |
|
static void |
|
compile_call_thread_scm_slow (scm_jit_state *j, uint32_t a, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_scm_u64 (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
#if INDIRECT_INT64_INTRINSICS |
|
emit_call_2 (j, intrinsic, sp_scm_operand (j, a), sp_slot_operand (j, b)); |
|
#else |
|
emit_call_2 (j, intrinsic, sp_scm_operand (j, a), sp_u64_operand (j, b)); |
|
#endif |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_scm_u64_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_thread (scm_jit_state *j, uint32_t dst, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_1 (j, intrinsic, thread_operand ()); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_thread_slow (scm_jit_state *j, uint32_t dst, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_scm (scm_jit_state *j, uint32_t a, uint32_t b, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_2 (j, intrinsic, sp_scm_operand (j, a), sp_scm_operand (j, b)); |
|
emit_reload_sp (j); |
|
} |
|
static void |
|
compile_call_scm_scm_slow (scm_jit_state *j, uint32_t a, uint32_t b, |
|
uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_scm_scm (scm_jit_state *j, uint32_t a, uint32_t b, uint32_t c, |
|
uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_3 (j, intrinsic, sp_scm_operand (j, a), sp_scm_operand (j, b), |
|
sp_scm_operand (j, c)); |
|
emit_reload_sp (j); |
|
} |
|
static void |
|
compile_call_scm_scm_scm_slow (scm_jit_state *j, uint32_t a, uint32_t b, |
|
uint32_t c, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_uimm_scm (scm_jit_state *j, uint32_t a, uint8_t b, uint32_t c, |
|
uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_3 (j, intrinsic, sp_scm_operand (j, a), |
|
jit_operand_imm (JIT_OPERAND_ABI_UINT8, b), |
|
sp_scm_operand (j, c)); |
|
emit_reload_sp (j); |
|
} |
|
static void |
|
compile_call_scm_uimm_scm_slow (scm_jit_state *j, uint32_t a, uint8_t b, |
|
uint32_t c, uint32_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_fadd (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
emit_sp_ref_f64 (j, JIT_F0, a); |
|
emit_sp_ref_f64 (j, JIT_F1, b); |
|
emit_addr_d (j, JIT_F0, JIT_F0, JIT_F1); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_fadd_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_fsub (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
emit_sp_ref_f64 (j, JIT_F0, a); |
|
emit_sp_ref_f64 (j, JIT_F1, b); |
|
emit_subr_d (j, JIT_F0, JIT_F0, JIT_F1); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_fsub_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_fmul (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
emit_sp_ref_f64 (j, JIT_F0, a); |
|
emit_sp_ref_f64 (j, JIT_F1, b); |
|
emit_mulr_d (j, JIT_F0, JIT_F0, JIT_F1); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_fmul_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_fdiv (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
emit_sp_ref_f64 (j, JIT_F0, a); |
|
emit_sp_ref_f64 (j, JIT_F1, b); |
|
emit_divr_d (j, JIT_F0, JIT_F0, JIT_F1); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_fdiv_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_uadd (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_addr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_addcr (j, T0, T0, T2); |
|
emit_addxr (j, T1, T1, T3_OR_FP); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_uadd_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_usub (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_subr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_subcr (j, T0, T0, T2); |
|
emit_subxr (j, T1, T1, T3_OR_FP); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_usub_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_umul (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_mulr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_mulr (j, T1, T1, T2); |
|
emit_mulr (j, T3_OR_FP, T3_OR_FP, T0); |
|
emit_addr (j, T1, T1, T3_OR_FP); |
|
emit_qmulr_u (j, T0, T2, T0, T2); |
|
emit_addr (j, T1, T1, T2); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_umul_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_uadd_immediate (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_addi (j, T0, T0, b); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_addci (j, T0, T0, b); |
|
emit_addxi (j, T1, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_uadd_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_usub_immediate (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_subi (j, T0, T0, b); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_subci (j, T0, T0, b); |
|
emit_subxi (j, T1, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_usub_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_umul_immediate (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_muli (j, T0, T0, b); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_muli (j, T1, T1, b); |
|
|
|
emit_movi (j, T2, b); |
|
emit_qmulr_u (j, T0, T2, T0, T2); |
|
emit_addr (j, T1, T1, T2); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_umul_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_load_f64 (scm_jit_state *j, uint32_t dst, double a) |
|
{ |
|
jit_movi_d (j->jit, JIT_F0, a); |
|
record_fpr_clobber (j, JIT_F0); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_load_f64_slow (scm_jit_state *j, uint32_t dst, double a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_load_u64 (scm_jit_state *j, uint32_t dst, uint64_t a) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_movi (j, T0, a); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_movi (j, T0, a & 0xffffffff); |
|
emit_movi (j, T1, a >> 32); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_load_u64_slow (scm_jit_state *j, uint32_t dst, uint64_t a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_load_s64 (scm_jit_state *j, uint32_t dst, int64_t a) |
|
{ |
|
compile_load_u64 (j, dst, a); |
|
} |
|
static void |
|
compile_load_s64_slow (scm_jit_state *j, uint32_t dst, int64_t a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_current_thread (scm_jit_state *j, uint32_t dst) |
|
{ |
|
emit_ldxi (j, T0, THREAD, thread_offset_handle); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_current_thread_slow (scm_jit_state *j, uint32_t dst) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ulogand (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_andr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_andr (j, T0, T0, T2); |
|
emit_andr (j, T1, T1, T3_OR_FP); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ulogand_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ulogand_immediate (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_andi (j, T0, T0, b); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_andi (j, T0, T0, b); |
|
emit_andi (j, T1, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ulogand_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ulogior (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_orr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_orr (j, T0, T0, T2); |
|
emit_orr (j, T1, T1, T3_OR_FP); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ulogior_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ulogsub (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_comr (j, T1, T1); |
|
emit_andr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_comr (j, T2, T2); |
|
emit_comr (j, T3_OR_FP, T3_OR_FP); |
|
emit_andr (j, T0, T0, T2); |
|
emit_andr (j, T1, T1, T3_OR_FP); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ulogsub_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ursh (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_andi (j, T1, T1, 63); |
|
emit_rshr_u (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
|
|
jit_reloc_t zero, both, done; |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_andi (j, T2, T2, 63); |
|
zero = jit_beqi (j->jit, T2, 0); |
|
both = jit_blti (j->jit, T2, 32); |
|
|
|
|
|
emit_subi (j, T2, T2, 32); |
|
emit_rshr_u (j, T0, T1, T2); |
|
emit_movi (j, T1, 0); |
|
done = jit_jmp (j->jit); |
|
|
|
jit_patch_here (j->jit, both); |
|
|
|
emit_negr (j, T3_OR_FP, T2); |
|
emit_addi (j, T3_OR_FP, T3_OR_FP, 32); |
|
emit_lshr (j, T3_OR_FP, T1, T3_OR_FP); |
|
emit_rshr_u (j, T1, T1, T2); |
|
emit_rshr_u (j, T0, T0, T2); |
|
emit_addr (j, T0, T0, T3_OR_FP); |
|
|
|
jit_patch_here (j->jit, done); |
|
jit_patch_here (j->jit, zero); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ursh_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ulsh (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_andi (j, T1, T1, 63); |
|
emit_lshr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
|
|
jit_reloc_t zero, both, done; |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_andi (j, T2, T2, 63); |
|
zero = jit_beqi (j->jit, T2, 0); |
|
both = jit_blti (j->jit, T2, 32); |
|
|
|
|
|
emit_subi (j, T2, T2, 32); |
|
emit_lshr (j, T1, T0, T2); |
|
emit_movi (j, T0, 0); |
|
done = jit_jmp (j->jit); |
|
|
|
jit_patch_here (j->jit, both); |
|
|
|
emit_negr (j, T3_OR_FP, T2); |
|
emit_addi (j, T3_OR_FP, T3_OR_FP, 32); |
|
emit_rshr_u (j, T3_OR_FP, T0, T3_OR_FP); |
|
emit_lshr (j, T1, T1, T2); |
|
emit_lshr (j, T0, T0, T2); |
|
emit_addr (j, T1, T1, T3_OR_FP); |
|
|
|
jit_patch_here (j->jit, done); |
|
jit_patch_here (j->jit, zero); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ulsh_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ursh_immediate (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
b &= 63; |
|
|
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_rshi_u (j, T0, T0, b); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
if (b == 0) |
|
{ |
|
|
|
} |
|
else if (b < 32) |
|
{ |
|
|
|
emit_lshi (j, T2, T1, 32 - b); |
|
emit_rshi_u (j, T1, T1, b); |
|
emit_rshi_u (j, T0, T0, b); |
|
emit_addr (j, T0, T0, T2); |
|
} |
|
else if (b == 32) |
|
{ |
|
|
|
emit_movr (j, T0, T1); |
|
emit_movi (j, T1, 0); |
|
} |
|
else |
|
{ |
|
|
|
emit_rshi_u (j, T0, T1, b - 32); |
|
emit_movi (j, T1, 0); |
|
} |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ursh_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ulsh_immediate (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
b &= 63; |
|
|
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_lshi (j, T0, T0, b); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
if (b == 0) |
|
{ |
|
|
|
} |
|
else if (b < 32) |
|
{ |
|
|
|
emit_rshi_u (j, T2, T0, 32 - b); |
|
emit_lshi (j, T1, T1, b); |
|
emit_lshi (j, T0, T0, b); |
|
emit_addr (j, T1, T1, T2); |
|
} |
|
else if (b == 32) |
|
{ |
|
|
|
emit_movr (j, T1, T0); |
|
emit_movi (j, T0, 0); |
|
} |
|
else |
|
{ |
|
|
|
emit_lshi (j, T1, T0, b - 32); |
|
emit_movi (j, T0, 0); |
|
} |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ulsh_immediate_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_ulogxor (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
emit_xorr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
emit_xorr (j, T0, T0, T2); |
|
emit_xorr (j, T1, T1, T3_OR_FP); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_ulogxor_slow (scm_jit_state *j, uint32_t dst, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_handle_interrupts (scm_jit_state *j) |
|
{ |
|
jit_addi (j->jit, T0, THREAD, thread_offset_pending_asyncs); |
|
jit_ldr_atomic (j->jit, T0, T0); |
|
add_slow_path_patch (j, jit_bnei (j->jit, T0, SCM_UNPACK (SCM_EOL))); |
|
} |
|
static void |
|
compile_handle_interrupts_slow (scm_jit_state *j) |
|
{ |
|
jit_ldxi_i (j->jit, T0, THREAD, thread_offset_block_asyncs); |
|
add_inter_instruction_patch (j, |
|
jit_bnei (j->jit, T0, 0), |
|
j->next_ip); |
|
|
|
emit_store_current_ip (j, T0); |
|
jit_jmpi_with_link (j->jit, handle_interrupts_trampoline); |
|
continue_after_slow_path (j, j->ip); |
|
} |
|
|
|
static void |
|
compile_return_from_interrupt (scm_jit_state *j) |
|
{ |
|
jit_gpr_t old_fp = T0, ra = T1; |
|
jit_reloc_t interp; |
|
|
|
emit_pop_fp (j, old_fp); |
|
|
|
emit_load_mra (j, ra, old_fp); |
|
interp = jit_beqi (j->jit, ra, 0); |
|
jit_addi (j->jit, SP, old_fp, frame_overhead_slots * sizeof (union scm_vm_stack_element)); |
|
set_register_state (j, SP_IN_REGISTER); |
|
emit_store_sp (j); |
|
jit_jmpr (j->jit, ra); |
|
|
|
jit_patch_here (j->jit, interp); |
|
emit_load_vra (j, ra, old_fp); |
|
emit_store_ip (j, ra); |
|
jit_addi (j->jit, SP, old_fp, frame_overhead_slots * sizeof (union scm_vm_stack_element)); |
|
set_register_state (j, SP_IN_REGISTER); |
|
emit_store_sp (j); |
|
emit_exit (j); |
|
|
|
clear_register_state (j, SP_CACHE_GPR | SP_CACHE_FPR); |
|
} |
|
static void |
|
compile_return_from_interrupt_slow (scm_jit_state *j) |
|
{ |
|
} |
|
|
|
static enum scm_opcode |
|
fuse_conditional_branch (scm_jit_state *j, uint32_t **target) |
|
{ |
|
uint8_t next = j->next_ip[0] & 0xff; |
|
|
|
switch (next) |
|
{ |
|
case scm_op_jl: |
|
case scm_op_je: |
|
case scm_op_jnl: |
|
case scm_op_jne: |
|
case scm_op_jge: |
|
case scm_op_jnge: |
|
*target = j->next_ip + (((int32_t) j->next_ip[0]) >> 8); |
|
j->next_ip += op_lengths[next]; |
|
return next; |
|
default: |
|
ASSERT (0); |
|
} |
|
} |
|
|
|
static void |
|
compile_u64_numerically_equal (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
uint32_t *target; |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_beqr (j->jit, T0, T1); |
|
break; |
|
case scm_op_jne: |
|
k = jit_bner (j->jit, T0, T1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2; |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k1 = jit_bner (j->jit, T0, T2); |
|
k2 = jit_beqr (j->jit, T1, T3_OR_FP); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
case scm_op_jne: |
|
k1 = jit_bner (j->jit, T0, T2); |
|
k2 = jit_bner (j->jit, T1, T3_OR_FP); |
|
add_inter_instruction_patch (j, k1, target); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_u64_numerically_equal_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u64_less (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
uint32_t *target; |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
emit_sp_ref_u64 (j, T0, a); |
|
emit_sp_ref_u64 (j, T1, b); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_bltr_u (j->jit, T0, T1); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_bger_u (j->jit, T0, T1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2, k3; |
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
emit_sp_ref_u64 (j, T2, T3_OR_FP, b); |
|
k1 = jit_bltr_u (j->jit, T1, T3_OR_FP); |
|
k2 = jit_bner (j->jit, T1, T3_OR_FP); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k3 = jit_bltr_u (j->jit, T0, T2); |
|
jit_patch_here (j->jit, k2); |
|
add_inter_instruction_patch (j, k1, target); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
case scm_op_jnl: |
|
k3 = jit_bger_u (j->jit, T0, T2); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_u64_less_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s64_less (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
uint32_t *target; |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
emit_sp_ref_s64 (j, T0, a); |
|
emit_sp_ref_s64 (j, T1, b); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_bltr (j->jit, T0, T1); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_bger (j->jit, T0, T1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2, k3; |
|
emit_sp_ref_s64 (j, T0, T1, a); |
|
emit_sp_ref_s64 (j, T2, T3_OR_FP, b); |
|
k1 = jit_bltr (j->jit, T1, T3_OR_FP); |
|
k2 = jit_bner (j->jit, T1, T3_OR_FP); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k3 = jit_bltr (j->jit, T0, T2); |
|
jit_patch_here (j->jit, k2); |
|
add_inter_instruction_patch (j, k1, target); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
case scm_op_jnl: |
|
k3 = jit_bger (j->jit, T0, T2); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_s64_less_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_f64_numerically_equal (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_f64 (j, JIT_F0, a); |
|
emit_sp_ref_f64 (j, JIT_F1, b); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_beqr_d (j->jit, JIT_F0, JIT_F1); |
|
break; |
|
case scm_op_jne: |
|
k = jit_bner_d (j->jit, JIT_F0, JIT_F1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_f64_numerically_equal_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_f64_less (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_f64 (j, JIT_F0, a); |
|
emit_sp_ref_f64 (j, JIT_F1, b); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_bltr_d (j->jit, JIT_F0, JIT_F1); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_bunger_d (j->jit, JIT_F0, JIT_F1); |
|
break; |
|
case scm_op_jge: |
|
k = jit_bger_d (j->jit, JIT_F0, JIT_F1); |
|
break; |
|
case scm_op_jnge: |
|
k = jit_bunltr_d (j->jit, JIT_F0, JIT_F1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_f64_less_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_numerically_equal (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_scm (j, T0, a); |
|
emit_sp_ref_scm (j, T1, b); |
|
|
|
emit_andr (j, T2, T0, T1); |
|
add_slow_path_patch (j, jit_bmci (j->jit, T2, scm_tc2_int)); |
|
|
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_beqr (j->jit, T0, T1); |
|
break; |
|
case scm_op_jne: |
|
k = jit_bner (j->jit, T0, T1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_numerically_equal_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_store_current_ip (j, T2); |
|
emit_call_2 (j, scm_vm_intrinsics.numerically_equal_p, |
|
jit_operand_gpr (JIT_OPERAND_ABI_WORD, T0), |
|
jit_operand_gpr (JIT_OPERAND_ABI_WORD, T1)); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_bnei (j->jit, T0, 0); |
|
break; |
|
case scm_op_jne: |
|
k = jit_beqi (j->jit, T0, 0); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
continue_after_slow_path (j, j->next_ip); |
|
} |
|
|
|
static void |
|
compile_less (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_scm (j, T0, a); |
|
emit_sp_ref_scm (j, T1, b); |
|
|
|
emit_andr (j, T2, T0, T1); |
|
add_slow_path_patch (j, jit_bmci (j->jit, T2, scm_tc2_int)); |
|
|
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
case scm_op_jnge: |
|
k = jit_bltr (j->jit, T0, T1); |
|
break; |
|
case scm_op_jnl: |
|
case scm_op_jge: |
|
k = jit_bger (j->jit, T0, T1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
|
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_less_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_store_current_ip (j, T2); |
|
emit_call_2 (j, scm_vm_intrinsics.less_p, |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, T0), |
|
jit_operand_gpr (JIT_OPERAND_ABI_POINTER, T1)); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
|
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_beqi (j->jit, T0, SCM_F_COMPARE_LESS_THAN); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_bnei (j->jit, T0, SCM_F_COMPARE_LESS_THAN); |
|
break; |
|
case scm_op_jge: |
|
k = jit_beqi (j->jit, T0, SCM_F_COMPARE_NONE); |
|
break; |
|
case scm_op_jnge: |
|
k = jit_bnei (j->jit, T0, SCM_F_COMPARE_NONE); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
|
|
add_inter_instruction_patch (j, k, target); |
|
continue_after_slow_path (j, j->next_ip); |
|
} |
|
|
|
static void |
|
compile_check_arguments (scm_jit_state *j, uint32_t expected) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
jit_gpr_t t = T0; |
|
|
|
emit_reload_fp (j); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jne: |
|
k = emit_branch_if_frame_locals_count_not_eq (j, t, expected); |
|
break; |
|
case scm_op_jl: |
|
k = emit_branch_if_frame_locals_count_less_than (j, t, expected); |
|
break; |
|
case scm_op_jge: |
|
|
|
|
|
|
|
|
|
|
|
k = emit_branch_if_frame_locals_count_greater_than (j, t, expected); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_check_arguments_slow (scm_jit_state *j, uint32_t expected) |
|
{ |
|
} |
|
|
|
static void |
|
compile_check_positional_arguments (scm_jit_state *j, uint32_t nreq, uint32_t expected) |
|
{ |
|
uint32_t *target; |
|
jit_reloc_t lt, gt; |
|
jit_gpr_t walk = T0, min = T1, obj = T2; |
|
|
|
ASSERT_HAS_REGISTER_STATE (FP_IN_REGISTER | SP_IN_REGISTER); |
|
|
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jge: |
|
|
|
|
|
|
|
|
|
|
|
|
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
|
|
emit_subtract_stack_slots (j, min, FP, expected); |
|
emit_subtract_stack_slots (j, walk, FP, nreq); |
|
|
|
void *head = jit_address (j->jit); |
|
|
|
gt = jit_bltr (j->jit, walk, min); |
|
emit_subtract_stack_slots (j, walk, walk, 1); |
|
lt = jit_bltr (j->jit, walk, SP); |
|
emit_ldr (j, obj, walk); |
|
jit_patch_there |
|
(j->jit, |
|
emit_branch_if_immediate (j, obj), |
|
head); |
|
jit_patch_there |
|
(j->jit, |
|
emit_branch_if_heap_object_not_tc7 (j, obj, obj, scm_tc7_keyword), |
|
head); |
|
jit_patch_here (j->jit, lt); |
|
add_inter_instruction_patch (j, gt, target); |
|
} |
|
static void |
|
compile_check_positional_arguments_slow (scm_jit_state *j, uint32_t nreq, uint32_t expected) |
|
{ |
|
} |
|
|
|
static void |
|
compile_immediate_tag_equals (scm_jit_state *j, uint32_t a, uint32_t mask, |
|
uint32_t expected) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_scm (j, T0, a); |
|
emit_andi (j, T0, T0, mask); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_beqi (j->jit, T0, expected); |
|
break; |
|
case scm_op_jne: |
|
k = jit_bnei (j->jit, T0, expected); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_immediate_tag_equals_slow (scm_jit_state *j, uint32_t a, uint32_t mask, |
|
uint32_t expected) |
|
{ |
|
} |
|
|
|
static void |
|
compile_heap_tag_equals (scm_jit_state *j, uint32_t obj, |
|
uint32_t mask, uint32_t expected) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_scm (j, T0, obj); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = emit_branch_if_heap_object_has_tc (j, T0, T0, mask, expected); |
|
break; |
|
case scm_op_jne: |
|
k = emit_branch_if_heap_object_not_tc (j, T0, T0, mask, expected); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_heap_tag_equals_slow (scm_jit_state *j, uint32_t obj, |
|
uint32_t mask, uint32_t expected) |
|
{ |
|
} |
|
|
|
static void |
|
compile_eq (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_scm (j, T0, a); |
|
emit_sp_ref_scm (j, T1, b); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_beqr (j->jit, T0, T1); |
|
break; |
|
case scm_op_jne: |
|
k = jit_bner (j->jit, T0, T1); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_eq_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_eq_immediate (scm_jit_state *j, uint32_t a, SCM b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_scm (j, T0, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_beqi (j->jit, T0, SCM_UNPACK (b)); |
|
break; |
|
case scm_op_jne: |
|
k = jit_bnei (j->jit, T0, SCM_UNPACK (b)); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_eq_immediate_slow (scm_jit_state *j, uint32_t a, SCM b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_j (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
jit_reloc_t jmp; |
|
jmp = jit_jmp (j->jit); |
|
add_inter_instruction_patch (j, jmp, vcode); |
|
} |
|
static void |
|
compile_j_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_jl (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_jl_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_je (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_je_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_jnl (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_jnl_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_jne (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_jne_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_jge (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_jge_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_jnge (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
UNREACHABLE (); |
|
} |
|
static void |
|
compile_jnge_slow (scm_jit_state *j, const uint32_t *vcode) |
|
{ |
|
} |
|
|
|
static void |
|
compile_jtable (scm_jit_state *j, uint32_t idx, uint32_t len, |
|
const uint32_t *offsets) |
|
{ |
|
ASSERT (len > 0); |
|
|
|
int32_t default_offset = offsets[len - 1]; |
|
default_offset >>= 8; |
|
uint32_t *default_target = j->ip + default_offset; |
|
|
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T0, idx); |
|
#else |
|
emit_sp_ref_u64 (j, T0, T1, idx); |
|
jit_reloc_t high_word_nonzero = jit_bnei (j->jit, T1, 0); |
|
add_inter_instruction_patch (j, high_word_nonzero, default_target); |
|
#endif |
|
|
|
jit_reloc_t out_of_range = jit_bgei_u (j->jit, T0, len - 1); |
|
add_inter_instruction_patch (j, out_of_range, default_target); |
|
|
|
|
|
|
|
emit_lshi (j, T0, T0, log2_sizeof_uintptr_t); |
|
jit_reloc_t table = emit_mov_addr (j, T1); |
|
jit_ldxr (j->jit, T0, T1, T0); |
|
jit_jmpr (j->jit, T0); |
|
|
|
|
|
jit_begin_data (j->jit, sizeof(intptr_t) * len); |
|
jit_align (j->jit, sizeof(intptr_t)); |
|
jit_patch_here (j->jit, table); |
|
for (size_t i = 0; i + 1 < len; i++) { |
|
int32_t offset = offsets[i]; |
|
offset >>= 8; |
|
uint32_t *target = j->ip + offset; |
|
jit_reloc_t addr = jit_emit_addr (j->jit); |
|
add_inter_instruction_patch (j, addr, target); |
|
} |
|
jit_end_data (j->jit); |
|
} |
|
static void |
|
compile_jtable_slow (scm_jit_state *j, uint32_t idx, uint32_t len, |
|
const uint32_t *offsets) |
|
{ |
|
} |
|
|
|
static void |
|
compile_heap_numbers_equal (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_store_current_ip (j, T0); |
|
emit_call_2 (j, scm_vm_intrinsics.heap_numbers_equal_p, sp_scm_operand (j, a), |
|
sp_scm_operand (j, b)); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_bnei (j->jit, T0, 0); |
|
break; |
|
case scm_op_jne: |
|
k = jit_beqi (j->jit, T0, 0); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
} |
|
static void |
|
compile_heap_numbers_equal_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_untag_fixnum (scm_jit_state *j, uint32_t dst, uint32_t a) |
|
{ |
|
emit_sp_ref_scm (j, T0, a); |
|
emit_rshi (j, T0, T0, 2); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_s64 (j, dst, T0); |
|
#else |
|
|
|
emit_rshi (j, T1, T0, 31); |
|
emit_sp_set_s64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_untag_fixnum_slow (scm_jit_state *j, uint32_t dst, uint32_t a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_tag_fixnum (scm_jit_state *j, uint32_t dst, uint32_t a) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_s64 (j, T0, a); |
|
#else |
|
emit_sp_ref_s32 (j, T0, a); |
|
#endif |
|
emit_lshi (j, T0, T0, 2); |
|
emit_addi (j, T0, T0, scm_tc2_int); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_tag_fixnum_slow (scm_jit_state *j, uint32_t dst, uint32_t a) |
|
{ |
|
} |
|
|
|
static void |
|
compile_srsh (scm_jit_state *j, uint8_t dst, uint8_t a, uint8_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_s64 (j, T0, a); |
|
emit_sp_ref_s64 (j, T1, b); |
|
emit_andi (j, T1, T1, 63); |
|
emit_rshr (j, T0, T0, T1); |
|
emit_sp_set_s64 (j, dst, T0); |
|
#else |
|
|
|
jit_reloc_t zero, both, done; |
|
|
|
emit_sp_ref_s64 (j, T0, T1, a); |
|
emit_sp_ref_s64 (j, T2, T3_OR_FP, b); |
|
emit_andi (j, T2, T2, 63); |
|
zero = jit_beqi (j->jit, T2, 0); |
|
both = jit_blti (j->jit, T2, 32); |
|
|
|
|
|
emit_subi (j, T2, T2, 32); |
|
emit_rshr (j, T0, T1, T2); |
|
emit_rshi (j, T1, T1, 31); |
|
done = jit_jmp (j->jit); |
|
|
|
jit_patch_here (j->jit, both); |
|
|
|
emit_negr (j, T3_OR_FP, T2); |
|
emit_addi (j, T3_OR_FP, T3_OR_FP, 32); |
|
emit_lshr (j, T3_OR_FP, T1, T3_OR_FP); |
|
emit_rshr (j, T1, T1, T2); |
|
emit_rshr_u (j, T0, T0, T2); |
|
emit_addr (j, T0, T0, T3_OR_FP); |
|
|
|
jit_patch_here (j->jit, done); |
|
jit_patch_here (j->jit, zero); |
|
emit_sp_set_s64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_srsh_slow (scm_jit_state *j, uint8_t dst, uint8_t a, uint8_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_srsh_immediate (scm_jit_state *j, uint8_t dst, uint8_t a, uint8_t b) |
|
{ |
|
b &= 63; |
|
|
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_s64 (j, T0, a); |
|
emit_rshi (j, T0, T0, b); |
|
emit_sp_set_s64 (j, dst, T0); |
|
#else |
|
|
|
emit_sp_ref_s64 (j, T0, T1, a); |
|
if (b == 0) |
|
{ |
|
|
|
} |
|
else if (b < 32) |
|
{ |
|
|
|
emit_lshi (j, T2, T1, 32 - b); |
|
emit_rshi (j, T1, T1, b); |
|
emit_rshi_u (j, T0, T0, b); |
|
emit_addr (j, T0, T0, T2); |
|
} |
|
else if (b == 32) |
|
{ |
|
|
|
emit_movr (j, T0, T1); |
|
emit_rshi (j, T1, T1, 31); |
|
} |
|
else |
|
{ |
|
|
|
emit_rshi (j, T0, T1, b - 32); |
|
emit_rshi (j, T1, T1, 31); |
|
} |
|
emit_sp_set_s64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_srsh_immediate_slow (scm_jit_state *j, uint8_t dst, uint8_t a, uint8_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s64_imm_numerically_equal (scm_jit_state *j, uint32_t a, int16_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_s64 (j, T0, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k = jit_beqi (j->jit, T0, b); |
|
break; |
|
case scm_op_jne: |
|
k = jit_bnei (j->jit, T0, b); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_s64 (j, T0, T1, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_je: |
|
k1 = jit_bnei (j->jit, T0, b); |
|
k2 = jit_beqi (j->jit, T1, b < 0 ? -1 : 0); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
case scm_op_jne: |
|
k1 = jit_bnei (j->jit, T0, b); |
|
k2 = jit_bnei (j->jit, T1, b < 0 ? -1 : 0); |
|
add_inter_instruction_patch (j, k1, target); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_s64_imm_numerically_equal_slow (scm_jit_state *j, uint32_t a, int16_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u64_imm_less (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_u64 (j, T0, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_blti_u (j->jit, T0, b); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_bgei_u (j->jit, T0, b); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k1 = jit_bnei (j->jit, T1, 0); |
|
k2 = jit_blti_u (j->jit, T0, b); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
case scm_op_jnl: |
|
k1 = jit_bnei (j->jit, T1, 0); |
|
k2 = jit_bgei_u (j->jit, T0, b); |
|
add_inter_instruction_patch (j, k1, target); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_u64_imm_less_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_imm_u64_less (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_u64 (j, T0, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_bgti_u (j->jit, T0, b); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_blei_u (j->jit, T0, b); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_u64 (j, T0, T1, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k1 = jit_bnei (j->jit, T1, 0); |
|
k2 = jit_bgti_u (j->jit, T0, b); |
|
add_inter_instruction_patch (j, k1, target); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
case scm_op_jnl: |
|
k1 = jit_bnei (j->jit, T1, 0); |
|
k2 = jit_blei_u (j->jit, T0, b); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_imm_u64_less_slow (scm_jit_state *j, uint32_t a, uint32_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s64_imm_less (scm_jit_state *j, uint32_t a, int16_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_s64 (j, T0, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_blti (j->jit, T0, b); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_bgei (j->jit, T0, b); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2, k3; |
|
int32_t sign = b < 0 ? -1 : 0; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_s64 (j, T0, T1, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k1 = jit_blti (j->jit, T1, sign); |
|
k2 = jit_bnei (j->jit, T1, sign); |
|
k3 = jit_blti (j->jit, T0, b); |
|
add_inter_instruction_patch (j, k1, target); |
|
jit_patch_here (j->jit, k2); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
case scm_op_jnl: |
|
k1 = jit_blti (j->jit, T1, sign); |
|
k2 = jit_bnei (j->jit, T1, sign); |
|
k3 = jit_bgei (j->jit, T0, b); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_s64_imm_less_slow (scm_jit_state *j, uint32_t a, int16_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_imm_s64_less (scm_jit_state *j, uint32_t a, int16_t b) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_reloc_t k; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_s64 (j, T0, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k = jit_bgti (j->jit, T0, b); |
|
break; |
|
case scm_op_jnl: |
|
k = jit_blei (j->jit, T0, b); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
add_inter_instruction_patch (j, k, target); |
|
#else |
|
jit_reloc_t k1, k2, k3; |
|
int32_t sign = b < 0 ? -1 : 0; |
|
uint32_t *target; |
|
|
|
emit_sp_ref_s64 (j, T0, T1, a); |
|
switch (fuse_conditional_branch (j, &target)) |
|
{ |
|
case scm_op_jl: |
|
k1 = jit_blti (j->jit, T1, sign); |
|
k2 = jit_bnei (j->jit, T1, sign); |
|
k3 = jit_bgti (j->jit, T0, b); |
|
jit_patch_here (j->jit, k1); |
|
add_inter_instruction_patch (j, k2, target); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
case scm_op_jnl: |
|
k1 = jit_blti (j->jit, T1, sign); |
|
k2 = jit_bnei (j->jit, T1, sign); |
|
k3 = jit_blei (j->jit, T0, b); |
|
add_inter_instruction_patch (j, k1, target); |
|
jit_patch_here (j->jit, k2); |
|
add_inter_instruction_patch (j, k3, target); |
|
break; |
|
default: |
|
UNREACHABLE (); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_imm_s64_less_slow (scm_jit_state *j, uint32_t a, int16_t b) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u8_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
jit_ldxr_uc (j->jit, T0, T0, T1); |
|
record_gpr_clobber (j, T0); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_movi (j, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_u8_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u16_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
jit_ldxr_us (j->jit, T0, T0, T1); |
|
record_gpr_clobber (j, T0); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_movi (j, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_u16_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u32_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
jit_ldxr_ui (j->jit, T0, T0, T1); |
|
record_gpr_clobber (j, T0); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_ldxr (j, T0, T0, T1); |
|
emit_movi (j, T1, 0); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_u32_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u64_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_ldxr (j, T0, T0, T1); |
|
emit_sp_set_u64 (j, dst, T0); |
|
#else |
|
emit_addr (j, T0, T0, T1); |
|
if (JIT_BIGENDIAN) |
|
{ |
|
emit_ldr (j, T1, T0); |
|
emit_ldxi (j, T0, T0, 4); |
|
} |
|
else |
|
{ |
|
emit_ldxi (j, T1, T0, 4); |
|
emit_ldr (j, T0, T0); |
|
} |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_u64_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u8_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T2, v); |
|
#else |
|
emit_sp_ref_u64_lower_half (j, T2, v); |
|
#endif |
|
jit_stxr_c (j->jit, T0, T1, T2); |
|
} |
|
static void |
|
compile_u8_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u16_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T2, v); |
|
#else |
|
emit_sp_ref_u64_lower_half (j, T2, v); |
|
#endif |
|
jit_stxr_s (j->jit, T0, T1, T2); |
|
} |
|
static void |
|
compile_u16_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u32_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T2, v); |
|
jit_stxr_i (j->jit, T0, T1, T2); |
|
#else |
|
emit_sp_ref_u64_lower_half (j, T2, v); |
|
jit_stxr (j->jit, T0, T1, T2); |
|
#endif |
|
} |
|
static void |
|
compile_u32_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_u64_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_u64 (j, T2, v); |
|
jit_stxr (j->jit, T0, T1, T2); |
|
#else |
|
jit_addr (j->jit, T0, T0, T1); |
|
emit_sp_ref_u64 (j, T1, T2, v); |
|
if (JIT_BIGENDIAN) |
|
{ |
|
jit_str (j->jit, T0, T2); |
|
jit_stxi (j->jit, 4, T0, T1); |
|
} |
|
else |
|
{ |
|
jit_str (j->jit, T0, T1); |
|
jit_stxi (j->jit, 4, T0, T2); |
|
} |
|
#endif |
|
} |
|
static void |
|
compile_u64_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s8_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
jit_ldxr_c (j->jit, T0, T0, T1); |
|
record_gpr_clobber (j, T0); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_s64 (j, dst, T0); |
|
#else |
|
emit_rshi (j, T1, T0, 7); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_s8_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s16_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
jit_ldxr_s (j->jit, T0, T0, T1); |
|
record_gpr_clobber (j, T0); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_s64 (j, dst, T0); |
|
#else |
|
emit_rshi (j, T1, T0, 15); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_s16_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s32_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
jit_ldxr_i (j->jit, T0, T0, T1); |
|
record_gpr_clobber (j, T0); |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_set_s64 (j, dst, T0); |
|
#else |
|
emit_rshi (j, T1, T0, 31); |
|
emit_sp_set_u64 (j, dst, T0, T1); |
|
#endif |
|
} |
|
static void |
|
compile_s32_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s64_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
compile_u64_ref (j, dst, ptr, idx); |
|
} |
|
static void |
|
compile_s64_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s8_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
compile_u8_set (j, ptr, idx, v); |
|
} |
|
static void |
|
compile_s8_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s16_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
compile_u16_set (j, ptr, idx, v); |
|
} |
|
static void |
|
compile_s16_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s32_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
compile_u32_set (j, ptr, idx, v); |
|
} |
|
static void |
|
compile_s32_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s64_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
compile_u64_set (j, ptr, idx, v); |
|
} |
|
static void |
|
compile_s64_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_f32_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
jit_ldxr_f (j->jit, JIT_F0, T0, T1); |
|
record_fpr_clobber (j, JIT_F0); |
|
jit_extr_f_d (j->jit, JIT_F0, JIT_F0); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_f32_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_f64_ref (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
jit_ldxr_d (j->jit, JIT_F0, T0, T1); |
|
record_fpr_clobber (j, JIT_F0); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_f64_ref_slow (scm_jit_state *j, uint8_t dst, uint8_t ptr, uint8_t idx) |
|
{ |
|
} |
|
|
|
static void |
|
compile_f32_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
emit_sp_ref_f64 (j, JIT_F0, v); |
|
jit_extr_d_f (j->jit, JIT_F0, JIT_F0); |
|
record_fpr_clobber (j, JIT_F0); |
|
jit_stxr_f (j->jit, T0, T1, JIT_F0); |
|
} |
|
static void |
|
compile_f32_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_f64_set (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
emit_sp_ref_ptr (j, T0, ptr); |
|
emit_sp_ref_sz (j, T1, idx); |
|
emit_sp_ref_f64 (j, JIT_F0, v); |
|
jit_stxr_d (j->jit, T0, T1, JIT_F0); |
|
} |
|
static void |
|
compile_f64_set_slow (scm_jit_state *j, uint8_t ptr, uint8_t idx, uint8_t v) |
|
{ |
|
} |
|
|
|
static void |
|
compile_s64_to_f64 (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
#if SIZEOF_UINTPTR_T >= 8 |
|
emit_sp_ref_s64 (j, T0, src); |
|
jit_extr_d (j->jit, JIT_F0, T0); |
|
#else |
|
emit_call_1 (j, scm_vm_intrinsics.s64_to_f64, sp_slot_operand (j, src)); |
|
jit_retval_d (j->jit, JIT_F0); |
|
emit_reload_sp (j); |
|
#endif |
|
record_fpr_clobber (j, JIT_F0); |
|
emit_sp_set_f64 (j, dst, JIT_F0); |
|
} |
|
static void |
|
compile_s64_to_f64_slow (scm_jit_state *j, uint32_t dst, uint32_t src) |
|
{ |
|
} |
|
|
|
static void |
|
compile_call_scm_from_scmn_scmn (scm_jit_state *j, uint32_t dst, |
|
void *a, void *b, uint32_t idx) |
|
{ |
|
void *intrinsic = ((void **) &scm_vm_intrinsics)[idx]; |
|
jit_operand_t op_a = jit_operand_imm (JIT_OPERAND_ABI_POINTER, (uintptr_t)a); |
|
jit_operand_t op_b = jit_operand_imm (JIT_OPERAND_ABI_POINTER, (uintptr_t)b); |
|
|
|
emit_store_current_ip (j, T2); |
|
emit_call_2 (j, intrinsic, op_a, op_b); |
|
emit_retval (j, T0); |
|
emit_reload_sp (j); |
|
emit_sp_set_scm (j, dst, T0); |
|
} |
|
static void |
|
compile_call_scm_from_scmn_scmn_slow (scm_jit_state *j, uint32_t dst, |
|
void *a, void *b, uint32_t idx) |
|
{ |
|
} |
|
|
|
|
|
#define UNPACK_8_8_8(op,a,b,c) \ |
|
do \ |
|
{ \ |
|
a = (op >> 8) & 0xff; \ |
|
b = (op >> 16) & 0xff; \ |
|
c = op >> 24; \ |
|
} \ |
|
while (0) |
|
|
|
#define UNPACK_8_16(op,a,b) \ |
|
do \ |
|
{ \ |
|
a = (op >> 8) & 0xff; \ |
|
b = op >> 16; \ |
|
} \ |
|
while (0) |
|
|
|
#define UNPACK_12_12(op,a,b) \ |
|
do \ |
|
{ \ |
|
a = (op >> 8) & 0xfff; \ |
|
b = op >> 20; \ |
|
} \ |
|
while (0) |
|
|
|
#define UNPACK_24(op,a) \ |
|
do \ |
|
{ \ |
|
a = op >> 8; \ |
|
} \ |
|
while (0) |
|
|
|
#define UNPACK_8_24(op,a,b) \ |
|
do \ |
|
{ \ |
|
a = op & 0xff; \ |
|
b = op >> 8; \ |
|
} \ |
|
while (0) |
|
|
|
#define UNPACK_16_16(op,a,b) \ |
|
do \ |
|
{ \ |
|
a = op & 0xffff; \ |
|
b = op >> 16; \ |
|
} \ |
|
while (0) |
|
|
|
#define COMPILE_OP1(t0) \ |
|
COMPILE_##t0 |
|
#define COMPILE_OP2(t0, t1) \ |
|
COMPILE_##t0##__##t1 |
|
#define COMPILE_OP3(t0, t1, t2) \ |
|
COMPILE_##t0##__##t1##__##t2 |
|
#define COMPILE_OP4(t0, t1, t2, t3) \ |
|
COMPILE_##t0##__##t1##__##t2##__##t3 |
|
#define COMPILE_OP5(t0, t1, t2, t3, t4) \ |
|
COMPILE_##t0##__##t1##__##t2##__##t3##__##t4 |
|
|
|
#define COMPILE_DOP1(t0) COMPILE_OP1(t0) |
|
#define COMPILE_DOP2(t0, t1) COMPILE_OP2(t0, t1) |
|
#define COMPILE_DOP3(t0, t1, t2) COMPILE_OP3(t0, t1, t2) |
|
#define COMPILE_DOP4(t0, t1, t2, t3) COMPILE_OP4(t0, t1, t2, t3) |
|
#define COMPILE_DOP5(t0, t1, t2, t3, t4) COMPILE_OP5(t0, t1, t2, t3, t4) |
|
|
|
#define COMPILE_WIDE_OP1(t0) \ |
|
COMPILE_WIDE_##t0 |
|
#define COMPILE_WIDE_OP2(t0, t1) \ |
|
COMPILE_WIDE_##t0##__##t1 |
|
#define COMPILE_WIDE_OP3(t0, t1, t2) \ |
|
COMPILE_WIDE_##t0##__##t1##__##t2 |
|
#define COMPILE_WIDE_OP4(t0, t1, t2, t3) \ |
|
COMPILE_WIDE_##t0##__##t1##__##t2##__##t3 |
|
#define COMPILE_WIDE_OP5(t0, t1, t2, t3, t4) \ |
|
COMPILE_WIDE_##t0##__##t1##__##t2##__##t3##__##t4 |
|
|
|
#define COMPILE_WIDE_DOP1(t0) COMPILE_WIDE_OP1(t0) |
|
#define COMPILE_WIDE_DOP2(t0, t1) COMPILE_WIDE_OP2(t0, t1) |
|
#define COMPILE_WIDE_DOP3(t0, t1, t2) COMPILE_WIDE_OP3(t0, t1, t2) |
|
#define COMPILE_WIDE_DOP4(t0, t1, t2, t3) COMPILE_WIDE_OP4(t0, t1, t2, t3) |
|
#define COMPILE_WIDE_DOP5(t0, t1, t2, t3, t4) COMPILE_WIDE_OP5(t0, t1, t2, t3, t4) |
|
|
|
#define COMPILE_NOP(j, comp) \ |
|
{ \ |
|
bad_instruction (j); \ |
|
} |
|
#define COMPILE_WIDE_NOP(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X32(j, comp) \ |
|
{ \ |
|
comp (j); \ |
|
} |
|
#define COMPILE_WIDE_X32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_C24(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
comp (j, a); \ |
|
} |
|
#define COMPILE_WIDE_X8_C24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_F24(j, comp) \ |
|
COMPILE_X8_C24 (j, comp) |
|
#define COMPILE_WIDE_X8_F24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24(j, comp) \ |
|
COMPILE_X8_C24 (j, comp) |
|
#define COMPILE_WIDE_X8_S24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_L24(j, comp) \ |
|
{ \ |
|
int32_t a = j->ip[0]; \ |
|
a >>= 8; \ |
|
comp (j, j->ip + a); \ |
|
} |
|
#define COMPILE_WIDE_X8_L24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_C12_C12(j, comp) \ |
|
{ \ |
|
uint16_t a, b; \ |
|
UNPACK_12_12 (j->ip[0], a, b); \ |
|
comp (j, a, b); \ |
|
} |
|
#define COMPILE_WIDE_X8_C12_C12(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S12_C12(j, comp) \ |
|
COMPILE_X8_C12_C12 (j, comp) |
|
#define COMPILE_WIDE_X8_S12_C12(j, comp) \ |
|
{ \ |
|
SCM_UNUSED uint16_t a; \ |
|
uint16_t b; \ |
|
UNPACK_12_12 (j->ip[0], a, b); \ |
|
comp (j, wide_operands[0], b); \ |
|
} |
|
|
|
#define COMPILE_X8_S12_S12(j, comp) \ |
|
COMPILE_X8_C12_C12 (j, comp) |
|
#define COMPILE_WIDE_X8_S12_S12(j, comp) \ |
|
{ \ |
|
comp (j, wide_operands[0], wide_operands[1]); \ |
|
} |
|
|
|
#define COMPILE_X8_F12_F12(j, comp) \ |
|
COMPILE_X8_C12_C12 (j, comp) |
|
#define COMPILE_WIDE_X8_F12_F12(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S12_Z12(j, comp) \ |
|
{ \ |
|
uint16_t a = (j->ip[0] >> 8) & 0xfff; \ |
|
int16_t b = ((int32_t) j->ip[0]) >> 20; \ |
|
comp (j, a, b); \ |
|
} |
|
#define COMPILE_WIDE_X8_S12_Z12(j, comp) \ |
|
{ \ |
|
int16_t b = ((int32_t) j->ip[0]) >> 20; \ |
|
comp (j, wide_operands[0], b); \ |
|
} |
|
|
|
#define COMPILE_X8_S8_C8_S8(j, comp) \ |
|
{ \ |
|
uint8_t a, b, c; \ |
|
UNPACK_8_8_8 (j->ip[0], a, b, c); \ |
|
comp (j, a, b, c); \ |
|
} |
|
#define COMPILE_WIDE_X8_S8_C8_S8(j, comp) \ |
|
{ \ |
|
SCM_UNUSED uint8_t a, c; \ |
|
uint8_t b; \ |
|
UNPACK_8_8_8 (j->ip[0], a, b, c); \ |
|
comp (j, wide_operands[0], b, wide_operands[1]); \ |
|
} |
|
|
|
#define COMPILE_X8_S8_S8_C8(j, comp) \ |
|
COMPILE_X8_S8_C8_S8 (j, comp) |
|
#define COMPILE_WIDE_X8_S8_S8_C8(j, comp) \ |
|
{ \ |
|
SCM_UNUSED uint8_t a, b; \ |
|
uint8_t c; \ |
|
UNPACK_8_8_8 (j->ip[0], a, b, c); \ |
|
comp (j, wide_operands[0], wide_operands[1], c); \ |
|
} |
|
|
|
#define COMPILE_X8_S8_S8_S8(j, comp) \ |
|
COMPILE_X8_S8_C8_S8 (j, comp) |
|
#define COMPILE_WIDE_X8_S8_S8_S8(j, comp) \ |
|
{ \ |
|
comp (j, wide_operands[0], wide_operands[1], wide_operands[2]); \ |
|
} |
|
|
|
#define COMPILE_X8_S8_I16(j, comp) \ |
|
{ \ |
|
uint8_t a; \ |
|
scm_t_bits b; \ |
|
UNPACK_8_16 (j->ip[0], a, b); \ |
|
comp (j, a, SCM_PACK (b)); \ |
|
} |
|
#define COMPILE_WIDE_X8_S8_I16(j, comp) \ |
|
{ \ |
|
SCM_UNUSED uint8_t a; \ |
|
scm_t_bits b; \ |
|
UNPACK_8_16 (j->ip[0], a, b); \ |
|
comp (j, wide_operands[0], SCM_PACK(b)); \ |
|
} |
|
|
|
#define COMPILE_X8_S8_ZI16(j, comp) \ |
|
{ \ |
|
uint8_t a; \ |
|
int16_t b; \ |
|
UNPACK_8_16 (j->ip[0], a, b); \ |
|
comp (j, a, SCM_PACK ((scm_t_signed_bits) b)); \ |
|
} |
|
#define COMPILE_WIDE_X8_S8_ZI16(j, comp) \ |
|
{ \ |
|
SCM_UNUSED uint8_t a; \ |
|
int16_t b; \ |
|
UNPACK_8_16 (j->ip[0], a, b); \ |
|
comp (j, wide_operands[0], SCM_PACK ((scm_t_signed_bits) b)); \ |
|
} |
|
|
|
#define COMPILE_X32__C32(j, comp) \ |
|
{ \ |
|
comp (j, j->ip[1]); \ |
|
} |
|
#define COMPILE_WIDE_X32__C32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X32__L32(j, comp) \ |
|
{ \ |
|
int32_t a = j->ip[1]; \ |
|
comp (j, j->ip + a); \ |
|
} |
|
#define COMPILE_WIDE_X32__L32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X32__N32(j, comp) \ |
|
COMPILE_X32__L32 (j, comp) |
|
#define COMPILE_WIDE_X32__N32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_C24__L32(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
int32_t b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = j->ip[1]; \ |
|
comp (j, a, j->ip + b); \ |
|
} |
|
#define COMPILE_WIDE_X8_C24__L32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__L32(j, comp) \ |
|
COMPILE_X8_C24__L32 (j, comp) |
|
#define COMPILE_WIDE_X8_S24__L32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__LO32(j, comp) \ |
|
COMPILE_X8_C24__L32 (j, comp) |
|
#define COMPILE_WIDE_X8_S24__LO32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__N32(j, comp) \ |
|
COMPILE_X8_C24__L32 (j, comp) |
|
#define COMPILE_WIDE_X8_S24__N32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__R32(j, comp) \ |
|
COMPILE_X8_C24__L32 (j, comp) |
|
#define COMPILE_WIDE_X8_S24__R32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_C24__X8_C24(j, comp) \ |
|
{ \ |
|
uint32_t a, b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
UNPACK_24 (j->ip[1], b); \ |
|
comp (j, a, b); \ |
|
} |
|
#define COMPILE_WIDE_X8_C24__X8_C24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_F24__X8_C24(j, comp) \ |
|
COMPILE_X8_C24__X8_C24(j, comp) |
|
#define COMPILE_WIDE_X8_F24__X8_C24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_F24__X8_F24(j, comp) \ |
|
COMPILE_X8_C24__X8_C24(j, comp) |
|
#define COMPILE_WIDE_X8_F24__X8_F24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__X8_S24(j, comp) \ |
|
COMPILE_X8_C24__X8_C24(j, comp) |
|
#define COMPILE_WIDE_X8_S24__X8_S24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_F12_F12__X8_C24(j, comp) \ |
|
{ \ |
|
uint16_t a, b; \ |
|
uint32_t c; \ |
|
UNPACK_12_12 (j->ip[0], a, b); \ |
|
UNPACK_24 (j->ip[1], c); \ |
|
comp (j, a, b, c); \ |
|
} |
|
#define COMPILE_WIDE_X8_F12_F12__X8_C24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_F24__B1_X7_C24(j, comp) \ |
|
{ \ |
|
uint32_t a, c; \ |
|
uint8_t b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = j->ip[1] & 0x1; \ |
|
UNPACK_24 (j->ip[1], c); \ |
|
comp (j, a, b, c); \ |
|
} |
|
#define COMPILE_WIDE_X8_F24__B1_X7_C24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S12_S12__C32(j, comp) \ |
|
{ \ |
|
uint16_t a, b; \ |
|
uint32_t c; \ |
|
UNPACK_12_12 (j->ip[0], a, b); \ |
|
c = j->ip[1]; \ |
|
comp (j, a, b, c); \ |
|
} |
|
#define COMPILE_WIDE_X8_S12_S12__C32(j, comp) \ |
|
{ \ |
|
uint32_t c = j->ip[1]; \ |
|
comp (j, wide_operands[0], wide_operands[1], c); \ |
|
} |
|
|
|
#define COMPILE_X8_S24__C16_C16(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
uint16_t b, c; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
UNPACK_16_16 (j->ip[1], b, c); \ |
|
comp (j, a, b, c); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__C16_C16(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__C32(j, comp) \ |
|
{ \ |
|
uint32_t a, b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = j->ip[1]; \ |
|
comp (j, a, b); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__C32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__I32(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
scm_t_bits b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = j->ip[1]; \ |
|
comp (j, a, SCM_PACK (b)); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__I32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S8_S8_C8__C32(j, comp) \ |
|
{ \ |
|
uint8_t a, b, c; \ |
|
UNPACK_8_8_8 (j->ip[0], a, b, c); \ |
|
uint32_t d; \ |
|
d = j->ip[1]; \ |
|
comp (j, a, b, c, d); \ |
|
} |
|
#define COMPILE_WIDE_X8_S8_S8_C8__C32(j, comp) \ |
|
{ \ |
|
SCM_UNUSED uint16_t a, b; \ |
|
uint8_t c; \ |
|
UNPACK_8_8_8 (j->ip[0], a, b, c); \ |
|
uint32_t d = j->ip[1]; \ |
|
comp (j, wide_operands[0], wide_operands[1], c, d); \ |
|
} |
|
|
|
#define COMPILE_X8_S8_S8_S8__C32(j, comp) \ |
|
COMPILE_X8_S8_S8_C8__C32(j, comp) |
|
#define COMPILE_WIDE_X8_S8_S8_S8__C32(j, comp) \ |
|
{ \ |
|
uint32_t d = j->ip[1]; \ |
|
comp (j, wide_operands[0], wide_operands[1], wide_operands[2], d); \ |
|
} |
|
|
|
#define COMPILE_X8_S8_C8_S8__C32(j, comp) \ |
|
COMPILE_X8_S8_S8_C8__C32(j, comp) |
|
#define COMPILE_WIDE_X8_S8_C8_S8__C32(j, comp) \ |
|
{ \ |
|
SCM_UNUSED uint8_t a, c; \ |
|
uint8_t b; \ |
|
UNPACK_8_8_8 (j->ip[0], a, b, c); \ |
|
uint32_t d = j->ip[1]; \ |
|
comp (j, wide_operands[0], b, wide_operands[1], d); \ |
|
} |
|
|
|
#define COMPILE_X8_S24__V32_X8_L24(j, comp) \ |
|
{ \ |
|
uint32_t a, len; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
len = j->ip[1]; \ |
|
j->next_ip += len; \ |
|
comp (j, a, len, j->ip + 2); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__V32_X8_L24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X32__LO32__L32(j, comp) \ |
|
{ \ |
|
int32_t a = j->ip[1], b = j->ip[2]; \ |
|
comp (j, j->ip + a, j->ip + b); \ |
|
} |
|
#define COMPILE_WIDE_X32__LO32__L32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_F24__X8_C24__L32(j, comp) \ |
|
{ \ |
|
uint32_t a, b; \ |
|
int32_t c; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
UNPACK_24 (j->ip[1], b); \ |
|
c = j->ip[2]; \ |
|
comp (j, a, b, j->ip + c); \ |
|
} |
|
#define COMPILE_WIDE_X8_F24__X8_C24__L32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__A32__B32(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
uint64_t b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = (((uint64_t) j->ip[1]) << 32) | ((uint64_t) j->ip[2]); \ |
|
ASSERT (b <= (uint64_t) UINTPTR_MAX); \ |
|
comp (j, a, SCM_PACK ((uintptr_t) b)); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__A32__B32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__AF32__BF32(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
union { uint64_t u; double d; } b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b.u = (((uint64_t) j->ip[1]) << 32) | ((uint64_t) j->ip[2]); \ |
|
comp (j, a, b.d); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__AF32__BF32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__AS32__BS32(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
uint64_t b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = (((uint64_t) j->ip[1]) << 32) | ((uint64_t) j->ip[2]); \ |
|
comp (j, a, (int64_t) b); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__AS32__BS32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__AU32__BU32(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
uint64_t b; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = (((uint64_t) j->ip[1]) << 32) | ((uint64_t) j->ip[2]); \ |
|
comp (j, a, b); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__AU32__BU32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__B1_X7_F24__X8_L24(j, comp) \ |
|
{ \ |
|
uint32_t a, c; \ |
|
uint8_t b; \ |
|
int32_t d; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
b = j->ip[1] & 0x1; \ |
|
UNPACK_24 (j->ip[1], c); \ |
|
d = j->ip[2]; d >>= 8; \ |
|
comp (j, a, b, c, j->ip + d); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__B1_X7_F24__X8_L24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__X8_S24__C8_S24(j, comp) \ |
|
{ \ |
|
uint32_t a, b, d; \ |
|
uint8_t c; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
UNPACK_24 (j->ip[1], b); \ |
|
UNPACK_8_24 (j->ip[2], c, d); \ |
|
comp (j, a, b, c, d); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__X8_S24__C8_S24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_C24__C8_C24__X8_C24__N32(j, comp) \ |
|
{ \ |
|
uint32_t a, c, d; \ |
|
uint8_t b; \ |
|
int32_t e; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
UNPACK_8_24 (j->ip[1], b, c); \ |
|
UNPACK_24 (j->ip[2], d); \ |
|
e = j->ip[3]; \ |
|
comp (j, a, b, c, d, j->ip + e); \ |
|
} |
|
#define COMPILE_WIDE_X8_C24__C8_C24__X8_C24__N32(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__X8_S24__C8_S24__X8_S24(j, comp) \ |
|
{ \ |
|
uint32_t a, b, d, e; \ |
|
uint8_t c; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
UNPACK_24 (j->ip[1], b); \ |
|
UNPACK_8_24 (j->ip[2], c, d); \ |
|
UNPACK_24 (j->ip[3], e); \ |
|
comp (j, a, b, c, d, e); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__X8_S24__C8_S24__X8_S24(j, comp) UNREACHABLE() |
|
|
|
#define COMPILE_X8_S24__N32__N32__C32(j, comp) \ |
|
{ \ |
|
uint32_t a; \ |
|
UNPACK_24 (j->ip[0], a); \ |
|
int32_t b = j->ip[1]; \ |
|
int32_t c = j->ip[2]; \ |
|
uint32_t d = j->ip[3]; \ |
|
comp (j, a, j->ip + b, j->ip + c, d); \ |
|
} |
|
#define COMPILE_WIDE_X8_S24__N32__N32__C32(j, comp) UNREACHABLE() |
|
|
|
static uint8_t |
|
parse_wide_operands (scm_jit_state *j, uint32_t wide_operands[3]) |
|
{ |
|
uint8_t opcode = j->ip[0] & 0xff; |
|
uint32_t push_count = 0; |
|
while (opcode == scm_op_push) |
|
{ |
|
ASSERT (push_count < 3); |
|
UNPACK_24 (j->ip[0], wide_operands[push_count]); |
|
wide_operands[push_count] -= push_count; |
|
push_count++; |
|
j->ip = j->next_ip; |
|
opcode = j->ip[0] & 0xff; |
|
j->next_ip = j->ip + op_lengths[opcode]; |
|
} |
|
ASSERT (push_count > 0); |
|
|
|
uint8_t finish_opcode = j->next_ip[0] & 0xff; |
|
uint32_t pop_count = 0; |
|
if (finish_opcode == scm_op_drop) |
|
{ |
|
uint32_t count; |
|
UNPACK_24 (j->next_ip[0], count); |
|
pop_count += count; |
|
ASSERT(pop_count <= push_count); |
|
j->next_ip = j->next_ip + op_lengths[finish_opcode]; |
|
finish_opcode = j->next_ip[0] & 0xff; |
|
} |
|
if (finish_opcode == scm_op_pop) |
|
{ |
|
ASSERT (push_count < 3); |
|
ASSERT (push_count - pop_count == 1); |
|
switch (push_count) { |
|
case 2: wide_operands[2] = wide_operands[1]; |
|
case 1: wide_operands[1] = wide_operands[0]; break; |
|
default: UNREACHABLE (); |
|
} |
|
UNPACK_24 (j->next_ip[0], wide_operands[0]); |
|
pop_count++; |
|
j->next_ip = j->next_ip + op_lengths[finish_opcode]; |
|
finish_opcode = j->next_ip[0] & 0xff; |
|
} |
|
|
|
ASSERT (pop_count == push_count); |
|
return opcode; |
|
} |
|
|
|
static uintptr_t opcodes_seen[256 / (SCM_SIZEOF_UINTPTR_T * 8)]; |
|
|
|
static uintptr_t |
|
bitvector_ref (const uintptr_t *bv, size_t n) |
|
{ |
|
uintptr_t word = bv[n / (SCM_SIZEOF_UINTPTR_T * 8)]; |
|
return word & (((uintptr_t) 1) << (n & (SCM_SIZEOF_UINTPTR_T * 8 - 1))); |
|
} |
|
|
|
static void |
|
bitvector_set (uintptr_t *bv, size_t n) |
|
{ |
|
uintptr_t *word_loc = &bv[n / (SCM_SIZEOF_UINTPTR_T * 8)]; |
|
*word_loc |= ((uintptr_t) 1) << (n & (SCM_SIZEOF_UINTPTR_T * 8 - 1)); |
|
} |
|
|
|
static void |
|
compile1 (scm_jit_state *j) |
|
{ |
|
uint8_t opcode = j->ip[0] & 0xff; |
|
|
|
if (jit_log_level >= LOG_LEVEL_DEBUG) |
|
{ |
|
const char *n; |
|
switch (opcode) |
|
{ |
|
#define NAME(code, cname, name, arity) case code: n = name; break; |
|
FOR_EACH_VM_OPERATION(NAME) |
|
#undef NAME |
|
default: |
|
UNREACHABLE (); |
|
} |
|
|
|
if (!bitvector_ref (opcodes_seen, opcode)) |
|
{ |
|
bitvector_set (opcodes_seen, opcode); |
|
DEBUG ("Instruction first seen at vcode %p: %s\n", j->ip, n); |
|
} |
|
|
|
LOG ("Instruction at vcode %p: %s\n", j->ip, n); |
|
} |
|
|
|
j->next_ip = j->ip + op_lengths[opcode]; |
|
|
|
if (opcode == scm_op_push) |
|
{ |
|
uint32_t wide_operands[3]; |
|
opcode = parse_wide_operands (j, wide_operands); |
|
switch (opcode) |
|
{ |
|
#define COMPILE1(code, cname, name, arity) \ |
|
case code: COMPILE_WIDE_##arity(j, compile_##cname); break; |
|
FOR_EACH_VM_OPERATION(COMPILE1) |
|
#undef COMPILE1 |
|
default: |
|
UNREACHABLE (); |
|
} |
|
} |
|
else |
|
switch (opcode) |
|
{ |
|
#define COMPILE1(code, cname, name, arity) \ |
|
case code: COMPILE_##arity(j, compile_##cname); break; |
|
FOR_EACH_VM_OPERATION(COMPILE1) |
|
#undef COMPILE1 |
|
default: |
|
UNREACHABLE (); |
|
} |
|
|
|
j->ip = j->next_ip; |
|
} |
|
|
|
static void |
|
compile_slow_path (scm_jit_state *j) |
|
{ |
|
uint8_t opcode = j->ip[0] & 0xff; |
|
j->next_ip = j->ip + op_lengths[opcode]; |
|
|
|
if (opcode == scm_op_push) |
|
{ |
|
uint32_t wide_operands[3]; |
|
opcode = parse_wide_operands (j, wide_operands); |
|
ptrdiff_t offset = j->ip - j->start; |
|
j->labels[slow_label_offset (offset)] = jit_address (j->jit); |
|
switch (opcode) |
|
{ |
|
#define COMPILE_SLOW(code, cname, name, arity) \ |
|
case code: COMPILE_WIDE_##arity(j, compile_##cname##_slow); break; |
|
FOR_EACH_VM_OPERATION(COMPILE_SLOW) |
|
#undef COMPILE_SLOW |
|
default: |
|
UNREACHABLE (); |
|
} |
|
} |
|
else |
|
{ |
|
ptrdiff_t offset = j->ip - j->start; |
|
j->labels[slow_label_offset (offset)] = jit_address (j->jit); |
|
switch (opcode) |
|
{ |
|
#define COMPILE_SLOW(code, cname, name, arity) \ |
|
case code: COMPILE_##arity(j, compile_##cname##_slow); break; |
|
FOR_EACH_VM_OPERATION(COMPILE_SLOW) |
|
#undef COMPILE_SLOW |
|
default: |
|
UNREACHABLE (); |
|
} |
|
} |
|
|
|
j->ip = j->next_ip; |
|
} |
|
|
|
static void |
|
analyze (scm_jit_state *j) |
|
{ |
|
memset (j->op_attrs, 0, j->end - j->start); |
|
|
|
j->op_attrs[0] = OP_ATTR_BLOCK | OP_ATTR_ENTRY; |
|
|
|
for (j->ip = (uint32_t *) j->start; j->ip < j->end; j->ip = j->next_ip) |
|
{ |
|
uint8_t opcode = j->ip[0] & 0xff; |
|
uint8_t attrs = 0; |
|
uint32_t *target; |
|
|
|
j->next_ip = j->ip + op_lengths[opcode]; |
|
|
|
switch (opcode) |
|
{ |
|
case scm_op_check_arguments: |
|
case scm_op_check_positional_arguments: |
|
attrs |= OP_ATTR_ENTRY; |
|
|
|
case scm_op_u64_numerically_equal: |
|
case scm_op_u64_less: |
|
case scm_op_s64_less: |
|
case scm_op_f64_numerically_equal: |
|
case scm_op_f64_less: |
|
case scm_op_numerically_equal: |
|
case scm_op_less: |
|
case scm_op_immediate_tag_equals: |
|
case scm_op_heap_tag_equals: |
|
case scm_op_eq: |
|
case scm_op_eq_immediate: |
|
case scm_op_heap_numbers_equal: |
|
case scm_op_s64_imm_numerically_equal: |
|
case scm_op_u64_imm_less: |
|
case scm_op_imm_u64_less: |
|
case scm_op_s64_imm_less: |
|
case scm_op_imm_s64_less: |
|
{ |
|
uint8_t next = j->next_ip[0] & 0xff; |
|
if (next == scm_op_drop) |
|
{ |
|
j->next_ip += op_lengths[next]; |
|
next = j->next_ip[0] & 0xff; |
|
} |
|
if (next == scm_op_pop) |
|
{ |
|
j->next_ip += op_lengths[next]; |
|
next = j->next_ip[0] & 0xff; |
|
} |
|
} |
|
attrs |= OP_ATTR_BLOCK; |
|
fuse_conditional_branch (j, &target); |
|
j->op_attrs[target - j->start] |= attrs; |
|
break; |
|
|
|
case scm_op_j: |
|
target = j->ip + (((int32_t)j->ip[0]) >> 8); |
|
j->op_attrs[target - j->start] |= OP_ATTR_BLOCK; |
|
break; |
|
|
|
case scm_op_jtable: |
|
{ |
|
uint32_t len = j->ip[1]; |
|
const uint32_t *offsets = j->ip + 2; |
|
for (uint32_t i = 0; i < len; i++) |
|
{ |
|
int32_t offset = offsets[i]; |
|
offset >>= 8; |
|
target = j->ip + offset; |
|
ASSERT(j->start <= target && target < j->end); |
|
j->op_attrs[target - j->start] |= OP_ATTR_BLOCK; |
|
} |
|
j->next_ip += len; |
|
break; |
|
} |
|
|
|
case scm_op_call: |
|
case scm_op_call_label: |
|
attrs = OP_ATTR_BLOCK; |
|
target = j->next_ip; |
|
j->op_attrs[target - j->start] |= OP_ATTR_BLOCK | OP_ATTR_ENTRY; |
|
break; |
|
|
|
case scm_op_prompt: |
|
target = j->ip + (((int32_t) j->ip[2]) >> 8); |
|
j->op_attrs[target - j->start] |= OP_ATTR_BLOCK | OP_ATTR_ENTRY; |
|
break; |
|
|
|
default: |
|
break; |
|
} |
|
} |
|
|
|
|
|
ASSERT (j->op_attrs[j->entry - j->start] & OP_ATTR_BLOCK); |
|
} |
|
|
|
static void |
|
compile (scm_jit_state *j) |
|
{ |
|
j->ip = (uint32_t *) j->start; |
|
set_register_state (j, SP_IN_REGISTER | FP_IN_REGISTER); |
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
|
|
for (ptrdiff_t offset = 0; j->ip + offset < j->end; offset++) { |
|
j->labels[inline_label_offset (offset)] = NULL; |
|
j->labels[slow_label_offset (offset)] = NULL; |
|
} |
|
|
|
j->reloc_idx = 0; |
|
|
|
while (j->ip < j->end) |
|
{ |
|
ptrdiff_t offset = j->ip - j->start; |
|
uint8_t attrs = j->op_attrs[offset]; |
|
j->labels[inline_label_offset (offset)] = jit_address (j->jit); |
|
if (attrs & OP_ATTR_BLOCK) |
|
{ |
|
uint32_t state = SP_IN_REGISTER; |
|
if (attrs & OP_ATTR_ENTRY) |
|
state |= FP_IN_REGISTER; |
|
j->register_state = state; |
|
} |
|
compile1 (j); |
|
|
|
if (jit_has_overflow (j->jit)) |
|
return; |
|
} |
|
|
|
jit_breakpoint (j->jit); |
|
|
|
j->ip = (uint32_t *) j->start; |
|
while (j->ip < j->end) |
|
{ |
|
|
|
reset_register_state (j, SP_IN_REGISTER); |
|
compile_slow_path (j); |
|
|
|
if (jit_has_overflow (j->jit)) |
|
return; |
|
} |
|
|
|
jit_breakpoint (j->jit); |
|
|
|
for (size_t i = 0; i < j->reloc_idx; i++) |
|
{ |
|
void *target = j->labels[j->relocs[i].target_label_offset]; |
|
ASSERT(target); |
|
jit_patch_there (j->jit, j->relocs[i].reloc, target); |
|
} |
|
} |
|
|
|
static scm_i_pthread_once_t initialize_jit_once = SCM_I_PTHREAD_ONCE_INIT; |
|
|
|
static void* |
|
jit_alloc_fn (size_t size) |
|
{ |
|
return scm_gc_malloc (size, "jit state"); |
|
} |
|
|
|
static void |
|
jit_free_fn (void *unused) |
|
{ |
|
} |
|
|
|
static scm_jit_state * |
|
initialize_thread_jit_state (scm_thread *thread) |
|
{ |
|
scm_jit_state *j; |
|
|
|
ASSERT (!thread->jit_state); |
|
|
|
j = scm_gc_malloc (sizeof (*j), "jit state"); |
|
memset (j, 0, sizeof (*j)); |
|
thread->jit_state = j; |
|
j->jit = jit_new_state (jit_alloc_fn, jit_free_fn); |
|
|
|
return j; |
|
} |
|
|
|
static void |
|
initialize_jit (void) |
|
{ |
|
scm_jit_state *j; |
|
|
|
if (!init_jit ()) |
|
{ |
|
scm_jit_counter_threshold = -1; |
|
fprintf (stderr, "JIT failed to initialize\n"); |
|
fprintf (stderr, "disabling automatic JIT compilation\n"); |
|
return; |
|
} |
|
|
|
|
|
|
|
j = initialize_thread_jit_state (SCM_I_CURRENT_THREAD); |
|
|
|
jit_pointer_t enter_mcode_addr = emit_code (j, emit_entry_trampoline); |
|
ASSERT (enter_mcode_addr); |
|
enter_mcode = jit_address_to_function_pointer (enter_mcode_addr); |
|
|
|
handle_interrupts_trampoline = |
|
emit_code (j, emit_handle_interrupts_trampoline); |
|
ASSERT (handle_interrupts_trampoline); |
|
|
|
scm_jit_return_to_interpreter_trampoline = |
|
emit_code (j, emit_return_to_interpreter_trampoline); |
|
ASSERT (scm_jit_return_to_interpreter_trampoline); |
|
scm_jit_return_to_interpreter_trampoline = jit_address_to_function_pointer |
|
(scm_jit_return_to_interpreter_trampoline); |
|
} |
|
|
|
static scm_i_pthread_once_t create_perf_map_once = SCM_I_PTHREAD_ONCE_INIT; |
|
static FILE *perf_map = NULL; |
|
static void |
|
create_perf_map (void) |
|
{ |
|
unsigned long pid = getpid (); |
|
char *file_name; |
|
if (asprintf (&file_name, "/tmp/perf-%lu.map", pid) < 0) |
|
return; |
|
perf_map = fopen (file_name, "w"); |
|
if (perf_map) |
|
DEBUG ("created %s\n", file_name); |
|
free (file_name); |
|
} |
|
|
|
static uint8_t * |
|
compute_mcode (scm_thread *thread, uint32_t *entry_ip, |
|
struct scm_jit_function_data *data) |
|
{ |
|
scm_jit_state *j = thread->jit_state; |
|
uint8_t *entry_mcode; |
|
|
|
if (!j) |
|
{ |
|
scm_i_pthread_once (&initialize_jit_once, initialize_jit); |
|
if (scm_jit_counter_threshold == -1) |
|
{ |
|
|
|
return NULL; |
|
} |
|
|
|
j = thread->jit_state; |
|
|
|
|
|
if (!j) |
|
j = initialize_thread_jit_state (thread); |
|
} |
|
|
|
j->thread = thread; |
|
j->start = (const uint32_t *) (((char *)data) + data->start); |
|
j->end = (const uint32_t *) (((char *)data) + data->end); |
|
j->entry = entry_ip; |
|
|
|
ASSERT (j->start < j->end); |
|
ASSERT (j->start <= j->entry); |
|
ASSERT (j->entry < j->end); |
|
|
|
j->op_attrs = calloc ((j->end - j->start), sizeof (*j->op_attrs)); |
|
ASSERT (j->op_attrs); |
|
j->labels = calloc ((j->end - j->start) * 2, sizeof (*j->labels)); |
|
ASSERT (j->labels); |
|
|
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
|
|
INFO ("vcode: start=%p,+%zu entry=+%zu\n", j->start, j->end - j->start, |
|
j->entry - j->start); |
|
|
|
analyze (j); |
|
|
|
uint8_t *mcode = emit_code (j, compile); |
|
if (mcode) |
|
{ |
|
entry_mcode = j->labels[inline_label_offset (j->entry - j->start)]; |
|
data->mcode = mcode; |
|
|
|
if (jit_log_level >= LOG_LEVEL_INFO) { |
|
scm_i_pthread_once (&create_perf_map_once, create_perf_map); |
|
if (perf_map) { |
|
uint8_t *end = j->code_arena->base + j->code_arena->used; |
|
fprintf (perf_map, "%lx %zx %p,+%zu\n", (long)mcode, end - mcode, |
|
j->start, j->end - j->start); |
|
fflush (perf_map); |
|
} |
|
} |
|
} |
|
else |
|
{ |
|
entry_mcode = NULL; |
|
} |
|
|
|
free (j->op_attrs); |
|
j->op_attrs = NULL; |
|
free (j->labels); |
|
j->labels = NULL; |
|
free (j->relocs); |
|
j->relocs = NULL; |
|
j->reloc_idx = 0; |
|
j->reloc_count = 0; |
|
|
|
j->start = j->end = j->ip = j->entry = NULL; |
|
j->frame_size_min = 0; |
|
j->frame_size_max = INT32_MAX; |
|
|
|
return entry_mcode; |
|
} |
|
|
|
const uint8_t * |
|
scm_jit_compute_mcode (scm_thread *thread, struct scm_jit_function_data *data) |
|
{ |
|
const uint32_t *vcode_start = (const uint32_t *) (((char *)data) + data->start); |
|
|
|
if (data->mcode) |
|
{ |
|
if (vcode_start == thread->vm.ip) |
|
return data->mcode; |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
} |
|
|
|
uint8_t *mcode = compute_mcode (thread, thread->vm.ip, data); |
|
|
|
if (!mcode) |
|
{ |
|
scm_jit_counter_threshold = -1; |
|
fprintf (stderr, "JIT failed due to resource exhaustion\n"); |
|
fprintf (stderr, "disabling automatic JIT compilation\n"); |
|
} |
|
else if (--jit_stop_after == 0) |
|
{ |
|
scm_jit_counter_threshold = -1; |
|
fprintf (stderr, "stopping automatic JIT compilation, as requested\n"); |
|
if (jit_pause_when_stopping) |
|
{ |
|
fprintf (stderr, "sleeping for 30s; to debug:\n"); |
|
fprintf (stderr, " gdb -p %d\n\n", getpid ()); |
|
sleep (30); |
|
} |
|
} |
|
|
|
return mcode; |
|
} |
|
|
|
void |
|
scm_jit_enter_mcode (scm_thread *thread, const uint8_t *mcode) |
|
{ |
|
LOG ("entering mcode: %p\n", mcode); |
|
if (!SCM_FRAME_MACHINE_RETURN_ADDRESS (thread->vm.fp)) |
|
SCM_FRAME_SET_MACHINE_RETURN_ADDRESS |
|
(thread->vm.fp, scm_jit_return_to_interpreter_trampoline); |
|
enter_mcode (thread, mcode); |
|
LOG ("exited mcode\n"); |
|
} |
|
|
|
|
|
|
|
void |
|
scm_jit_clear_mcode_return_addresses (scm_thread *thread) |
|
{ |
|
union scm_vm_stack_element *fp; |
|
struct scm_vm *vp = &thread->vm; |
|
|
|
for (fp = vp->fp; fp < vp->stack_top; fp = SCM_FRAME_DYNAMIC_LINK (fp)) |
|
SCM_FRAME_SET_MACHINE_RETURN_ADDRESS |
|
(fp, scm_jit_return_to_interpreter_trampoline); |
|
} |
|
|
|
void |
|
scm_jit_state_free (scm_jit_state *j) |
|
{ |
|
|
|
} |
|
|
|
void |
|
scm_init_jit (void) |
|
{ |
|
scm_jit_counter_threshold = scm_getenv_int ("GUILE_JIT_THRESHOLD", |
|
default_jit_threshold); |
|
jit_stop_after = scm_getenv_int ("GUILE_JIT_STOP_AFTER", -1); |
|
jit_pause_when_stopping = scm_getenv_int ("GUILE_JIT_PAUSE_WHEN_STOPPING", 0); |
|
jit_log_level = scm_getenv_int ("GUILE_JIT_LOG", 0); |
|
} |
|
|
|
#endif |
|
|