525 lines
14 KiB
C
525 lines
14 KiB
C
/* swap_macros.h - helper macros for context switch */
|
|
|
|
/*
|
|
* Copyright (c) 2014 Wind River Systems, Inc.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#ifndef ZEPHYR_ARCH_ARC_INCLUDE_SWAP_MACROS_H_
|
|
#define ZEPHYR_ARCH_ARC_INCLUDE_SWAP_MACROS_H_
|
|
|
|
#include <kernel_structs.h>
|
|
#include <offsets_short.h>
|
|
#include <toolchain.h>
|
|
#include <arch/cpu.h>
|
|
|
|
#ifdef _ASMLANGUAGE
|
|
|
|
/* save callee regs of current thread in r2 */
|
|
.macro _save_callee_saved_regs
|
|
|
|
sub_s sp, sp, ___callee_saved_stack_t_SIZEOF
|
|
|
|
/* save regs on stack */
|
|
st_s r13, [sp, ___callee_saved_stack_t_r13_OFFSET]
|
|
st_s r14, [sp, ___callee_saved_stack_t_r14_OFFSET]
|
|
st_s r15, [sp, ___callee_saved_stack_t_r15_OFFSET]
|
|
st r16, [sp, ___callee_saved_stack_t_r16_OFFSET]
|
|
st r17, [sp, ___callee_saved_stack_t_r17_OFFSET]
|
|
st r18, [sp, ___callee_saved_stack_t_r18_OFFSET]
|
|
st r19, [sp, ___callee_saved_stack_t_r19_OFFSET]
|
|
st r20, [sp, ___callee_saved_stack_t_r20_OFFSET]
|
|
st r21, [sp, ___callee_saved_stack_t_r21_OFFSET]
|
|
st r22, [sp, ___callee_saved_stack_t_r22_OFFSET]
|
|
st r23, [sp, ___callee_saved_stack_t_r23_OFFSET]
|
|
st r24, [sp, ___callee_saved_stack_t_r24_OFFSET]
|
|
st r25, [sp, ___callee_saved_stack_t_r25_OFFSET]
|
|
st r26, [sp, ___callee_saved_stack_t_r26_OFFSET]
|
|
st fp, [sp, ___callee_saved_stack_t_fp_OFFSET]
|
|
|
|
#ifdef CONFIG_USERSPACE
|
|
#ifdef CONFIG_ARC_HAS_SECURE
|
|
#ifdef CONFIG_ARC_SECURE_FIRMWARE
|
|
lr r13, [_ARC_V2_SEC_U_SP]
|
|
st_s r13, [sp, ___callee_saved_stack_t_user_sp_OFFSET]
|
|
lr r13, [_ARC_V2_SEC_K_SP]
|
|
st_s r13, [sp, ___callee_saved_stack_t_kernel_sp_OFFSET]
|
|
#else
|
|
lr r13, [_ARC_V2_USER_SP]
|
|
st_s r13, [sp, ___callee_saved_stack_t_user_sp_OFFSET]
|
|
lr r13, [_ARC_V2_KERNEL_SP]
|
|
st_s r13, [sp, ___callee_saved_stack_t_kernel_sp_OFFSET]
|
|
#endif /* CONFIG_ARC_SECURE_FIRMWARE */
|
|
#else
|
|
lr r13, [_ARC_V2_USER_SP]
|
|
st_s r13, [sp, ___callee_saved_stack_t_user_sp_OFFSET]
|
|
#endif
|
|
#endif
|
|
st r30, [sp, ___callee_saved_stack_t_r30_OFFSET]
|
|
|
|
#ifdef CONFIG_ARC_HAS_ACCL_REGS
|
|
st r58, [sp, ___callee_saved_stack_t_r58_OFFSET]
|
|
st r59, [sp, ___callee_saved_stack_t_r59_OFFSET]
|
|
#endif
|
|
|
|
#ifdef CONFIG_FP_SHARING
|
|
ld_s r13, [r2, ___thread_base_t_user_options_OFFSET]
|
|
/* K_FP_REGS is bit 1 */
|
|
bbit0 r13, 1, 1f
|
|
lr r13, [_ARC_V2_FPU_STATUS]
|
|
st_s r13, [sp, ___callee_saved_stack_t_fpu_status_OFFSET]
|
|
lr r13, [_ARC_V2_FPU_CTRL]
|
|
st_s r13, [sp, ___callee_saved_stack_t_fpu_ctrl_OFFSET]
|
|
|
|
#ifdef CONFIG_FP_FPU_DA
|
|
lr r13, [_ARC_V2_FPU_DPFP1L]
|
|
st_s r13, [sp, ___callee_saved_stack_t_dpfp1l_OFFSET]
|
|
lr r13, [_ARC_V2_FPU_DPFP1H]
|
|
st_s r13, [sp, ___callee_saved_stack_t_dpfp1h_OFFSET]
|
|
lr r13, [_ARC_V2_FPU_DPFP2L]
|
|
st_s r13, [sp, ___callee_saved_stack_t_dpfp2l_OFFSET]
|
|
lr r13, [_ARC_V2_FPU_DPFP2H]
|
|
st_s r13, [sp, ___callee_saved_stack_t_dpfp2h_OFFSET]
|
|
#endif
|
|
1 :
|
|
#endif
|
|
|
|
/* save stack pointer in struct k_thread */
|
|
st sp, [r2, _thread_offset_to_sp]
|
|
.endm
|
|
|
|
/* load the callee regs of thread (in r2)*/
|
|
.macro _load_callee_saved_regs
|
|
/* restore stack pointer from struct k_thread */
|
|
ld sp, [r2, _thread_offset_to_sp]
|
|
|
|
#ifdef CONFIG_ARC_HAS_ACCL_REGS
|
|
ld r58, [sp, ___callee_saved_stack_t_r58_OFFSET]
|
|
ld r59, [sp, ___callee_saved_stack_t_r59_OFFSET]
|
|
#endif
|
|
|
|
#ifdef CONFIG_FP_SHARING
|
|
ld_s r13, [r2, ___thread_base_t_user_options_OFFSET]
|
|
/* K_FP_REGS is bit 1 */
|
|
bbit0 r13, 1, 2f
|
|
|
|
ld_s r13, [sp, ___callee_saved_stack_t_fpu_status_OFFSET]
|
|
sr r13, [_ARC_V2_FPU_STATUS]
|
|
ld_s r13, [sp, ___callee_saved_stack_t_fpu_ctrl_OFFSET]
|
|
sr r13, [_ARC_V2_FPU_CTRL]
|
|
|
|
#ifdef CONFIG_FP_FPU_DA
|
|
ld_s r13, [sp, ___callee_saved_stack_t_dpfp1l_OFFSET]
|
|
sr r13, [_ARC_V2_FPU_DPFP1L]
|
|
ld_s r13, [sp, ___callee_saved_stack_t_dpfp1h_OFFSET]
|
|
sr r13, [_ARC_V2_FPU_DPFP1H]
|
|
ld_s r13, [sp, ___callee_saved_stack_t_dpfp2l_OFFSET]
|
|
sr r13, [_ARC_V2_FPU_DPFP2L]
|
|
ld_s r13, [sp, ___callee_saved_stack_t_dpfp2h_OFFSET]
|
|
sr r13, [_ARC_V2_FPU_DPFP2H]
|
|
#endif
|
|
2 :
|
|
#endif
|
|
|
|
#ifdef CONFIG_USERSPACE
|
|
#ifdef CONFIG_ARC_HAS_SECURE
|
|
#ifdef CONFIG_ARC_SECURE_FIRMWARE
|
|
ld_s r13, [sp, ___callee_saved_stack_t_user_sp_OFFSET]
|
|
sr r13, [_ARC_V2_SEC_U_SP]
|
|
ld_s r13, [sp, ___callee_saved_stack_t_kernel_sp_OFFSET]
|
|
sr r13, [_ARC_V2_SEC_K_SP]
|
|
#else
|
|
ld_s r13, [sp, ___callee_saved_stack_t_user_sp_OFFSET]
|
|
sr r13, [_ARC_V2_USER_SP]
|
|
ld_s r13, [sp, ___callee_saved_stack_t_kernel_sp_OFFSET]
|
|
sr r13, [_ARC_V2_KERNEL_SP]
|
|
#endif /* CONFIG_ARC_SECURE_FIRMWARE */
|
|
#else
|
|
ld_s r13, [sp, ___callee_saved_stack_t_user_sp_OFFSET]
|
|
sr r13, [_ARC_V2_USER_SP]
|
|
#endif
|
|
#endif
|
|
|
|
ld_s r13, [sp, ___callee_saved_stack_t_r13_OFFSET]
|
|
ld_s r14, [sp, ___callee_saved_stack_t_r14_OFFSET]
|
|
ld_s r15, [sp, ___callee_saved_stack_t_r15_OFFSET]
|
|
ld r16, [sp, ___callee_saved_stack_t_r16_OFFSET]
|
|
ld r17, [sp, ___callee_saved_stack_t_r17_OFFSET]
|
|
ld r18, [sp, ___callee_saved_stack_t_r18_OFFSET]
|
|
ld r19, [sp, ___callee_saved_stack_t_r19_OFFSET]
|
|
ld r20, [sp, ___callee_saved_stack_t_r20_OFFSET]
|
|
ld r21, [sp, ___callee_saved_stack_t_r21_OFFSET]
|
|
ld r22, [sp, ___callee_saved_stack_t_r22_OFFSET]
|
|
ld r23, [sp, ___callee_saved_stack_t_r23_OFFSET]
|
|
ld r24, [sp, ___callee_saved_stack_t_r24_OFFSET]
|
|
ld r25, [sp, ___callee_saved_stack_t_r25_OFFSET]
|
|
ld r26, [sp, ___callee_saved_stack_t_r26_OFFSET]
|
|
ld fp, [sp, ___callee_saved_stack_t_fp_OFFSET]
|
|
ld r30, [sp, ___callee_saved_stack_t_r30_OFFSET]
|
|
|
|
add_s sp, sp, ___callee_saved_stack_t_SIZEOF
|
|
|
|
.endm
|
|
|
|
/* discard callee regs */
|
|
.macro _discard_callee_saved_regs
|
|
add_s sp, sp, ___callee_saved_stack_t_SIZEOF
|
|
.endm
|
|
|
|
/*
|
|
* Must be called with interrupts locked or in P0.
|
|
* Upon exit, sp will be pointing to the stack frame.
|
|
*/
|
|
.macro _create_irq_stack_frame
|
|
|
|
sub_s sp, sp, ___isf_t_SIZEOF
|
|
|
|
st blink, [sp, ___isf_t_blink_OFFSET]
|
|
|
|
/* store these right away so we can use them if needed */
|
|
|
|
st_s r13, [sp, ___isf_t_r13_OFFSET]
|
|
st_s r12, [sp, ___isf_t_r12_OFFSET]
|
|
st r11, [sp, ___isf_t_r11_OFFSET]
|
|
st r10, [sp, ___isf_t_r10_OFFSET]
|
|
st r9, [sp, ___isf_t_r9_OFFSET]
|
|
st r8, [sp, ___isf_t_r8_OFFSET]
|
|
st r7, [sp, ___isf_t_r7_OFFSET]
|
|
st r6, [sp, ___isf_t_r6_OFFSET]
|
|
st r5, [sp, ___isf_t_r5_OFFSET]
|
|
st r4, [sp, ___isf_t_r4_OFFSET]
|
|
st_s r3, [sp, ___isf_t_r3_OFFSET]
|
|
st_s r2, [sp, ___isf_t_r2_OFFSET]
|
|
st_s r1, [sp, ___isf_t_r1_OFFSET]
|
|
st_s r0, [sp, ___isf_t_r0_OFFSET]
|
|
|
|
mov r0, lp_count
|
|
st_s r0, [sp, ___isf_t_lp_count_OFFSET]
|
|
lr r1, [_ARC_V2_LP_START]
|
|
lr r0, [_ARC_V2_LP_END]
|
|
st_s r1, [sp, ___isf_t_lp_start_OFFSET]
|
|
st_s r0, [sp, ___isf_t_lp_end_OFFSET]
|
|
|
|
#ifdef CONFIG_CODE_DENSITY
|
|
lr r1, [_ARC_V2_JLI_BASE]
|
|
lr r0, [_ARC_V2_LDI_BASE]
|
|
lr r2, [_ARC_V2_EI_BASE]
|
|
st_s r1, [sp, ___isf_t_jli_base_OFFSET]
|
|
st_s r0, [sp, ___isf_t_ldi_base_OFFSET]
|
|
st_s r2, [sp, ___isf_t_ei_base_OFFSET]
|
|
#endif
|
|
|
|
.endm
|
|
|
|
/*
|
|
* Must be called with interrupts locked or in P0.
|
|
* sp must be pointing the to stack frame.
|
|
*/
|
|
.macro _pop_irq_stack_frame
|
|
|
|
ld blink, [sp, ___isf_t_blink_OFFSET]
|
|
|
|
#ifdef CONFIG_CODE_DENSITY
|
|
ld_s r1, [sp, ___isf_t_jli_base_OFFSET]
|
|
ld_s r0, [sp, ___isf_t_ldi_base_OFFSET]
|
|
ld_s r2, [sp, ___isf_t_ei_base_OFFSET]
|
|
sr r1, [_ARC_V2_JLI_BASE]
|
|
sr r0, [_ARC_V2_LDI_BASE]
|
|
sr r2, [_ARC_V2_EI_BASE]
|
|
#endif
|
|
|
|
ld_s r0, [sp, ___isf_t_lp_count_OFFSET]
|
|
mov lp_count, r0
|
|
ld_s r1, [sp, ___isf_t_lp_start_OFFSET]
|
|
ld_s r0, [sp, ___isf_t_lp_end_OFFSET]
|
|
sr r1, [_ARC_V2_LP_START]
|
|
sr r0, [_ARC_V2_LP_END]
|
|
|
|
ld_s r13, [sp, ___isf_t_r13_OFFSET]
|
|
ld_s r12, [sp, ___isf_t_r12_OFFSET]
|
|
ld r11, [sp, ___isf_t_r11_OFFSET]
|
|
ld r10, [sp, ___isf_t_r10_OFFSET]
|
|
ld r9, [sp, ___isf_t_r9_OFFSET]
|
|
ld r8, [sp, ___isf_t_r8_OFFSET]
|
|
ld r7, [sp, ___isf_t_r7_OFFSET]
|
|
ld r6, [sp, ___isf_t_r6_OFFSET]
|
|
ld r5, [sp, ___isf_t_r5_OFFSET]
|
|
ld r4, [sp, ___isf_t_r4_OFFSET]
|
|
ld_s r3, [sp, ___isf_t_r3_OFFSET]
|
|
ld_s r2, [sp, ___isf_t_r2_OFFSET]
|
|
ld_s r1, [sp, ___isf_t_r1_OFFSET]
|
|
ld_s r0, [sp, ___isf_t_r0_OFFSET]
|
|
|
|
|
|
/*
|
|
* All gprs have been reloaded, the only one that is still usable is
|
|
* ilink.
|
|
*
|
|
* The pc and status32 values will still be on the stack. We cannot
|
|
* pop them yet because the callers of _pop_irq_stack_frame must reload
|
|
* status32 differently depending on the execution context they are
|
|
* running in (arch_switch(), firq or exception).
|
|
*/
|
|
add_s sp, sp, ___isf_t_SIZEOF
|
|
|
|
.endm
|
|
|
|
/*
|
|
* To use this macro, r2 should have the value of thread struct pointer to
|
|
* _kernel.current. r3 is a scratch reg.
|
|
*/
|
|
.macro _load_stack_check_regs
|
|
#if defined(CONFIG_ARC_SECURE_FIRMWARE)
|
|
ld r3, [r2, _thread_offset_to_k_stack_base]
|
|
sr r3, [_ARC_V2_S_KSTACK_BASE]
|
|
ld r3, [r2, _thread_offset_to_k_stack_top]
|
|
sr r3, [_ARC_V2_S_KSTACK_TOP]
|
|
#ifdef CONFIG_USERSPACE
|
|
ld r3, [r2, _thread_offset_to_u_stack_base]
|
|
sr r3, [_ARC_V2_S_USTACK_BASE]
|
|
ld r3, [r2, _thread_offset_to_u_stack_top]
|
|
sr r3, [_ARC_V2_S_USTACK_TOP]
|
|
#endif
|
|
#else /* CONFIG_ARC_HAS_SECURE */
|
|
ld r3, [r2, _thread_offset_to_k_stack_base]
|
|
sr r3, [_ARC_V2_KSTACK_BASE]
|
|
ld r3, [r2, _thread_offset_to_k_stack_top]
|
|
sr r3, [_ARC_V2_KSTACK_TOP]
|
|
#ifdef CONFIG_USERSPACE
|
|
ld r3, [r2, _thread_offset_to_u_stack_base]
|
|
sr r3, [_ARC_V2_USTACK_BASE]
|
|
ld r3, [r2, _thread_offset_to_u_stack_top]
|
|
sr r3, [_ARC_V2_USTACK_TOP]
|
|
#endif
|
|
#endif /* CONFIG_ARC_SECURE_FIRMWARE */
|
|
.endm
|
|
|
|
/* check and increase the interrupt nest counter
|
|
* after increase, check whether nest counter == 1
|
|
* the result will be EQ bit of status32
|
|
* two temp regs are needed
|
|
*/
|
|
.macro _check_and_inc_int_nest_counter reg1 reg2
|
|
#ifdef CONFIG_SMP
|
|
_get_cpu_id \reg1
|
|
ld.as \reg1, [@_curr_cpu, \reg1]
|
|
ld \reg2, [\reg1, ___cpu_t_nested_OFFSET]
|
|
#else
|
|
mov \reg1, _kernel
|
|
ld \reg2, [\reg1, ___kernel_t_nested_OFFSET]
|
|
#endif
|
|
add \reg2, \reg2, 1
|
|
#ifdef CONFIG_SMP
|
|
st \reg2, [\reg1, ___cpu_t_nested_OFFSET]
|
|
#else
|
|
st \reg2, [\reg1, ___kernel_t_nested_OFFSET]
|
|
#endif
|
|
cmp \reg2, 1
|
|
.endm
|
|
|
|
/* decrease interrupt stack nest counter
|
|
* the counter > 0, interrupt stack is used, or
|
|
* not used
|
|
*/
|
|
.macro _dec_int_nest_counter reg1 reg2
|
|
#ifdef CONFIG_SMP
|
|
_get_cpu_id \reg1
|
|
ld.as \reg1, [@_curr_cpu, \reg1]
|
|
ld \reg2, [\reg1, ___cpu_t_nested_OFFSET]
|
|
#else
|
|
mov \reg1, _kernel
|
|
ld \reg2, [\reg1, ___kernel_t_nested_OFFSET]
|
|
#endif
|
|
sub \reg2, \reg2, 1
|
|
#ifdef CONFIG_SMP
|
|
st \reg2, [\reg1, ___cpu_t_nested_OFFSET]
|
|
#else
|
|
st \reg2, [\reg1, ___kernel_t_nested_OFFSET]
|
|
#endif
|
|
.endm
|
|
|
|
/* If multi bits in IRQ_ACT are set, i.e. last bit != fist bit, it's
|
|
* in nest interrupt. The result will be EQ bit of status32
|
|
* need two temp reg to do this
|
|
*/
|
|
.macro _check_nest_int_by_irq_act reg1, reg2
|
|
lr \reg1, [_ARC_V2_AUX_IRQ_ACT]
|
|
#ifdef CONFIG_ARC_SECURE_FIRMWARE
|
|
and \reg1, \reg1, ((1 << ARC_N_IRQ_START_LEVEL) - 1)
|
|
#else
|
|
and \reg1, \reg1, 0xffff
|
|
#endif
|
|
ffs \reg2, \reg1
|
|
fls \reg1, \reg1
|
|
cmp \reg1, \reg2
|
|
.endm
|
|
|
|
|
|
/* macro to get id of current cpu
|
|
* the result will be in reg (a reg)
|
|
*/
|
|
.macro _get_cpu_id reg
|
|
lr \reg, [_ARC_V2_IDENTITY]
|
|
xbfu \reg, \reg, 0xe8
|
|
.endm
|
|
|
|
/* macro to get the interrupt stack of current cpu
|
|
* the result will be in irq_sp (a reg)
|
|
*/
|
|
.macro _get_curr_cpu_irq_stack irq_sp
|
|
#ifdef CONFIG_SMP
|
|
_get_cpu_id \irq_sp
|
|
ld.as \irq_sp, [@_curr_cpu, \irq_sp]
|
|
ld \irq_sp, [\irq_sp, ___cpu_t_irq_stack_OFFSET]
|
|
#else
|
|
mov \irq_sp, _kernel
|
|
ld \irq_sp, [\irq_sp, _kernel_offset_to_irq_stack]
|
|
#endif
|
|
.endm
|
|
|
|
/* macro to push aux reg through reg */
|
|
.macro PUSHAX reg aux
|
|
lr \reg, [\aux]
|
|
st.a \reg, [sp, -4]
|
|
.endm
|
|
|
|
/* macro to pop aux reg through reg */
|
|
.macro POPAX reg aux
|
|
ld.ab \reg, [sp, 4]
|
|
sr \reg, [\aux]
|
|
.endm
|
|
|
|
|
|
/* macro to store old thread call regs */
|
|
.macro _store_old_thread_callee_regs
|
|
|
|
_save_callee_saved_regs
|
|
#ifdef CONFIG_SMP
|
|
/* save old thread into switch handle which is required by
|
|
* wait_for_switch
|
|
*/
|
|
st r2, [r2, ___thread_t_switch_handle_OFFSET]
|
|
#endif
|
|
.endm
|
|
|
|
/* macro to store old thread call regs in interrupt*/
|
|
.macro _irq_store_old_thread_callee_regs
|
|
#if defined(CONFIG_USERSPACE)
|
|
/*
|
|
* when USERSPACE is enabled, according to ARCv2 ISA, SP will be switched
|
|
* if interrupt comes out in user mode, and will be recorded in bit 31
|
|
* (U bit) of IRQ_ACT. when interrupt exits, SP will be switched back
|
|
* according to U bit.
|
|
*
|
|
* need to remember the user/kernel status of interrupted thread, will be
|
|
* restored when thread switched back
|
|
*
|
|
*/
|
|
lr r1, [_ARC_V2_AUX_IRQ_ACT]
|
|
and r3, r1, 0x80000000
|
|
push_s r3
|
|
|
|
bclr r1, r1, 31
|
|
sr r1, [_ARC_V2_AUX_IRQ_ACT]
|
|
#endif
|
|
_store_old_thread_callee_regs
|
|
.endm
|
|
|
|
/* macro to load new thread callee regs */
|
|
.macro _load_new_thread_callee_regs
|
|
#ifdef CONFIG_ARC_STACK_CHECKING
|
|
_load_stack_check_regs
|
|
#endif
|
|
/*
|
|
* _load_callee_saved_regs expects incoming thread in r2.
|
|
* _load_callee_saved_regs restores the stack pointer.
|
|
*/
|
|
_load_callee_saved_regs
|
|
|
|
#if defined(CONFIG_MPU_STACK_GUARD) || defined(CONFIG_USERSPACE)
|
|
push_s r2
|
|
bl configure_mpu_thread
|
|
pop_s r2
|
|
#endif
|
|
|
|
ld r3, [r2, _thread_offset_to_relinquish_cause]
|
|
.endm
|
|
|
|
|
|
/* when switch to thread caused by coop, some status regs need to set */
|
|
.macro _set_misc_regs_irq_switch_from_coop
|
|
#ifdef CONFIG_ARC_SECURE_FIRMWARE
|
|
/* must return to secure mode, so set IRM bit to 1 */
|
|
lr r0, [_ARC_V2_SEC_STAT]
|
|
bset r0, r0, _ARC_V2_SEC_STAT_IRM_BIT
|
|
sflag r0
|
|
#endif
|
|
.endm
|
|
|
|
/* when switch to thread caused by irq, some status regs need to set */
|
|
.macro _set_misc_regs_irq_switch_from_irq
|
|
#if defined(CONFIG_USERSPACE)
|
|
/*
|
|
* need to recover the user/kernel status of interrupted thread
|
|
*/
|
|
pop_s r3
|
|
lr r2, [_ARC_V2_AUX_IRQ_ACT]
|
|
or r2, r2, r3
|
|
sr r2, [_ARC_V2_AUX_IRQ_ACT]
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARC_SECURE_FIRMWARE
|
|
/* here need to recover SEC_STAT.IRM bit */
|
|
pop_s r3
|
|
sflag r3
|
|
#endif
|
|
.endm
|
|
|
|
/* macro to get next switch handle in assembly */
|
|
.macro _get_next_switch_handle
|
|
push_s r2
|
|
mov r0, sp
|
|
bl z_arch_get_next_switch_handle
|
|
pop_s r2
|
|
.endm
|
|
|
|
/* macro to disable stack checking in assembly, need a GPR
|
|
* to do this
|
|
*/
|
|
.macro _disable_stack_checking reg
|
|
#ifdef CONFIG_ARC_STACK_CHECKING
|
|
#ifdef CONFIG_ARC_SECURE_FIRMWARE
|
|
lr \reg, [_ARC_V2_SEC_STAT]
|
|
bclr \reg, \reg, _ARC_V2_SEC_STAT_SSC_BIT
|
|
sflag \reg
|
|
|
|
#else
|
|
lr \reg, [_ARC_V2_STATUS32]
|
|
bclr \reg, \reg, _ARC_V2_STATUS32_SC_BIT
|
|
kflag \reg
|
|
#endif
|
|
#endif
|
|
.endm
|
|
|
|
/* macro to enable stack checking in assembly, need a GPR
|
|
* to do this
|
|
*/
|
|
.macro _enable_stack_checking reg
|
|
#ifdef CONFIG_ARC_STACK_CHECKING
|
|
#ifdef CONFIG_ARC_SECURE_FIRMWARE
|
|
lr \reg, [_ARC_V2_SEC_STAT]
|
|
bset \reg, \reg, _ARC_V2_SEC_STAT_SSC_BIT
|
|
sflag \reg
|
|
#else
|
|
lr \reg, [_ARC_V2_STATUS32]
|
|
bset \reg, \reg, _ARC_V2_STATUS32_SC_BIT
|
|
kflag \reg
|
|
#endif
|
|
#endif
|
|
.endm
|
|
|
|
#endif /* _ASMLANGUAGE */
|
|
|
|
#endif /* ZEPHYR_ARCH_ARC_INCLUDE_SWAP_MACROS_H_ */
|