mirror of https://github.com/ARMmbed/mbed-os.git
Modify to support GCC. (CMSIS-RTOS RTX for Cortex-A9)
parent
aba8a90630
commit
c46f5b894b
|
|
@ -258,40 +258,6 @@ void _main_init (void) {
|
||||||
#pragma pop
|
#pragma pop
|
||||||
#endif
|
#endif
|
||||||
#else
|
#else
|
||||||
#if 0
|
|
||||||
//#ifdef __MBED_CMSIS_RTOS_CA9
|
|
||||||
__asm void __rt_entry (void) {
|
|
||||||
|
|
||||||
IMPORT __user_setup_stackheap
|
|
||||||
IMPORT __rt_lib_init
|
|
||||||
IMPORT os_thread_def_main
|
|
||||||
IMPORT osKernelInitialize
|
|
||||||
IMPORT osKernelStart
|
|
||||||
IMPORT osThreadCreate
|
|
||||||
IMPORT InterruptHandlerRegister
|
|
||||||
IMPORT PendSV_Handler
|
|
||||||
IMPORT OS_Tick_Handler
|
|
||||||
IMPORT exit
|
|
||||||
|
|
||||||
BL __user_setup_stackheap
|
|
||||||
MOV R1,R2
|
|
||||||
BL __rt_lib_init
|
|
||||||
BL osKernelInitialize
|
|
||||||
LDR R0,=os_thread_def_main
|
|
||||||
MOVS R1,#0
|
|
||||||
BL osThreadCreate
|
|
||||||
BL osKernelStart
|
|
||||||
MOVS R0,#0
|
|
||||||
LDR R1,=PendSV_Handler
|
|
||||||
BL InterruptHandlerRegister
|
|
||||||
MOVS R0,#134
|
|
||||||
LDR R1,=OS_Tick_Handler
|
|
||||||
BL InterruptHandlerRegister
|
|
||||||
BL exit
|
|
||||||
|
|
||||||
ALIGN
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
__asm void __rt_entry (void) {
|
__asm void __rt_entry (void) {
|
||||||
|
|
||||||
IMPORT __user_setup_stackheap
|
IMPORT __user_setup_stackheap
|
||||||
|
|
@ -315,7 +281,6 @@ __asm void __rt_entry (void) {
|
||||||
ALIGN
|
ALIGN
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
#endif
|
|
||||||
|
|
||||||
#elif defined (__GNUC__)
|
#elif defined (__GNUC__)
|
||||||
|
|
||||||
|
|
@ -375,7 +340,7 @@ __attribute ((noreturn)) void __cs3_start_c (void){
|
||||||
__attribute__((naked)) void software_init_hook (void) {
|
__attribute__((naked)) void software_init_hook (void) {
|
||||||
__asm (
|
__asm (
|
||||||
".syntax unified\n"
|
".syntax unified\n"
|
||||||
".thumb\n"
|
".arm\n"
|
||||||
"movs r0,#0\n"
|
"movs r0,#0\n"
|
||||||
"movs r1,#0\n"
|
"movs r1,#0\n"
|
||||||
"mov r4,r0\n"
|
"mov r4,r0\n"
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,474 @@
|
||||||
|
/*----------------------------------------------------------------------------
|
||||||
|
* RL-ARM - RTX
|
||||||
|
*----------------------------------------------------------------------------
|
||||||
|
* Name: HAL_CA9.c
|
||||||
|
* Purpose: Hardware Abstraction Layer for Cortex-A9
|
||||||
|
* Rev.: 3 Sept 2013
|
||||||
|
*----------------------------------------------------------------------------
|
||||||
|
*
|
||||||
|
* Copyright (c) 2012 - 2013 ARM Limited
|
||||||
|
* All rights reserved.
|
||||||
|
* Redistribution and use in source and binary forms, with or without
|
||||||
|
* modification, are permitted provided that the following conditions are met:
|
||||||
|
* - Redistributions of source code must retain the above copyright
|
||||||
|
* notice, this list of conditions and the following disclaimer.
|
||||||
|
* - Redistributions in binary form must reproduce the above copyright
|
||||||
|
* notice, this list of conditions and the following disclaimer in the
|
||||||
|
* documentation and/or other materials provided with the distribution.
|
||||||
|
* - Neither the name of ARM nor the names of its contributors may be used
|
||||||
|
* to endorse or promote products derived from this software without
|
||||||
|
* specific prior written permission.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
||||||
|
* ARE DISCLAIMED. IN NO EVENT SHALL COPYRIGHT HOLDERS AND CONTRIBUTORS BE
|
||||||
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
||||||
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
||||||
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
||||||
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
||||||
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
||||||
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
||||||
|
* POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*---------------------------------------------------------------------------*/
|
||||||
|
|
||||||
|
.global rt_set_PSP
|
||||||
|
.global rt_get_PSP
|
||||||
|
.global _alloc_box
|
||||||
|
.global _free_box
|
||||||
|
.global PendSV_Handler
|
||||||
|
.global OS_Tick_Handler
|
||||||
|
|
||||||
|
.EQU CPSR_T_BIT, 0x20
|
||||||
|
.EQU CPSR_I_BIT, 0x80
|
||||||
|
.EQU CPSR_F_BIT, 0x40
|
||||||
|
|
||||||
|
.EQU MODE_USR, 0x10
|
||||||
|
.EQU MODE_FIQ, 0x11
|
||||||
|
.EQU MODE_IRQ, 0x12
|
||||||
|
.EQU MODE_SVC, 0x13
|
||||||
|
.EQU MODE_ABT, 0x17
|
||||||
|
.EQU MODE_UND, 0x1B
|
||||||
|
.EQU MODE_SYS, 0x1F
|
||||||
|
|
||||||
|
.EQU TCB_TID, 3 /* 'task id' offset */
|
||||||
|
.EQU TCB_STACKF, 32 /* 'stack_frame' offset */
|
||||||
|
.EQU TCB_TSTACK, 36 /* 'tsk_stack' offset */
|
||||||
|
|
||||||
|
.extern rt_alloc_box
|
||||||
|
.extern os_tsk
|
||||||
|
.extern GICInterface_BASE
|
||||||
|
.extern rt_pop_req
|
||||||
|
.extern os_tick_irqack
|
||||||
|
.extern rt_systick
|
||||||
|
|
||||||
|
/*----------------------------------------------------------------------------
|
||||||
|
* Functions
|
||||||
|
*---------------------------------------------------------------------------*/
|
||||||
|
.text
|
||||||
|
@ For A-class, set USR/SYS stack
|
||||||
|
@ __asm void rt_set_PSP (U32 stack) {
|
||||||
|
rt_set_PSP:
|
||||||
|
.arm
|
||||||
|
|
||||||
|
MRS R1, CPSR
|
||||||
|
CPS #MODE_SYS @no effect in USR mode
|
||||||
|
ISB
|
||||||
|
MOV SP, R0
|
||||||
|
MSR CPSR_c, R1 @no effect in USR mode
|
||||||
|
ISB
|
||||||
|
BX LR
|
||||||
|
|
||||||
|
@ }
|
||||||
|
|
||||||
|
@ For A-class, get USR/SYS stack
|
||||||
|
@ __asm U32 rt_get_PSP (void) {
|
||||||
|
rt_get_PSP:
|
||||||
|
.arm
|
||||||
|
|
||||||
|
MRS R1, CPSR
|
||||||
|
CPS #MODE_SYS @no effect in USR mode
|
||||||
|
ISB
|
||||||
|
MOV R0, SP
|
||||||
|
MSR CPSR_c, R1 @no effect in USR mode
|
||||||
|
ISB
|
||||||
|
BX LR
|
||||||
|
|
||||||
|
@ }
|
||||||
|
|
||||||
|
/*--------------------------- _alloc_box ------------------------------------*/
|
||||||
|
@ __asm void *_alloc_box (void *box_mem) {
|
||||||
|
_alloc_box:
|
||||||
|
/* Function wrapper for Unprivileged/Privileged mode. */
|
||||||
|
.arm
|
||||||
|
|
||||||
|
LDR R12,=rt_alloc_box @ __cpp(rt_alloc_box)
|
||||||
|
MRS R2, CPSR
|
||||||
|
LSLS R2, #28
|
||||||
|
BXNE R12
|
||||||
|
SVC 0
|
||||||
|
BX LR
|
||||||
|
@ }
|
||||||
|
|
||||||
|
|
||||||
|
/*--------------------------- _free_box -------------------------------------*/
|
||||||
|
@ __asm int _free_box (void *box_mem, void *box) {
|
||||||
|
_free_box:
|
||||||
|
/* Function wrapper for Unprivileged/Privileged mode. */
|
||||||
|
.arm
|
||||||
|
|
||||||
|
LDR R12,=rt_free_box @ __cpp(rt_free_box)
|
||||||
|
MRS R2, CPSR
|
||||||
|
LSLS R2, #28
|
||||||
|
BXNE R12
|
||||||
|
SVC 0
|
||||||
|
BX LR
|
||||||
|
|
||||||
|
@ }
|
||||||
|
|
||||||
|
/*-------------------------- SVC_Handler -----------------------------------*/
|
||||||
|
|
||||||
|
@ #pragma push
|
||||||
|
@ #pragma arm
|
||||||
|
@ __asm void SVC_Handler (void) {
|
||||||
|
|
||||||
|
.type SVC_Handler, %function
|
||||||
|
.global SVC_Handler
|
||||||
|
SVC_Handler:
|
||||||
|
@ PRESERVE8
|
||||||
|
.arm
|
||||||
|
.extern rt_tsk_lock
|
||||||
|
.extern rt_tsk_unlock
|
||||||
|
.extern SVC_Count
|
||||||
|
.extern SVC_Table
|
||||||
|
.extern rt_stk_check
|
||||||
|
.extern FPUEnable
|
||||||
|
|
||||||
|
.EQU Mode_SVC, 0x13
|
||||||
|
|
||||||
|
SRSDB SP!, #Mode_SVC @ Push LR_SVC and SPRS_SVC onto SVC mode stack
|
||||||
|
PUSH {R4} @ Push R4 so we can use it as a temp
|
||||||
|
|
||||||
|
|
||||||
|
MRS R4,SPSR @ Get SPSR
|
||||||
|
TST R4,#CPSR_T_BIT @ Check Thumb Bit
|
||||||
|
LDRNEH R4,[LR,#-2] @ Thumb: Load Halfword
|
||||||
|
BICNE R4,R4,#0xFF00 @ Extract SVC Number
|
||||||
|
LDREQ R4,[LR,#-4] @ ARM: Load Word
|
||||||
|
BICEQ R4,R4,#0xFF000000 @ Extract SVC Number
|
||||||
|
|
||||||
|
/* Lock out systick and re-enable interrupts */
|
||||||
|
PUSH {R0-R3,R12,LR}
|
||||||
|
|
||||||
|
AND R12, SP, #4 @ Ensure stack is 8-byte aligned
|
||||||
|
SUB SP, SP, R12 @ Adjust stack
|
||||||
|
PUSH {R12, LR} @ Store stack adjustment and dummy LR to SVC stack
|
||||||
|
|
||||||
|
BLX rt_tsk_lock
|
||||||
|
CPSIE i
|
||||||
|
|
||||||
|
POP {R12, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R12 @ Unadjust stack
|
||||||
|
|
||||||
|
POP {R0-R3,R12,LR}
|
||||||
|
|
||||||
|
CMP R4,#0
|
||||||
|
BNE SVC_User
|
||||||
|
|
||||||
|
MRS R4,SPSR
|
||||||
|
PUSH {R4} @ Push R4 so we can use it as a temp
|
||||||
|
AND R4, SP, #4 @ Ensure stack is 8-byte aligned
|
||||||
|
SUB SP, SP, R4 @ Adjust stack
|
||||||
|
PUSH {R4, LR} @ Store stack adjustment and dummy LR
|
||||||
|
BLX R12
|
||||||
|
POP {R4, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R4 @ Unadjust stack
|
||||||
|
POP {R4} @ Restore R4
|
||||||
|
MSR SPSR_cxsf,R4
|
||||||
|
|
||||||
|
/* Here we will be in SVC mode (even if coming in from PendSV_Handler or OS_Tick_Handler) */
|
||||||
|
Sys_Switch:
|
||||||
|
LDR LR,=os_tsk @ __cpp(&os_tsk)
|
||||||
|
LDM LR,{R4,LR} @ os_tsk.run, os_tsk.new
|
||||||
|
CMP R4,LR
|
||||||
|
BNE switching
|
||||||
|
|
||||||
|
PUSH {R0-R3,R12,LR}
|
||||||
|
|
||||||
|
AND R12, SP, #4 @ Ensure stack is 8-byte aligned
|
||||||
|
SUB SP, SP, R12 @ Adjust stack
|
||||||
|
PUSH {R12, LR} @ Store stack adjustment and dummy LR to SVC stack
|
||||||
|
|
||||||
|
CPSID i
|
||||||
|
BLX rt_tsk_unlock
|
||||||
|
|
||||||
|
POP {R12, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R12 @ Unadjust stack
|
||||||
|
|
||||||
|
POP {R0-R3,R12,LR}
|
||||||
|
POP {R4}
|
||||||
|
RFEFD SP! @ Return from exception, no task switch
|
||||||
|
|
||||||
|
switching:
|
||||||
|
CLREX
|
||||||
|
CMP R4,#0
|
||||||
|
ADDEQ SP,SP,#12 @ Original R4, LR & SPSR do not need to be popped when we are paging in a different task
|
||||||
|
BEQ SVC_Next @ Runtask deleted?
|
||||||
|
|
||||||
|
|
||||||
|
PUSH {R8-R11} @ R4 and LR already stacked
|
||||||
|
MOV R10,R4 @ Preserve os_tsk.run
|
||||||
|
MOV R11,LR @ Preserve os_tsk.new
|
||||||
|
|
||||||
|
ADD R8,SP,#16 @ Unstack R4,LR
|
||||||
|
LDMIA R8,{R4,LR}
|
||||||
|
|
||||||
|
SUB SP,SP,#4 @ Make space on the stack for the next instn
|
||||||
|
STMIA SP,{SP}^ @ Put User SP onto stack
|
||||||
|
POP {R8} @ Pop User SP into R8
|
||||||
|
|
||||||
|
MRS R9,SPSR
|
||||||
|
STMDB R8!,{R9} @ User CPSR
|
||||||
|
STMDB R8!,{LR} @ User PC
|
||||||
|
STMDB R8,{LR}^ @ User LR
|
||||||
|
SUB R8,R8,#4 @ No writeback for store of User LR
|
||||||
|
STMDB R8!,{R0-R3,R12} @ User R0-R3,R12
|
||||||
|
MOV R3,R10 @ os_tsk.run
|
||||||
|
MOV LR,R11 @ os_tsk.new
|
||||||
|
POP {R9-R12}
|
||||||
|
ADD SP,SP,#12 @ Fix up SP for unstack of R4, LR & SPSR
|
||||||
|
STMDB R8!,{R4-R7,R9-R12} @ User R4-R11
|
||||||
|
|
||||||
|
@ If applicable, stack VFP state
|
||||||
|
MRC p15,0,R1,c1,c0,2 @ VFP/NEON access enabled? (CPACR)
|
||||||
|
AND R2,R1,#0x00F00000
|
||||||
|
CMP R2,#0x00F00000
|
||||||
|
BNE no_outgoing_vfp
|
||||||
|
VMRS R2,FPSCR
|
||||||
|
STMDB R8!,{R2,R4} @ Push FPSCR, maintain 8-byte alignment
|
||||||
|
VSTMDB R8!,{S0-S31}
|
||||||
|
LDRB R2,[R3,#TCB_STACKF] @ Record in TCB that VFP state is stacked
|
||||||
|
ORR R2,#2
|
||||||
|
STRB R2,[R3,#TCB_STACKF]
|
||||||
|
|
||||||
|
no_outgoing_vfp:
|
||||||
|
STR R8,[R3,#TCB_TSTACK]
|
||||||
|
MOV R4,LR
|
||||||
|
|
||||||
|
PUSH {R4} @ Push R4 so we can use it as a temp
|
||||||
|
AND R4, SP, #4 @ Ensure stack is 8-byte aligned
|
||||||
|
SUB SP, SP, R4 @ Adjust stack
|
||||||
|
PUSH {R4, LR} @ Store stack adjustment and dummy LR to SVC stack
|
||||||
|
|
||||||
|
BLX rt_stk_check
|
||||||
|
|
||||||
|
POP {R4, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R4 @ Unadjust stack
|
||||||
|
POP {R4} @ Restore R4
|
||||||
|
|
||||||
|
MOV LR,R4
|
||||||
|
|
||||||
|
SVC_Next: @ R4 == os_tsk.run, LR == os_tsk.new, R0-R3, R5-R12 corruptible
|
||||||
|
LDR R1,=os_tsk @ __cpp(&os_tsk), os_tsk.run = os_tsk.new
|
||||||
|
STR LR,[R1]
|
||||||
|
LDRB R1,[LR,#TCB_TID] @ os_tsk.run->task_id
|
||||||
|
LSL R1,#8 @ Store PROCID
|
||||||
|
MCR p15,0,R1,c13,c0,1 @ Write CONTEXTIDR
|
||||||
|
|
||||||
|
LDR R0,[LR,#TCB_TSTACK] @ os_tsk.run->tsk_stack
|
||||||
|
|
||||||
|
@ Does incoming task have VFP state in stack?
|
||||||
|
LDRB R3,[LR,#TCB_STACKF]
|
||||||
|
TST R3,#0x2
|
||||||
|
MRC p15,0,R1,c1,c0,2 @ Read CPACR
|
||||||
|
ANDEQ R1,R1,#0xFF0FFFFF @ Disable VFP access if incoming task does not have stacked VFP state
|
||||||
|
ORRNE R1,R1,#0x00F00000 @ Enable VFP access if incoming task does have stacked VFP state
|
||||||
|
MCR p15,0,R1,c1,c0,2 @ Write CPACR
|
||||||
|
BEQ no_incoming_vfp
|
||||||
|
ISB @ We only need the sync if we enabled, otherwise we will context switch before next VFP instruction anyway
|
||||||
|
VLDMIA R0!,{S0-S31}
|
||||||
|
LDR R2,[R0]
|
||||||
|
VMSR FPSCR,R2
|
||||||
|
ADD R0,R0,#8
|
||||||
|
|
||||||
|
no_incoming_vfp:
|
||||||
|
LDR R1,[R0,#60] @ Restore User CPSR
|
||||||
|
MSR SPSR_cxsf,R1
|
||||||
|
LDMIA R0!,{R4-R11} @ Restore User R4-R11
|
||||||
|
ADD R0,R0,#4 @ Restore User R1-R3,R12
|
||||||
|
LDMIA R0!,{R1-R3,R12}
|
||||||
|
LDMIA R0,{LR}^ @ Restore User LR
|
||||||
|
ADD R0,R0,#4 @ No writeback for load to user LR
|
||||||
|
LDMIA R0!,{LR} @ Restore User PC
|
||||||
|
ADD R0,R0,#4 @ Correct User SP for unstacked user CPSR
|
||||||
|
|
||||||
|
PUSH {R0} @ Push R0 onto stack
|
||||||
|
LDMIA SP,{SP}^ @ Get R0 off stack into User SP
|
||||||
|
ADD SP,SP,#4 @ Put SP back
|
||||||
|
|
||||||
|
LDR R0,[R0,#-32] @ Restore R0
|
||||||
|
|
||||||
|
PUSH {R0-R3,R12,LR}
|
||||||
|
|
||||||
|
AND R12, SP, #4 @ Ensure stack is 8-byte aligned
|
||||||
|
SUB SP, SP, R12 @ Adjust stack
|
||||||
|
PUSH {R12, LR} @ Store stack adjustment and dummy LR to SVC stack
|
||||||
|
|
||||||
|
CPSID i
|
||||||
|
BLX rt_tsk_unlock
|
||||||
|
|
||||||
|
POP {R12, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R12 @ Unadjust stack
|
||||||
|
|
||||||
|
POP {R0-R3,R12,LR}
|
||||||
|
|
||||||
|
MOVS PC,LR @ Return from exception
|
||||||
|
|
||||||
|
|
||||||
|
/*------------------- User SVC -------------------------------*/
|
||||||
|
|
||||||
|
SVC_User:
|
||||||
|
LDR R12,=SVC_Count
|
||||||
|
LDR R12,[R12]
|
||||||
|
CMP R4,R12 @ Check for overflow
|
||||||
|
BHI SVC_Done
|
||||||
|
|
||||||
|
LDR R12,=SVC_Table-4
|
||||||
|
LDR R12,[R12,R4,LSL #2] @ Load SVC Function Address
|
||||||
|
MRS R4,SPSR @ Save SPSR
|
||||||
|
PUSH {R4} @ Push R4 so we can use it as a temp
|
||||||
|
AND R4, SP, #4 @ Ensure stack is 8-byte aligned
|
||||||
|
SUB SP, SP, R4 @ Adjust stack
|
||||||
|
PUSH {R4, LR} @ Store stack adjustment and dummy LR
|
||||||
|
BLX R12 @ Call SVC Function
|
||||||
|
POP {R4, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R4 @ Unadjust stack
|
||||||
|
POP {R4} @ Restore R4
|
||||||
|
MSR SPSR_cxsf,R4 @ Restore SPSR
|
||||||
|
|
||||||
|
SVC_Done:
|
||||||
|
PUSH {R0-R3,R12,LR}
|
||||||
|
|
||||||
|
PUSH {R4} @ Push R4 so we can use it as a temp
|
||||||
|
AND R4, SP, #4 @ Ensure stack is 8-byte aligned
|
||||||
|
SUB SP, SP, R4 @ Adjust stack
|
||||||
|
PUSH {R4, LR} @ Store stack adjustment and dummy LR
|
||||||
|
|
||||||
|
CPSID i
|
||||||
|
BLX rt_tsk_unlock
|
||||||
|
|
||||||
|
POP {R4, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R4 @ Unadjust stack
|
||||||
|
POP {R4} @ Restore R4
|
||||||
|
|
||||||
|
POP {R0-R3,R12,LR}
|
||||||
|
POP {R4}
|
||||||
|
RFEFD SP! @ Return from exception
|
||||||
|
|
||||||
|
@ }
|
||||||
|
|
||||||
|
@ #pragma pop
|
||||||
|
|
||||||
|
|
||||||
|
@ #pragma push
|
||||||
|
@ #pragma arm
|
||||||
|
@ __asm void PendSV_Handler (U32 IRQn) {
|
||||||
|
PendSV_Handler:
|
||||||
|
.arm
|
||||||
|
|
||||||
|
.extern rt_tsk_lock
|
||||||
|
.extern IRQNestLevel
|
||||||
|
|
||||||
|
ADD SP,SP,#8 @ fix up stack pointer (R0 has been pushed and will never be popped, R1 was pushed for stack alignment)
|
||||||
|
|
||||||
|
@ Disable systick interrupts, then write EOIR. We want interrupts disabled before we enter the context switcher.
|
||||||
|
PUSH {R0, R1}
|
||||||
|
BLX rt_tsk_lock
|
||||||
|
POP {R0, R1}
|
||||||
|
LDR R1, =GICInterface_BASE @ __cpp(&GICInterface_BASE)
|
||||||
|
LDR R1, [R1, #0]
|
||||||
|
STR R0, [R1, #0x10]
|
||||||
|
|
||||||
|
LDR R0, =IRQNestLevel @ Get address of nesting counter
|
||||||
|
LDR R1, [R0]
|
||||||
|
SUB R1, R1, #1 @ Decrement nesting counter
|
||||||
|
STR R1, [R0]
|
||||||
|
|
||||||
|
BLX rt_pop_req @ __cpp(rt_pop_req)
|
||||||
|
|
||||||
|
POP {R1, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R1 @ Unadjust stack
|
||||||
|
|
||||||
|
LDR R0,[SP,#24]
|
||||||
|
MSR SPSR_cxsf,R0
|
||||||
|
POP {R0-R3,R12} @ Leave SPSR & LR on the stack
|
||||||
|
PUSH {R4}
|
||||||
|
B Sys_Switch
|
||||||
|
@ }
|
||||||
|
@ #pragma pop
|
||||||
|
|
||||||
|
@ #pragma push
|
||||||
|
@ #pragma arm
|
||||||
|
@ __asm void OS_Tick_Handler (U32 IRQn) {
|
||||||
|
OS_Tick_Handler:
|
||||||
|
.arm
|
||||||
|
|
||||||
|
ADD SP,SP,#8 @ fix up stack pointer (R0 has been pushed and will never be popped, R1 was pushed for stack alignment)
|
||||||
|
|
||||||
|
PUSH {R0, R1}
|
||||||
|
BLX rt_tsk_lock
|
||||||
|
POP {R0, R1}
|
||||||
|
LDR R1, =GICInterface_BASE @ __cpp(&GICInterface_BASE)
|
||||||
|
LDR R1, [R1, #0]
|
||||||
|
STR R0, [R1, #0x10]
|
||||||
|
|
||||||
|
LDR R0, =IRQNestLevel @ Get address of nesting counter
|
||||||
|
LDR R1, [R0]
|
||||||
|
SUB R1, R1, #1 @ Decrement nesting counter
|
||||||
|
STR R1, [R0]
|
||||||
|
|
||||||
|
BLX os_tick_irqack @ __cpp(os_tick_irqack)
|
||||||
|
BLX rt_systick @ __cpp(rt_systick)
|
||||||
|
|
||||||
|
POP {R1, LR} @ Get stack adjustment & discard dummy LR
|
||||||
|
ADD SP, SP, R1 @ Unadjust stack
|
||||||
|
|
||||||
|
LDR R0,[SP,#24]
|
||||||
|
MSR SPSR_cxsf,R0
|
||||||
|
POP {R0-R3,R12} @ Leave SPSR & LR on the stack
|
||||||
|
PUSH {R4}
|
||||||
|
B Sys_Switch
|
||||||
|
@ }
|
||||||
|
@ #pragma pop
|
||||||
|
|
||||||
|
.global __set_PSP
|
||||||
|
@ __STATIC_ASM void __set_PSP(uint32_t topOfProcStack)
|
||||||
|
@ {
|
||||||
|
__set_PSP:
|
||||||
|
@ PRESERVE8
|
||||||
|
.arm
|
||||||
|
|
||||||
|
BIC R0, R0, #7 @ensure stack is 8-byte aligned
|
||||||
|
MRS R1, CPSR
|
||||||
|
CPS #MODE_SYS @no effect in USR mode
|
||||||
|
MOV SP, R0
|
||||||
|
MSR CPSR_c, R1 @no effect in USR mode
|
||||||
|
ISB
|
||||||
|
BX LR
|
||||||
|
|
||||||
|
@ }
|
||||||
|
|
||||||
|
.global __set_CPS_USR
|
||||||
|
@ __STATIC_ASM void __set_CPS_USR(void)
|
||||||
|
@ {
|
||||||
|
__set_CPS_USR:
|
||||||
|
.arm
|
||||||
|
|
||||||
|
CPS #MODE_USR
|
||||||
|
BX LR
|
||||||
|
@ }
|
||||||
|
|
||||||
|
.END
|
||||||
|
/*----------------------------------------------------------------------------
|
||||||
|
* end of file
|
||||||
|
*---------------------------------------------------------------------------*/
|
||||||
|
|
@ -0,0 +1,60 @@
|
||||||
|
/*----------------------------------------------------------------------------
|
||||||
|
* RL-ARM - RTX
|
||||||
|
*----------------------------------------------------------------------------
|
||||||
|
* Name: SVC_TABLE.S
|
||||||
|
* Purpose: Pre-defined SVC Table for Cortex-M
|
||||||
|
* Rev.: V4.70
|
||||||
|
*----------------------------------------------------------------------------
|
||||||
|
*
|
||||||
|
* Copyright (c) 1999-2009 KEIL, 2009-2013 ARM Germany GmbH
|
||||||
|
* All rights reserved.
|
||||||
|
* Redistribution and use in source and binary forms, with or without
|
||||||
|
* modification, are permitted provided that the following conditions are met:
|
||||||
|
* - Redistributions of source code must retain the above copyright
|
||||||
|
* notice, this list of conditions and the following disclaimer.
|
||||||
|
* - Redistributions in binary form must reproduce the above copyright
|
||||||
|
* notice, this list of conditions and the following disclaimer in the
|
||||||
|
* documentation and/or other materials provided with the distribution.
|
||||||
|
* - Neither the name of ARM nor the names of its contributors may be used
|
||||||
|
* to endorse or promote products derived from this software without
|
||||||
|
* specific prior written permission.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
||||||
|
* ARE DISCLAIMED. IN NO EVENT SHALL COPYRIGHT HOLDERS AND CONTRIBUTORS BE
|
||||||
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
||||||
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
||||||
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
||||||
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
||||||
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
||||||
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
||||||
|
* POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*---------------------------------------------------------------------------*/
|
||||||
|
|
||||||
|
|
||||||
|
.section SVC_TABLE @, CODE, READONLY
|
||||||
|
.align 5
|
||||||
|
|
||||||
|
.global SVC_Count
|
||||||
|
|
||||||
|
.EQU SVC_Cnt, (SVC_End-SVC_Table)/4
|
||||||
|
|
||||||
|
SVC_Count:
|
||||||
|
.word SVC_Cnt
|
||||||
|
|
||||||
|
@ Import user SVC functions here.
|
||||||
|
@ .extern __SVC_1
|
||||||
|
.global SVC_Table
|
||||||
|
SVC_Table:
|
||||||
|
@ Insert user SVC functions here. SVC 0 used by RTL Kernel.
|
||||||
|
@ .word __SVC_1 @ InitMemorySubsystem
|
||||||
|
|
||||||
|
@SVC_End
|
||||||
|
SVC_End:
|
||||||
|
|
||||||
|
.END
|
||||||
|
|
||||||
|
/*----------------------------------------------------------------------------
|
||||||
|
* end of file
|
||||||
|
*---------------------------------------------------------------------------*/
|
||||||
|
|
@ -145,11 +145,63 @@ typedef uint32_t __attribute__((vector_size(16))) ret128;
|
||||||
|
|
||||||
#define RET_pointer __r0
|
#define RET_pointer __r0
|
||||||
#define RET_int32_t __r0
|
#define RET_int32_t __r0
|
||||||
|
#define RET_uint32_t __r0
|
||||||
#define RET_osStatus __r0
|
#define RET_osStatus __r0
|
||||||
#define RET_osPriority __r0
|
#define RET_osPriority __r0
|
||||||
#define RET_osEvent {(osStatus)__r0, {(uint32_t)__r1}, {(void *)__r2}}
|
#define RET_osEvent {(osStatus)__r0, {(uint32_t)__r1}, {(void *)__r2}}
|
||||||
#define RET_osCallback {(void *)__r0, (void *)__r1}
|
#define RET_osCallback {(void *)__r0, (void *)__r1}
|
||||||
|
|
||||||
|
#if defined (__ARM_PCS_VFP)
|
||||||
|
|
||||||
|
#define osEvent_type void
|
||||||
|
#define osEvent_ret_status { __asm ("MOV r0, %0;" \
|
||||||
|
: /* no outputs */ \
|
||||||
|
: "r"(ret.status) \
|
||||||
|
: "r0" \
|
||||||
|
); \
|
||||||
|
}
|
||||||
|
#define osEvent_ret_value { __asm ("MOV r1, %0;" \
|
||||||
|
"MOV r0, %1;" \
|
||||||
|
: /* no outputs */ \
|
||||||
|
: "r"(ret.value.v), \
|
||||||
|
"r"(ret.status) \
|
||||||
|
: "r0", "r1" \
|
||||||
|
); \
|
||||||
|
}
|
||||||
|
#define osEvent_ret_msg { __asm ("MOV r2, %0;" \
|
||||||
|
"MOV r1, %1;" \
|
||||||
|
"MOV r0, %2;" \
|
||||||
|
: /* no outputs */ \
|
||||||
|
: "r"(ret.def.message_id), \
|
||||||
|
"r"(ret.value.v), \
|
||||||
|
"r"(ret.status) \
|
||||||
|
: "r0", "r1" , "r2" \
|
||||||
|
); \
|
||||||
|
}
|
||||||
|
|
||||||
|
#define osEvent_ret_mail { __asm ("MOV r2, %0;" \
|
||||||
|
"MOV r1, %1;" \
|
||||||
|
"MOV r0, %2;" \
|
||||||
|
: /* no outputs */ \
|
||||||
|
: "r"(ret.def.mail_id), \
|
||||||
|
"r"(ret.value.v), \
|
||||||
|
"r"(ret.status) \
|
||||||
|
: "r0", "r1" , "r2" \
|
||||||
|
); \
|
||||||
|
}
|
||||||
|
|
||||||
|
#define osCallback_type void
|
||||||
|
#define osCallback_ret { __asm ("MOV r1, %0;" \
|
||||||
|
"MOV r0, %1;" \
|
||||||
|
: /* no outputs */ \
|
||||||
|
: "r"(ret.arg), \
|
||||||
|
"r"(ret.fp) \
|
||||||
|
: "r0", "r1" \
|
||||||
|
); \
|
||||||
|
}
|
||||||
|
|
||||||
|
#else /* defined (__ARM_PCS_VFP) */
|
||||||
|
|
||||||
#define osEvent_type ret128
|
#define osEvent_type ret128
|
||||||
#define osEvent_ret_status (ret128){ret.status}
|
#define osEvent_ret_status (ret128){ret.status}
|
||||||
#define osEvent_ret_value (ret128){ret.status, ret.value.v}
|
#define osEvent_ret_value (ret128){ret.status, ret.value.v}
|
||||||
|
|
@ -159,6 +211,8 @@ typedef uint32_t __attribute__((vector_size(16))) ret128;
|
||||||
#define osCallback_type ret64
|
#define osCallback_type ret64
|
||||||
#define osCallback_ret (ret64) {(uint32_t)ret.fp, (uint32_t)ret.arg}
|
#define osCallback_ret (ret64) {(uint32_t)ret.fp, (uint32_t)ret.arg}
|
||||||
|
|
||||||
|
#endif /* defined (__ARM_PCS_VFP) */
|
||||||
|
|
||||||
#define SVC_ArgN(n) \
|
#define SVC_ArgN(n) \
|
||||||
register int __r##n __asm("r"#n);
|
register int __r##n __asm("r"#n);
|
||||||
|
|
||||||
|
|
@ -808,14 +862,24 @@ os_InRegs osEvent_type svcWait (uint32_t millisec) {
|
||||||
|
|
||||||
if (millisec == 0) {
|
if (millisec == 0) {
|
||||||
ret.status = osOK;
|
ret.status = osOK;
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_status;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_status;
|
return osEvent_ret_status;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
/* To Do: osEventSignal, osEventMessage, osEventMail */
|
/* To Do: osEventSignal, osEventMessage, osEventMail */
|
||||||
rt_dly_wait(rt_ms2tick(millisec));
|
rt_dly_wait(rt_ms2tick(millisec));
|
||||||
ret.status = osEventTimeout;
|
ret.status = osEventTimeout;
|
||||||
|
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_status;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_status;
|
return osEvent_ret_status;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
@ -1046,13 +1110,23 @@ os_InRegs osCallback_type svcTimerCall (osTimerId timer_id) {
|
||||||
if (pt == NULL) {
|
if (pt == NULL) {
|
||||||
ret.fp = NULL;
|
ret.fp = NULL;
|
||||||
ret.arg = NULL;
|
ret.arg = NULL;
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osCallback_ret;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osCallback_ret;
|
return osCallback_ret;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
ret.fp = (void *)pt->timer->ptimer;
|
ret.fp = (void *)pt->timer->ptimer;
|
||||||
ret.arg = pt->arg;
|
ret.arg = pt->arg;
|
||||||
|
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osCallback_ret;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osCallback_ret;
|
return osCallback_ret;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
static __INLINE osStatus isrMessagePut (osMessageQId queue_id, uint32_t info, uint32_t millisec);
|
static __INLINE osStatus isrMessagePut (osMessageQId queue_id, uint32_t info, uint32_t millisec);
|
||||||
|
|
@ -1195,7 +1269,12 @@ os_InRegs osEvent_type svcSignalWait (int32_t signals, uint32_t millisec) {
|
||||||
|
|
||||||
if (signals & (0xFFFFFFFF << osFeature_Signals)) {
|
if (signals & (0xFFFFFFFF << osFeature_Signals)) {
|
||||||
ret.status = osErrorValue;
|
ret.status = osErrorValue;
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_status;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_status;
|
return osEvent_ret_status;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
if (signals != 0) { // Wait for all specified signals
|
if (signals != 0) { // Wait for all specified signals
|
||||||
|
|
@ -1212,7 +1291,12 @@ os_InRegs osEvent_type svcSignalWait (int32_t signals, uint32_t millisec) {
|
||||||
ret.value.signals = 0;
|
ret.value.signals = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_value;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_value;
|
return osEvent_ret_value;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
@ -1694,24 +1778,44 @@ os_InRegs osEvent_type svcMessageGet (osMessageQId queue_id, uint32_t millisec)
|
||||||
|
|
||||||
if (queue_id == NULL) {
|
if (queue_id == NULL) {
|
||||||
ret.status = osErrorParameter;
|
ret.status = osErrorParameter;
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_status;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_status;
|
return osEvent_ret_status;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
if (((P_MCB)queue_id)->cb_type != MCB) {
|
if (((P_MCB)queue_id)->cb_type != MCB) {
|
||||||
ret.status = osErrorParameter;
|
ret.status = osErrorParameter;
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_status;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_status;
|
return osEvent_ret_status;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
res = rt_mbx_wait(queue_id, &ret.value.p, rt_ms2tick(millisec));
|
res = rt_mbx_wait(queue_id, &ret.value.p, rt_ms2tick(millisec));
|
||||||
|
|
||||||
if (res == OS_R_TMO) {
|
if (res == OS_R_TMO) {
|
||||||
ret.status = millisec ? osEventTimeout : osOK;
|
ret.status = millisec ? osEventTimeout : osOK;
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_value;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_value;
|
return osEvent_ret_value;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
ret.status = osEventMessage;
|
ret.status = osEventMessage;
|
||||||
|
|
||||||
|
#if defined (__GNUC__) && defined (__ARM_PCS_VFP)
|
||||||
|
osEvent_ret_value;
|
||||||
|
return;
|
||||||
|
#else
|
||||||
return osEvent_ret_value;
|
return osEvent_ret_value;
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -54,7 +54,7 @@
|
||||||
|
|
||||||
#if defined (__CC_ARM) /* ARM Compiler */
|
#if defined (__CC_ARM) /* ARM Compiler */
|
||||||
|
|
||||||
#if ((__TARGET_ARCH_7_M || __TARGET_ARCH_7E_M || __TARGET_ARCH_7_A) && !NO_EXCLUSIVE_ACCESS)
|
#if ((__TARGET_ARCH_7_M || __TARGET_ARCH_7E_M || __TARGET_ARCH_7_A) && !defined(NO_EXCLUSIVE_ACCESS))
|
||||||
#define __USE_EXCLUSIVE_ACCESS
|
#define __USE_EXCLUSIVE_ACCESS
|
||||||
#else
|
#else
|
||||||
#undef __USE_EXCLUSIVE_ACCESS
|
#undef __USE_EXCLUSIVE_ACCESS
|
||||||
|
|
@ -62,7 +62,16 @@
|
||||||
|
|
||||||
#elif defined (__GNUC__) /* GNU Compiler */
|
#elif defined (__GNUC__) /* GNU Compiler */
|
||||||
|
|
||||||
#error GNU Compiler support not implemented for Cortex-A
|
#undef __USE_EXCLUSIVE_ACCESS
|
||||||
|
|
||||||
|
#if defined (__VFP_FP__) && !defined(__SOFTFP__)
|
||||||
|
#define __TARGET_FPU_VFP 1
|
||||||
|
#else
|
||||||
|
#define __TARGET_FPU_VFP 0
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#define __inline inline
|
||||||
|
#define __weak __attribute__((weak))
|
||||||
|
|
||||||
#elif defined (__ICCARM__) /* IAR Compiler */
|
#elif defined (__ICCARM__) /* IAR Compiler */
|
||||||
|
|
||||||
|
|
@ -94,7 +103,6 @@ extern const U32 GICInterface_BASE;
|
||||||
priority = GICI_ICCPMR; \
|
priority = GICI_ICCPMR; \
|
||||||
GICI_ICCPMR = 0xff; \
|
GICI_ICCPMR = 0xff; \
|
||||||
GICI_ICCPMR = GICI_ICCPMR - 1; \
|
GICI_ICCPMR = GICI_ICCPMR - 1; \
|
||||||
while(GICI_ICCPMR > priority);\
|
|
||||||
__DSB();\
|
__DSB();\
|
||||||
if(!irq_dis) __enable_irq(); \
|
if(!irq_dis) __enable_irq(); \
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue