mirror of
				https://github.com/smaeul/u-boot.git
				synced 2025-10-31 12:08:19 +00:00 
			
		
		
		
	Today, we have cases where we wish to build all of U-Boot in Thumb2 mode for various reasons. We also have cases where we only build SPL in Thumb2 mode due to size constraints and wish to build the rest of the system in ARM mode. So in this migration we introduce a new symbol as well, SPL_SYS_THUMB_BUILD to control if we build everything or just SPL (or in theory, just U-Boot) in Thumb2 mode. Signed-off-by: Tom Rini <trini@konsulko.com> Acked-by: Siarhei Siamashka <siarhei.siamashka@gmail.com>
		
			
				
	
	
		
			85 lines
		
	
	
		
			2.0 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			85 lines
		
	
	
		
			2.0 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  *  arch/arm/include/asm/assembler.h
 | |
|  *
 | |
|  *  Copyright (C) 1996-2000 Russell King
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License version 2 as
 | |
|  * published by the Free Software Foundation.
 | |
|  *
 | |
|  *  This file contains arm architecture specific defines
 | |
|  *  for the different processors.
 | |
|  *
 | |
|  *  Do not include any C declarations in this file - it is included by
 | |
|  *  assembler source.
 | |
|  */
 | |
| 
 | |
| #include <config.h>
 | |
| #include <asm/unified.h>
 | |
| 
 | |
| /*
 | |
|  * Endian independent macros for shifting bytes within registers.
 | |
|  */
 | |
| #ifndef __ARMEB__
 | |
| #define lspull		lsr
 | |
| #define lspush		lsl
 | |
| #define get_byte_0	lsl #0
 | |
| #define get_byte_1	lsr #8
 | |
| #define get_byte_2	lsr #16
 | |
| #define get_byte_3	lsr #24
 | |
| #define put_byte_0	lsl #0
 | |
| #define put_byte_1	lsl #8
 | |
| #define put_byte_2	lsl #16
 | |
| #define put_byte_3	lsl #24
 | |
| #else
 | |
| #define lspull		lsl
 | |
| #define lspush		lsr
 | |
| #define get_byte_0	lsr #24
 | |
| #define get_byte_1	lsr #16
 | |
| #define get_byte_2	lsr #8
 | |
| #define get_byte_3      lsl #0
 | |
| #define put_byte_0	lsl #24
 | |
| #define put_byte_1	lsl #16
 | |
| #define put_byte_2	lsl #8
 | |
| #define put_byte_3      lsl #0
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * Data preload for architectures that support it
 | |
|  */
 | |
| #if defined(__ARM_ARCH_5E__) || defined(__ARM_ARCH_5TE__) || \
 | |
| 	defined(__ARM_ARCH_6__) || defined(__ARM_ARCH_6J__) || \
 | |
| 	defined(__ARM_ARCH_6T2__) || defined(__ARM_ARCH_6Z__) || \
 | |
| 	defined(__ARM_ARCH_6ZK__) || defined(__ARM_ARCH_7A__) || \
 | |
| 	defined(__ARM_ARCH_7R__)
 | |
| #define PLD(code...)	code
 | |
| #else
 | |
| #define PLD(code...)
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * We only support cores that support at least Thumb-1 and thus we use
 | |
|  * 'bx lr'
 | |
|  */
 | |
| 	.irp	c,,eq,ne,cs,cc,mi,pl,vs,vc,hi,ls,ge,lt,gt,le,hs,lo
 | |
| 	.macro	ret\c, reg
 | |
| 	.ifeqs	"\reg", "lr"
 | |
| 	bx\c	\reg
 | |
| 	.else
 | |
| 	mov\c	pc, \reg
 | |
| 	.endif
 | |
| 	.endm
 | |
| 	.endr
 | |
| 
 | |
| /*
 | |
|  * Cache aligned, used for optimized memcpy/memset
 | |
|  * In the kernel this is only enabled for Feroceon CPU's...
 | |
|  * We disable it especially for Thumb builds since those instructions
 | |
|  * are not made in a Thumb ready way...
 | |
|  */
 | |
| #if CONFIG_IS_ENABLED(SYS_THUMB_BUILD)
 | |
| #define CALGN(code...)
 | |
| #else
 | |
| #define CALGN(code...) code
 | |
| #endif
 |