blob: a20702e612b1b0a9fd31cc5a5eb141e1fa0894b1 [file] [log] [blame]
Aneesh V2c451f72011-06-16 23:30:47 +00001/*
2 * (C) Copyright 2010
3 * Texas Instruments, <www.ti.com>
4 * Aneesh V <aneesh@ti.com>
5 *
Wolfgang Denk1a459662013-07-08 09:37:19 +02006 * SPDX-License-Identifier: GPL-2.0+
Aneesh V2c451f72011-06-16 23:30:47 +00007 */
8#ifndef ARMV7_H
9#define ARMV7_H
Aneesh V2c451f72011-06-16 23:30:47 +000010
Aneesh Vad577c82011-07-21 09:10:04 -040011/* Cortex-A9 revisions */
12#define MIDR_CORTEX_A9_R0P1 0x410FC091
13#define MIDR_CORTEX_A9_R1P2 0x411FC092
14#define MIDR_CORTEX_A9_R1P3 0x411FC093
Aneesh V5ab12a92011-07-21 09:29:23 -040015#define MIDR_CORTEX_A9_R2P10 0x412FC09A
Aneesh Vad577c82011-07-21 09:10:04 -040016
Sricharan508a58f2011-11-15 09:49:55 -050017/* Cortex-A15 revisions */
18#define MIDR_CORTEX_A15_R0P0 0x410FC0F0
SRICHARAN Reed7c0f2013-02-12 01:33:41 +000019#define MIDR_CORTEX_A15_R2P2 0x412FC0F2
Sricharan508a58f2011-11-15 09:49:55 -050020
Andre Przywara16212b52013-09-19 18:06:41 +020021/* Cortex-A7 revisions */
22#define MIDR_CORTEX_A7_R0P0 0x410FC070
23
24#define MIDR_PRIMARY_PART_MASK 0xFF0FFFF0
25
26/* ID_PFR1 feature fields */
27#define CPUID_ARM_SEC_SHIFT 4
28#define CPUID_ARM_SEC_MASK (0xF << CPUID_ARM_SEC_SHIFT)
29#define CPUID_ARM_VIRT_SHIFT 12
30#define CPUID_ARM_VIRT_MASK (0xF << CPUID_ARM_VIRT_SHIFT)
31#define CPUID_ARM_GENTIMER_SHIFT 16
32#define CPUID_ARM_GENTIMER_MASK (0xF << CPUID_ARM_GENTIMER_SHIFT)
33
34/* valid bits in CBAR register / PERIPHBASE value */
35#define CBAR_MASK 0xFFFF8000
36
Aneesh V2c451f72011-06-16 23:30:47 +000037/* CCSIDR */
38#define CCSIDR_LINE_SIZE_OFFSET 0
39#define CCSIDR_LINE_SIZE_MASK 0x7
40#define CCSIDR_ASSOCIATIVITY_OFFSET 3
41#define CCSIDR_ASSOCIATIVITY_MASK (0x3FF << 3)
42#define CCSIDR_NUM_SETS_OFFSET 13
43#define CCSIDR_NUM_SETS_MASK (0x7FFF << 13)
44
45/*
46 * Values for InD field in CSSELR
47 * Selects the type of cache
48 */
49#define ARMV7_CSSELR_IND_DATA_UNIFIED 0
50#define ARMV7_CSSELR_IND_INSTRUCTION 1
51
52/* Values for Ctype fields in CLIDR */
53#define ARMV7_CLIDR_CTYPE_NO_CACHE 0
54#define ARMV7_CLIDR_CTYPE_INSTRUCTION_ONLY 1
55#define ARMV7_CLIDR_CTYPE_DATA_ONLY 2
56#define ARMV7_CLIDR_CTYPE_INSTRUCTION_DATA 3
57#define ARMV7_CLIDR_CTYPE_UNIFIED 4
58
Andre Przywarad75ba502013-09-19 18:06:39 +020059#ifndef __ASSEMBLY__
60#include <linux/types.h>
Tom Rini301c1282015-03-02 08:24:45 -050061#include <asm/io.h>
Andre Przywara1ea4fac2016-05-12 12:14:41 +010062#include <asm/barriers.h>
Valentine Barshak9ba379a2015-03-20 18:16:17 +030063
Akshay Saraswat0c08baf2015-02-20 13:27:13 +053064/*
65 * Workaround for ARM errata # 798870
66 * Set L2ACTLR[7] to reissue any memory transaction in the L2 that has been
67 * stalled for 1024 cycles to verify that its hazard condition still exists.
68 */
69static inline void v7_enable_l2_hazard_detect(void)
70{
71 uint32_t val;
72
73 /* L2ACTLR[7]: Enable hazard detect timeout */
74 asm volatile ("mrc p15, 1, %0, c15, c0, 0\n\t" : "=r"(val));
75 val |= (1 << 7);
76 asm volatile ("mcr p15, 1, %0, c15, c0, 0\n\t" : : "r"(val));
77}
78
Akshay Saraswata3895312015-02-20 13:27:14 +053079/*
80 * Workaround for ARM errata # 799270
81 * Ensure that the L2 logic has been used within the previous 256 cycles
82 * before modifying the ACTLR.SMP bit. This is required during boot before
83 * MMU has been enabled, or during a specified reset or power down sequence.
84 */
85static inline void v7_enable_smp(uint32_t address)
86{
87 uint32_t temp, val;
88
89 /* Read auxiliary control register */
90 asm volatile ("mrc p15, 0, %0, c1, c0, 1\n\t" : "=r"(val));
91
92 /* Enable SMP */
93 val |= (1 << 6);
94
95 /* Dummy read to assure L2 access */
96 temp = readl(address);
97 temp &= 0;
98 val |= temp;
99
100 /* Write auxiliary control register */
101 asm volatile ("mcr p15, 0, %0, c1, c0, 1\n\t" : : "r"(val));
102
103 CP15DSB;
104 CP15ISB;
105}
106
Akshay Saraswat0c08baf2015-02-20 13:27:13 +0530107void v7_en_l2_hazard_detect(void);
Aneesh V2c451f72011-06-16 23:30:47 +0000108void v7_outer_cache_enable(void);
109void v7_outer_cache_disable(void);
110void v7_outer_cache_flush_all(void);
111void v7_outer_cache_inval_all(void);
112void v7_outer_cache_flush_range(u32 start, u32 end);
113void v7_outer_cache_inval_range(u32 start, u32 end);
114
Jan Kiszka104d6fb2015-04-21 07:18:24 +0200115#ifdef CONFIG_ARMV7_NONSEC
Andre Przywara1ef92382013-09-19 18:06:42 +0200116
Marc Zyngierf510aea2014-07-12 14:24:03 +0100117int armv7_init_nonsec(void);
Jan Kiszkad6b72da2015-04-21 07:18:32 +0200118int armv7_apply_memory_carveout(u64 *start, u64 *size);
Ian Campbell97a81962014-12-21 09:45:11 +0000119bool armv7_boot_nonsec(void);
Andre Przywara1ef92382013-09-19 18:06:42 +0200120
Andre Przywara16212b52013-09-19 18:06:41 +0200121/* defined in assembly file */
122unsigned int _nonsec_init(void);
Marc Zyngierf510aea2014-07-12 14:24:03 +0100123void _do_nonsec_entry(void *target_pc, unsigned long r0,
124 unsigned long r1, unsigned long r2);
Andre Przywaraba6a1692013-09-19 18:06:44 +0200125void _smp_pen(void);
Marc Zyngierf510aea2014-07-12 14:24:03 +0100126
127extern char __secure_start[];
128extern char __secure_end[];
Chen-Yu Tsai980d6a52016-06-19 12:38:36 +0800129extern char __secure_stack_start[];
130extern char __secure_stack_end[];
Marc Zyngierf510aea2014-07-12 14:24:03 +0100131
Jan Kiszka104d6fb2015-04-21 07:18:24 +0200132#endif /* CONFIG_ARMV7_NONSEC */
Andre Przywara16212b52013-09-19 18:06:41 +0200133
Nishanth Menonc616a0d2015-03-09 17:11:59 -0500134void v7_arch_cp15_set_l2aux_ctrl(u32 l2auxctrl, u32 cpu_midr,
135 u32 cpu_rev_comb, u32 cpu_variant,
136 u32 cpu_rev);
Nishanth Menonb45c48a2015-03-09 17:12:00 -0500137void v7_arch_cp15_set_acr(u32 acr, u32 cpu_midr, u32 cpu_rev_comb,
138 u32 cpu_variant, u32 cpu_rev);
Andre Przywarad75ba502013-09-19 18:06:39 +0200139#endif /* ! __ASSEMBLY__ */
140
Aneesh V2c451f72011-06-16 23:30:47 +0000141#endif