Tom Rini | 83d290c | 2018-05-06 17:58:06 -0400 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0+ |
wdenk | edc48b6 | 2002-09-08 17:56:50 +0000 | [diff] [blame] | 2 | /* |
| 3 | * (C) Copyright 2002 |
| 4 | * Wolfgang Denk, DENX Software Engineering, wd@denx.de. |
wdenk | edc48b6 | 2002-09-08 17:56:50 +0000 | [diff] [blame] | 5 | */ |
| 6 | |
| 7 | /* for now: just dummy functions to satisfy the linker */ |
| 8 | |
wdenk | 8ed9604 | 2005-01-09 23:16:25 +0000 | [diff] [blame] | 9 | #include <common.h> |
Simon Glass | 1eb69ae | 2019-11-14 12:57:39 -0700 | [diff] [blame] | 10 | #include <cpu_func.h> |
Simon Glass | f7ae49f | 2020-05-10 11:40:05 -0600 | [diff] [blame] | 11 | #include <log.h> |
Thierry Reding | 1dfdd9b | 2014-12-09 22:25:22 -0700 | [diff] [blame] | 12 | #include <malloc.h> |
Simon Glass | 90526e9 | 2020-05-10 11:39:56 -0600 | [diff] [blame] | 13 | #include <asm/cache.h> |
wdenk | 8ed9604 | 2005-01-09 23:16:25 +0000 | [diff] [blame] | 14 | |
Ovidiu Panait | 586b15b | 2020-03-29 20:57:39 +0300 | [diff] [blame] | 15 | DECLARE_GLOBAL_DATA_PTR; |
| 16 | |
Wu, Josh | 633b6cc | 2015-07-27 11:40:17 +0800 | [diff] [blame] | 17 | /* |
| 18 | * Flush range from all levels of d-cache/unified-cache. |
| 19 | * Affects the range [start, start + size - 1]. |
| 20 | */ |
Jeroen Hofstee | fcfddfd | 2014-06-23 22:07:04 +0200 | [diff] [blame] | 21 | __weak void flush_cache(unsigned long start, unsigned long size) |
wdenk | edc48b6 | 2002-09-08 17:56:50 +0000 | [diff] [blame] | 22 | { |
Wu, Josh | 633b6cc | 2015-07-27 11:40:17 +0800 | [diff] [blame] | 23 | flush_dcache_range(start, start + size); |
wdenk | edc48b6 | 2002-09-08 17:56:50 +0000 | [diff] [blame] | 24 | } |
Aneesh V | e05f007 | 2011-06-16 23:30:50 +0000 | [diff] [blame] | 25 | |
| 26 | /* |
| 27 | * Default implementation: |
| 28 | * do a range flush for the entire range |
| 29 | */ |
Jeroen Hofstee | fcfddfd | 2014-06-23 22:07:04 +0200 | [diff] [blame] | 30 | __weak void flush_dcache_all(void) |
Aneesh V | e05f007 | 2011-06-16 23:30:50 +0000 | [diff] [blame] | 31 | { |
| 32 | flush_cache(0, ~0); |
| 33 | } |
Aneesh V | cba4b18 | 2011-08-16 04:33:05 +0000 | [diff] [blame] | 34 | |
| 35 | /* |
| 36 | * Default implementation of enable_caches() |
| 37 | * Real implementation should be in platform code |
| 38 | */ |
Jeroen Hofstee | fcfddfd | 2014-06-23 22:07:04 +0200 | [diff] [blame] | 39 | __weak void enable_caches(void) |
Aneesh V | cba4b18 | 2011-08-16 04:33:05 +0000 | [diff] [blame] | 40 | { |
| 41 | puts("WARNING: Caches not enabled\n"); |
| 42 | } |
Thierry Reding | 1dfdd9b | 2014-12-09 22:25:22 -0700 | [diff] [blame] | 43 | |
Wu, Josh | 387871a | 2015-07-27 11:40:16 +0800 | [diff] [blame] | 44 | __weak void invalidate_dcache_range(unsigned long start, unsigned long stop) |
| 45 | { |
| 46 | /* An empty stub, real implementation should be in platform code */ |
| 47 | } |
| 48 | __weak void flush_dcache_range(unsigned long start, unsigned long stop) |
| 49 | { |
| 50 | /* An empty stub, real implementation should be in platform code */ |
| 51 | } |
| 52 | |
Simon Glass | 397b569 | 2016-06-19 19:43:01 -0600 | [diff] [blame] | 53 | int check_cache_range(unsigned long start, unsigned long stop) |
| 54 | { |
| 55 | int ok = 1; |
| 56 | |
| 57 | if (start & (CONFIG_SYS_CACHELINE_SIZE - 1)) |
| 58 | ok = 0; |
| 59 | |
| 60 | if (stop & (CONFIG_SYS_CACHELINE_SIZE - 1)) |
| 61 | ok = 0; |
| 62 | |
| 63 | if (!ok) { |
Simon Glass | bcc53bf | 2016-06-19 19:43:05 -0600 | [diff] [blame] | 64 | warn_non_spl("CACHE: Misaligned operation at range [%08lx, %08lx]\n", |
| 65 | start, stop); |
Simon Glass | 397b569 | 2016-06-19 19:43:01 -0600 | [diff] [blame] | 66 | } |
| 67 | |
| 68 | return ok; |
| 69 | } |
| 70 | |
Thierry Reding | 1dfdd9b | 2014-12-09 22:25:22 -0700 | [diff] [blame] | 71 | #ifdef CONFIG_SYS_NONCACHED_MEMORY |
| 72 | /* |
| 73 | * Reserve one MMU section worth of address space below the malloc() area that |
| 74 | * will be mapped uncached. |
| 75 | */ |
| 76 | static unsigned long noncached_start; |
| 77 | static unsigned long noncached_end; |
| 78 | static unsigned long noncached_next; |
| 79 | |
Patrice Chotard | c2a2123 | 2020-04-28 11:38:03 +0200 | [diff] [blame] | 80 | void noncached_set_region(void) |
| 81 | { |
| 82 | #if !CONFIG_IS_ENABLED(SYS_DCACHE_OFF) |
| 83 | mmu_set_region_dcache_behaviour(noncached_start, |
| 84 | noncached_end - noncached_start, |
| 85 | DCACHE_OFF); |
| 86 | #endif |
| 87 | } |
| 88 | |
Thierry Reding | 1dfdd9b | 2014-12-09 22:25:22 -0700 | [diff] [blame] | 89 | void noncached_init(void) |
| 90 | { |
| 91 | phys_addr_t start, end; |
| 92 | size_t size; |
| 93 | |
Stephen Warren | 5e0404f | 2019-08-27 11:54:31 -0600 | [diff] [blame] | 94 | /* If this calculation changes, update board_f.c:reserve_noncached() */ |
Thierry Reding | 1dfdd9b | 2014-12-09 22:25:22 -0700 | [diff] [blame] | 95 | end = ALIGN(mem_malloc_start, MMU_SECTION_SIZE) - MMU_SECTION_SIZE; |
| 96 | size = ALIGN(CONFIG_SYS_NONCACHED_MEMORY, MMU_SECTION_SIZE); |
| 97 | start = end - size; |
| 98 | |
| 99 | debug("mapping memory %pa-%pa non-cached\n", &start, &end); |
| 100 | |
| 101 | noncached_start = start; |
| 102 | noncached_end = end; |
| 103 | noncached_next = start; |
| 104 | |
Patrice Chotard | c2a2123 | 2020-04-28 11:38:03 +0200 | [diff] [blame] | 105 | noncached_set_region(); |
Thierry Reding | 1dfdd9b | 2014-12-09 22:25:22 -0700 | [diff] [blame] | 106 | } |
| 107 | |
| 108 | phys_addr_t noncached_alloc(size_t size, size_t align) |
| 109 | { |
| 110 | phys_addr_t next = ALIGN(noncached_next, align); |
| 111 | |
| 112 | if (next >= noncached_end || (noncached_end - next) < size) |
| 113 | return 0; |
| 114 | |
| 115 | debug("allocated %zu bytes of uncached memory @%pa\n", size, &next); |
| 116 | noncached_next = next + size; |
| 117 | |
| 118 | return next; |
| 119 | } |
| 120 | #endif /* CONFIG_SYS_NONCACHED_MEMORY */ |
Albert ARIBAUD | 62e9207 | 2015-10-23 18:06:40 +0200 | [diff] [blame] | 121 | |
Tom Rini | 3a64940 | 2017-03-18 09:01:44 -0400 | [diff] [blame] | 122 | #if CONFIG_IS_ENABLED(SYS_THUMB_BUILD) |
Albert ARIBAUD | 62e9207 | 2015-10-23 18:06:40 +0200 | [diff] [blame] | 123 | void invalidate_l2_cache(void) |
| 124 | { |
| 125 | unsigned int val = 0; |
| 126 | |
| 127 | asm volatile("mcr p15, 1, %0, c15, c11, 0 @ invl l2 cache" |
| 128 | : : "r" (val) : "cc"); |
| 129 | isb(); |
| 130 | } |
| 131 | #endif |
Ovidiu Panait | 586b15b | 2020-03-29 20:57:39 +0300 | [diff] [blame] | 132 | |
Ovidiu Panait | 79926e4 | 2020-03-29 20:57:41 +0300 | [diff] [blame] | 133 | int arch_reserve_mmu(void) |
Ovidiu Panait | 586b15b | 2020-03-29 20:57:39 +0300 | [diff] [blame] | 134 | { |
Ovidiu Panait | 6184858 | 2020-03-29 20:57:40 +0300 | [diff] [blame] | 135 | return arm_reserve_mmu(); |
| 136 | } |
| 137 | |
| 138 | __weak int arm_reserve_mmu(void) |
| 139 | { |
Ovidiu Panait | 586b15b | 2020-03-29 20:57:39 +0300 | [diff] [blame] | 140 | #if !(CONFIG_IS_ENABLED(SYS_ICACHE_OFF) && CONFIG_IS_ENABLED(SYS_DCACHE_OFF)) |
| 141 | /* reserve TLB table */ |
| 142 | gd->arch.tlb_size = PGTABLE_SIZE; |
| 143 | gd->relocaddr -= gd->arch.tlb_size; |
| 144 | |
| 145 | /* round down to next 64 kB limit */ |
| 146 | gd->relocaddr &= ~(0x10000 - 1); |
| 147 | |
| 148 | gd->arch.tlb_addr = gd->relocaddr; |
| 149 | debug("TLB table from %08lx to %08lx\n", gd->arch.tlb_addr, |
| 150 | gd->arch.tlb_addr + gd->arch.tlb_size); |
| 151 | |
| 152 | #ifdef CONFIG_SYS_MEM_RESERVE_SECURE |
| 153 | /* |
| 154 | * Record allocated tlb_addr in case gd->tlb_addr to be overwritten |
| 155 | * with location within secure ram. |
| 156 | */ |
| 157 | gd->arch.tlb_allocated = gd->arch.tlb_addr; |
| 158 | #endif |
| 159 | #endif |
| 160 | |
| 161 | return 0; |
| 162 | } |