cache.c 3.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143
  1. // SPDX-License-Identifier: GPL-2.0+
  2. /*
  3. * (C) Copyright 2002
  4. * Wolfgang Denk, DENX Software Engineering, wd@denx.de.
  5. */
  6. #include <common.h>
  7. #include <asm/immap.h>
  8. #include <asm/cache.h>
  9. volatile int *cf_icache_status = (int *)ICACHE_STATUS;
  10. volatile int *cf_dcache_status = (int *)DCACHE_STATUS;
  11. void flush_cache(ulong start_addr, ulong size)
  12. {
  13. /* Must be implemented for all M68k processors with copy-back data cache */
  14. }
  15. int icache_status(void)
  16. {
  17. return *cf_icache_status;
  18. }
  19. int dcache_status(void)
  20. {
  21. return *cf_dcache_status;
  22. }
  23. void icache_enable(void)
  24. {
  25. icache_invalid();
  26. *cf_icache_status = 1;
  27. #if defined(CONFIG_CF_V4) || defined(CONFIG_CF_V4E)
  28. __asm__ __volatile__("movec %0, %%acr2"::"r"(CONFIG_SYS_CACHE_ACR2));
  29. __asm__ __volatile__("movec %0, %%acr3"::"r"(CONFIG_SYS_CACHE_ACR3));
  30. #if defined(CONFIG_CF_V4E)
  31. __asm__ __volatile__("movec %0, %%acr6"::"r"(CONFIG_SYS_CACHE_ACR6));
  32. __asm__ __volatile__("movec %0, %%acr7"::"r"(CONFIG_SYS_CACHE_ACR7));
  33. #endif
  34. #else
  35. __asm__ __volatile__("movec %0, %%acr0"::"r"(CONFIG_SYS_CACHE_ACR0));
  36. __asm__ __volatile__("movec %0, %%acr1"::"r"(CONFIG_SYS_CACHE_ACR1));
  37. #endif
  38. __asm__ __volatile__("movec %0, %%cacr"::"r"(CONFIG_SYS_CACHE_ICACR));
  39. }
  40. void icache_disable(void)
  41. {
  42. u32 temp = 0;
  43. *cf_icache_status = 0;
  44. icache_invalid();
  45. #if defined(CONFIG_CF_V4) || defined(CONFIG_CF_V4E)
  46. __asm__ __volatile__("movec %0, %%acr2"::"r"(temp));
  47. __asm__ __volatile__("movec %0, %%acr3"::"r"(temp));
  48. #if defined(CONFIG_CF_V4E)
  49. __asm__ __volatile__("movec %0, %%acr6"::"r"(temp));
  50. __asm__ __volatile__("movec %0, %%acr7"::"r"(temp));
  51. #endif
  52. #else
  53. __asm__ __volatile__("movec %0, %%acr0"::"r"(temp));
  54. __asm__ __volatile__("movec %0, %%acr1"::"r"(temp));
  55. #endif
  56. }
  57. void icache_invalid(void)
  58. {
  59. u32 temp;
  60. temp = CONFIG_SYS_ICACHE_INV;
  61. if (*cf_icache_status)
  62. temp |= CONFIG_SYS_CACHE_ICACR;
  63. __asm__ __volatile__("movec %0, %%cacr"::"r"(temp));
  64. }
  65. /*
  66. * data cache only for ColdFire V4 such as MCF547x_8x, MCF5445x
  67. * the dcache will be dummy in ColdFire V2 and V3
  68. */
  69. void dcache_enable(void)
  70. {
  71. dcache_invalid();
  72. *cf_dcache_status = 1;
  73. #if defined(CONFIG_CF_V4) || defined(CONFIG_CF_V4E)
  74. __asm__ __volatile__("movec %0, %%acr0"::"r"(CONFIG_SYS_CACHE_ACR0));
  75. __asm__ __volatile__("movec %0, %%acr1"::"r"(CONFIG_SYS_CACHE_ACR1));
  76. #if defined(CONFIG_CF_V4E)
  77. __asm__ __volatile__("movec %0, %%acr4"::"r"(CONFIG_SYS_CACHE_ACR4));
  78. __asm__ __volatile__("movec %0, %%acr5"::"r"(CONFIG_SYS_CACHE_ACR5));
  79. #endif
  80. #endif
  81. __asm__ __volatile__("movec %0, %%cacr"::"r"(CONFIG_SYS_CACHE_DCACR));
  82. }
  83. void dcache_disable(void)
  84. {
  85. u32 temp = 0;
  86. *cf_dcache_status = 0;
  87. dcache_invalid();
  88. __asm__ __volatile__("movec %0, %%cacr"::"r"(temp));
  89. #if defined(CONFIG_CF_V4) || defined(CONFIG_CF_V4E)
  90. __asm__ __volatile__("movec %0, %%acr0"::"r"(temp));
  91. __asm__ __volatile__("movec %0, %%acr1"::"r"(temp));
  92. #if defined(CONFIG_CF_V4E)
  93. __asm__ __volatile__("movec %0, %%acr4"::"r"(temp));
  94. __asm__ __volatile__("movec %0, %%acr5"::"r"(temp));
  95. #endif
  96. #endif
  97. }
  98. void dcache_invalid(void)
  99. {
  100. #if defined(CONFIG_CF_V4) || defined(CONFIG_CF_V4E)
  101. u32 temp;
  102. temp = CONFIG_SYS_DCACHE_INV;
  103. if (*cf_dcache_status)
  104. temp |= CONFIG_SYS_CACHE_DCACR;
  105. if (*cf_icache_status)
  106. temp |= CONFIG_SYS_CACHE_ICACR;
  107. __asm__ __volatile__("movec %0, %%cacr"::"r"(temp));
  108. #endif
  109. }
  110. __weak void invalidate_dcache_range(unsigned long start, unsigned long stop)
  111. {
  112. /* An empty stub, real implementation should be in platform code */
  113. }
  114. __weak void flush_dcache_range(unsigned long start, unsigned long stop)
  115. {
  116. /* An empty stub, real implementation should be in platform code */
  117. }