arch_timer.h 2.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121
  1. /* SPDX-License-Identifier: GPL-2.0 */
  2. #ifndef __ASMARM_ARCH_TIMER_H
  3. #define __ASMARM_ARCH_TIMER_H
  4. #include <asm/barrier.h>
  5. #include <asm/errno.h>
  6. #include <linux/clocksource.h>
  7. #include <linux/init.h>
  8. #include <linux/types.h>
  9. #include <clocksource/arm_arch_timer.h>
  10. #ifdef CONFIG_ARM_ARCH_TIMER
  11. int arch_timer_arch_init(void);
  12. /*
  13. * These register accessors are marked inline so the compiler can
  14. * nicely work out which register we want, and chuck away the rest of
  15. * the code. At least it does so with a recent GCC (4.6.3).
  16. */
  17. static __always_inline
  18. void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u32 val)
  19. {
  20. if (access == ARCH_TIMER_PHYS_ACCESS) {
  21. switch (reg) {
  22. case ARCH_TIMER_REG_CTRL:
  23. asm volatile("mcr p15, 0, %0, c14, c2, 1" : : "r" (val));
  24. break;
  25. case ARCH_TIMER_REG_TVAL:
  26. asm volatile("mcr p15, 0, %0, c14, c2, 0" : : "r" (val));
  27. break;
  28. }
  29. } else if (access == ARCH_TIMER_VIRT_ACCESS) {
  30. switch (reg) {
  31. case ARCH_TIMER_REG_CTRL:
  32. asm volatile("mcr p15, 0, %0, c14, c3, 1" : : "r" (val));
  33. break;
  34. case ARCH_TIMER_REG_TVAL:
  35. asm volatile("mcr p15, 0, %0, c14, c3, 0" : : "r" (val));
  36. break;
  37. }
  38. }
  39. isb();
  40. }
  41. static __always_inline
  42. u32 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg)
  43. {
  44. u32 val = 0;
  45. if (access == ARCH_TIMER_PHYS_ACCESS) {
  46. switch (reg) {
  47. case ARCH_TIMER_REG_CTRL:
  48. asm volatile("mrc p15, 0, %0, c14, c2, 1" : "=r" (val));
  49. break;
  50. case ARCH_TIMER_REG_TVAL:
  51. asm volatile("mrc p15, 0, %0, c14, c2, 0" : "=r" (val));
  52. break;
  53. }
  54. } else if (access == ARCH_TIMER_VIRT_ACCESS) {
  55. switch (reg) {
  56. case ARCH_TIMER_REG_CTRL:
  57. asm volatile("mrc p15, 0, %0, c14, c3, 1" : "=r" (val));
  58. break;
  59. case ARCH_TIMER_REG_TVAL:
  60. asm volatile("mrc p15, 0, %0, c14, c3, 0" : "=r" (val));
  61. break;
  62. }
  63. }
  64. return val;
  65. }
  66. static inline u32 arch_timer_get_cntfrq(void)
  67. {
  68. u32 val;
  69. asm volatile("mrc p15, 0, %0, c14, c0, 0" : "=r" (val));
  70. return val;
  71. }
  72. extern u64 arch_counter_last_cntpct;
  73. static inline u64 arch_counter_get_cntpct(void)
  74. {
  75. u64 cval;
  76. int timeout = 3;
  77. isb();
  78. asm volatile("mrrc p15, 0, %Q0, %R0, c14" : "=r" (cval));
  79. while (cval < arch_counter_last_cntpct && timeout--)
  80. asm volatile("mrrc p15, 0, %Q0, %R0, c14" : "=r" (cval));
  81. if (cval < arch_counter_last_cntpct) cval = arch_counter_last_cntpct;
  82. arch_counter_last_cntpct = cval;
  83. return cval;
  84. }
  85. static inline u64 arch_counter_get_cntvct(void)
  86. {
  87. u64 cval;
  88. isb();
  89. asm volatile("mrrc p15, 1, %Q0, %R0, c14" : "=r" (cval));
  90. return cval;
  91. }
  92. static inline u32 arch_timer_get_cntkctl(void)
  93. {
  94. u32 cntkctl;
  95. asm volatile("mrc p15, 0, %0, c14, c1, 0" : "=r" (cntkctl));
  96. return cntkctl;
  97. }
  98. static inline void arch_timer_set_cntkctl(u32 cntkctl)
  99. {
  100. asm volatile("mcr p15, 0, %0, c14, c1, 0" : : "r" (cntkctl));
  101. isb();
  102. }
  103. #endif
  104. #endif