aia_device.c 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Copyright (C) 2021 Western Digital Corporation or its affiliates.
  4. * Copyright (C) 2022 Ventana Micro Systems Inc.
  5. *
  6. * Authors:
  7. * Anup Patel <apatel@ventanamicro.com>
  8. */
  9. #include <linux/bits.h>
  10. #include <linux/irqchip/riscv-imsic.h>
  11. #include <linux/kvm_host.h>
  12. #include <linux/uaccess.h>
  13. static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx)
  14. {
  15. struct kvm_vcpu *tmp_vcpu;
  16. for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) {
  17. tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx);
  18. mutex_unlock(&tmp_vcpu->mutex);
  19. }
  20. }
  21. static void unlock_all_vcpus(struct kvm *kvm)
  22. {
  23. unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1);
  24. }
  25. static bool lock_all_vcpus(struct kvm *kvm)
  26. {
  27. struct kvm_vcpu *tmp_vcpu;
  28. unsigned long c;
  29. kvm_for_each_vcpu(c, tmp_vcpu, kvm) {
  30. if (!mutex_trylock(&tmp_vcpu->mutex)) {
  31. unlock_vcpus(kvm, c - 1);
  32. return false;
  33. }
  34. }
  35. return true;
  36. }
  37. static int aia_create(struct kvm_device *dev, u32 type)
  38. {
  39. int ret;
  40. unsigned long i;
  41. struct kvm *kvm = dev->kvm;
  42. struct kvm_vcpu *vcpu;
  43. if (irqchip_in_kernel(kvm))
  44. return -EEXIST;
  45. ret = -EBUSY;
  46. if (!lock_all_vcpus(kvm))
  47. return ret;
  48. kvm_for_each_vcpu(i, vcpu, kvm) {
  49. if (vcpu->arch.ran_atleast_once)
  50. goto out_unlock;
  51. }
  52. ret = 0;
  53. kvm->arch.aia.in_kernel = true;
  54. out_unlock:
  55. unlock_all_vcpus(kvm);
  56. return ret;
  57. }
  58. static void aia_destroy(struct kvm_device *dev)
  59. {
  60. kfree(dev);
  61. }
  62. static int aia_config(struct kvm *kvm, unsigned long type,
  63. u32 *nr, bool write)
  64. {
  65. struct kvm_aia *aia = &kvm->arch.aia;
  66. /* Writes can only be done before irqchip is initialized */
  67. if (write && kvm_riscv_aia_initialized(kvm))
  68. return -EBUSY;
  69. switch (type) {
  70. case KVM_DEV_RISCV_AIA_CONFIG_MODE:
  71. if (write) {
  72. switch (*nr) {
  73. case KVM_DEV_RISCV_AIA_MODE_EMUL:
  74. break;
  75. case KVM_DEV_RISCV_AIA_MODE_HWACCEL:
  76. case KVM_DEV_RISCV_AIA_MODE_AUTO:
  77. /*
  78. * HW Acceleration and Auto modes only
  79. * supported on host with non-zero guest
  80. * external interrupts (i.e. non-zero
  81. * VS-level IMSIC pages).
  82. */
  83. if (!kvm_riscv_aia_nr_hgei)
  84. return -EINVAL;
  85. break;
  86. default:
  87. return -EINVAL;
  88. }
  89. aia->mode = *nr;
  90. } else
  91. *nr = aia->mode;
  92. break;
  93. case KVM_DEV_RISCV_AIA_CONFIG_IDS:
  94. if (write) {
  95. if ((*nr < KVM_DEV_RISCV_AIA_IDS_MIN) ||
  96. (*nr >= KVM_DEV_RISCV_AIA_IDS_MAX) ||
  97. ((*nr & KVM_DEV_RISCV_AIA_IDS_MIN) !=
  98. KVM_DEV_RISCV_AIA_IDS_MIN) ||
  99. (kvm_riscv_aia_max_ids <= *nr))
  100. return -EINVAL;
  101. aia->nr_ids = *nr;
  102. } else
  103. *nr = aia->nr_ids;
  104. break;
  105. case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
  106. if (write) {
  107. if ((*nr >= KVM_DEV_RISCV_AIA_SRCS_MAX) ||
  108. (*nr >= kvm_riscv_aia_max_ids))
  109. return -EINVAL;
  110. aia->nr_sources = *nr;
  111. } else
  112. *nr = aia->nr_sources;
  113. break;
  114. case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
  115. if (write) {
  116. if (*nr >= KVM_DEV_RISCV_AIA_GROUP_BITS_MAX)
  117. return -EINVAL;
  118. aia->nr_group_bits = *nr;
  119. } else
  120. *nr = aia->nr_group_bits;
  121. break;
  122. case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
  123. if (write) {
  124. if ((*nr < KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN) ||
  125. (*nr >= KVM_DEV_RISCV_AIA_GROUP_SHIFT_MAX))
  126. return -EINVAL;
  127. aia->nr_group_shift = *nr;
  128. } else
  129. *nr = aia->nr_group_shift;
  130. break;
  131. case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
  132. if (write) {
  133. if (*nr >= KVM_DEV_RISCV_AIA_HART_BITS_MAX)
  134. return -EINVAL;
  135. aia->nr_hart_bits = *nr;
  136. } else
  137. *nr = aia->nr_hart_bits;
  138. break;
  139. case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
  140. if (write) {
  141. if (*nr >= KVM_DEV_RISCV_AIA_GUEST_BITS_MAX)
  142. return -EINVAL;
  143. aia->nr_guest_bits = *nr;
  144. } else
  145. *nr = aia->nr_guest_bits;
  146. break;
  147. default:
  148. return -ENXIO;
  149. }
  150. return 0;
  151. }
  152. static int aia_aplic_addr(struct kvm *kvm, u64 *addr, bool write)
  153. {
  154. struct kvm_aia *aia = &kvm->arch.aia;
  155. if (write) {
  156. /* Writes can only be done before irqchip is initialized */
  157. if (kvm_riscv_aia_initialized(kvm))
  158. return -EBUSY;
  159. if (*addr & (KVM_DEV_RISCV_APLIC_ALIGN - 1))
  160. return -EINVAL;
  161. aia->aplic_addr = *addr;
  162. } else
  163. *addr = aia->aplic_addr;
  164. return 0;
  165. }
  166. static int aia_imsic_addr(struct kvm *kvm, u64 *addr,
  167. unsigned long vcpu_idx, bool write)
  168. {
  169. struct kvm_vcpu *vcpu;
  170. struct kvm_vcpu_aia *vcpu_aia;
  171. vcpu = kvm_get_vcpu(kvm, vcpu_idx);
  172. if (!vcpu)
  173. return -EINVAL;
  174. vcpu_aia = &vcpu->arch.aia_context;
  175. if (write) {
  176. /* Writes can only be done before irqchip is initialized */
  177. if (kvm_riscv_aia_initialized(kvm))
  178. return -EBUSY;
  179. if (*addr & (KVM_DEV_RISCV_IMSIC_ALIGN - 1))
  180. return -EINVAL;
  181. }
  182. mutex_lock(&vcpu->mutex);
  183. if (write)
  184. vcpu_aia->imsic_addr = *addr;
  185. else
  186. *addr = vcpu_aia->imsic_addr;
  187. mutex_unlock(&vcpu->mutex);
  188. return 0;
  189. }
  190. static gpa_t aia_imsic_ppn(struct kvm_aia *aia, gpa_t addr)
  191. {
  192. u32 h, l;
  193. gpa_t mask = 0;
  194. h = aia->nr_hart_bits + aia->nr_guest_bits +
  195. IMSIC_MMIO_PAGE_SHIFT - 1;
  196. mask = GENMASK_ULL(h, 0);
  197. if (aia->nr_group_bits) {
  198. h = aia->nr_group_bits + aia->nr_group_shift - 1;
  199. l = aia->nr_group_shift;
  200. mask |= GENMASK_ULL(h, l);
  201. }
  202. return (addr & ~mask) >> IMSIC_MMIO_PAGE_SHIFT;
  203. }
  204. static u32 aia_imsic_hart_index(struct kvm_aia *aia, gpa_t addr)
  205. {
  206. u32 hart = 0, group = 0;
  207. if (aia->nr_hart_bits)
  208. hart = (addr >> (aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT)) &
  209. GENMASK_ULL(aia->nr_hart_bits - 1, 0);
  210. if (aia->nr_group_bits)
  211. group = (addr >> aia->nr_group_shift) &
  212. GENMASK_ULL(aia->nr_group_bits - 1, 0);
  213. return (group << aia->nr_hart_bits) | hart;
  214. }
  215. static int aia_init(struct kvm *kvm)
  216. {
  217. int ret, i;
  218. unsigned long idx;
  219. struct kvm_vcpu *vcpu;
  220. struct kvm_vcpu_aia *vaia;
  221. struct kvm_aia *aia = &kvm->arch.aia;
  222. gpa_t base_ppn = KVM_RISCV_AIA_UNDEF_ADDR;
  223. /* Irqchip can be initialized only once */
  224. if (kvm_riscv_aia_initialized(kvm))
  225. return -EBUSY;
  226. /* We might be in the middle of creating a VCPU? */
  227. if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus))
  228. return -EBUSY;
  229. /* Number of sources should be less than or equals number of IDs */
  230. if (aia->nr_ids < aia->nr_sources)
  231. return -EINVAL;
  232. /* APLIC base is required for non-zero number of sources */
  233. if (aia->nr_sources && aia->aplic_addr == KVM_RISCV_AIA_UNDEF_ADDR)
  234. return -EINVAL;
  235. /* Initialize APLIC */
  236. ret = kvm_riscv_aia_aplic_init(kvm);
  237. if (ret)
  238. return ret;
  239. /* Iterate over each VCPU */
  240. kvm_for_each_vcpu(idx, vcpu, kvm) {
  241. vaia = &vcpu->arch.aia_context;
  242. /* IMSIC base is required */
  243. if (vaia->imsic_addr == KVM_RISCV_AIA_UNDEF_ADDR) {
  244. ret = -EINVAL;
  245. goto fail_cleanup_imsics;
  246. }
  247. /* All IMSICs should have matching base PPN */
  248. if (base_ppn == KVM_RISCV_AIA_UNDEF_ADDR)
  249. base_ppn = aia_imsic_ppn(aia, vaia->imsic_addr);
  250. if (base_ppn != aia_imsic_ppn(aia, vaia->imsic_addr)) {
  251. ret = -EINVAL;
  252. goto fail_cleanup_imsics;
  253. }
  254. /* Update HART index of the IMSIC based on IMSIC base */
  255. vaia->hart_index = aia_imsic_hart_index(aia,
  256. vaia->imsic_addr);
  257. /* Initialize IMSIC for this VCPU */
  258. ret = kvm_riscv_vcpu_aia_imsic_init(vcpu);
  259. if (ret)
  260. goto fail_cleanup_imsics;
  261. }
  262. /* Set the initialized flag */
  263. kvm->arch.aia.initialized = true;
  264. return 0;
  265. fail_cleanup_imsics:
  266. for (i = idx - 1; i >= 0; i--) {
  267. vcpu = kvm_get_vcpu(kvm, i);
  268. if (!vcpu)
  269. continue;
  270. kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
  271. }
  272. kvm_riscv_aia_aplic_cleanup(kvm);
  273. return ret;
  274. }
  275. static int aia_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
  276. {
  277. u32 nr;
  278. u64 addr;
  279. int nr_vcpus, r = -ENXIO;
  280. unsigned long v, type = (unsigned long)attr->attr;
  281. void __user *uaddr = (void __user *)(long)attr->addr;
  282. switch (attr->group) {
  283. case KVM_DEV_RISCV_AIA_GRP_CONFIG:
  284. if (copy_from_user(&nr, uaddr, sizeof(nr)))
  285. return -EFAULT;
  286. mutex_lock(&dev->kvm->lock);
  287. r = aia_config(dev->kvm, type, &nr, true);
  288. mutex_unlock(&dev->kvm->lock);
  289. break;
  290. case KVM_DEV_RISCV_AIA_GRP_ADDR:
  291. if (copy_from_user(&addr, uaddr, sizeof(addr)))
  292. return -EFAULT;
  293. nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
  294. mutex_lock(&dev->kvm->lock);
  295. if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
  296. r = aia_aplic_addr(dev->kvm, &addr, true);
  297. else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
  298. r = aia_imsic_addr(dev->kvm, &addr,
  299. type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), true);
  300. mutex_unlock(&dev->kvm->lock);
  301. break;
  302. case KVM_DEV_RISCV_AIA_GRP_CTRL:
  303. switch (type) {
  304. case KVM_DEV_RISCV_AIA_CTRL_INIT:
  305. mutex_lock(&dev->kvm->lock);
  306. r = aia_init(dev->kvm);
  307. mutex_unlock(&dev->kvm->lock);
  308. break;
  309. }
  310. break;
  311. case KVM_DEV_RISCV_AIA_GRP_APLIC:
  312. if (copy_from_user(&nr, uaddr, sizeof(nr)))
  313. return -EFAULT;
  314. mutex_lock(&dev->kvm->lock);
  315. r = kvm_riscv_aia_aplic_set_attr(dev->kvm, type, nr);
  316. mutex_unlock(&dev->kvm->lock);
  317. break;
  318. case KVM_DEV_RISCV_AIA_GRP_IMSIC:
  319. if (copy_from_user(&v, uaddr, sizeof(v)))
  320. return -EFAULT;
  321. mutex_lock(&dev->kvm->lock);
  322. r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, true, &v);
  323. mutex_unlock(&dev->kvm->lock);
  324. break;
  325. }
  326. return r;
  327. }
  328. static int aia_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
  329. {
  330. u32 nr;
  331. u64 addr;
  332. int nr_vcpus, r = -ENXIO;
  333. void __user *uaddr = (void __user *)(long)attr->addr;
  334. unsigned long v, type = (unsigned long)attr->attr;
  335. switch (attr->group) {
  336. case KVM_DEV_RISCV_AIA_GRP_CONFIG:
  337. if (copy_from_user(&nr, uaddr, sizeof(nr)))
  338. return -EFAULT;
  339. mutex_lock(&dev->kvm->lock);
  340. r = aia_config(dev->kvm, type, &nr, false);
  341. mutex_unlock(&dev->kvm->lock);
  342. if (r)
  343. return r;
  344. if (copy_to_user(uaddr, &nr, sizeof(nr)))
  345. return -EFAULT;
  346. break;
  347. case KVM_DEV_RISCV_AIA_GRP_ADDR:
  348. if (copy_from_user(&addr, uaddr, sizeof(addr)))
  349. return -EFAULT;
  350. nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
  351. mutex_lock(&dev->kvm->lock);
  352. if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
  353. r = aia_aplic_addr(dev->kvm, &addr, false);
  354. else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
  355. r = aia_imsic_addr(dev->kvm, &addr,
  356. type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), false);
  357. mutex_unlock(&dev->kvm->lock);
  358. if (r)
  359. return r;
  360. if (copy_to_user(uaddr, &addr, sizeof(addr)))
  361. return -EFAULT;
  362. break;
  363. case KVM_DEV_RISCV_AIA_GRP_APLIC:
  364. if (copy_from_user(&nr, uaddr, sizeof(nr)))
  365. return -EFAULT;
  366. mutex_lock(&dev->kvm->lock);
  367. r = kvm_riscv_aia_aplic_get_attr(dev->kvm, type, &nr);
  368. mutex_unlock(&dev->kvm->lock);
  369. if (r)
  370. return r;
  371. if (copy_to_user(uaddr, &nr, sizeof(nr)))
  372. return -EFAULT;
  373. break;
  374. case KVM_DEV_RISCV_AIA_GRP_IMSIC:
  375. if (copy_from_user(&v, uaddr, sizeof(v)))
  376. return -EFAULT;
  377. mutex_lock(&dev->kvm->lock);
  378. r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, false, &v);
  379. mutex_unlock(&dev->kvm->lock);
  380. if (r)
  381. return r;
  382. if (copy_to_user(uaddr, &v, sizeof(v)))
  383. return -EFAULT;
  384. break;
  385. }
  386. return r;
  387. }
  388. static int aia_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
  389. {
  390. int nr_vcpus;
  391. switch (attr->group) {
  392. case KVM_DEV_RISCV_AIA_GRP_CONFIG:
  393. switch (attr->attr) {
  394. case KVM_DEV_RISCV_AIA_CONFIG_MODE:
  395. case KVM_DEV_RISCV_AIA_CONFIG_IDS:
  396. case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
  397. case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
  398. case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
  399. case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
  400. case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
  401. return 0;
  402. }
  403. break;
  404. case KVM_DEV_RISCV_AIA_GRP_ADDR:
  405. nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
  406. if (attr->attr == KVM_DEV_RISCV_AIA_ADDR_APLIC)
  407. return 0;
  408. else if (attr->attr < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
  409. return 0;
  410. break;
  411. case KVM_DEV_RISCV_AIA_GRP_CTRL:
  412. switch (attr->attr) {
  413. case KVM_DEV_RISCV_AIA_CTRL_INIT:
  414. return 0;
  415. }
  416. break;
  417. case KVM_DEV_RISCV_AIA_GRP_APLIC:
  418. return kvm_riscv_aia_aplic_has_attr(dev->kvm, attr->attr);
  419. case KVM_DEV_RISCV_AIA_GRP_IMSIC:
  420. return kvm_riscv_aia_imsic_has_attr(dev->kvm, attr->attr);
  421. }
  422. return -ENXIO;
  423. }
  424. struct kvm_device_ops kvm_riscv_aia_device_ops = {
  425. .name = "kvm-riscv-aia",
  426. .create = aia_create,
  427. .destroy = aia_destroy,
  428. .set_attr = aia_set_attr,
  429. .get_attr = aia_get_attr,
  430. .has_attr = aia_has_attr,
  431. };
  432. int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu)
  433. {
  434. /* Proceed only if AIA was initialized successfully */
  435. if (!kvm_riscv_aia_initialized(vcpu->kvm))
  436. return 1;
  437. /* Update the IMSIC HW state before entering guest mode */
  438. return kvm_riscv_vcpu_aia_imsic_update(vcpu);
  439. }
  440. void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu)
  441. {
  442. struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
  443. struct kvm_vcpu_aia_csr *reset_csr =
  444. &vcpu->arch.aia_context.guest_reset_csr;
  445. if (!kvm_riscv_aia_available())
  446. return;
  447. memcpy(csr, reset_csr, sizeof(*csr));
  448. /* Proceed only if AIA was initialized successfully */
  449. if (!kvm_riscv_aia_initialized(vcpu->kvm))
  450. return;
  451. /* Reset the IMSIC context */
  452. kvm_riscv_vcpu_aia_imsic_reset(vcpu);
  453. }
  454. int kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu)
  455. {
  456. struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
  457. if (!kvm_riscv_aia_available())
  458. return 0;
  459. /*
  460. * We don't do any memory allocations over here because these
  461. * will be done after AIA device is initialized by the user-space.
  462. *
  463. * Refer, aia_init() implementation for more details.
  464. */
  465. /* Initialize default values in AIA vcpu context */
  466. vaia->imsic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
  467. vaia->hart_index = vcpu->vcpu_idx;
  468. return 0;
  469. }
  470. void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu)
  471. {
  472. /* Proceed only if AIA was initialized successfully */
  473. if (!kvm_riscv_aia_initialized(vcpu->kvm))
  474. return;
  475. /* Cleanup IMSIC context */
  476. kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
  477. }
  478. int kvm_riscv_aia_inject_msi_by_id(struct kvm *kvm, u32 hart_index,
  479. u32 guest_index, u32 iid)
  480. {
  481. unsigned long idx;
  482. struct kvm_vcpu *vcpu;
  483. /* Proceed only if AIA was initialized successfully */
  484. if (!kvm_riscv_aia_initialized(kvm))
  485. return -EBUSY;
  486. /* Inject MSI to matching VCPU */
  487. kvm_for_each_vcpu(idx, vcpu, kvm) {
  488. if (vcpu->arch.aia_context.hart_index == hart_index)
  489. return kvm_riscv_vcpu_aia_imsic_inject(vcpu,
  490. guest_index,
  491. 0, iid);
  492. }
  493. return 0;
  494. }
  495. int kvm_riscv_aia_inject_msi(struct kvm *kvm, struct kvm_msi *msi)
  496. {
  497. gpa_t tppn, ippn;
  498. unsigned long idx;
  499. struct kvm_vcpu *vcpu;
  500. u32 g, toff, iid = msi->data;
  501. struct kvm_aia *aia = &kvm->arch.aia;
  502. gpa_t target = (((gpa_t)msi->address_hi) << 32) | msi->address_lo;
  503. /* Proceed only if AIA was initialized successfully */
  504. if (!kvm_riscv_aia_initialized(kvm))
  505. return -EBUSY;
  506. /* Convert target address to target PPN */
  507. tppn = target >> IMSIC_MMIO_PAGE_SHIFT;
  508. /* Extract and clear Guest ID from target PPN */
  509. g = tppn & (BIT(aia->nr_guest_bits) - 1);
  510. tppn &= ~((gpa_t)(BIT(aia->nr_guest_bits) - 1));
  511. /* Inject MSI to matching VCPU */
  512. kvm_for_each_vcpu(idx, vcpu, kvm) {
  513. ippn = vcpu->arch.aia_context.imsic_addr >>
  514. IMSIC_MMIO_PAGE_SHIFT;
  515. if (ippn == tppn) {
  516. toff = target & (IMSIC_MMIO_PAGE_SZ - 1);
  517. return kvm_riscv_vcpu_aia_imsic_inject(vcpu, g,
  518. toff, iid);
  519. }
  520. }
  521. return 0;
  522. }
  523. int kvm_riscv_aia_inject_irq(struct kvm *kvm, unsigned int irq, bool level)
  524. {
  525. /* Proceed only if AIA was initialized successfully */
  526. if (!kvm_riscv_aia_initialized(kvm))
  527. return -EBUSY;
  528. /* Inject interrupt level change in APLIC */
  529. return kvm_riscv_aia_aplic_inject(kvm, irq, level);
  530. }
  531. void kvm_riscv_aia_init_vm(struct kvm *kvm)
  532. {
  533. struct kvm_aia *aia = &kvm->arch.aia;
  534. if (!kvm_riscv_aia_available())
  535. return;
  536. /*
  537. * We don't do any memory allocations over here because these
  538. * will be done after AIA device is initialized by the user-space.
  539. *
  540. * Refer, aia_init() implementation for more details.
  541. */
  542. /* Initialize default values in AIA global context */
  543. aia->mode = (kvm_riscv_aia_nr_hgei) ?
  544. KVM_DEV_RISCV_AIA_MODE_AUTO : KVM_DEV_RISCV_AIA_MODE_EMUL;
  545. aia->nr_ids = kvm_riscv_aia_max_ids - 1;
  546. aia->nr_sources = 0;
  547. aia->nr_group_bits = 0;
  548. aia->nr_group_shift = KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN;
  549. aia->nr_hart_bits = 0;
  550. aia->nr_guest_bits = 0;
  551. aia->aplic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
  552. }
  553. void kvm_riscv_aia_destroy_vm(struct kvm *kvm)
  554. {
  555. /* Proceed only if AIA was initialized successfully */
  556. if (!kvm_riscv_aia_initialized(kvm))
  557. return;
  558. /* Cleanup APLIC context */
  559. kvm_riscv_aia_aplic_cleanup(kvm);
  560. }