123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611 |
- /* SPDX-License-Identifier: (GPL-2.0 OR CDDL-1.0) */
- /*
- * vboxguest core guest-device handling code, VBoxGuest.cpp in upstream svn.
- *
- * Copyright (C) 2007-2016 Oracle Corporation
- */
- #include <linux/device.h>
- #include <linux/mm.h>
- #include <linux/sched.h>
- #include <linux/sizes.h>
- #include <linux/slab.h>
- #include <linux/vbox_err.h>
- #include <linux/vbox_utils.h>
- #include <linux/vmalloc.h>
- #include "vboxguest_core.h"
- #include "vboxguest_version.h"
- /* Get the pointer to the first HGCM parameter. */
- #define VBG_IOCTL_HGCM_CALL_PARMS(a) \
- ((struct vmmdev_hgcm_function_parameter *)( \
- (u8 *)(a) + sizeof(struct vbg_ioctl_hgcm_call)))
- /* Get the pointer to the first HGCM parameter in a 32-bit request. */
- #define VBG_IOCTL_HGCM_CALL_PARMS32(a) \
- ((struct vmmdev_hgcm_function_parameter32 *)( \
- (u8 *)(a) + sizeof(struct vbg_ioctl_hgcm_call)))
- #define GUEST_MAPPINGS_TRIES 5
- /**
- * Reserves memory in which the VMM can relocate any guest mappings
- * that are floating around.
- *
- * This operation is a little bit tricky since the VMM might not accept
- * just any address because of address clashes between the three contexts
- * it operates in, so we try several times.
- *
- * Failure to reserve the guest mappings is ignored.
- *
- * @gdev: The Guest extension device.
- */
- static void vbg_guest_mappings_init(struct vbg_dev *gdev)
- {
- struct vmmdev_hypervisorinfo *req;
- void *guest_mappings[GUEST_MAPPINGS_TRIES];
- struct page **pages = NULL;
- u32 size, hypervisor_size;
- int i, rc;
- /* Query the required space. */
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_GET_HYPERVISOR_INFO);
- if (!req)
- return;
- req->hypervisor_start = 0;
- req->hypervisor_size = 0;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0)
- goto out;
- /*
- * The VMM will report back if there is nothing it wants to map, like
- * for instance in VT-x and AMD-V mode.
- */
- if (req->hypervisor_size == 0)
- goto out;
- hypervisor_size = req->hypervisor_size;
- /* Add 4M so that we can align the vmap to 4MiB as the host requires. */
- size = PAGE_ALIGN(req->hypervisor_size) + SZ_4M;
- pages = kmalloc_array(size >> PAGE_SHIFT, sizeof(*pages), GFP_KERNEL);
- if (!pages)
- goto out;
- gdev->guest_mappings_dummy_page = alloc_page(GFP_HIGHUSER);
- if (!gdev->guest_mappings_dummy_page)
- goto out;
- for (i = 0; i < (size >> PAGE_SHIFT); i++)
- pages[i] = gdev->guest_mappings_dummy_page;
- /*
- * Try several times, the VMM might not accept some addresses because
- * of address clashes between the three contexts.
- */
- for (i = 0; i < GUEST_MAPPINGS_TRIES; i++) {
- guest_mappings[i] = vmap(pages, (size >> PAGE_SHIFT),
- VM_MAP, PAGE_KERNEL_RO);
- if (!guest_mappings[i])
- break;
- req->header.request_type = VMMDEVREQ_SET_HYPERVISOR_INFO;
- req->header.rc = VERR_INTERNAL_ERROR;
- req->hypervisor_size = hypervisor_size;
- req->hypervisor_start =
- (unsigned long)PTR_ALIGN(guest_mappings[i], SZ_4M);
- rc = vbg_req_perform(gdev, req);
- if (rc >= 0) {
- gdev->guest_mappings = guest_mappings[i];
- break;
- }
- }
- /* Free vmap's from failed attempts. */
- while (--i >= 0)
- vunmap(guest_mappings[i]);
- /* On failure free the dummy-page backing the vmap */
- if (!gdev->guest_mappings) {
- __free_page(gdev->guest_mappings_dummy_page);
- gdev->guest_mappings_dummy_page = NULL;
- }
- out:
- vbg_req_free(req, sizeof(*req));
- kfree(pages);
- }
- /**
- * Undo what vbg_guest_mappings_init did.
- *
- * @gdev: The Guest extension device.
- */
- static void vbg_guest_mappings_exit(struct vbg_dev *gdev)
- {
- struct vmmdev_hypervisorinfo *req;
- int rc;
- if (!gdev->guest_mappings)
- return;
- /*
- * Tell the host that we're going to free the memory we reserved for
- * it, the free it up. (Leak the memory if anything goes wrong here.)
- */
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_HYPERVISOR_INFO);
- if (!req)
- return;
- req->hypervisor_start = 0;
- req->hypervisor_size = 0;
- rc = vbg_req_perform(gdev, req);
- vbg_req_free(req, sizeof(*req));
- if (rc < 0) {
- vbg_err("%s error: %d\n", __func__, rc);
- return;
- }
- vunmap(gdev->guest_mappings);
- gdev->guest_mappings = NULL;
- __free_page(gdev->guest_mappings_dummy_page);
- gdev->guest_mappings_dummy_page = NULL;
- }
- /**
- * Report the guest information to the host.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- */
- static int vbg_report_guest_info(struct vbg_dev *gdev)
- {
- /*
- * Allocate and fill in the two guest info reports.
- */
- struct vmmdev_guest_info *req1 = NULL;
- struct vmmdev_guest_info2 *req2 = NULL;
- int rc, ret = -ENOMEM;
- req1 = vbg_req_alloc(sizeof(*req1), VMMDEVREQ_REPORT_GUEST_INFO);
- req2 = vbg_req_alloc(sizeof(*req2), VMMDEVREQ_REPORT_GUEST_INFO2);
- if (!req1 || !req2)
- goto out_free;
- req1->interface_version = VMMDEV_VERSION;
- req1->os_type = VMMDEV_OSTYPE_LINUX26;
- #if __BITS_PER_LONG == 64
- req1->os_type |= VMMDEV_OSTYPE_X64;
- #endif
- req2->additions_major = VBG_VERSION_MAJOR;
- req2->additions_minor = VBG_VERSION_MINOR;
- req2->additions_build = VBG_VERSION_BUILD;
- req2->additions_revision = VBG_SVN_REV;
- /* (no features defined yet) */
- req2->additions_features = 0;
- strlcpy(req2->name, VBG_VERSION_STRING,
- sizeof(req2->name));
- /*
- * There are two protocols here:
- * 1. INFO2 + INFO1. Supported by >=3.2.51.
- * 2. INFO1 and optionally INFO2. The old protocol.
- *
- * We try protocol 2 first. It will fail with VERR_NOT_SUPPORTED
- * if not supported by the VMMDev (message ordering requirement).
- */
- rc = vbg_req_perform(gdev, req2);
- if (rc >= 0) {
- rc = vbg_req_perform(gdev, req1);
- } else if (rc == VERR_NOT_SUPPORTED || rc == VERR_NOT_IMPLEMENTED) {
- rc = vbg_req_perform(gdev, req1);
- if (rc >= 0) {
- rc = vbg_req_perform(gdev, req2);
- if (rc == VERR_NOT_IMPLEMENTED)
- rc = VINF_SUCCESS;
- }
- }
- ret = vbg_status_code_to_errno(rc);
- out_free:
- vbg_req_free(req2, sizeof(*req2));
- vbg_req_free(req1, sizeof(*req1));
- return ret;
- }
- /**
- * Report the guest driver status to the host.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @active: Flag whether the driver is now active or not.
- */
- static int vbg_report_driver_status(struct vbg_dev *gdev, bool active)
- {
- struct vmmdev_guest_status *req;
- int rc;
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_REPORT_GUEST_STATUS);
- if (!req)
- return -ENOMEM;
- req->facility = VBOXGUEST_FACILITY_TYPE_VBOXGUEST_DRIVER;
- if (active)
- req->status = VBOXGUEST_FACILITY_STATUS_ACTIVE;
- else
- req->status = VBOXGUEST_FACILITY_STATUS_INACTIVE;
- req->flags = 0;
- rc = vbg_req_perform(gdev, req);
- if (rc == VERR_NOT_IMPLEMENTED) /* Compatibility with older hosts. */
- rc = VINF_SUCCESS;
- vbg_req_free(req, sizeof(*req));
- return vbg_status_code_to_errno(rc);
- }
- /**
- * Inflate the balloon by one chunk. The caller owns the balloon mutex.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @chunk_idx: Index of the chunk.
- */
- static int vbg_balloon_inflate(struct vbg_dev *gdev, u32 chunk_idx)
- {
- struct vmmdev_memballoon_change *req = gdev->mem_balloon.change_req;
- struct page **pages;
- int i, rc, ret;
- pages = kmalloc_array(VMMDEV_MEMORY_BALLOON_CHUNK_PAGES,
- sizeof(*pages),
- GFP_KERNEL | __GFP_NOWARN);
- if (!pages)
- return -ENOMEM;
- req->header.size = sizeof(*req);
- req->inflate = true;
- req->pages = VMMDEV_MEMORY_BALLOON_CHUNK_PAGES;
- for (i = 0; i < VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; i++) {
- pages[i] = alloc_page(GFP_KERNEL | __GFP_NOWARN);
- if (!pages[i]) {
- ret = -ENOMEM;
- goto out_error;
- }
- req->phys_page[i] = page_to_phys(pages[i]);
- }
- rc = vbg_req_perform(gdev, req);
- if (rc < 0) {
- vbg_err("%s error, rc: %d\n", __func__, rc);
- ret = vbg_status_code_to_errno(rc);
- goto out_error;
- }
- gdev->mem_balloon.pages[chunk_idx] = pages;
- return 0;
- out_error:
- while (--i >= 0)
- __free_page(pages[i]);
- kfree(pages);
- return ret;
- }
- /**
- * Deflate the balloon by one chunk. The caller owns the balloon mutex.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @chunk_idx: Index of the chunk.
- */
- static int vbg_balloon_deflate(struct vbg_dev *gdev, u32 chunk_idx)
- {
- struct vmmdev_memballoon_change *req = gdev->mem_balloon.change_req;
- struct page **pages = gdev->mem_balloon.pages[chunk_idx];
- int i, rc;
- req->header.size = sizeof(*req);
- req->inflate = false;
- req->pages = VMMDEV_MEMORY_BALLOON_CHUNK_PAGES;
- for (i = 0; i < VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; i++)
- req->phys_page[i] = page_to_phys(pages[i]);
- rc = vbg_req_perform(gdev, req);
- if (rc < 0) {
- vbg_err("%s error, rc: %d\n", __func__, rc);
- return vbg_status_code_to_errno(rc);
- }
- for (i = 0; i < VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; i++)
- __free_page(pages[i]);
- kfree(pages);
- gdev->mem_balloon.pages[chunk_idx] = NULL;
- return 0;
- }
- /**
- * Respond to VMMDEV_EVENT_BALLOON_CHANGE_REQUEST events, query the size
- * the host wants the balloon to be and adjust accordingly.
- */
- static void vbg_balloon_work(struct work_struct *work)
- {
- struct vbg_dev *gdev =
- container_of(work, struct vbg_dev, mem_balloon.work);
- struct vmmdev_memballoon_info *req = gdev->mem_balloon.get_req;
- u32 i, chunks;
- int rc, ret;
- /*
- * Setting this bit means that we request the value from the host and
- * change the guest memory balloon according to the returned value.
- */
- req->event_ack = VMMDEV_EVENT_BALLOON_CHANGE_REQUEST;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0) {
- vbg_err("%s error, rc: %d)\n", __func__, rc);
- return;
- }
- /*
- * The host always returns the same maximum amount of chunks, so
- * we do this once.
- */
- if (!gdev->mem_balloon.max_chunks) {
- gdev->mem_balloon.pages =
- devm_kcalloc(gdev->dev, req->phys_mem_chunks,
- sizeof(struct page **), GFP_KERNEL);
- if (!gdev->mem_balloon.pages)
- return;
- gdev->mem_balloon.max_chunks = req->phys_mem_chunks;
- }
- chunks = req->balloon_chunks;
- if (chunks > gdev->mem_balloon.max_chunks) {
- vbg_err("%s: illegal balloon size %u (max=%u)\n",
- __func__, chunks, gdev->mem_balloon.max_chunks);
- return;
- }
- if (chunks > gdev->mem_balloon.chunks) {
- /* inflate */
- for (i = gdev->mem_balloon.chunks; i < chunks; i++) {
- ret = vbg_balloon_inflate(gdev, i);
- if (ret < 0)
- return;
- gdev->mem_balloon.chunks++;
- }
- } else {
- /* deflate */
- for (i = gdev->mem_balloon.chunks; i-- > chunks;) {
- ret = vbg_balloon_deflate(gdev, i);
- if (ret < 0)
- return;
- gdev->mem_balloon.chunks--;
- }
- }
- }
- /**
- * Callback for heartbeat timer.
- */
- static void vbg_heartbeat_timer(struct timer_list *t)
- {
- struct vbg_dev *gdev = from_timer(gdev, t, heartbeat_timer);
- vbg_req_perform(gdev, gdev->guest_heartbeat_req);
- mod_timer(&gdev->heartbeat_timer,
- msecs_to_jiffies(gdev->heartbeat_interval_ms));
- }
- /**
- * Configure the host to check guest's heartbeat
- * and get heartbeat interval from the host.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @enabled: Set true to enable guest heartbeat checks on host.
- */
- static int vbg_heartbeat_host_config(struct vbg_dev *gdev, bool enabled)
- {
- struct vmmdev_heartbeat *req;
- int rc;
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_HEARTBEAT_CONFIGURE);
- if (!req)
- return -ENOMEM;
- req->enabled = enabled;
- req->interval_ns = 0;
- rc = vbg_req_perform(gdev, req);
- do_div(req->interval_ns, 1000000); /* ns -> ms */
- gdev->heartbeat_interval_ms = req->interval_ns;
- vbg_req_free(req, sizeof(*req));
- return vbg_status_code_to_errno(rc);
- }
- /**
- * Initializes the heartbeat timer. This feature may be disabled by the host.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- */
- static int vbg_heartbeat_init(struct vbg_dev *gdev)
- {
- int ret;
- /* Make sure that heartbeat checking is disabled if we fail. */
- ret = vbg_heartbeat_host_config(gdev, false);
- if (ret < 0)
- return ret;
- ret = vbg_heartbeat_host_config(gdev, true);
- if (ret < 0)
- return ret;
- gdev->guest_heartbeat_req = vbg_req_alloc(
- sizeof(*gdev->guest_heartbeat_req),
- VMMDEVREQ_GUEST_HEARTBEAT);
- if (!gdev->guest_heartbeat_req)
- return -ENOMEM;
- vbg_info("%s: Setting up heartbeat to trigger every %d milliseconds\n",
- __func__, gdev->heartbeat_interval_ms);
- mod_timer(&gdev->heartbeat_timer, 0);
- return 0;
- }
- /**
- * Cleanup hearbeat code, stop HB timer and disable host heartbeat checking.
- * @gdev: The Guest extension device.
- */
- static void vbg_heartbeat_exit(struct vbg_dev *gdev)
- {
- del_timer_sync(&gdev->heartbeat_timer);
- vbg_heartbeat_host_config(gdev, false);
- vbg_req_free(gdev->guest_heartbeat_req,
- sizeof(*gdev->guest_heartbeat_req));
- }
- /**
- * Applies a change to the bit usage tracker.
- * Return: true if the mask changed, false if not.
- * @tracker: The bit usage tracker.
- * @changed: The bits to change.
- * @previous: The previous value of the bits.
- */
- static bool vbg_track_bit_usage(struct vbg_bit_usage_tracker *tracker,
- u32 changed, u32 previous)
- {
- bool global_change = false;
- while (changed) {
- u32 bit = ffs(changed) - 1;
- u32 bitmask = BIT(bit);
- if (bitmask & previous) {
- tracker->per_bit_usage[bit] -= 1;
- if (tracker->per_bit_usage[bit] == 0) {
- global_change = true;
- tracker->mask &= ~bitmask;
- }
- } else {
- tracker->per_bit_usage[bit] += 1;
- if (tracker->per_bit_usage[bit] == 1) {
- global_change = true;
- tracker->mask |= bitmask;
- }
- }
- changed &= ~bitmask;
- }
- return global_change;
- }
- /**
- * Init and termination worker for resetting the (host) event filter on the host
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @fixed_events: Fixed events (init time).
- */
- static int vbg_reset_host_event_filter(struct vbg_dev *gdev,
- u32 fixed_events)
- {
- struct vmmdev_mask *req;
- int rc;
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_CTL_GUEST_FILTER_MASK);
- if (!req)
- return -ENOMEM;
- req->not_mask = U32_MAX & ~fixed_events;
- req->or_mask = fixed_events;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0)
- vbg_err("%s error, rc: %d\n", __func__, rc);
- vbg_req_free(req, sizeof(*req));
- return vbg_status_code_to_errno(rc);
- }
- /**
- * Changes the event filter mask for the given session.
- *
- * This is called in response to VBG_IOCTL_CHANGE_FILTER_MASK as well as to
- * do session cleanup. Takes the session spinlock.
- *
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @session: The session.
- * @or_mask: The events to add.
- * @not_mask: The events to remove.
- * @session_termination: Set if we're called by the session cleanup code.
- * This tweaks the error handling so we perform
- * proper session cleanup even if the host
- * misbehaves.
- */
- static int vbg_set_session_event_filter(struct vbg_dev *gdev,
- struct vbg_session *session,
- u32 or_mask, u32 not_mask,
- bool session_termination)
- {
- struct vmmdev_mask *req;
- u32 changed, previous;
- int rc, ret = 0;
- /* Allocate a request buffer before taking the spinlock */
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_CTL_GUEST_FILTER_MASK);
- if (!req) {
- if (!session_termination)
- return -ENOMEM;
- /* Ignore allocation failure, we must do session cleanup. */
- }
- mutex_lock(&gdev->session_mutex);
- /* Apply the changes to the session mask. */
- previous = session->event_filter;
- session->event_filter |= or_mask;
- session->event_filter &= ~not_mask;
- /* If anything actually changed, update the global usage counters. */
- changed = previous ^ session->event_filter;
- if (!changed)
- goto out;
- vbg_track_bit_usage(&gdev->event_filter_tracker, changed, previous);
- or_mask = gdev->fixed_events | gdev->event_filter_tracker.mask;
- if (gdev->event_filter_host == or_mask || !req)
- goto out;
- gdev->event_filter_host = or_mask;
- req->or_mask = or_mask;
- req->not_mask = ~or_mask;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0) {
- ret = vbg_status_code_to_errno(rc);
- /* Failed, roll back (unless it's session termination time). */
- gdev->event_filter_host = U32_MAX;
- if (session_termination)
- goto out;
- vbg_track_bit_usage(&gdev->event_filter_tracker, changed,
- session->event_filter);
- session->event_filter = previous;
- }
- out:
- mutex_unlock(&gdev->session_mutex);
- vbg_req_free(req, sizeof(*req));
- return ret;
- }
- /**
- * Init and termination worker for set guest capabilities to zero on the host.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- */
- static int vbg_reset_host_capabilities(struct vbg_dev *gdev)
- {
- struct vmmdev_mask *req;
- int rc;
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_GUEST_CAPABILITIES);
- if (!req)
- return -ENOMEM;
- req->not_mask = U32_MAX;
- req->or_mask = 0;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0)
- vbg_err("%s error, rc: %d\n", __func__, rc);
- vbg_req_free(req, sizeof(*req));
- return vbg_status_code_to_errno(rc);
- }
- /**
- * Sets the guest capabilities for a session. Takes the session spinlock.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @session: The session.
- * @or_mask: The capabilities to add.
- * @not_mask: The capabilities to remove.
- * @session_termination: Set if we're called by the session cleanup code.
- * This tweaks the error handling so we perform
- * proper session cleanup even if the host
- * misbehaves.
- */
- static int vbg_set_session_capabilities(struct vbg_dev *gdev,
- struct vbg_session *session,
- u32 or_mask, u32 not_mask,
- bool session_termination)
- {
- struct vmmdev_mask *req;
- u32 changed, previous;
- int rc, ret = 0;
- /* Allocate a request buffer before taking the spinlock */
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_GUEST_CAPABILITIES);
- if (!req) {
- if (!session_termination)
- return -ENOMEM;
- /* Ignore allocation failure, we must do session cleanup. */
- }
- mutex_lock(&gdev->session_mutex);
- /* Apply the changes to the session mask. */
- previous = session->guest_caps;
- session->guest_caps |= or_mask;
- session->guest_caps &= ~not_mask;
- /* If anything actually changed, update the global usage counters. */
- changed = previous ^ session->guest_caps;
- if (!changed)
- goto out;
- vbg_track_bit_usage(&gdev->guest_caps_tracker, changed, previous);
- or_mask = gdev->guest_caps_tracker.mask;
- if (gdev->guest_caps_host == or_mask || !req)
- goto out;
- gdev->guest_caps_host = or_mask;
- req->or_mask = or_mask;
- req->not_mask = ~or_mask;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0) {
- ret = vbg_status_code_to_errno(rc);
- /* Failed, roll back (unless it's session termination time). */
- gdev->guest_caps_host = U32_MAX;
- if (session_termination)
- goto out;
- vbg_track_bit_usage(&gdev->guest_caps_tracker, changed,
- session->guest_caps);
- session->guest_caps = previous;
- }
- out:
- mutex_unlock(&gdev->session_mutex);
- vbg_req_free(req, sizeof(*req));
- return ret;
- }
- /**
- * vbg_query_host_version get the host feature mask and version information.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- */
- static int vbg_query_host_version(struct vbg_dev *gdev)
- {
- struct vmmdev_host_version *req;
- int rc, ret;
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_GET_HOST_VERSION);
- if (!req)
- return -ENOMEM;
- rc = vbg_req_perform(gdev, req);
- ret = vbg_status_code_to_errno(rc);
- if (ret) {
- vbg_err("%s error: %d\n", __func__, rc);
- goto out;
- }
- snprintf(gdev->host_version, sizeof(gdev->host_version), "%u.%u.%ur%u",
- req->major, req->minor, req->build, req->revision);
- gdev->host_features = req->features;
- vbg_info("vboxguest: host-version: %s %#x\n", gdev->host_version,
- gdev->host_features);
- if (!(req->features & VMMDEV_HVF_HGCM_PHYS_PAGE_LIST)) {
- vbg_err("vboxguest: Error host too old (does not support page-lists)\n");
- ret = -ENODEV;
- }
- out:
- vbg_req_free(req, sizeof(*req));
- return ret;
- }
- /**
- * Initializes the VBoxGuest device extension when the
- * device driver is loaded.
- *
- * The native code locates the VMMDev on the PCI bus and retrieve
- * the MMIO and I/O port ranges, this function will take care of
- * mapping the MMIO memory (if present). Upon successful return
- * the native code should set up the interrupt handler.
- *
- * Return: 0 or negative errno value.
- *
- * @gdev: The Guest extension device.
- * @fixed_events: Events that will be enabled upon init and no client
- * will ever be allowed to mask.
- */
- int vbg_core_init(struct vbg_dev *gdev, u32 fixed_events)
- {
- int ret = -ENOMEM;
- gdev->fixed_events = fixed_events | VMMDEV_EVENT_HGCM;
- gdev->event_filter_host = U32_MAX; /* forces a report */
- gdev->guest_caps_host = U32_MAX; /* forces a report */
- init_waitqueue_head(&gdev->event_wq);
- init_waitqueue_head(&gdev->hgcm_wq);
- spin_lock_init(&gdev->event_spinlock);
- mutex_init(&gdev->session_mutex);
- mutex_init(&gdev->cancel_req_mutex);
- timer_setup(&gdev->heartbeat_timer, vbg_heartbeat_timer, 0);
- INIT_WORK(&gdev->mem_balloon.work, vbg_balloon_work);
- gdev->mem_balloon.get_req =
- vbg_req_alloc(sizeof(*gdev->mem_balloon.get_req),
- VMMDEVREQ_GET_MEMBALLOON_CHANGE_REQ);
- gdev->mem_balloon.change_req =
- vbg_req_alloc(sizeof(*gdev->mem_balloon.change_req),
- VMMDEVREQ_CHANGE_MEMBALLOON);
- gdev->cancel_req =
- vbg_req_alloc(sizeof(*(gdev->cancel_req)),
- VMMDEVREQ_HGCM_CANCEL2);
- gdev->ack_events_req =
- vbg_req_alloc(sizeof(*gdev->ack_events_req),
- VMMDEVREQ_ACKNOWLEDGE_EVENTS);
- gdev->mouse_status_req =
- vbg_req_alloc(sizeof(*gdev->mouse_status_req),
- VMMDEVREQ_GET_MOUSE_STATUS);
- if (!gdev->mem_balloon.get_req || !gdev->mem_balloon.change_req ||
- !gdev->cancel_req || !gdev->ack_events_req ||
- !gdev->mouse_status_req)
- goto err_free_reqs;
- ret = vbg_query_host_version(gdev);
- if (ret)
- goto err_free_reqs;
- ret = vbg_report_guest_info(gdev);
- if (ret) {
- vbg_err("vboxguest: vbg_report_guest_info error: %d\n", ret);
- goto err_free_reqs;
- }
- ret = vbg_reset_host_event_filter(gdev, gdev->fixed_events);
- if (ret) {
- vbg_err("vboxguest: Error setting fixed event filter: %d\n",
- ret);
- goto err_free_reqs;
- }
- ret = vbg_reset_host_capabilities(gdev);
- if (ret) {
- vbg_err("vboxguest: Error clearing guest capabilities: %d\n",
- ret);
- goto err_free_reqs;
- }
- ret = vbg_core_set_mouse_status(gdev, 0);
- if (ret) {
- vbg_err("vboxguest: Error clearing mouse status: %d\n", ret);
- goto err_free_reqs;
- }
- /* These may fail without requiring the driver init to fail. */
- vbg_guest_mappings_init(gdev);
- vbg_heartbeat_init(gdev);
- /* All Done! */
- ret = vbg_report_driver_status(gdev, true);
- if (ret < 0)
- vbg_err("vboxguest: Error reporting driver status: %d\n", ret);
- return 0;
- err_free_reqs:
- vbg_req_free(gdev->mouse_status_req,
- sizeof(*gdev->mouse_status_req));
- vbg_req_free(gdev->ack_events_req,
- sizeof(*gdev->ack_events_req));
- vbg_req_free(gdev->cancel_req,
- sizeof(*gdev->cancel_req));
- vbg_req_free(gdev->mem_balloon.change_req,
- sizeof(*gdev->mem_balloon.change_req));
- vbg_req_free(gdev->mem_balloon.get_req,
- sizeof(*gdev->mem_balloon.get_req));
- return ret;
- }
- /**
- * Call this on exit to clean-up vboxguest-core managed resources.
- *
- * The native code should call this before the driver is loaded,
- * but don't call this on shutdown.
- * @gdev: The Guest extension device.
- */
- void vbg_core_exit(struct vbg_dev *gdev)
- {
- vbg_heartbeat_exit(gdev);
- vbg_guest_mappings_exit(gdev);
- /* Clear the host flags (mouse status etc). */
- vbg_reset_host_event_filter(gdev, 0);
- vbg_reset_host_capabilities(gdev);
- vbg_core_set_mouse_status(gdev, 0);
- vbg_req_free(gdev->mouse_status_req,
- sizeof(*gdev->mouse_status_req));
- vbg_req_free(gdev->ack_events_req,
- sizeof(*gdev->ack_events_req));
- vbg_req_free(gdev->cancel_req,
- sizeof(*gdev->cancel_req));
- vbg_req_free(gdev->mem_balloon.change_req,
- sizeof(*gdev->mem_balloon.change_req));
- vbg_req_free(gdev->mem_balloon.get_req,
- sizeof(*gdev->mem_balloon.get_req));
- }
- /**
- * Creates a VBoxGuest user session.
- *
- * vboxguest_linux.c calls this when userspace opens the char-device.
- * Return: A pointer to the new session or an ERR_PTR on error.
- * @gdev: The Guest extension device.
- * @user: Set if this is a session for the vboxuser device.
- */
- struct vbg_session *vbg_core_open_session(struct vbg_dev *gdev, bool user)
- {
- struct vbg_session *session;
- session = kzalloc(sizeof(*session), GFP_KERNEL);
- if (!session)
- return ERR_PTR(-ENOMEM);
- session->gdev = gdev;
- session->user_session = user;
- return session;
- }
- /**
- * Closes a VBoxGuest session.
- * @session: The session to close (and free).
- */
- void vbg_core_close_session(struct vbg_session *session)
- {
- struct vbg_dev *gdev = session->gdev;
- int i, rc;
- vbg_set_session_capabilities(gdev, session, 0, U32_MAX, true);
- vbg_set_session_event_filter(gdev, session, 0, U32_MAX, true);
- for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++) {
- if (!session->hgcm_client_ids[i])
- continue;
- vbg_hgcm_disconnect(gdev, session->hgcm_client_ids[i], &rc);
- }
- kfree(session);
- }
- static int vbg_ioctl_chk(struct vbg_ioctl_hdr *hdr, size_t in_size,
- size_t out_size)
- {
- if (hdr->size_in != (sizeof(*hdr) + in_size) ||
- hdr->size_out != (sizeof(*hdr) + out_size))
- return -EINVAL;
- return 0;
- }
- static int vbg_ioctl_driver_version_info(
- struct vbg_ioctl_driver_version_info *info)
- {
- const u16 vbg_maj_version = VBG_IOC_VERSION >> 16;
- u16 min_maj_version, req_maj_version;
- if (vbg_ioctl_chk(&info->hdr, sizeof(info->u.in), sizeof(info->u.out)))
- return -EINVAL;
- req_maj_version = info->u.in.req_version >> 16;
- min_maj_version = info->u.in.min_version >> 16;
- if (info->u.in.min_version > info->u.in.req_version ||
- min_maj_version != req_maj_version)
- return -EINVAL;
- if (info->u.in.min_version <= VBG_IOC_VERSION &&
- min_maj_version == vbg_maj_version) {
- info->u.out.session_version = VBG_IOC_VERSION;
- } else {
- info->u.out.session_version = U32_MAX;
- info->hdr.rc = VERR_VERSION_MISMATCH;
- }
- info->u.out.driver_version = VBG_IOC_VERSION;
- info->u.out.driver_revision = 0;
- info->u.out.reserved1 = 0;
- info->u.out.reserved2 = 0;
- return 0;
- }
- static bool vbg_wait_event_cond(struct vbg_dev *gdev,
- struct vbg_session *session,
- u32 event_mask)
- {
- unsigned long flags;
- bool wakeup;
- u32 events;
- spin_lock_irqsave(&gdev->event_spinlock, flags);
- events = gdev->pending_events & event_mask;
- wakeup = events || session->cancel_waiters;
- spin_unlock_irqrestore(&gdev->event_spinlock, flags);
- return wakeup;
- }
- /* Must be called with the event_lock held */
- static u32 vbg_consume_events_locked(struct vbg_dev *gdev,
- struct vbg_session *session,
- u32 event_mask)
- {
- u32 events = gdev->pending_events & event_mask;
- gdev->pending_events &= ~events;
- return events;
- }
- static int vbg_ioctl_wait_for_events(struct vbg_dev *gdev,
- struct vbg_session *session,
- struct vbg_ioctl_wait_for_events *wait)
- {
- u32 timeout_ms = wait->u.in.timeout_ms;
- u32 event_mask = wait->u.in.events;
- unsigned long flags;
- long timeout;
- int ret = 0;
- if (vbg_ioctl_chk(&wait->hdr, sizeof(wait->u.in), sizeof(wait->u.out)))
- return -EINVAL;
- if (timeout_ms == U32_MAX)
- timeout = MAX_SCHEDULE_TIMEOUT;
- else
- timeout = msecs_to_jiffies(timeout_ms);
- wait->u.out.events = 0;
- do {
- timeout = wait_event_interruptible_timeout(
- gdev->event_wq,
- vbg_wait_event_cond(gdev, session, event_mask),
- timeout);
- spin_lock_irqsave(&gdev->event_spinlock, flags);
- if (timeout < 0 || session->cancel_waiters) {
- ret = -EINTR;
- } else if (timeout == 0) {
- ret = -ETIMEDOUT;
- } else {
- wait->u.out.events =
- vbg_consume_events_locked(gdev, session, event_mask);
- }
- spin_unlock_irqrestore(&gdev->event_spinlock, flags);
- /*
- * Someone else may have consumed the event(s) first, in
- * which case we go back to waiting.
- */
- } while (ret == 0 && wait->u.out.events == 0);
- return ret;
- }
- static int vbg_ioctl_interrupt_all_wait_events(struct vbg_dev *gdev,
- struct vbg_session *session,
- struct vbg_ioctl_hdr *hdr)
- {
- unsigned long flags;
- if (hdr->size_in != sizeof(*hdr) || hdr->size_out != sizeof(*hdr))
- return -EINVAL;
- spin_lock_irqsave(&gdev->event_spinlock, flags);
- session->cancel_waiters = true;
- spin_unlock_irqrestore(&gdev->event_spinlock, flags);
- wake_up(&gdev->event_wq);
- return 0;
- }
- /**
- * Checks if the VMM request is allowed in the context of the given session.
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @session: The calling session.
- * @req: The request.
- */
- static int vbg_req_allowed(struct vbg_dev *gdev, struct vbg_session *session,
- const struct vmmdev_request_header *req)
- {
- const struct vmmdev_guest_status *guest_status;
- bool trusted_apps_only;
- switch (req->request_type) {
- /* Trusted users apps only. */
- case VMMDEVREQ_QUERY_CREDENTIALS:
- case VMMDEVREQ_REPORT_CREDENTIALS_JUDGEMENT:
- case VMMDEVREQ_REGISTER_SHARED_MODULE:
- case VMMDEVREQ_UNREGISTER_SHARED_MODULE:
- case VMMDEVREQ_WRITE_COREDUMP:
- case VMMDEVREQ_GET_CPU_HOTPLUG_REQ:
- case VMMDEVREQ_SET_CPU_HOTPLUG_STATUS:
- case VMMDEVREQ_CHECK_SHARED_MODULES:
- case VMMDEVREQ_GET_PAGE_SHARING_STATUS:
- case VMMDEVREQ_DEBUG_IS_PAGE_SHARED:
- case VMMDEVREQ_REPORT_GUEST_STATS:
- case VMMDEVREQ_REPORT_GUEST_USER_STATE:
- case VMMDEVREQ_GET_STATISTICS_CHANGE_REQ:
- trusted_apps_only = true;
- break;
- /* Anyone. */
- case VMMDEVREQ_GET_MOUSE_STATUS:
- case VMMDEVREQ_SET_MOUSE_STATUS:
- case VMMDEVREQ_SET_POINTER_SHAPE:
- case VMMDEVREQ_GET_HOST_VERSION:
- case VMMDEVREQ_IDLE:
- case VMMDEVREQ_GET_HOST_TIME:
- case VMMDEVREQ_SET_POWER_STATUS:
- case VMMDEVREQ_ACKNOWLEDGE_EVENTS:
- case VMMDEVREQ_CTL_GUEST_FILTER_MASK:
- case VMMDEVREQ_REPORT_GUEST_STATUS:
- case VMMDEVREQ_GET_DISPLAY_CHANGE_REQ:
- case VMMDEVREQ_VIDEMODE_SUPPORTED:
- case VMMDEVREQ_GET_HEIGHT_REDUCTION:
- case VMMDEVREQ_GET_DISPLAY_CHANGE_REQ2:
- case VMMDEVREQ_VIDEMODE_SUPPORTED2:
- case VMMDEVREQ_VIDEO_ACCEL_ENABLE:
- case VMMDEVREQ_VIDEO_ACCEL_FLUSH:
- case VMMDEVREQ_VIDEO_SET_VISIBLE_REGION:
- case VMMDEVREQ_GET_DISPLAY_CHANGE_REQEX:
- case VMMDEVREQ_GET_SEAMLESS_CHANGE_REQ:
- case VMMDEVREQ_GET_VRDPCHANGE_REQ:
- case VMMDEVREQ_LOG_STRING:
- case VMMDEVREQ_GET_SESSION_ID:
- trusted_apps_only = false;
- break;
- /* Depends on the request parameters... */
- case VMMDEVREQ_REPORT_GUEST_CAPABILITIES:
- guest_status = (const struct vmmdev_guest_status *)req;
- switch (guest_status->facility) {
- case VBOXGUEST_FACILITY_TYPE_ALL:
- case VBOXGUEST_FACILITY_TYPE_VBOXGUEST_DRIVER:
- vbg_err("Denying userspace vmm report guest cap. call facility %#08x\n",
- guest_status->facility);
- return -EPERM;
- case VBOXGUEST_FACILITY_TYPE_VBOX_SERVICE:
- trusted_apps_only = true;
- break;
- case VBOXGUEST_FACILITY_TYPE_VBOX_TRAY_CLIENT:
- case VBOXGUEST_FACILITY_TYPE_SEAMLESS:
- case VBOXGUEST_FACILITY_TYPE_GRAPHICS:
- default:
- trusted_apps_only = false;
- break;
- }
- break;
- /* Anything else is not allowed. */
- default:
- vbg_err("Denying userspace vmm call type %#08x\n",
- req->request_type);
- return -EPERM;
- }
- if (trusted_apps_only && session->user_session) {
- vbg_err("Denying userspace vmm call type %#08x through vboxuser device node\n",
- req->request_type);
- return -EPERM;
- }
- return 0;
- }
- static int vbg_ioctl_vmmrequest(struct vbg_dev *gdev,
- struct vbg_session *session, void *data)
- {
- struct vbg_ioctl_hdr *hdr = data;
- int ret;
- if (hdr->size_in != hdr->size_out)
- return -EINVAL;
- if (hdr->size_in > VMMDEV_MAX_VMMDEVREQ_SIZE)
- return -E2BIG;
- if (hdr->type == VBG_IOCTL_HDR_TYPE_DEFAULT)
- return -EINVAL;
- ret = vbg_req_allowed(gdev, session, data);
- if (ret < 0)
- return ret;
- vbg_req_perform(gdev, data);
- WARN_ON(hdr->rc == VINF_HGCM_ASYNC_EXECUTE);
- return 0;
- }
- static int vbg_ioctl_hgcm_connect(struct vbg_dev *gdev,
- struct vbg_session *session,
- struct vbg_ioctl_hgcm_connect *conn)
- {
- u32 client_id;
- int i, ret;
- if (vbg_ioctl_chk(&conn->hdr, sizeof(conn->u.in), sizeof(conn->u.out)))
- return -EINVAL;
- /* Find a free place in the sessions clients array and claim it */
- mutex_lock(&gdev->session_mutex);
- for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++) {
- if (!session->hgcm_client_ids[i]) {
- session->hgcm_client_ids[i] = U32_MAX;
- break;
- }
- }
- mutex_unlock(&gdev->session_mutex);
- if (i >= ARRAY_SIZE(session->hgcm_client_ids))
- return -EMFILE;
- ret = vbg_hgcm_connect(gdev, &conn->u.in.loc, &client_id,
- &conn->hdr.rc);
- mutex_lock(&gdev->session_mutex);
- if (ret == 0 && conn->hdr.rc >= 0) {
- conn->u.out.client_id = client_id;
- session->hgcm_client_ids[i] = client_id;
- } else {
- conn->u.out.client_id = 0;
- session->hgcm_client_ids[i] = 0;
- }
- mutex_unlock(&gdev->session_mutex);
- return ret;
- }
- static int vbg_ioctl_hgcm_disconnect(struct vbg_dev *gdev,
- struct vbg_session *session,
- struct vbg_ioctl_hgcm_disconnect *disconn)
- {
- u32 client_id;
- int i, ret;
- if (vbg_ioctl_chk(&disconn->hdr, sizeof(disconn->u.in), 0))
- return -EINVAL;
- client_id = disconn->u.in.client_id;
- if (client_id == 0 || client_id == U32_MAX)
- return -EINVAL;
- mutex_lock(&gdev->session_mutex);
- for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++) {
- if (session->hgcm_client_ids[i] == client_id) {
- session->hgcm_client_ids[i] = U32_MAX;
- break;
- }
- }
- mutex_unlock(&gdev->session_mutex);
- if (i >= ARRAY_SIZE(session->hgcm_client_ids))
- return -EINVAL;
- ret = vbg_hgcm_disconnect(gdev, client_id, &disconn->hdr.rc);
- mutex_lock(&gdev->session_mutex);
- if (ret == 0 && disconn->hdr.rc >= 0)
- session->hgcm_client_ids[i] = 0;
- else
- session->hgcm_client_ids[i] = client_id;
- mutex_unlock(&gdev->session_mutex);
- return ret;
- }
- static bool vbg_param_valid(enum vmmdev_hgcm_function_parameter_type type)
- {
- switch (type) {
- case VMMDEV_HGCM_PARM_TYPE_32BIT:
- case VMMDEV_HGCM_PARM_TYPE_64BIT:
- case VMMDEV_HGCM_PARM_TYPE_LINADDR:
- case VMMDEV_HGCM_PARM_TYPE_LINADDR_IN:
- case VMMDEV_HGCM_PARM_TYPE_LINADDR_OUT:
- return true;
- default:
- return false;
- }
- }
- static int vbg_ioctl_hgcm_call(struct vbg_dev *gdev,
- struct vbg_session *session, bool f32bit,
- struct vbg_ioctl_hgcm_call *call)
- {
- size_t actual_size;
- u32 client_id;
- int i, ret;
- if (call->hdr.size_in < sizeof(*call))
- return -EINVAL;
- if (call->hdr.size_in != call->hdr.size_out)
- return -EINVAL;
- if (call->parm_count > VMMDEV_HGCM_MAX_PARMS)
- return -E2BIG;
- client_id = call->client_id;
- if (client_id == 0 || client_id == U32_MAX)
- return -EINVAL;
- actual_size = sizeof(*call);
- if (f32bit)
- actual_size += call->parm_count *
- sizeof(struct vmmdev_hgcm_function_parameter32);
- else
- actual_size += call->parm_count *
- sizeof(struct vmmdev_hgcm_function_parameter);
- if (call->hdr.size_in < actual_size) {
- vbg_debug("VBG_IOCTL_HGCM_CALL: hdr.size_in %d required size is %zd\n",
- call->hdr.size_in, actual_size);
- return -EINVAL;
- }
- call->hdr.size_out = actual_size;
- /* Validate parameter types */
- if (f32bit) {
- struct vmmdev_hgcm_function_parameter32 *parm =
- VBG_IOCTL_HGCM_CALL_PARMS32(call);
- for (i = 0; i < call->parm_count; i++)
- if (!vbg_param_valid(parm[i].type))
- return -EINVAL;
- } else {
- struct vmmdev_hgcm_function_parameter *parm =
- VBG_IOCTL_HGCM_CALL_PARMS(call);
- for (i = 0; i < call->parm_count; i++)
- if (!vbg_param_valid(parm[i].type))
- return -EINVAL;
- }
- /*
- * Validate the client id.
- */
- mutex_lock(&gdev->session_mutex);
- for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++)
- if (session->hgcm_client_ids[i] == client_id)
- break;
- mutex_unlock(&gdev->session_mutex);
- if (i >= ARRAY_SIZE(session->hgcm_client_ids)) {
- vbg_debug("VBG_IOCTL_HGCM_CALL: INVALID handle. u32Client=%#08x\n",
- client_id);
- return -EINVAL;
- }
- if (IS_ENABLED(CONFIG_COMPAT) && f32bit)
- ret = vbg_hgcm_call32(gdev, client_id,
- call->function, call->timeout_ms,
- VBG_IOCTL_HGCM_CALL_PARMS32(call),
- call->parm_count, &call->hdr.rc);
- else
- ret = vbg_hgcm_call(gdev, client_id,
- call->function, call->timeout_ms,
- VBG_IOCTL_HGCM_CALL_PARMS(call),
- call->parm_count, &call->hdr.rc);
- if (ret == -E2BIG) {
- /* E2BIG needs to be reported through the hdr.rc field. */
- call->hdr.rc = VERR_OUT_OF_RANGE;
- ret = 0;
- }
- if (ret && ret != -EINTR && ret != -ETIMEDOUT)
- vbg_err("VBG_IOCTL_HGCM_CALL error: %d\n", ret);
- return ret;
- }
- static int vbg_ioctl_log(struct vbg_ioctl_log *log)
- {
- if (log->hdr.size_out != sizeof(log->hdr))
- return -EINVAL;
- vbg_info("%.*s", (int)(log->hdr.size_in - sizeof(log->hdr)),
- log->u.in.msg);
- return 0;
- }
- static int vbg_ioctl_change_filter_mask(struct vbg_dev *gdev,
- struct vbg_session *session,
- struct vbg_ioctl_change_filter *filter)
- {
- u32 or_mask, not_mask;
- if (vbg_ioctl_chk(&filter->hdr, sizeof(filter->u.in), 0))
- return -EINVAL;
- or_mask = filter->u.in.or_mask;
- not_mask = filter->u.in.not_mask;
- if ((or_mask | not_mask) & ~VMMDEV_EVENT_VALID_EVENT_MASK)
- return -EINVAL;
- return vbg_set_session_event_filter(gdev, session, or_mask, not_mask,
- false);
- }
- static int vbg_ioctl_change_guest_capabilities(struct vbg_dev *gdev,
- struct vbg_session *session, struct vbg_ioctl_set_guest_caps *caps)
- {
- u32 or_mask, not_mask;
- int ret;
- if (vbg_ioctl_chk(&caps->hdr, sizeof(caps->u.in), sizeof(caps->u.out)))
- return -EINVAL;
- or_mask = caps->u.in.or_mask;
- not_mask = caps->u.in.not_mask;
- if ((or_mask | not_mask) & ~VMMDEV_GUEST_CAPABILITIES_MASK)
- return -EINVAL;
- ret = vbg_set_session_capabilities(gdev, session, or_mask, not_mask,
- false);
- if (ret)
- return ret;
- caps->u.out.session_caps = session->guest_caps;
- caps->u.out.global_caps = gdev->guest_caps_host;
- return 0;
- }
- static int vbg_ioctl_check_balloon(struct vbg_dev *gdev,
- struct vbg_ioctl_check_balloon *balloon_info)
- {
- if (vbg_ioctl_chk(&balloon_info->hdr, 0, sizeof(balloon_info->u.out)))
- return -EINVAL;
- balloon_info->u.out.balloon_chunks = gdev->mem_balloon.chunks;
- /*
- * Under Linux we handle VMMDEV_EVENT_BALLOON_CHANGE_REQUEST
- * events entirely in the kernel, see vbg_core_isr().
- */
- balloon_info->u.out.handle_in_r3 = false;
- return 0;
- }
- static int vbg_ioctl_write_core_dump(struct vbg_dev *gdev,
- struct vbg_ioctl_write_coredump *dump)
- {
- struct vmmdev_write_core_dump *req;
- if (vbg_ioctl_chk(&dump->hdr, sizeof(dump->u.in), 0))
- return -EINVAL;
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_WRITE_COREDUMP);
- if (!req)
- return -ENOMEM;
- req->flags = dump->u.in.flags;
- dump->hdr.rc = vbg_req_perform(gdev, req);
- vbg_req_free(req, sizeof(*req));
- return 0;
- }
- /**
- * Common IOCtl for user to kernel communication.
- * Return: 0 or negative errno value.
- * @session: The client session.
- * @req: The requested function.
- * @data: The i/o data buffer, minimum size sizeof(struct vbg_ioctl_hdr).
- */
- int vbg_core_ioctl(struct vbg_session *session, unsigned int req, void *data)
- {
- unsigned int req_no_size = req & ~IOCSIZE_MASK;
- struct vbg_dev *gdev = session->gdev;
- struct vbg_ioctl_hdr *hdr = data;
- bool f32bit = false;
- hdr->rc = VINF_SUCCESS;
- if (!hdr->size_out)
- hdr->size_out = hdr->size_in;
- /*
- * hdr->version and hdr->size_in / hdr->size_out minimum size are
- * already checked by vbg_misc_device_ioctl().
- */
- /* For VMMDEV_REQUEST hdr->type != VBG_IOCTL_HDR_TYPE_DEFAULT */
- if (req_no_size == VBG_IOCTL_VMMDEV_REQUEST(0) ||
- req == VBG_IOCTL_VMMDEV_REQUEST_BIG ||
- req == VBG_IOCTL_VMMDEV_REQUEST_BIG_ALT)
- return vbg_ioctl_vmmrequest(gdev, session, data);
- if (hdr->type != VBG_IOCTL_HDR_TYPE_DEFAULT)
- return -EINVAL;
- /* Fixed size requests. */
- switch (req) {
- case VBG_IOCTL_DRIVER_VERSION_INFO:
- return vbg_ioctl_driver_version_info(data);
- case VBG_IOCTL_HGCM_CONNECT:
- return vbg_ioctl_hgcm_connect(gdev, session, data);
- case VBG_IOCTL_HGCM_DISCONNECT:
- return vbg_ioctl_hgcm_disconnect(gdev, session, data);
- case VBG_IOCTL_WAIT_FOR_EVENTS:
- return vbg_ioctl_wait_for_events(gdev, session, data);
- case VBG_IOCTL_INTERRUPT_ALL_WAIT_FOR_EVENTS:
- return vbg_ioctl_interrupt_all_wait_events(gdev, session, data);
- case VBG_IOCTL_CHANGE_FILTER_MASK:
- return vbg_ioctl_change_filter_mask(gdev, session, data);
- case VBG_IOCTL_CHANGE_GUEST_CAPABILITIES:
- return vbg_ioctl_change_guest_capabilities(gdev, session, data);
- case VBG_IOCTL_CHECK_BALLOON:
- return vbg_ioctl_check_balloon(gdev, data);
- case VBG_IOCTL_WRITE_CORE_DUMP:
- return vbg_ioctl_write_core_dump(gdev, data);
- }
- /* Variable sized requests. */
- switch (req_no_size) {
- #ifdef CONFIG_COMPAT
- case VBG_IOCTL_HGCM_CALL_32(0):
- f32bit = true;
- /* Fall through */
- #endif
- case VBG_IOCTL_HGCM_CALL(0):
- return vbg_ioctl_hgcm_call(gdev, session, f32bit, data);
- case VBG_IOCTL_LOG(0):
- case VBG_IOCTL_LOG_ALT(0):
- return vbg_ioctl_log(data);
- }
- vbg_debug("VGDrvCommonIoCtl: Unknown req %#08x\n", req);
- return -ENOTTY;
- }
- /**
- * Report guest supported mouse-features to the host.
- *
- * Return: 0 or negative errno value.
- * @gdev: The Guest extension device.
- * @features: The set of features to report to the host.
- */
- int vbg_core_set_mouse_status(struct vbg_dev *gdev, u32 features)
- {
- struct vmmdev_mouse_status *req;
- int rc;
- req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_MOUSE_STATUS);
- if (!req)
- return -ENOMEM;
- req->mouse_features = features;
- req->pointer_pos_x = 0;
- req->pointer_pos_y = 0;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0)
- vbg_err("%s error, rc: %d\n", __func__, rc);
- vbg_req_free(req, sizeof(*req));
- return vbg_status_code_to_errno(rc);
- }
- /** Core interrupt service routine. */
- irqreturn_t vbg_core_isr(int irq, void *dev_id)
- {
- struct vbg_dev *gdev = dev_id;
- struct vmmdev_events *req = gdev->ack_events_req;
- bool mouse_position_changed = false;
- unsigned long flags;
- u32 events = 0;
- int rc;
- if (!gdev->mmio->V.V1_04.have_events)
- return IRQ_NONE;
- /* Get and acknowlegde events. */
- req->header.rc = VERR_INTERNAL_ERROR;
- req->events = 0;
- rc = vbg_req_perform(gdev, req);
- if (rc < 0) {
- vbg_err("Error performing events req, rc: %d\n", rc);
- return IRQ_NONE;
- }
- events = req->events;
- if (events & VMMDEV_EVENT_MOUSE_POSITION_CHANGED) {
- mouse_position_changed = true;
- events &= ~VMMDEV_EVENT_MOUSE_POSITION_CHANGED;
- }
- if (events & VMMDEV_EVENT_HGCM) {
- wake_up(&gdev->hgcm_wq);
- events &= ~VMMDEV_EVENT_HGCM;
- }
- if (events & VMMDEV_EVENT_BALLOON_CHANGE_REQUEST) {
- schedule_work(&gdev->mem_balloon.work);
- events &= ~VMMDEV_EVENT_BALLOON_CHANGE_REQUEST;
- }
- if (events) {
- spin_lock_irqsave(&gdev->event_spinlock, flags);
- gdev->pending_events |= events;
- spin_unlock_irqrestore(&gdev->event_spinlock, flags);
- wake_up(&gdev->event_wq);
- }
- if (mouse_position_changed)
- vbg_linux_mouse_event(gdev);
- return IRQ_HANDLED;
- }
|