| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557 |
- // SPDX-License-Identifier: GPL-2.0-only
- /*
- * Copyright (C) 2020-2024 Intel Corporation
- */
- #include "ivpu_drv.h"
- #include "ivpu_hw.h"
- #include "ivpu_ipc.h"
- #include "ivpu_jsm_msg.h"
- const char *ivpu_jsm_msg_type_to_str(enum vpu_ipc_msg_type type)
- {
- #define IVPU_CASE_TO_STR(x) case x: return #x
- switch (type) {
- IVPU_CASE_TO_STR(VPU_JSM_MSG_UNKNOWN);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_RESET);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_PREEMPT);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_REGISTER_DB);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_UNREGISTER_DB);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_QUERY_ENGINE_HB);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL_COUNT);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_POWER_LEVEL);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_OPEN);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_CLOSE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_SET_CONFIG);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CONFIG);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CAPABILITY);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_NAME);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SSID_RELEASE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_START);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_STOP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_UPDATE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_INFO);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_CREATE_CMD_QUEUE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DESTROY_CMD_QUEUE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_REGISTER_DB);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_RESUME_CMDQ);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SUSPEND_CMDQ);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_RESUME_CMDQ_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SUSPEND_CMDQ_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SCHEDULING_LOG_NOTIFICATION);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_ENGINE_RESUME);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_RESUME_ENGINE_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_STATE_DUMP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_STATE_DUMP_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_BLOB_DEINIT);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DYNDBG_CONTROL);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_JOB_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_RESET_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_PREEMPT_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_REGISTER_DB_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_UNREGISTER_DB_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_QUERY_ENGINE_HB_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL_COUNT_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_POWER_LEVEL_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_OPEN_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_CLOSE_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_SET_CONFIG_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CONFIG_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CAPABILITY_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_NAME_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SSID_RELEASE_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_START_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_STOP_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_UPDATE_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_INFO_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_NOTIFICATION);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_CREATE_CMD_QUEUE_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DESTROY_CMD_QUEUE_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_BLOB_DEINIT_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DYNDBG_CONTROL_RSP);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_PWR_D0I3_ENTER);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_PWR_D0I3_ENTER_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_ENABLE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_ENABLE_DONE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_DISABLE);
- IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_DISABLE_DONE);
- }
- #undef IVPU_CASE_TO_STR
- return "Unknown JSM message type";
- }
- int ivpu_jsm_register_db(struct ivpu_device *vdev, u32 ctx_id, u32 db_id,
- u64 jobq_base, u32 jobq_size)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_REGISTER_DB };
- struct vpu_jsm_msg resp;
- int ret = 0;
- req.payload.register_db.db_idx = db_id;
- req.payload.register_db.jobq_base = jobq_base;
- req.payload.register_db.jobq_size = jobq_size;
- req.payload.register_db.host_ssid = ctx_id;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_REGISTER_DB_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_err_ratelimited(vdev, "Failed to register doorbell %u: %d\n", db_id, ret);
- return ret;
- }
- int ivpu_jsm_unregister_db(struct ivpu_device *vdev, u32 db_id)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_UNREGISTER_DB };
- struct vpu_jsm_msg resp;
- int ret = 0;
- req.payload.unregister_db.db_idx = db_id;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_UNREGISTER_DB_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to unregister doorbell %u: %d\n", db_id, ret);
- return ret;
- }
- int ivpu_jsm_get_heartbeat(struct ivpu_device *vdev, u32 engine, u64 *heartbeat)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_QUERY_ENGINE_HB };
- struct vpu_jsm_msg resp;
- int ret;
- if (engine > VPU_ENGINE_COPY)
- return -EINVAL;
- req.payload.query_engine_hb.engine_idx = engine;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_QUERY_ENGINE_HB_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret) {
- ivpu_err_ratelimited(vdev, "Failed to get heartbeat from engine %d: %d\n",
- engine, ret);
- return ret;
- }
- *heartbeat = resp.payload.query_engine_hb_done.heartbeat;
- return ret;
- }
- int ivpu_jsm_reset_engine(struct ivpu_device *vdev, u32 engine)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_ENGINE_RESET };
- struct vpu_jsm_msg resp;
- int ret;
- if (engine > VPU_ENGINE_COPY)
- return -EINVAL;
- req.payload.engine_reset.engine_idx = engine;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_ENGINE_RESET_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_err_ratelimited(vdev, "Failed to reset engine %d: %d\n", engine, ret);
- return ret;
- }
- int ivpu_jsm_preempt_engine(struct ivpu_device *vdev, u32 engine, u32 preempt_id)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_ENGINE_PREEMPT };
- struct vpu_jsm_msg resp;
- int ret;
- if (engine > VPU_ENGINE_COPY)
- return -EINVAL;
- req.payload.engine_preempt.engine_idx = engine;
- req.payload.engine_preempt.preempt_id = preempt_id;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_ENGINE_PREEMPT_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_err_ratelimited(vdev, "Failed to preempt engine %d: %d\n", engine, ret);
- return ret;
- }
- int ivpu_jsm_dyndbg_control(struct ivpu_device *vdev, char *command, size_t size)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DYNDBG_CONTROL };
- struct vpu_jsm_msg resp;
- int ret;
- strscpy(req.payload.dyndbg_control.dyndbg_cmd, command, VPU_DYNDBG_CMD_MAX_LEN);
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_DYNDBG_CONTROL_RSP, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to send command \"%s\": ret %d\n",
- command, ret);
- return ret;
- }
- int ivpu_jsm_trace_get_capability(struct ivpu_device *vdev, u32 *trace_destination_mask,
- u64 *trace_hw_component_mask)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_TRACE_GET_CAPABILITY };
- struct vpu_jsm_msg resp;
- int ret;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_TRACE_GET_CAPABILITY_RSP, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret) {
- ivpu_warn_ratelimited(vdev, "Failed to get trace capability: %d\n", ret);
- return ret;
- }
- *trace_destination_mask = resp.payload.trace_capability.trace_destination_mask;
- *trace_hw_component_mask = resp.payload.trace_capability.trace_hw_component_mask;
- return ret;
- }
- int ivpu_jsm_trace_set_config(struct ivpu_device *vdev, u32 trace_level, u32 trace_destination_mask,
- u64 trace_hw_component_mask)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_TRACE_SET_CONFIG };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.trace_config.trace_level = trace_level;
- req.payload.trace_config.trace_destination_mask = trace_destination_mask;
- req.payload.trace_config.trace_hw_component_mask = trace_hw_component_mask;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_TRACE_SET_CONFIG_RSP, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to set config: %d\n", ret);
- return ret;
- }
- int ivpu_jsm_context_release(struct ivpu_device *vdev, u32 host_ssid)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_SSID_RELEASE };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.ssid_release.host_ssid = host_ssid;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_SSID_RELEASE_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to release context: %d\n", ret);
- return ret;
- }
- int ivpu_jsm_pwr_d0i3_enter(struct ivpu_device *vdev)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_PWR_D0I3_ENTER };
- struct vpu_jsm_msg resp;
- int ret;
- if (IVPU_WA(disable_d0i3_msg))
- return 0;
- req.payload.pwr_d0i3_enter.send_response = 1;
- ret = ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_PWR_D0I3_ENTER_DONE, &resp,
- VPU_IPC_CHAN_GEN_CMD, vdev->timeout.d0i3_entry_msg);
- if (ret)
- return ret;
- return ivpu_hw_wait_for_idle(vdev);
- }
- int ivpu_jsm_hws_create_cmdq(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_group, u32 cmdq_id,
- u32 pid, u32 engine, u64 cmdq_base, u32 cmdq_size)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_CREATE_CMD_QUEUE };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.hws_create_cmdq.host_ssid = ctx_id;
- req.payload.hws_create_cmdq.process_id = pid;
- req.payload.hws_create_cmdq.engine_idx = engine;
- req.payload.hws_create_cmdq.cmdq_group = cmdq_group;
- req.payload.hws_create_cmdq.cmdq_id = cmdq_id;
- req.payload.hws_create_cmdq.cmdq_base = cmdq_base;
- req.payload.hws_create_cmdq.cmdq_size = cmdq_size;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_CREATE_CMD_QUEUE_RSP, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to create command queue: %d\n", ret);
- return ret;
- }
- int ivpu_jsm_hws_destroy_cmdq(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DESTROY_CMD_QUEUE };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.hws_destroy_cmdq.host_ssid = ctx_id;
- req.payload.hws_destroy_cmdq.cmdq_id = cmdq_id;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_DESTROY_CMD_QUEUE_RSP, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to destroy command queue: %d\n", ret);
- return ret;
- }
- int ivpu_jsm_hws_register_db(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id, u32 db_id,
- u64 cmdq_base, u32 cmdq_size)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_HWS_REGISTER_DB };
- struct vpu_jsm_msg resp;
- int ret = 0;
- req.payload.hws_register_db.db_id = db_id;
- req.payload.hws_register_db.host_ssid = ctx_id;
- req.payload.hws_register_db.cmdq_id = cmdq_id;
- req.payload.hws_register_db.cmdq_base = cmdq_base;
- req.payload.hws_register_db.cmdq_size = cmdq_size;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_REGISTER_DB_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_err_ratelimited(vdev, "Failed to register doorbell %u: %d\n", db_id, ret);
- return ret;
- }
- int ivpu_jsm_hws_resume_engine(struct ivpu_device *vdev, u32 engine)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_HWS_ENGINE_RESUME };
- struct vpu_jsm_msg resp;
- int ret;
- if (engine >= VPU_ENGINE_NB)
- return -EINVAL;
- req.payload.hws_resume_engine.engine_idx = engine;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_HWS_RESUME_ENGINE_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_err_ratelimited(vdev, "Failed to resume engine %d: %d\n", engine, ret);
- return ret;
- }
- int ivpu_jsm_hws_set_context_sched_properties(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id,
- u32 priority)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.hws_set_context_sched_properties.host_ssid = ctx_id;
- req.payload.hws_set_context_sched_properties.cmdq_id = cmdq_id;
- req.payload.hws_set_context_sched_properties.priority_band = priority;
- req.payload.hws_set_context_sched_properties.realtime_priority_level = 0;
- req.payload.hws_set_context_sched_properties.in_process_priority = 0;
- req.payload.hws_set_context_sched_properties.context_quantum = 20000;
- req.payload.hws_set_context_sched_properties.grace_period_same_priority = 10000;
- req.payload.hws_set_context_sched_properties.grace_period_lower_priority = 0;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES_RSP, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to set context sched properties: %d\n", ret);
- return ret;
- }
- int ivpu_jsm_hws_set_scheduling_log(struct ivpu_device *vdev, u32 engine_idx, u32 host_ssid,
- u64 vpu_log_buffer_va)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.hws_set_scheduling_log.engine_idx = engine_idx;
- req.payload.hws_set_scheduling_log.host_ssid = host_ssid;
- req.payload.hws_set_scheduling_log.vpu_log_buffer_va = vpu_log_buffer_va;
- req.payload.hws_set_scheduling_log.notify_index = 0;
- req.payload.hws_set_scheduling_log.enable_extra_events =
- ivpu_test_mode & IVPU_TEST_MODE_HWS_EXTRA_EVENTS;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG_RSP, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to set scheduling log: %d\n", ret);
- return ret;
- }
- int ivpu_jsm_hws_setup_priority_bands(struct ivpu_device *vdev)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP };
- struct vpu_jsm_msg resp;
- int ret;
- /* Idle */
- req.payload.hws_priority_band_setup.grace_period[0] = 0;
- req.payload.hws_priority_band_setup.process_grace_period[0] = 50000;
- req.payload.hws_priority_band_setup.process_quantum[0] = 160000;
- /* Normal */
- req.payload.hws_priority_band_setup.grace_period[1] = 50000;
- req.payload.hws_priority_band_setup.process_grace_period[1] = 50000;
- req.payload.hws_priority_band_setup.process_quantum[1] = 300000;
- /* Focus */
- req.payload.hws_priority_band_setup.grace_period[2] = 50000;
- req.payload.hws_priority_band_setup.process_grace_period[2] = 50000;
- req.payload.hws_priority_band_setup.process_quantum[2] = 200000;
- /* Realtime */
- req.payload.hws_priority_band_setup.grace_period[3] = 0;
- req.payload.hws_priority_band_setup.process_grace_period[3] = 50000;
- req.payload.hws_priority_band_setup.process_quantum[3] = 200000;
- req.payload.hws_priority_band_setup.normal_band_percentage = 10;
- ret = ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP_RSP,
- &resp, VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to set priority bands: %d\n", ret);
- return ret;
- }
- int ivpu_jsm_metric_streamer_start(struct ivpu_device *vdev, u64 metric_group_mask,
- u64 sampling_rate, u64 buffer_addr, u64 buffer_size)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_START };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.metric_streamer_start.metric_group_mask = metric_group_mask;
- req.payload.metric_streamer_start.sampling_rate = sampling_rate;
- req.payload.metric_streamer_start.buffer_addr = buffer_addr;
- req.payload.metric_streamer_start.buffer_size = buffer_size;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_START_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret) {
- ivpu_warn_ratelimited(vdev, "Failed to start metric streamer: ret %d\n", ret);
- return ret;
- }
- return ret;
- }
- int ivpu_jsm_metric_streamer_stop(struct ivpu_device *vdev, u64 metric_group_mask)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_STOP };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.metric_streamer_stop.metric_group_mask = metric_group_mask;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_STOP_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret)
- ivpu_warn_ratelimited(vdev, "Failed to stop metric streamer: ret %d\n", ret);
- return ret;
- }
- int ivpu_jsm_metric_streamer_update(struct ivpu_device *vdev, u64 metric_group_mask,
- u64 buffer_addr, u64 buffer_size, u64 *bytes_written)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_UPDATE };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.metric_streamer_update.metric_group_mask = metric_group_mask;
- req.payload.metric_streamer_update.buffer_addr = buffer_addr;
- req.payload.metric_streamer_update.buffer_size = buffer_size;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_UPDATE_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret) {
- ivpu_warn_ratelimited(vdev, "Failed to update metric streamer: ret %d\n", ret);
- return ret;
- }
- if (buffer_size && resp.payload.metric_streamer_done.bytes_written > buffer_size) {
- ivpu_warn_ratelimited(vdev, "MS buffer overflow: bytes_written %#llx > buffer_size %#llx\n",
- resp.payload.metric_streamer_done.bytes_written, buffer_size);
- return -EOVERFLOW;
- }
- *bytes_written = resp.payload.metric_streamer_done.bytes_written;
- return ret;
- }
- int ivpu_jsm_metric_streamer_info(struct ivpu_device *vdev, u64 metric_group_mask, u64 buffer_addr,
- u64 buffer_size, u32 *sample_size, u64 *info_size)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_INFO };
- struct vpu_jsm_msg resp;
- int ret;
- req.payload.metric_streamer_start.metric_group_mask = metric_group_mask;
- req.payload.metric_streamer_start.buffer_addr = buffer_addr;
- req.payload.metric_streamer_start.buffer_size = buffer_size;
- ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_INFO_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- if (ret) {
- ivpu_warn_ratelimited(vdev, "Failed to get metric streamer info: ret %d\n", ret);
- return ret;
- }
- if (!resp.payload.metric_streamer_done.sample_size) {
- ivpu_warn_ratelimited(vdev, "Invalid sample size\n");
- return -EBADMSG;
- }
- if (sample_size)
- *sample_size = resp.payload.metric_streamer_done.sample_size;
- if (info_size)
- *info_size = resp.payload.metric_streamer_done.bytes_written;
- return ret;
- }
- int ivpu_jsm_dct_enable(struct ivpu_device *vdev, u32 active_us, u32 inactive_us)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DCT_ENABLE };
- struct vpu_jsm_msg resp;
- req.payload.pwr_dct_control.dct_active_us = active_us;
- req.payload.pwr_dct_control.dct_inactive_us = inactive_us;
- return ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_DCT_ENABLE_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- }
- int ivpu_jsm_dct_disable(struct ivpu_device *vdev)
- {
- struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DCT_DISABLE };
- struct vpu_jsm_msg resp;
- return ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_DCT_DISABLE_DONE, &resp,
- VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
- }
|