Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 | // SPDX-License-Identifier: GPL-2.0 /* * Intel Turbo Boost Max Technology 3.0 legacy (non HWP) enumeration driver * Copyright (c) 2017, Intel Corporation. * All rights reserved. * * Author: Srinivas Pandruvada <srinivas.pandruvada@linux.intel.com> */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include <linux/cpufeature.h> #include <linux/cpuhotplug.h> #include <linux/init.h> #include <linux/kernel.h> #include <linux/topology.h> #include <linux/workqueue.h> #include <asm/cpu_device_id.h> #include <asm/intel-family.h> #define MSR_OC_MAILBOX 0x150 #define MSR_OC_MAILBOX_CMD_OFFSET 32 #define MSR_OC_MAILBOX_RSP_OFFSET 32 #define MSR_OC_MAILBOX_BUSY_BIT 63 #define OC_MAILBOX_FC_CONTROL_CMD 0x1C /* * Typical latency to get mail box response is ~3us, It takes +3 us to * process reading mailbox after issuing mailbox write on a Broadwell 3.4 GHz * system. So for most of the time, the first mailbox read should have the * response, but to avoid some boundary cases retry twice. */ #define OC_MAILBOX_RETRY_COUNT 2 static int get_oc_core_priority(unsigned int cpu) { u64 value, cmd = OC_MAILBOX_FC_CONTROL_CMD; int ret, i; /* Issue favored core read command */ value = cmd << MSR_OC_MAILBOX_CMD_OFFSET; /* Set the busy bit to indicate OS is trying to issue command */ value |= BIT_ULL(MSR_OC_MAILBOX_BUSY_BIT); ret = wrmsrl_safe(MSR_OC_MAILBOX, value); if (ret) { pr_debug("cpu %d OC mailbox write failed\n", cpu); return ret; } for (i = 0; i < OC_MAILBOX_RETRY_COUNT; ++i) { ret = rdmsrl_safe(MSR_OC_MAILBOX, &value); if (ret) { pr_debug("cpu %d OC mailbox read failed\n", cpu); break; } if (value & BIT_ULL(MSR_OC_MAILBOX_BUSY_BIT)) { pr_debug("cpu %d OC mailbox still processing\n", cpu); ret = -EBUSY; continue; } if ((value >> MSR_OC_MAILBOX_RSP_OFFSET) & 0xff) { pr_debug("cpu %d OC mailbox cmd failed\n", cpu); ret = -ENXIO; break; } ret = value & 0xff; pr_debug("cpu %d max_ratio %d\n", cpu, ret); break; } return ret; } /* * The work item is needed to avoid CPU hotplug locking issues. The function * itmt_legacy_set_priority() is called from CPU online callback, so can't * call sched_set_itmt_support() from there as this function will aquire * hotplug locks in its path. */ static void itmt_legacy_work_fn(struct work_struct *work) { sched_set_itmt_support(); } static DECLARE_WORK(sched_itmt_work, itmt_legacy_work_fn); static int itmt_legacy_cpu_online(unsigned int cpu) { static u32 max_highest_perf = 0, min_highest_perf = U32_MAX; int priority; priority = get_oc_core_priority(cpu); if (priority < 0) return 0; sched_set_itmt_core_prio(priority, cpu); /* Enable ITMT feature when a core with different priority is found */ if (max_highest_perf <= min_highest_perf) { if (priority > max_highest_perf) max_highest_perf = priority; if (priority < min_highest_perf) min_highest_perf = priority; if (max_highest_perf > min_highest_perf) schedule_work(&sched_itmt_work); } return 0; } static const struct x86_cpu_id itmt_legacy_cpu_ids[] = { X86_MATCH_INTEL_FAM6_MODEL(BROADWELL_X, NULL), X86_MATCH_INTEL_FAM6_MODEL(SKYLAKE_X, NULL), {} }; static int __init itmt_legacy_init(void) { const struct x86_cpu_id *id; int ret; id = x86_match_cpu(itmt_legacy_cpu_ids); if (!id) return -ENODEV; ret = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "platform/x86/turbo_max_3:online", itmt_legacy_cpu_online, NULL); if (ret < 0) return ret; return 0; } late_initcall(itmt_legacy_init) |