intel_ips.c 44 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739
  1. /*
  2. * Copyright (c) 2009-2010 Intel Corporation
  3. *
  4. * This program is free software; you can redistribute it and/or modify it
  5. * under the terms and conditions of the GNU General Public License,
  6. * version 2, as published by the Free Software Foundation.
  7. *
  8. * This program is distributed in the hope it will be useful, but WITHOUT
  9. * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  10. * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
  11. * more details.
  12. *
  13. * You should have received a copy of the GNU General Public License along with
  14. * this program; if not, write to the Free Software Foundation, Inc.,
  15. * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
  16. *
  17. * The full GNU General Public License is included in this distribution in
  18. * the file called "COPYING".
  19. *
  20. * Authors:
  21. * Jesse Barnes <jbarnes@virtuousgeek.org>
  22. */
  23. /*
  24. * Some Intel Ibex Peak based platforms support so-called "intelligent
  25. * power sharing", which allows the CPU and GPU to cooperate to maximize
  26. * performance within a given TDP (thermal design point). This driver
  27. * performs the coordination between the CPU and GPU, monitors thermal and
  28. * power statistics in the platform, and initializes power monitoring
  29. * hardware. It also provides a few tunables to control behavior. Its
  30. * primary purpose is to safely allow CPU and GPU turbo modes to be enabled
  31. * by tracking power and thermal budget; secondarily it can boost turbo
  32. * performance by allocating more power or thermal budget to the CPU or GPU
  33. * based on available headroom and activity.
  34. *
  35. * The basic algorithm is driven by a 5s moving average of temperature. If
  36. * thermal headroom is available, the CPU and/or GPU power clamps may be
  37. * adjusted upwards. If we hit the thermal ceiling or a thermal trigger,
  38. * we scale back the clamp. Aside from trigger events (when we're critically
  39. * close or over our TDP) we don't adjust the clamps more than once every
  40. * five seconds.
  41. *
  42. * The thermal device (device 31, function 6) has a set of registers that
  43. * are updated by the ME firmware. The ME should also take the clamp values
  44. * written to those registers and write them to the CPU, but we currently
  45. * bypass that functionality and write the CPU MSR directly.
  46. *
  47. * UNSUPPORTED:
  48. * - dual MCP configs
  49. *
  50. * TODO:
  51. * - handle CPU hotplug
  52. * - provide turbo enable/disable api
  53. *
  54. * Related documents:
  55. * - CDI 403777, 403778 - Auburndale EDS vol 1 & 2
  56. * - CDI 401376 - Ibex Peak EDS
  57. * - ref 26037, 26641 - IPS BIOS spec
  58. * - ref 26489 - Nehalem BIOS writer's guide
  59. * - ref 26921 - Ibex Peak BIOS Specification
  60. */
  61. #include <linux/debugfs.h>
  62. #include <linux/delay.h>
  63. #include <linux/interrupt.h>
  64. #include <linux/kernel.h>
  65. #include <linux/kthread.h>
  66. #include <linux/module.h>
  67. #include <linux/pci.h>
  68. #include <linux/sched.h>
  69. #include <linux/sched/loadavg.h>
  70. #include <linux/seq_file.h>
  71. #include <linux/string.h>
  72. #include <linux/tick.h>
  73. #include <linux/timer.h>
  74. #include <linux/dmi.h>
  75. #include <drm/i915_drm.h>
  76. #include <asm/msr.h>
  77. #include <asm/processor.h>
  78. #include "intel_ips.h"
  79. #include <linux/io-64-nonatomic-lo-hi.h>
  80. #define PCI_DEVICE_ID_INTEL_THERMAL_SENSOR 0x3b32
  81. /*
  82. * Package level MSRs for monitor/control
  83. */
  84. #define PLATFORM_INFO 0xce
  85. #define PLATFORM_TDP (1<<29)
  86. #define PLATFORM_RATIO (1<<28)
  87. #define IA32_MISC_ENABLE 0x1a0
  88. #define IA32_MISC_TURBO_EN (1ULL<<38)
  89. #define TURBO_POWER_CURRENT_LIMIT 0x1ac
  90. #define TURBO_TDC_OVR_EN (1UL<<31)
  91. #define TURBO_TDC_MASK (0x000000007fff0000UL)
  92. #define TURBO_TDC_SHIFT (16)
  93. #define TURBO_TDP_OVR_EN (1UL<<15)
  94. #define TURBO_TDP_MASK (0x0000000000003fffUL)
  95. /*
  96. * Core/thread MSRs for monitoring
  97. */
  98. #define IA32_PERF_CTL 0x199
  99. #define IA32_PERF_TURBO_DIS (1ULL<<32)
  100. /*
  101. * Thermal PCI device regs
  102. */
  103. #define THM_CFG_TBAR 0x10
  104. #define THM_CFG_TBAR_HI 0x14
  105. #define THM_TSIU 0x00
  106. #define THM_TSE 0x01
  107. #define TSE_EN 0xb8
  108. #define THM_TSS 0x02
  109. #define THM_TSTR 0x03
  110. #define THM_TSTTP 0x04
  111. #define THM_TSCO 0x08
  112. #define THM_TSES 0x0c
  113. #define THM_TSGPEN 0x0d
  114. #define TSGPEN_HOT_LOHI (1<<1)
  115. #define TSGPEN_CRIT_LOHI (1<<2)
  116. #define THM_TSPC 0x0e
  117. #define THM_PPEC 0x10
  118. #define THM_CTA 0x12
  119. #define THM_PTA 0x14
  120. #define PTA_SLOPE_MASK (0xff00)
  121. #define PTA_SLOPE_SHIFT 8
  122. #define PTA_OFFSET_MASK (0x00ff)
  123. #define THM_MGTA 0x16
  124. #define MGTA_SLOPE_MASK (0xff00)
  125. #define MGTA_SLOPE_SHIFT 8
  126. #define MGTA_OFFSET_MASK (0x00ff)
  127. #define THM_TRC 0x1a
  128. #define TRC_CORE2_EN (1<<15)
  129. #define TRC_THM_EN (1<<12)
  130. #define TRC_C6_WAR (1<<8)
  131. #define TRC_CORE1_EN (1<<7)
  132. #define TRC_CORE_PWR (1<<6)
  133. #define TRC_PCH_EN (1<<5)
  134. #define TRC_MCH_EN (1<<4)
  135. #define TRC_DIMM4 (1<<3)
  136. #define TRC_DIMM3 (1<<2)
  137. #define TRC_DIMM2 (1<<1)
  138. #define TRC_DIMM1 (1<<0)
  139. #define THM_TES 0x20
  140. #define THM_TEN 0x21
  141. #define TEN_UPDATE_EN 1
  142. #define THM_PSC 0x24
  143. #define PSC_NTG (1<<0) /* No GFX turbo support */
  144. #define PSC_NTPC (1<<1) /* No CPU turbo support */
  145. #define PSC_PP_DEF (0<<2) /* Perf policy up to driver */
  146. #define PSP_PP_PC (1<<2) /* BIOS prefers CPU perf */
  147. #define PSP_PP_BAL (2<<2) /* BIOS wants balanced perf */
  148. #define PSP_PP_GFX (3<<2) /* BIOS prefers GFX perf */
  149. #define PSP_PBRT (1<<4) /* BIOS run time support */
  150. #define THM_CTV1 0x30
  151. #define CTV_TEMP_ERROR (1<<15)
  152. #define CTV_TEMP_MASK 0x3f
  153. #define CTV_
  154. #define THM_CTV2 0x32
  155. #define THM_CEC 0x34 /* undocumented power accumulator in joules */
  156. #define THM_AE 0x3f
  157. #define THM_HTS 0x50 /* 32 bits */
  158. #define HTS_PCPL_MASK (0x7fe00000)
  159. #define HTS_PCPL_SHIFT 21
  160. #define HTS_GPL_MASK (0x001ff000)
  161. #define HTS_GPL_SHIFT 12
  162. #define HTS_PP_MASK (0x00000c00)
  163. #define HTS_PP_SHIFT 10
  164. #define HTS_PP_DEF 0
  165. #define HTS_PP_PROC 1
  166. #define HTS_PP_BAL 2
  167. #define HTS_PP_GFX 3
  168. #define HTS_PCTD_DIS (1<<9)
  169. #define HTS_GTD_DIS (1<<8)
  170. #define HTS_PTL_MASK (0x000000fe)
  171. #define HTS_PTL_SHIFT 1
  172. #define HTS_NVV (1<<0)
  173. #define THM_HTSHI 0x54 /* 16 bits */
  174. #define HTS2_PPL_MASK (0x03ff)
  175. #define HTS2_PRST_MASK (0x3c00)
  176. #define HTS2_PRST_SHIFT 10
  177. #define HTS2_PRST_UNLOADED 0
  178. #define HTS2_PRST_RUNNING 1
  179. #define HTS2_PRST_TDISOP 2 /* turbo disabled due to power */
  180. #define HTS2_PRST_TDISHT 3 /* turbo disabled due to high temp */
  181. #define HTS2_PRST_TDISUSR 4 /* user disabled turbo */
  182. #define HTS2_PRST_TDISPLAT 5 /* platform disabled turbo */
  183. #define HTS2_PRST_TDISPM 6 /* power management disabled turbo */
  184. #define HTS2_PRST_TDISERR 7 /* some kind of error disabled turbo */
  185. #define THM_PTL 0x56
  186. #define THM_MGTV 0x58
  187. #define TV_MASK 0x000000000000ff00
  188. #define TV_SHIFT 8
  189. #define THM_PTV 0x60
  190. #define PTV_MASK 0x00ff
  191. #define THM_MMGPC 0x64
  192. #define THM_MPPC 0x66
  193. #define THM_MPCPC 0x68
  194. #define THM_TSPIEN 0x82
  195. #define TSPIEN_AUX_LOHI (1<<0)
  196. #define TSPIEN_HOT_LOHI (1<<1)
  197. #define TSPIEN_CRIT_LOHI (1<<2)
  198. #define TSPIEN_AUX2_LOHI (1<<3)
  199. #define THM_TSLOCK 0x83
  200. #define THM_ATR 0x84
  201. #define THM_TOF 0x87
  202. #define THM_STS 0x98
  203. #define STS_PCPL_MASK (0x7fe00000)
  204. #define STS_PCPL_SHIFT 21
  205. #define STS_GPL_MASK (0x001ff000)
  206. #define STS_GPL_SHIFT 12
  207. #define STS_PP_MASK (0x00000c00)
  208. #define STS_PP_SHIFT 10
  209. #define STS_PP_DEF 0
  210. #define STS_PP_PROC 1
  211. #define STS_PP_BAL 2
  212. #define STS_PP_GFX 3
  213. #define STS_PCTD_DIS (1<<9)
  214. #define STS_GTD_DIS (1<<8)
  215. #define STS_PTL_MASK (0x000000fe)
  216. #define STS_PTL_SHIFT 1
  217. #define STS_NVV (1<<0)
  218. #define THM_SEC 0x9c
  219. #define SEC_ACK (1<<0)
  220. #define THM_TC3 0xa4
  221. #define THM_TC1 0xa8
  222. #define STS_PPL_MASK (0x0003ff00)
  223. #define STS_PPL_SHIFT 16
  224. #define THM_TC2 0xac
  225. #define THM_DTV 0xb0
  226. #define THM_ITV 0xd8
  227. #define ITV_ME_SEQNO_MASK 0x00ff0000 /* ME should update every ~200ms */
  228. #define ITV_ME_SEQNO_SHIFT (16)
  229. #define ITV_MCH_TEMP_MASK 0x0000ff00
  230. #define ITV_MCH_TEMP_SHIFT (8)
  231. #define ITV_PCH_TEMP_MASK 0x000000ff
  232. #define thm_readb(off) readb(ips->regmap + (off))
  233. #define thm_readw(off) readw(ips->regmap + (off))
  234. #define thm_readl(off) readl(ips->regmap + (off))
  235. #define thm_readq(off) readq(ips->regmap + (off))
  236. #define thm_writeb(off, val) writeb((val), ips->regmap + (off))
  237. #define thm_writew(off, val) writew((val), ips->regmap + (off))
  238. #define thm_writel(off, val) writel((val), ips->regmap + (off))
  239. static const int IPS_ADJUST_PERIOD = 5000; /* ms */
  240. static bool late_i915_load = false;
  241. /* For initial average collection */
  242. static const int IPS_SAMPLE_PERIOD = 200; /* ms */
  243. static const int IPS_SAMPLE_WINDOW = 5000; /* 5s moving window of samples */
  244. #define IPS_SAMPLE_COUNT (IPS_SAMPLE_WINDOW / IPS_SAMPLE_PERIOD)
  245. /* Per-SKU limits */
  246. struct ips_mcp_limits {
  247. int cpu_family;
  248. int cpu_model; /* includes extended model... */
  249. int mcp_power_limit; /* mW units */
  250. int core_power_limit;
  251. int mch_power_limit;
  252. int core_temp_limit; /* degrees C */
  253. int mch_temp_limit;
  254. };
  255. /* Max temps are -10 degrees C to avoid PROCHOT# */
  256. static struct ips_mcp_limits ips_sv_limits = {
  257. .mcp_power_limit = 35000,
  258. .core_power_limit = 29000,
  259. .mch_power_limit = 20000,
  260. .core_temp_limit = 95,
  261. .mch_temp_limit = 90
  262. };
  263. static struct ips_mcp_limits ips_lv_limits = {
  264. .mcp_power_limit = 25000,
  265. .core_power_limit = 21000,
  266. .mch_power_limit = 13000,
  267. .core_temp_limit = 95,
  268. .mch_temp_limit = 90
  269. };
  270. static struct ips_mcp_limits ips_ulv_limits = {
  271. .mcp_power_limit = 18000,
  272. .core_power_limit = 14000,
  273. .mch_power_limit = 11000,
  274. .core_temp_limit = 95,
  275. .mch_temp_limit = 90
  276. };
  277. struct ips_driver {
  278. struct pci_dev *dev;
  279. void *regmap;
  280. struct task_struct *monitor;
  281. struct task_struct *adjust;
  282. struct dentry *debug_root;
  283. /* Average CPU core temps (all averages in .01 degrees C for precision) */
  284. u16 ctv1_avg_temp;
  285. u16 ctv2_avg_temp;
  286. /* GMCH average */
  287. u16 mch_avg_temp;
  288. /* Average for the CPU (both cores?) */
  289. u16 mcp_avg_temp;
  290. /* Average power consumption (in mW) */
  291. u32 cpu_avg_power;
  292. u32 mch_avg_power;
  293. /* Offset values */
  294. u16 cta_val;
  295. u16 pta_val;
  296. u16 mgta_val;
  297. /* Maximums & prefs, protected by turbo status lock */
  298. spinlock_t turbo_status_lock;
  299. u16 mcp_temp_limit;
  300. u16 mcp_power_limit;
  301. u16 core_power_limit;
  302. u16 mch_power_limit;
  303. bool cpu_turbo_enabled;
  304. bool __cpu_turbo_on;
  305. bool gpu_turbo_enabled;
  306. bool __gpu_turbo_on;
  307. bool gpu_preferred;
  308. bool poll_turbo_status;
  309. bool second_cpu;
  310. bool turbo_toggle_allowed;
  311. struct ips_mcp_limits *limits;
  312. /* Optional MCH interfaces for if i915 is in use */
  313. unsigned long (*read_mch_val)(void);
  314. bool (*gpu_raise)(void);
  315. bool (*gpu_lower)(void);
  316. bool (*gpu_busy)(void);
  317. bool (*gpu_turbo_disable)(void);
  318. /* For restoration at unload */
  319. u64 orig_turbo_limit;
  320. u64 orig_turbo_ratios;
  321. };
  322. static bool
  323. ips_gpu_turbo_enabled(struct ips_driver *ips);
  324. /**
  325. * ips_cpu_busy - is CPU busy?
  326. * @ips: IPS driver struct
  327. *
  328. * Check CPU for load to see whether we should increase its thermal budget.
  329. *
  330. * RETURNS:
  331. * True if the CPU could use more power, false otherwise.
  332. */
  333. static bool ips_cpu_busy(struct ips_driver *ips)
  334. {
  335. if ((avenrun[0] >> FSHIFT) > 1)
  336. return true;
  337. return false;
  338. }
  339. /**
  340. * ips_cpu_raise - raise CPU power clamp
  341. * @ips: IPS driver struct
  342. *
  343. * Raise the CPU power clamp by %IPS_CPU_STEP, in accordance with TDP for
  344. * this platform.
  345. *
  346. * We do this by adjusting the TURBO_POWER_CURRENT_LIMIT MSR upwards (as
  347. * long as we haven't hit the TDP limit for the SKU).
  348. */
  349. static void ips_cpu_raise(struct ips_driver *ips)
  350. {
  351. u64 turbo_override;
  352. u16 cur_tdp_limit, new_tdp_limit;
  353. if (!ips->cpu_turbo_enabled)
  354. return;
  355. rdmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  356. cur_tdp_limit = turbo_override & TURBO_TDP_MASK;
  357. new_tdp_limit = cur_tdp_limit + 8; /* 1W increase */
  358. /* Clamp to SKU TDP limit */
  359. if (((new_tdp_limit * 10) / 8) > ips->core_power_limit)
  360. new_tdp_limit = cur_tdp_limit;
  361. thm_writew(THM_MPCPC, (new_tdp_limit * 10) / 8);
  362. turbo_override |= TURBO_TDC_OVR_EN | TURBO_TDP_OVR_EN;
  363. wrmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  364. turbo_override &= ~TURBO_TDP_MASK;
  365. turbo_override |= new_tdp_limit;
  366. wrmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  367. }
  368. /**
  369. * ips_cpu_lower - lower CPU power clamp
  370. * @ips: IPS driver struct
  371. *
  372. * Lower CPU power clamp b %IPS_CPU_STEP if possible.
  373. *
  374. * We do this by adjusting the TURBO_POWER_CURRENT_LIMIT MSR down, going
  375. * as low as the platform limits will allow (though we could go lower there
  376. * wouldn't be much point).
  377. */
  378. static void ips_cpu_lower(struct ips_driver *ips)
  379. {
  380. u64 turbo_override;
  381. u16 cur_limit, new_limit;
  382. rdmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  383. cur_limit = turbo_override & TURBO_TDP_MASK;
  384. new_limit = cur_limit - 8; /* 1W decrease */
  385. /* Clamp to SKU TDP limit */
  386. if (new_limit < (ips->orig_turbo_limit & TURBO_TDP_MASK))
  387. new_limit = ips->orig_turbo_limit & TURBO_TDP_MASK;
  388. thm_writew(THM_MPCPC, (new_limit * 10) / 8);
  389. turbo_override |= TURBO_TDC_OVR_EN | TURBO_TDP_OVR_EN;
  390. wrmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  391. turbo_override &= ~TURBO_TDP_MASK;
  392. turbo_override |= new_limit;
  393. wrmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  394. }
  395. /**
  396. * do_enable_cpu_turbo - internal turbo enable function
  397. * @data: unused
  398. *
  399. * Internal function for actually updating MSRs. When we enable/disable
  400. * turbo, we need to do it on each CPU; this function is the one called
  401. * by on_each_cpu() when needed.
  402. */
  403. static void do_enable_cpu_turbo(void *data)
  404. {
  405. u64 perf_ctl;
  406. rdmsrl(IA32_PERF_CTL, perf_ctl);
  407. if (perf_ctl & IA32_PERF_TURBO_DIS) {
  408. perf_ctl &= ~IA32_PERF_TURBO_DIS;
  409. wrmsrl(IA32_PERF_CTL, perf_ctl);
  410. }
  411. }
  412. /**
  413. * ips_enable_cpu_turbo - enable turbo mode on all CPUs
  414. * @ips: IPS driver struct
  415. *
  416. * Enable turbo mode by clearing the disable bit in IA32_PERF_CTL on
  417. * all logical threads.
  418. */
  419. static void ips_enable_cpu_turbo(struct ips_driver *ips)
  420. {
  421. /* Already on, no need to mess with MSRs */
  422. if (ips->__cpu_turbo_on)
  423. return;
  424. if (ips->turbo_toggle_allowed)
  425. on_each_cpu(do_enable_cpu_turbo, ips, 1);
  426. ips->__cpu_turbo_on = true;
  427. }
  428. /**
  429. * do_disable_cpu_turbo - internal turbo disable function
  430. * @data: unused
  431. *
  432. * Internal function for actually updating MSRs. When we enable/disable
  433. * turbo, we need to do it on each CPU; this function is the one called
  434. * by on_each_cpu() when needed.
  435. */
  436. static void do_disable_cpu_turbo(void *data)
  437. {
  438. u64 perf_ctl;
  439. rdmsrl(IA32_PERF_CTL, perf_ctl);
  440. if (!(perf_ctl & IA32_PERF_TURBO_DIS)) {
  441. perf_ctl |= IA32_PERF_TURBO_DIS;
  442. wrmsrl(IA32_PERF_CTL, perf_ctl);
  443. }
  444. }
  445. /**
  446. * ips_disable_cpu_turbo - disable turbo mode on all CPUs
  447. * @ips: IPS driver struct
  448. *
  449. * Disable turbo mode by setting the disable bit in IA32_PERF_CTL on
  450. * all logical threads.
  451. */
  452. static void ips_disable_cpu_turbo(struct ips_driver *ips)
  453. {
  454. /* Already off, leave it */
  455. if (!ips->__cpu_turbo_on)
  456. return;
  457. if (ips->turbo_toggle_allowed)
  458. on_each_cpu(do_disable_cpu_turbo, ips, 1);
  459. ips->__cpu_turbo_on = false;
  460. }
  461. /**
  462. * ips_gpu_busy - is GPU busy?
  463. * @ips: IPS driver struct
  464. *
  465. * Check GPU for load to see whether we should increase its thermal budget.
  466. * We need to call into the i915 driver in this case.
  467. *
  468. * RETURNS:
  469. * True if the GPU could use more power, false otherwise.
  470. */
  471. static bool ips_gpu_busy(struct ips_driver *ips)
  472. {
  473. if (!ips_gpu_turbo_enabled(ips))
  474. return false;
  475. return ips->gpu_busy();
  476. }
  477. /**
  478. * ips_gpu_raise - raise GPU power clamp
  479. * @ips: IPS driver struct
  480. *
  481. * Raise the GPU frequency/power if possible. We need to call into the
  482. * i915 driver in this case.
  483. */
  484. static void ips_gpu_raise(struct ips_driver *ips)
  485. {
  486. if (!ips_gpu_turbo_enabled(ips))
  487. return;
  488. if (!ips->gpu_raise())
  489. ips->gpu_turbo_enabled = false;
  490. return;
  491. }
  492. /**
  493. * ips_gpu_lower - lower GPU power clamp
  494. * @ips: IPS driver struct
  495. *
  496. * Lower GPU frequency/power if possible. Need to call i915.
  497. */
  498. static void ips_gpu_lower(struct ips_driver *ips)
  499. {
  500. if (!ips_gpu_turbo_enabled(ips))
  501. return;
  502. if (!ips->gpu_lower())
  503. ips->gpu_turbo_enabled = false;
  504. return;
  505. }
  506. /**
  507. * ips_enable_gpu_turbo - notify the gfx driver turbo is available
  508. * @ips: IPS driver struct
  509. *
  510. * Call into the graphics driver indicating that it can safely use
  511. * turbo mode.
  512. */
  513. static void ips_enable_gpu_turbo(struct ips_driver *ips)
  514. {
  515. if (ips->__gpu_turbo_on)
  516. return;
  517. ips->__gpu_turbo_on = true;
  518. }
  519. /**
  520. * ips_disable_gpu_turbo - notify the gfx driver to disable turbo mode
  521. * @ips: IPS driver struct
  522. *
  523. * Request that the graphics driver disable turbo mode.
  524. */
  525. static void ips_disable_gpu_turbo(struct ips_driver *ips)
  526. {
  527. /* Avoid calling i915 if turbo is already disabled */
  528. if (!ips->__gpu_turbo_on)
  529. return;
  530. if (!ips->gpu_turbo_disable())
  531. dev_err(&ips->dev->dev, "failed to disable graphics turbo\n");
  532. else
  533. ips->__gpu_turbo_on = false;
  534. }
  535. /**
  536. * mcp_exceeded - check whether we're outside our thermal & power limits
  537. * @ips: IPS driver struct
  538. *
  539. * Check whether the MCP is over its thermal or power budget.
  540. */
  541. static bool mcp_exceeded(struct ips_driver *ips)
  542. {
  543. unsigned long flags;
  544. bool ret = false;
  545. u32 temp_limit;
  546. u32 avg_power;
  547. spin_lock_irqsave(&ips->turbo_status_lock, flags);
  548. temp_limit = ips->mcp_temp_limit * 100;
  549. if (ips->mcp_avg_temp > temp_limit)
  550. ret = true;
  551. avg_power = ips->cpu_avg_power + ips->mch_avg_power;
  552. if (avg_power > ips->mcp_power_limit)
  553. ret = true;
  554. spin_unlock_irqrestore(&ips->turbo_status_lock, flags);
  555. return ret;
  556. }
  557. /**
  558. * cpu_exceeded - check whether a CPU core is outside its limits
  559. * @ips: IPS driver struct
  560. * @cpu: CPU number to check
  561. *
  562. * Check a given CPU's average temp or power is over its limit.
  563. */
  564. static bool cpu_exceeded(struct ips_driver *ips, int cpu)
  565. {
  566. unsigned long flags;
  567. int avg;
  568. bool ret = false;
  569. spin_lock_irqsave(&ips->turbo_status_lock, flags);
  570. avg = cpu ? ips->ctv2_avg_temp : ips->ctv1_avg_temp;
  571. if (avg > (ips->limits->core_temp_limit * 100))
  572. ret = true;
  573. if (ips->cpu_avg_power > ips->core_power_limit * 100)
  574. ret = true;
  575. spin_unlock_irqrestore(&ips->turbo_status_lock, flags);
  576. if (ret)
  577. dev_info(&ips->dev->dev,
  578. "CPU power or thermal limit exceeded\n");
  579. return ret;
  580. }
  581. /**
  582. * mch_exceeded - check whether the GPU is over budget
  583. * @ips: IPS driver struct
  584. *
  585. * Check the MCH temp & power against their maximums.
  586. */
  587. static bool mch_exceeded(struct ips_driver *ips)
  588. {
  589. unsigned long flags;
  590. bool ret = false;
  591. spin_lock_irqsave(&ips->turbo_status_lock, flags);
  592. if (ips->mch_avg_temp > (ips->limits->mch_temp_limit * 100))
  593. ret = true;
  594. if (ips->mch_avg_power > ips->mch_power_limit)
  595. ret = true;
  596. spin_unlock_irqrestore(&ips->turbo_status_lock, flags);
  597. return ret;
  598. }
  599. /**
  600. * verify_limits - verify BIOS provided limits
  601. * @ips: IPS structure
  602. *
  603. * BIOS can optionally provide non-default limits for power and temp. Check
  604. * them here and use the defaults if the BIOS values are not provided or
  605. * are otherwise unusable.
  606. */
  607. static void verify_limits(struct ips_driver *ips)
  608. {
  609. if (ips->mcp_power_limit < ips->limits->mcp_power_limit ||
  610. ips->mcp_power_limit > 35000)
  611. ips->mcp_power_limit = ips->limits->mcp_power_limit;
  612. if (ips->mcp_temp_limit < ips->limits->core_temp_limit ||
  613. ips->mcp_temp_limit < ips->limits->mch_temp_limit ||
  614. ips->mcp_temp_limit > 150)
  615. ips->mcp_temp_limit = min(ips->limits->core_temp_limit,
  616. ips->limits->mch_temp_limit);
  617. }
  618. /**
  619. * update_turbo_limits - get various limits & settings from regs
  620. * @ips: IPS driver struct
  621. *
  622. * Update the IPS power & temp limits, along with turbo enable flags,
  623. * based on latest register contents.
  624. *
  625. * Used at init time and for runtime BIOS support, which requires polling
  626. * the regs for updates (as a result of AC->DC transition for example).
  627. *
  628. * LOCKING:
  629. * Caller must hold turbo_status_lock (outside of init)
  630. */
  631. static void update_turbo_limits(struct ips_driver *ips)
  632. {
  633. u32 hts = thm_readl(THM_HTS);
  634. ips->cpu_turbo_enabled = !(hts & HTS_PCTD_DIS);
  635. /*
  636. * Disable turbo for now, until we can figure out why the power figures
  637. * are wrong
  638. */
  639. ips->cpu_turbo_enabled = false;
  640. if (ips->gpu_busy)
  641. ips->gpu_turbo_enabled = !(hts & HTS_GTD_DIS);
  642. ips->core_power_limit = thm_readw(THM_MPCPC);
  643. ips->mch_power_limit = thm_readw(THM_MMGPC);
  644. ips->mcp_temp_limit = thm_readw(THM_PTL);
  645. ips->mcp_power_limit = thm_readw(THM_MPPC);
  646. verify_limits(ips);
  647. /* Ignore BIOS CPU vs GPU pref */
  648. }
  649. /**
  650. * ips_adjust - adjust power clamp based on thermal state
  651. * @data: ips driver structure
  652. *
  653. * Wake up every 5s or so and check whether we should adjust the power clamp.
  654. * Check CPU and GPU load to determine which needs adjustment. There are
  655. * several things to consider here:
  656. * - do we need to adjust up or down?
  657. * - is CPU busy?
  658. * - is GPU busy?
  659. * - is CPU in turbo?
  660. * - is GPU in turbo?
  661. * - is CPU or GPU preferred? (CPU is default)
  662. *
  663. * So, given the above, we do the following:
  664. * - up (TDP available)
  665. * - CPU not busy, GPU not busy - nothing
  666. * - CPU busy, GPU not busy - adjust CPU up
  667. * - CPU not busy, GPU busy - adjust GPU up
  668. * - CPU busy, GPU busy - adjust preferred unit up, taking headroom from
  669. * non-preferred unit if necessary
  670. * - down (at TDP limit)
  671. * - adjust both CPU and GPU down if possible
  672. *
  673. cpu+ gpu+ cpu+gpu- cpu-gpu+ cpu-gpu-
  674. cpu < gpu < cpu+gpu+ cpu+ gpu+ nothing
  675. cpu < gpu >= cpu+gpu-(mcp<) cpu+gpu-(mcp<) gpu- gpu-
  676. cpu >= gpu < cpu-gpu+(mcp<) cpu- cpu-gpu+(mcp<) cpu-
  677. cpu >= gpu >= cpu-gpu- cpu-gpu- cpu-gpu- cpu-gpu-
  678. *
  679. */
  680. static int ips_adjust(void *data)
  681. {
  682. struct ips_driver *ips = data;
  683. unsigned long flags;
  684. dev_dbg(&ips->dev->dev, "starting ips-adjust thread\n");
  685. /*
  686. * Adjust CPU and GPU clamps every 5s if needed. Doing it more
  687. * often isn't recommended due to ME interaction.
  688. */
  689. do {
  690. bool cpu_busy = ips_cpu_busy(ips);
  691. bool gpu_busy = ips_gpu_busy(ips);
  692. spin_lock_irqsave(&ips->turbo_status_lock, flags);
  693. if (ips->poll_turbo_status)
  694. update_turbo_limits(ips);
  695. spin_unlock_irqrestore(&ips->turbo_status_lock, flags);
  696. /* Update turbo status if necessary */
  697. if (ips->cpu_turbo_enabled)
  698. ips_enable_cpu_turbo(ips);
  699. else
  700. ips_disable_cpu_turbo(ips);
  701. if (ips->gpu_turbo_enabled)
  702. ips_enable_gpu_turbo(ips);
  703. else
  704. ips_disable_gpu_turbo(ips);
  705. /* We're outside our comfort zone, crank them down */
  706. if (mcp_exceeded(ips)) {
  707. ips_cpu_lower(ips);
  708. ips_gpu_lower(ips);
  709. goto sleep;
  710. }
  711. if (!cpu_exceeded(ips, 0) && cpu_busy)
  712. ips_cpu_raise(ips);
  713. else
  714. ips_cpu_lower(ips);
  715. if (!mch_exceeded(ips) && gpu_busy)
  716. ips_gpu_raise(ips);
  717. else
  718. ips_gpu_lower(ips);
  719. sleep:
  720. schedule_timeout_interruptible(msecs_to_jiffies(IPS_ADJUST_PERIOD));
  721. } while (!kthread_should_stop());
  722. dev_dbg(&ips->dev->dev, "ips-adjust thread stopped\n");
  723. return 0;
  724. }
  725. /*
  726. * Helpers for reading out temp/power values and calculating their
  727. * averages for the decision making and monitoring functions.
  728. */
  729. static u16 calc_avg_temp(struct ips_driver *ips, u16 *array)
  730. {
  731. u64 total = 0;
  732. int i;
  733. u16 avg;
  734. for (i = 0; i < IPS_SAMPLE_COUNT; i++)
  735. total += (u64)(array[i] * 100);
  736. do_div(total, IPS_SAMPLE_COUNT);
  737. avg = (u16)total;
  738. return avg;
  739. }
  740. static u16 read_mgtv(struct ips_driver *ips)
  741. {
  742. u16 ret;
  743. u64 slope, offset;
  744. u64 val;
  745. val = thm_readq(THM_MGTV);
  746. val = (val & TV_MASK) >> TV_SHIFT;
  747. slope = offset = thm_readw(THM_MGTA);
  748. slope = (slope & MGTA_SLOPE_MASK) >> MGTA_SLOPE_SHIFT;
  749. offset = offset & MGTA_OFFSET_MASK;
  750. ret = ((val * slope + 0x40) >> 7) + offset;
  751. return 0; /* MCH temp reporting buggy */
  752. }
  753. static u16 read_ptv(struct ips_driver *ips)
  754. {
  755. u16 val, slope, offset;
  756. slope = (ips->pta_val & PTA_SLOPE_MASK) >> PTA_SLOPE_SHIFT;
  757. offset = ips->pta_val & PTA_OFFSET_MASK;
  758. val = thm_readw(THM_PTV) & PTV_MASK;
  759. return val;
  760. }
  761. static u16 read_ctv(struct ips_driver *ips, int cpu)
  762. {
  763. int reg = cpu ? THM_CTV2 : THM_CTV1;
  764. u16 val;
  765. val = thm_readw(reg);
  766. if (!(val & CTV_TEMP_ERROR))
  767. val = (val) >> 6; /* discard fractional component */
  768. else
  769. val = 0;
  770. return val;
  771. }
  772. static u32 get_cpu_power(struct ips_driver *ips, u32 *last, int period)
  773. {
  774. u32 val;
  775. u32 ret;
  776. /*
  777. * CEC is in joules/65535. Take difference over time to
  778. * get watts.
  779. */
  780. val = thm_readl(THM_CEC);
  781. /* period is in ms and we want mW */
  782. ret = (((val - *last) * 1000) / period);
  783. ret = (ret * 1000) / 65535;
  784. *last = val;
  785. return 0;
  786. }
  787. static const u16 temp_decay_factor = 2;
  788. static u16 update_average_temp(u16 avg, u16 val)
  789. {
  790. u16 ret;
  791. /* Multiply by 100 for extra precision */
  792. ret = (val * 100 / temp_decay_factor) +
  793. (((temp_decay_factor - 1) * avg) / temp_decay_factor);
  794. return ret;
  795. }
  796. static const u16 power_decay_factor = 2;
  797. static u16 update_average_power(u32 avg, u32 val)
  798. {
  799. u32 ret;
  800. ret = (val / power_decay_factor) +
  801. (((power_decay_factor - 1) * avg) / power_decay_factor);
  802. return ret;
  803. }
  804. static u32 calc_avg_power(struct ips_driver *ips, u32 *array)
  805. {
  806. u64 total = 0;
  807. u32 avg;
  808. int i;
  809. for (i = 0; i < IPS_SAMPLE_COUNT; i++)
  810. total += array[i];
  811. do_div(total, IPS_SAMPLE_COUNT);
  812. avg = (u32)total;
  813. return avg;
  814. }
  815. static void monitor_timeout(unsigned long arg)
  816. {
  817. wake_up_process((struct task_struct *)arg);
  818. }
  819. /**
  820. * ips_monitor - temp/power monitoring thread
  821. * @data: ips driver structure
  822. *
  823. * This is the main function for the IPS driver. It monitors power and
  824. * tempurature in the MCP and adjusts CPU and GPU power clams accordingly.
  825. *
  826. * We keep a 5s moving average of power consumption and tempurature. Using
  827. * that data, along with CPU vs GPU preference, we adjust the power clamps
  828. * up or down.
  829. */
  830. static int ips_monitor(void *data)
  831. {
  832. struct ips_driver *ips = data;
  833. struct timer_list timer;
  834. unsigned long seqno_timestamp, expire, last_msecs, last_sample_period;
  835. int i;
  836. u32 *cpu_samples, *mchp_samples, old_cpu_power;
  837. u16 *mcp_samples, *ctv1_samples, *ctv2_samples, *mch_samples;
  838. u8 cur_seqno, last_seqno;
  839. mcp_samples = kzalloc(sizeof(u16) * IPS_SAMPLE_COUNT, GFP_KERNEL);
  840. ctv1_samples = kzalloc(sizeof(u16) * IPS_SAMPLE_COUNT, GFP_KERNEL);
  841. ctv2_samples = kzalloc(sizeof(u16) * IPS_SAMPLE_COUNT, GFP_KERNEL);
  842. mch_samples = kzalloc(sizeof(u16) * IPS_SAMPLE_COUNT, GFP_KERNEL);
  843. cpu_samples = kzalloc(sizeof(u32) * IPS_SAMPLE_COUNT, GFP_KERNEL);
  844. mchp_samples = kzalloc(sizeof(u32) * IPS_SAMPLE_COUNT, GFP_KERNEL);
  845. if (!mcp_samples || !ctv1_samples || !ctv2_samples || !mch_samples ||
  846. !cpu_samples || !mchp_samples) {
  847. dev_err(&ips->dev->dev,
  848. "failed to allocate sample array, ips disabled\n");
  849. kfree(mcp_samples);
  850. kfree(ctv1_samples);
  851. kfree(ctv2_samples);
  852. kfree(mch_samples);
  853. kfree(cpu_samples);
  854. kfree(mchp_samples);
  855. return -ENOMEM;
  856. }
  857. last_seqno = (thm_readl(THM_ITV) & ITV_ME_SEQNO_MASK) >>
  858. ITV_ME_SEQNO_SHIFT;
  859. seqno_timestamp = get_jiffies_64();
  860. old_cpu_power = thm_readl(THM_CEC);
  861. schedule_timeout_interruptible(msecs_to_jiffies(IPS_SAMPLE_PERIOD));
  862. /* Collect an initial average */
  863. for (i = 0; i < IPS_SAMPLE_COUNT; i++) {
  864. u32 mchp, cpu_power;
  865. u16 val;
  866. mcp_samples[i] = read_ptv(ips);
  867. val = read_ctv(ips, 0);
  868. ctv1_samples[i] = val;
  869. val = read_ctv(ips, 1);
  870. ctv2_samples[i] = val;
  871. val = read_mgtv(ips);
  872. mch_samples[i] = val;
  873. cpu_power = get_cpu_power(ips, &old_cpu_power,
  874. IPS_SAMPLE_PERIOD);
  875. cpu_samples[i] = cpu_power;
  876. if (ips->read_mch_val) {
  877. mchp = ips->read_mch_val();
  878. mchp_samples[i] = mchp;
  879. }
  880. schedule_timeout_interruptible(msecs_to_jiffies(IPS_SAMPLE_PERIOD));
  881. if (kthread_should_stop())
  882. break;
  883. }
  884. ips->mcp_avg_temp = calc_avg_temp(ips, mcp_samples);
  885. ips->ctv1_avg_temp = calc_avg_temp(ips, ctv1_samples);
  886. ips->ctv2_avg_temp = calc_avg_temp(ips, ctv2_samples);
  887. ips->mch_avg_temp = calc_avg_temp(ips, mch_samples);
  888. ips->cpu_avg_power = calc_avg_power(ips, cpu_samples);
  889. ips->mch_avg_power = calc_avg_power(ips, mchp_samples);
  890. kfree(mcp_samples);
  891. kfree(ctv1_samples);
  892. kfree(ctv2_samples);
  893. kfree(mch_samples);
  894. kfree(cpu_samples);
  895. kfree(mchp_samples);
  896. /* Start the adjustment thread now that we have data */
  897. wake_up_process(ips->adjust);
  898. /*
  899. * Ok, now we have an initial avg. From here on out, we track the
  900. * running avg using a decaying average calculation. This allows
  901. * us to reduce the sample frequency if the CPU and GPU are idle.
  902. */
  903. old_cpu_power = thm_readl(THM_CEC);
  904. schedule_timeout_interruptible(msecs_to_jiffies(IPS_SAMPLE_PERIOD));
  905. last_sample_period = IPS_SAMPLE_PERIOD;
  906. setup_deferrable_timer_on_stack(&timer, monitor_timeout,
  907. (unsigned long)current);
  908. do {
  909. u32 cpu_val, mch_val;
  910. u16 val;
  911. /* MCP itself */
  912. val = read_ptv(ips);
  913. ips->mcp_avg_temp = update_average_temp(ips->mcp_avg_temp, val);
  914. /* Processor 0 */
  915. val = read_ctv(ips, 0);
  916. ips->ctv1_avg_temp =
  917. update_average_temp(ips->ctv1_avg_temp, val);
  918. /* Power */
  919. cpu_val = get_cpu_power(ips, &old_cpu_power,
  920. last_sample_period);
  921. ips->cpu_avg_power =
  922. update_average_power(ips->cpu_avg_power, cpu_val);
  923. if (ips->second_cpu) {
  924. /* Processor 1 */
  925. val = read_ctv(ips, 1);
  926. ips->ctv2_avg_temp =
  927. update_average_temp(ips->ctv2_avg_temp, val);
  928. }
  929. /* MCH */
  930. val = read_mgtv(ips);
  931. ips->mch_avg_temp = update_average_temp(ips->mch_avg_temp, val);
  932. /* Power */
  933. if (ips->read_mch_val) {
  934. mch_val = ips->read_mch_val();
  935. ips->mch_avg_power =
  936. update_average_power(ips->mch_avg_power,
  937. mch_val);
  938. }
  939. /*
  940. * Make sure ME is updating thermal regs.
  941. * Note:
  942. * If it's been more than a second since the last update,
  943. * the ME is probably hung.
  944. */
  945. cur_seqno = (thm_readl(THM_ITV) & ITV_ME_SEQNO_MASK) >>
  946. ITV_ME_SEQNO_SHIFT;
  947. if (cur_seqno == last_seqno &&
  948. time_after(jiffies, seqno_timestamp + HZ)) {
  949. dev_warn(&ips->dev->dev, "ME failed to update for more than 1s, likely hung\n");
  950. } else {
  951. seqno_timestamp = get_jiffies_64();
  952. last_seqno = cur_seqno;
  953. }
  954. last_msecs = jiffies_to_msecs(jiffies);
  955. expire = jiffies + msecs_to_jiffies(IPS_SAMPLE_PERIOD);
  956. __set_current_state(TASK_INTERRUPTIBLE);
  957. mod_timer(&timer, expire);
  958. schedule();
  959. /* Calculate actual sample period for power averaging */
  960. last_sample_period = jiffies_to_msecs(jiffies) - last_msecs;
  961. if (!last_sample_period)
  962. last_sample_period = 1;
  963. } while (!kthread_should_stop());
  964. del_timer_sync(&timer);
  965. destroy_timer_on_stack(&timer);
  966. dev_dbg(&ips->dev->dev, "ips-monitor thread stopped\n");
  967. return 0;
  968. }
  969. #if 0
  970. #define THM_DUMPW(reg) \
  971. { \
  972. u16 val = thm_readw(reg); \
  973. dev_dbg(&ips->dev->dev, #reg ": 0x%04x\n", val); \
  974. }
  975. #define THM_DUMPL(reg) \
  976. { \
  977. u32 val = thm_readl(reg); \
  978. dev_dbg(&ips->dev->dev, #reg ": 0x%08x\n", val); \
  979. }
  980. #define THM_DUMPQ(reg) \
  981. { \
  982. u64 val = thm_readq(reg); \
  983. dev_dbg(&ips->dev->dev, #reg ": 0x%016x\n", val); \
  984. }
  985. static void dump_thermal_info(struct ips_driver *ips)
  986. {
  987. u16 ptl;
  988. ptl = thm_readw(THM_PTL);
  989. dev_dbg(&ips->dev->dev, "Processor temp limit: %d\n", ptl);
  990. THM_DUMPW(THM_CTA);
  991. THM_DUMPW(THM_TRC);
  992. THM_DUMPW(THM_CTV1);
  993. THM_DUMPL(THM_STS);
  994. THM_DUMPW(THM_PTV);
  995. THM_DUMPQ(THM_MGTV);
  996. }
  997. #endif
  998. /**
  999. * ips_irq_handler - handle temperature triggers and other IPS events
  1000. * @irq: irq number
  1001. * @arg: unused
  1002. *
  1003. * Handle temperature limit trigger events, generally by lowering the clamps.
  1004. * If we're at a critical limit, we clamp back to the lowest possible value
  1005. * to prevent emergency shutdown.
  1006. */
  1007. static irqreturn_t ips_irq_handler(int irq, void *arg)
  1008. {
  1009. struct ips_driver *ips = arg;
  1010. u8 tses = thm_readb(THM_TSES);
  1011. u8 tes = thm_readb(THM_TES);
  1012. if (!tses && !tes)
  1013. return IRQ_NONE;
  1014. dev_info(&ips->dev->dev, "TSES: 0x%02x\n", tses);
  1015. dev_info(&ips->dev->dev, "TES: 0x%02x\n", tes);
  1016. /* STS update from EC? */
  1017. if (tes & 1) {
  1018. u32 sts, tc1;
  1019. sts = thm_readl(THM_STS);
  1020. tc1 = thm_readl(THM_TC1);
  1021. if (sts & STS_NVV) {
  1022. spin_lock(&ips->turbo_status_lock);
  1023. ips->core_power_limit = (sts & STS_PCPL_MASK) >>
  1024. STS_PCPL_SHIFT;
  1025. ips->mch_power_limit = (sts & STS_GPL_MASK) >>
  1026. STS_GPL_SHIFT;
  1027. /* ignore EC CPU vs GPU pref */
  1028. ips->cpu_turbo_enabled = !(sts & STS_PCTD_DIS);
  1029. /*
  1030. * Disable turbo for now, until we can figure
  1031. * out why the power figures are wrong
  1032. */
  1033. ips->cpu_turbo_enabled = false;
  1034. if (ips->gpu_busy)
  1035. ips->gpu_turbo_enabled = !(sts & STS_GTD_DIS);
  1036. ips->mcp_temp_limit = (sts & STS_PTL_MASK) >>
  1037. STS_PTL_SHIFT;
  1038. ips->mcp_power_limit = (tc1 & STS_PPL_MASK) >>
  1039. STS_PPL_SHIFT;
  1040. verify_limits(ips);
  1041. spin_unlock(&ips->turbo_status_lock);
  1042. thm_writeb(THM_SEC, SEC_ACK);
  1043. }
  1044. thm_writeb(THM_TES, tes);
  1045. }
  1046. /* Thermal trip */
  1047. if (tses) {
  1048. dev_warn(&ips->dev->dev,
  1049. "thermal trip occurred, tses: 0x%04x\n", tses);
  1050. thm_writeb(THM_TSES, tses);
  1051. }
  1052. return IRQ_HANDLED;
  1053. }
  1054. #ifndef CONFIG_DEBUG_FS
  1055. static void ips_debugfs_init(struct ips_driver *ips) { return; }
  1056. static void ips_debugfs_cleanup(struct ips_driver *ips) { return; }
  1057. #else
  1058. /* Expose current state and limits in debugfs if possible */
  1059. struct ips_debugfs_node {
  1060. struct ips_driver *ips;
  1061. char *name;
  1062. int (*show)(struct seq_file *m, void *data);
  1063. };
  1064. static int show_cpu_temp(struct seq_file *m, void *data)
  1065. {
  1066. struct ips_driver *ips = m->private;
  1067. seq_printf(m, "%d.%02d\n", ips->ctv1_avg_temp / 100,
  1068. ips->ctv1_avg_temp % 100);
  1069. return 0;
  1070. }
  1071. static int show_cpu_power(struct seq_file *m, void *data)
  1072. {
  1073. struct ips_driver *ips = m->private;
  1074. seq_printf(m, "%dmW\n", ips->cpu_avg_power);
  1075. return 0;
  1076. }
  1077. static int show_cpu_clamp(struct seq_file *m, void *data)
  1078. {
  1079. u64 turbo_override;
  1080. int tdp, tdc;
  1081. rdmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  1082. tdp = (int)(turbo_override & TURBO_TDP_MASK);
  1083. tdc = (int)((turbo_override & TURBO_TDC_MASK) >> TURBO_TDC_SHIFT);
  1084. /* Convert to .1W/A units */
  1085. tdp = tdp * 10 / 8;
  1086. tdc = tdc * 10 / 8;
  1087. /* Watts Amperes */
  1088. seq_printf(m, "%d.%dW %d.%dA\n", tdp / 10, tdp % 10,
  1089. tdc / 10, tdc % 10);
  1090. return 0;
  1091. }
  1092. static int show_mch_temp(struct seq_file *m, void *data)
  1093. {
  1094. struct ips_driver *ips = m->private;
  1095. seq_printf(m, "%d.%02d\n", ips->mch_avg_temp / 100,
  1096. ips->mch_avg_temp % 100);
  1097. return 0;
  1098. }
  1099. static int show_mch_power(struct seq_file *m, void *data)
  1100. {
  1101. struct ips_driver *ips = m->private;
  1102. seq_printf(m, "%dmW\n", ips->mch_avg_power);
  1103. return 0;
  1104. }
  1105. static struct ips_debugfs_node ips_debug_files[] = {
  1106. { NULL, "cpu_temp", show_cpu_temp },
  1107. { NULL, "cpu_power", show_cpu_power },
  1108. { NULL, "cpu_clamp", show_cpu_clamp },
  1109. { NULL, "mch_temp", show_mch_temp },
  1110. { NULL, "mch_power", show_mch_power },
  1111. };
  1112. static int ips_debugfs_open(struct inode *inode, struct file *file)
  1113. {
  1114. struct ips_debugfs_node *node = inode->i_private;
  1115. return single_open(file, node->show, node->ips);
  1116. }
  1117. static const struct file_operations ips_debugfs_ops = {
  1118. .owner = THIS_MODULE,
  1119. .open = ips_debugfs_open,
  1120. .read = seq_read,
  1121. .llseek = seq_lseek,
  1122. .release = single_release,
  1123. };
  1124. static void ips_debugfs_cleanup(struct ips_driver *ips)
  1125. {
  1126. if (ips->debug_root)
  1127. debugfs_remove_recursive(ips->debug_root);
  1128. return;
  1129. }
  1130. static void ips_debugfs_init(struct ips_driver *ips)
  1131. {
  1132. int i;
  1133. ips->debug_root = debugfs_create_dir("ips", NULL);
  1134. if (!ips->debug_root) {
  1135. dev_err(&ips->dev->dev,
  1136. "failed to create debugfs entries: %ld\n",
  1137. PTR_ERR(ips->debug_root));
  1138. return;
  1139. }
  1140. for (i = 0; i < ARRAY_SIZE(ips_debug_files); i++) {
  1141. struct dentry *ent;
  1142. struct ips_debugfs_node *node = &ips_debug_files[i];
  1143. node->ips = ips;
  1144. ent = debugfs_create_file(node->name, S_IFREG | S_IRUGO,
  1145. ips->debug_root, node,
  1146. &ips_debugfs_ops);
  1147. if (!ent) {
  1148. dev_err(&ips->dev->dev,
  1149. "failed to create debug file: %ld\n",
  1150. PTR_ERR(ent));
  1151. goto err_cleanup;
  1152. }
  1153. }
  1154. return;
  1155. err_cleanup:
  1156. ips_debugfs_cleanup(ips);
  1157. return;
  1158. }
  1159. #endif /* CONFIG_DEBUG_FS */
  1160. /**
  1161. * ips_detect_cpu - detect whether CPU supports IPS
  1162. *
  1163. * Walk our list and see if we're on a supported CPU. If we find one,
  1164. * return the limits for it.
  1165. */
  1166. static struct ips_mcp_limits *ips_detect_cpu(struct ips_driver *ips)
  1167. {
  1168. u64 turbo_power, misc_en;
  1169. struct ips_mcp_limits *limits = NULL;
  1170. u16 tdp;
  1171. if (!(boot_cpu_data.x86 == 6 && boot_cpu_data.x86_model == 37)) {
  1172. dev_info(&ips->dev->dev, "Non-IPS CPU detected.\n");
  1173. goto out;
  1174. }
  1175. rdmsrl(IA32_MISC_ENABLE, misc_en);
  1176. /*
  1177. * If the turbo enable bit isn't set, we shouldn't try to enable/disable
  1178. * turbo manually or we'll get an illegal MSR access, even though
  1179. * turbo will still be available.
  1180. */
  1181. if (misc_en & IA32_MISC_TURBO_EN)
  1182. ips->turbo_toggle_allowed = true;
  1183. else
  1184. ips->turbo_toggle_allowed = false;
  1185. if (strstr(boot_cpu_data.x86_model_id, "CPU M"))
  1186. limits = &ips_sv_limits;
  1187. else if (strstr(boot_cpu_data.x86_model_id, "CPU L"))
  1188. limits = &ips_lv_limits;
  1189. else if (strstr(boot_cpu_data.x86_model_id, "CPU U"))
  1190. limits = &ips_ulv_limits;
  1191. else {
  1192. dev_info(&ips->dev->dev, "No CPUID match found.\n");
  1193. goto out;
  1194. }
  1195. rdmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_power);
  1196. tdp = turbo_power & TURBO_TDP_MASK;
  1197. /* Sanity check TDP against CPU */
  1198. if (limits->core_power_limit != (tdp / 8) * 1000) {
  1199. dev_info(&ips->dev->dev, "CPU TDP doesn't match expected value (found %d, expected %d)\n",
  1200. tdp / 8, limits->core_power_limit / 1000);
  1201. limits->core_power_limit = (tdp / 8) * 1000;
  1202. }
  1203. out:
  1204. return limits;
  1205. }
  1206. /**
  1207. * ips_get_i915_syms - try to get GPU control methods from i915 driver
  1208. * @ips: IPS driver
  1209. *
  1210. * The i915 driver exports several interfaces to allow the IPS driver to
  1211. * monitor and control graphics turbo mode. If we can find them, we can
  1212. * enable graphics turbo, otherwise we must disable it to avoid exceeding
  1213. * thermal and power limits in the MCP.
  1214. */
  1215. static bool ips_get_i915_syms(struct ips_driver *ips)
  1216. {
  1217. ips->read_mch_val = symbol_get(i915_read_mch_val);
  1218. if (!ips->read_mch_val)
  1219. goto out_err;
  1220. ips->gpu_raise = symbol_get(i915_gpu_raise);
  1221. if (!ips->gpu_raise)
  1222. goto out_put_mch;
  1223. ips->gpu_lower = symbol_get(i915_gpu_lower);
  1224. if (!ips->gpu_lower)
  1225. goto out_put_raise;
  1226. ips->gpu_busy = symbol_get(i915_gpu_busy);
  1227. if (!ips->gpu_busy)
  1228. goto out_put_lower;
  1229. ips->gpu_turbo_disable = symbol_get(i915_gpu_turbo_disable);
  1230. if (!ips->gpu_turbo_disable)
  1231. goto out_put_busy;
  1232. return true;
  1233. out_put_busy:
  1234. symbol_put(i915_gpu_busy);
  1235. out_put_lower:
  1236. symbol_put(i915_gpu_lower);
  1237. out_put_raise:
  1238. symbol_put(i915_gpu_raise);
  1239. out_put_mch:
  1240. symbol_put(i915_read_mch_val);
  1241. out_err:
  1242. return false;
  1243. }
  1244. static bool
  1245. ips_gpu_turbo_enabled(struct ips_driver *ips)
  1246. {
  1247. if (!ips->gpu_busy && late_i915_load) {
  1248. if (ips_get_i915_syms(ips)) {
  1249. dev_info(&ips->dev->dev,
  1250. "i915 driver attached, reenabling gpu turbo\n");
  1251. ips->gpu_turbo_enabled = !(thm_readl(THM_HTS) & HTS_GTD_DIS);
  1252. }
  1253. }
  1254. return ips->gpu_turbo_enabled;
  1255. }
  1256. void
  1257. ips_link_to_i915_driver(void)
  1258. {
  1259. /* We can't cleanly get at the various ips_driver structs from
  1260. * this caller (the i915 driver), so just set a flag saying
  1261. * that it's time to try getting the symbols again.
  1262. */
  1263. late_i915_load = true;
  1264. }
  1265. EXPORT_SYMBOL_GPL(ips_link_to_i915_driver);
  1266. static const struct pci_device_id ips_id_table[] = {
  1267. { PCI_DEVICE(PCI_VENDOR_ID_INTEL,
  1268. PCI_DEVICE_ID_INTEL_THERMAL_SENSOR), },
  1269. { 0, }
  1270. };
  1271. MODULE_DEVICE_TABLE(pci, ips_id_table);
  1272. static int ips_blacklist_callback(const struct dmi_system_id *id)
  1273. {
  1274. pr_info("Blacklisted intel_ips for %s\n", id->ident);
  1275. return 1;
  1276. }
  1277. static const struct dmi_system_id ips_blacklist[] = {
  1278. {
  1279. .callback = ips_blacklist_callback,
  1280. .ident = "HP ProBook",
  1281. .matches = {
  1282. DMI_MATCH(DMI_SYS_VENDOR, "Hewlett-Packard"),
  1283. DMI_MATCH(DMI_PRODUCT_NAME, "HP ProBook"),
  1284. },
  1285. },
  1286. { } /* terminating entry */
  1287. };
  1288. static int ips_probe(struct pci_dev *dev, const struct pci_device_id *id)
  1289. {
  1290. u64 platform_info;
  1291. struct ips_driver *ips;
  1292. u32 hts;
  1293. int ret = 0;
  1294. u16 htshi, trc, trc_required_mask;
  1295. u8 tse;
  1296. if (dmi_check_system(ips_blacklist))
  1297. return -ENODEV;
  1298. ips = kzalloc(sizeof(struct ips_driver), GFP_KERNEL);
  1299. if (!ips)
  1300. return -ENOMEM;
  1301. pci_set_drvdata(dev, ips);
  1302. ips->dev = dev;
  1303. ips->limits = ips_detect_cpu(ips);
  1304. if (!ips->limits) {
  1305. dev_info(&dev->dev, "IPS not supported on this CPU\n");
  1306. ret = -ENXIO;
  1307. goto error_free;
  1308. }
  1309. spin_lock_init(&ips->turbo_status_lock);
  1310. ret = pci_enable_device(dev);
  1311. if (ret) {
  1312. dev_err(&dev->dev, "can't enable PCI device, aborting\n");
  1313. goto error_free;
  1314. }
  1315. if (!pci_resource_start(dev, 0)) {
  1316. dev_err(&dev->dev, "TBAR not assigned, aborting\n");
  1317. ret = -ENXIO;
  1318. goto error_free;
  1319. }
  1320. ret = pci_request_regions(dev, "ips thermal sensor");
  1321. if (ret) {
  1322. dev_err(&dev->dev, "thermal resource busy, aborting\n");
  1323. goto error_free;
  1324. }
  1325. ips->regmap = ioremap(pci_resource_start(dev, 0),
  1326. pci_resource_len(dev, 0));
  1327. if (!ips->regmap) {
  1328. dev_err(&dev->dev, "failed to map thermal regs, aborting\n");
  1329. ret = -EBUSY;
  1330. goto error_release;
  1331. }
  1332. tse = thm_readb(THM_TSE);
  1333. if (tse != TSE_EN) {
  1334. dev_err(&dev->dev, "thermal device not enabled (0x%02x), aborting\n", tse);
  1335. ret = -ENXIO;
  1336. goto error_unmap;
  1337. }
  1338. trc = thm_readw(THM_TRC);
  1339. trc_required_mask = TRC_CORE1_EN | TRC_CORE_PWR | TRC_MCH_EN;
  1340. if ((trc & trc_required_mask) != trc_required_mask) {
  1341. dev_err(&dev->dev, "thermal reporting for required devices not enabled, aborting\n");
  1342. ret = -ENXIO;
  1343. goto error_unmap;
  1344. }
  1345. if (trc & TRC_CORE2_EN)
  1346. ips->second_cpu = true;
  1347. update_turbo_limits(ips);
  1348. dev_dbg(&dev->dev, "max cpu power clamp: %dW\n",
  1349. ips->mcp_power_limit / 10);
  1350. dev_dbg(&dev->dev, "max core power clamp: %dW\n",
  1351. ips->core_power_limit / 10);
  1352. /* BIOS may update limits at runtime */
  1353. if (thm_readl(THM_PSC) & PSP_PBRT)
  1354. ips->poll_turbo_status = true;
  1355. if (!ips_get_i915_syms(ips)) {
  1356. dev_info(&dev->dev, "failed to get i915 symbols, graphics turbo disabled until i915 loads\n");
  1357. ips->gpu_turbo_enabled = false;
  1358. } else {
  1359. dev_dbg(&dev->dev, "graphics turbo enabled\n");
  1360. ips->gpu_turbo_enabled = true;
  1361. }
  1362. /*
  1363. * Check PLATFORM_INFO MSR to make sure this chip is
  1364. * turbo capable.
  1365. */
  1366. rdmsrl(PLATFORM_INFO, platform_info);
  1367. if (!(platform_info & PLATFORM_TDP)) {
  1368. dev_err(&dev->dev, "platform indicates TDP override unavailable, aborting\n");
  1369. ret = -ENODEV;
  1370. goto error_unmap;
  1371. }
  1372. /*
  1373. * IRQ handler for ME interaction
  1374. * Note: don't use MSI here as the PCH has bugs.
  1375. */
  1376. pci_disable_msi(dev);
  1377. ret = request_irq(dev->irq, ips_irq_handler, IRQF_SHARED, "ips",
  1378. ips);
  1379. if (ret) {
  1380. dev_err(&dev->dev, "request irq failed, aborting\n");
  1381. goto error_unmap;
  1382. }
  1383. /* Enable aux, hot & critical interrupts */
  1384. thm_writeb(THM_TSPIEN, TSPIEN_AUX2_LOHI | TSPIEN_CRIT_LOHI |
  1385. TSPIEN_HOT_LOHI | TSPIEN_AUX_LOHI);
  1386. thm_writeb(THM_TEN, TEN_UPDATE_EN);
  1387. /* Collect adjustment values */
  1388. ips->cta_val = thm_readw(THM_CTA);
  1389. ips->pta_val = thm_readw(THM_PTA);
  1390. ips->mgta_val = thm_readw(THM_MGTA);
  1391. /* Save turbo limits & ratios */
  1392. rdmsrl(TURBO_POWER_CURRENT_LIMIT, ips->orig_turbo_limit);
  1393. ips_disable_cpu_turbo(ips);
  1394. ips->cpu_turbo_enabled = false;
  1395. /* Create thermal adjust thread */
  1396. ips->adjust = kthread_create(ips_adjust, ips, "ips-adjust");
  1397. if (IS_ERR(ips->adjust)) {
  1398. dev_err(&dev->dev,
  1399. "failed to create thermal adjust thread, aborting\n");
  1400. ret = -ENOMEM;
  1401. goto error_free_irq;
  1402. }
  1403. /*
  1404. * Set up the work queue and monitor thread. The monitor thread
  1405. * will wake up ips_adjust thread.
  1406. */
  1407. ips->monitor = kthread_run(ips_monitor, ips, "ips-monitor");
  1408. if (IS_ERR(ips->monitor)) {
  1409. dev_err(&dev->dev,
  1410. "failed to create thermal monitor thread, aborting\n");
  1411. ret = -ENOMEM;
  1412. goto error_thread_cleanup;
  1413. }
  1414. hts = (ips->core_power_limit << HTS_PCPL_SHIFT) |
  1415. (ips->mcp_temp_limit << HTS_PTL_SHIFT) | HTS_NVV;
  1416. htshi = HTS2_PRST_RUNNING << HTS2_PRST_SHIFT;
  1417. thm_writew(THM_HTSHI, htshi);
  1418. thm_writel(THM_HTS, hts);
  1419. ips_debugfs_init(ips);
  1420. dev_info(&dev->dev, "IPS driver initialized, MCP temp limit %d\n",
  1421. ips->mcp_temp_limit);
  1422. return ret;
  1423. error_thread_cleanup:
  1424. kthread_stop(ips->adjust);
  1425. error_free_irq:
  1426. free_irq(ips->dev->irq, ips);
  1427. error_unmap:
  1428. iounmap(ips->regmap);
  1429. error_release:
  1430. pci_release_regions(dev);
  1431. error_free:
  1432. kfree(ips);
  1433. return ret;
  1434. }
  1435. static void ips_remove(struct pci_dev *dev)
  1436. {
  1437. struct ips_driver *ips = pci_get_drvdata(dev);
  1438. u64 turbo_override;
  1439. if (!ips)
  1440. return;
  1441. ips_debugfs_cleanup(ips);
  1442. /* Release i915 driver */
  1443. if (ips->read_mch_val)
  1444. symbol_put(i915_read_mch_val);
  1445. if (ips->gpu_raise)
  1446. symbol_put(i915_gpu_raise);
  1447. if (ips->gpu_lower)
  1448. symbol_put(i915_gpu_lower);
  1449. if (ips->gpu_busy)
  1450. symbol_put(i915_gpu_busy);
  1451. if (ips->gpu_turbo_disable)
  1452. symbol_put(i915_gpu_turbo_disable);
  1453. rdmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  1454. turbo_override &= ~(TURBO_TDC_OVR_EN | TURBO_TDP_OVR_EN);
  1455. wrmsrl(TURBO_POWER_CURRENT_LIMIT, turbo_override);
  1456. wrmsrl(TURBO_POWER_CURRENT_LIMIT, ips->orig_turbo_limit);
  1457. free_irq(ips->dev->irq, ips);
  1458. if (ips->adjust)
  1459. kthread_stop(ips->adjust);
  1460. if (ips->monitor)
  1461. kthread_stop(ips->monitor);
  1462. iounmap(ips->regmap);
  1463. pci_release_regions(dev);
  1464. kfree(ips);
  1465. dev_dbg(&dev->dev, "IPS driver removed\n");
  1466. }
  1467. static void ips_shutdown(struct pci_dev *dev)
  1468. {
  1469. }
  1470. static struct pci_driver ips_pci_driver = {
  1471. .name = "intel ips",
  1472. .id_table = ips_id_table,
  1473. .probe = ips_probe,
  1474. .remove = ips_remove,
  1475. .shutdown = ips_shutdown,
  1476. };
  1477. module_pci_driver(ips_pci_driver);
  1478. MODULE_LICENSE("GPL");
  1479. MODULE_AUTHOR("Jesse Barnes <jbarnes@virtuousgeek.org>");
  1480. MODULE_DESCRIPTION("Intelligent Power Sharing Driver");