cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

smu8_smumgr.c (25754B)


      1/*
      2 * Copyright 2015 Advanced Micro Devices, Inc.
      3 *
      4 * Permission is hereby granted, free of charge, to any person obtaining a
      5 * copy of this software and associated documentation files (the "Software"),
      6 * to deal in the Software without restriction, including without limitation
      7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
      8 * and/or sell copies of the Software, and to permit persons to whom the
      9 * Software is furnished to do so, subject to the following conditions:
     10 *
     11 * The above copyright notice and this permission notice shall be included in
     12 * all copies or substantial portions of the Software.
     13 *
     14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
     18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
     19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
     20 * OTHER DEALINGS IN THE SOFTWARE.
     21 *
     22 */
     23
     24#include <linux/delay.h>
     25#include <linux/gfp.h>
     26#include <linux/kernel.h>
     27#include <linux/ktime.h>
     28#include <linux/slab.h>
     29#include <linux/types.h>
     30
     31#include "cgs_common.h"
     32#include "smu/smu_8_0_d.h"
     33#include "smu/smu_8_0_sh_mask.h"
     34#include "smu8.h"
     35#include "smu8_fusion.h"
     36#include "smu8_smumgr.h"
     37#include "cz_ppsmc.h"
     38#include "smu_ucode_xfer_cz.h"
     39#include "gca/gfx_8_0_d.h"
     40#include "gca/gfx_8_0_sh_mask.h"
     41#include "smumgr.h"
     42
     43#define SIZE_ALIGN_32(x)    (((x) + 31) / 32 * 32)
     44
     45static const enum smu8_scratch_entry firmware_list[] = {
     46	SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA0,
     47	SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA1,
     48	SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE,
     49	SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP,
     50	SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME,
     51	SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1,
     52	SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2,
     53	SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G,
     54};
     55
     56static uint32_t smu8_get_argument(struct pp_hwmgr *hwmgr)
     57{
     58	if (hwmgr == NULL || hwmgr->device == NULL)
     59		return 0;
     60
     61	return cgs_read_register(hwmgr->device,
     62					mmSMU_MP1_SRBM2P_ARG_0);
     63}
     64
     65/* Send a message to the SMC, and wait for its response.*/
     66static int smu8_send_msg_to_smc_with_parameter(struct pp_hwmgr *hwmgr,
     67					    uint16_t msg, uint32_t parameter)
     68{
     69	int result = 0;
     70	ktime_t t_start;
     71	s64 elapsed_us;
     72
     73	if (hwmgr == NULL || hwmgr->device == NULL)
     74		return -EINVAL;
     75
     76	result = PHM_WAIT_FIELD_UNEQUAL(hwmgr,
     77					SMU_MP1_SRBM2P_RESP_0, CONTENT, 0);
     78	if (result != 0) {
     79		/* Read the last message to SMU, to report actual cause */
     80		uint32_t val = cgs_read_register(hwmgr->device,
     81						 mmSMU_MP1_SRBM2P_MSG_0);
     82		pr_err("%s(0x%04x) aborted; SMU still servicing msg (0x%04x)\n",
     83			__func__, msg, val);
     84		return result;
     85	}
     86	t_start = ktime_get();
     87
     88	cgs_write_register(hwmgr->device, mmSMU_MP1_SRBM2P_ARG_0, parameter);
     89
     90	cgs_write_register(hwmgr->device, mmSMU_MP1_SRBM2P_RESP_0, 0);
     91	cgs_write_register(hwmgr->device, mmSMU_MP1_SRBM2P_MSG_0, msg);
     92
     93	result = PHM_WAIT_FIELD_UNEQUAL(hwmgr,
     94					SMU_MP1_SRBM2P_RESP_0, CONTENT, 0);
     95
     96	elapsed_us = ktime_us_delta(ktime_get(), t_start);
     97
     98	WARN(result, "%s(0x%04x, %#x) timed out after %lld us\n",
     99			__func__, msg, parameter, elapsed_us);
    100
    101	return result;
    102}
    103
    104static int smu8_send_msg_to_smc(struct pp_hwmgr *hwmgr, uint16_t msg)
    105{
    106	return smu8_send_msg_to_smc_with_parameter(hwmgr, msg, 0);
    107}
    108
    109static int smu8_set_smc_sram_address(struct pp_hwmgr *hwmgr,
    110				     uint32_t smc_address, uint32_t limit)
    111{
    112	if (hwmgr == NULL || hwmgr->device == NULL)
    113		return -EINVAL;
    114
    115	if (0 != (3 & smc_address)) {
    116		pr_err("SMC address must be 4 byte aligned\n");
    117		return -EINVAL;
    118	}
    119
    120	if (limit <= (smc_address + 3)) {
    121		pr_err("SMC address beyond the SMC RAM area\n");
    122		return -EINVAL;
    123	}
    124
    125	cgs_write_register(hwmgr->device, mmMP0PUB_IND_INDEX_0,
    126				SMN_MP1_SRAM_START_ADDR + smc_address);
    127
    128	return 0;
    129}
    130
    131static int smu8_write_smc_sram_dword(struct pp_hwmgr *hwmgr,
    132		uint32_t smc_address, uint32_t value, uint32_t limit)
    133{
    134	int result;
    135
    136	if (hwmgr == NULL || hwmgr->device == NULL)
    137		return -EINVAL;
    138
    139	result = smu8_set_smc_sram_address(hwmgr, smc_address, limit);
    140	if (!result)
    141		cgs_write_register(hwmgr->device, mmMP0PUB_IND_DATA_0, value);
    142
    143	return result;
    144}
    145
    146static int smu8_check_fw_load_finish(struct pp_hwmgr *hwmgr,
    147				   uint32_t firmware)
    148{
    149	int i;
    150	uint32_t index = SMN_MP1_SRAM_START_ADDR +
    151			 SMU8_FIRMWARE_HEADER_LOCATION +
    152			 offsetof(struct SMU8_Firmware_Header, UcodeLoadStatus);
    153
    154	if (hwmgr == NULL || hwmgr->device == NULL)
    155		return -EINVAL;
    156
    157	cgs_write_register(hwmgr->device, mmMP0PUB_IND_INDEX, index);
    158
    159	for (i = 0; i < hwmgr->usec_timeout; i++) {
    160		if (firmware ==
    161			(cgs_read_register(hwmgr->device, mmMP0PUB_IND_DATA) & firmware))
    162			break;
    163		udelay(1);
    164	}
    165
    166	if (i >= hwmgr->usec_timeout) {
    167		pr_err("SMU check loaded firmware failed.\n");
    168		return -EINVAL;
    169	}
    170
    171	return 0;
    172}
    173
    174static int smu8_load_mec_firmware(struct pp_hwmgr *hwmgr)
    175{
    176	uint32_t reg_data;
    177	uint32_t tmp;
    178	int ret = 0;
    179	struct cgs_firmware_info info = {0};
    180
    181	if (hwmgr == NULL || hwmgr->device == NULL)
    182		return -EINVAL;
    183
    184	ret = cgs_get_firmware_info(hwmgr->device,
    185						CGS_UCODE_ID_CP_MEC, &info);
    186
    187	if (ret)
    188		return -EINVAL;
    189
    190	/* Disable MEC parsing/prefetching */
    191	tmp = cgs_read_register(hwmgr->device,
    192					mmCP_MEC_CNTL);
    193	tmp = PHM_SET_FIELD(tmp, CP_MEC_CNTL, MEC_ME1_HALT, 1);
    194	tmp = PHM_SET_FIELD(tmp, CP_MEC_CNTL, MEC_ME2_HALT, 1);
    195	cgs_write_register(hwmgr->device, mmCP_MEC_CNTL, tmp);
    196
    197	tmp = cgs_read_register(hwmgr->device,
    198					mmCP_CPC_IC_BASE_CNTL);
    199
    200	tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, VMID, 0);
    201	tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, ATC, 0);
    202	tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, CACHE_POLICY, 0);
    203	tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, MTYPE, 1);
    204	cgs_write_register(hwmgr->device, mmCP_CPC_IC_BASE_CNTL, tmp);
    205
    206	reg_data = lower_32_bits(info.mc_addr) &
    207			PHM_FIELD_MASK(CP_CPC_IC_BASE_LO, IC_BASE_LO);
    208	cgs_write_register(hwmgr->device, mmCP_CPC_IC_BASE_LO, reg_data);
    209
    210	reg_data = upper_32_bits(info.mc_addr) &
    211			PHM_FIELD_MASK(CP_CPC_IC_BASE_HI, IC_BASE_HI);
    212	cgs_write_register(hwmgr->device, mmCP_CPC_IC_BASE_HI, reg_data);
    213
    214	return 0;
    215}
    216
    217static uint8_t smu8_translate_firmware_enum_to_arg(struct pp_hwmgr *hwmgr,
    218			enum smu8_scratch_entry firmware_enum)
    219{
    220	uint8_t ret = 0;
    221
    222	switch (firmware_enum) {
    223	case SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA0:
    224		ret = UCODE_ID_SDMA0;
    225		break;
    226	case SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA1:
    227		if (hwmgr->chip_id == CHIP_STONEY)
    228			ret = UCODE_ID_SDMA0;
    229		else
    230			ret = UCODE_ID_SDMA1;
    231		break;
    232	case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE:
    233		ret = UCODE_ID_CP_CE;
    234		break;
    235	case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP:
    236		ret = UCODE_ID_CP_PFP;
    237		break;
    238	case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME:
    239		ret = UCODE_ID_CP_ME;
    240		break;
    241	case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1:
    242		ret = UCODE_ID_CP_MEC_JT1;
    243		break;
    244	case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2:
    245		if (hwmgr->chip_id == CHIP_STONEY)
    246			ret = UCODE_ID_CP_MEC_JT1;
    247		else
    248			ret = UCODE_ID_CP_MEC_JT2;
    249		break;
    250	case SMU8_SCRATCH_ENTRY_UCODE_ID_GMCON_RENG:
    251		ret = UCODE_ID_GMCON_RENG;
    252		break;
    253	case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G:
    254		ret = UCODE_ID_RLC_G;
    255		break;
    256	case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH:
    257		ret = UCODE_ID_RLC_SCRATCH;
    258		break;
    259	case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM:
    260		ret = UCODE_ID_RLC_SRM_ARAM;
    261		break;
    262	case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM:
    263		ret = UCODE_ID_RLC_SRM_DRAM;
    264		break;
    265	case SMU8_SCRATCH_ENTRY_UCODE_ID_DMCU_ERAM:
    266		ret = UCODE_ID_DMCU_ERAM;
    267		break;
    268	case SMU8_SCRATCH_ENTRY_UCODE_ID_DMCU_IRAM:
    269		ret = UCODE_ID_DMCU_IRAM;
    270		break;
    271	case SMU8_SCRATCH_ENTRY_UCODE_ID_POWER_PROFILING:
    272		ret = TASK_ARG_INIT_MM_PWR_LOG;
    273		break;
    274	case SMU8_SCRATCH_ENTRY_DATA_ID_SDMA_HALT:
    275	case SMU8_SCRATCH_ENTRY_DATA_ID_SYS_CLOCKGATING:
    276	case SMU8_SCRATCH_ENTRY_DATA_ID_SDMA_RING_REGS:
    277	case SMU8_SCRATCH_ENTRY_DATA_ID_NONGFX_REINIT:
    278	case SMU8_SCRATCH_ENTRY_DATA_ID_SDMA_START:
    279	case SMU8_SCRATCH_ENTRY_DATA_ID_IH_REGISTERS:
    280		ret = TASK_ARG_REG_MMIO;
    281		break;
    282	case SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE:
    283		ret = TASK_ARG_INIT_CLK_TABLE;
    284		break;
    285	}
    286
    287	return ret;
    288}
    289
    290static enum cgs_ucode_id smu8_convert_fw_type_to_cgs(uint32_t fw_type)
    291{
    292	enum cgs_ucode_id result = CGS_UCODE_ID_MAXIMUM;
    293
    294	switch (fw_type) {
    295	case UCODE_ID_SDMA0:
    296		result = CGS_UCODE_ID_SDMA0;
    297		break;
    298	case UCODE_ID_SDMA1:
    299		result = CGS_UCODE_ID_SDMA1;
    300		break;
    301	case UCODE_ID_CP_CE:
    302		result = CGS_UCODE_ID_CP_CE;
    303		break;
    304	case UCODE_ID_CP_PFP:
    305		result = CGS_UCODE_ID_CP_PFP;
    306		break;
    307	case UCODE_ID_CP_ME:
    308		result = CGS_UCODE_ID_CP_ME;
    309		break;
    310	case UCODE_ID_CP_MEC_JT1:
    311		result = CGS_UCODE_ID_CP_MEC_JT1;
    312		break;
    313	case UCODE_ID_CP_MEC_JT2:
    314		result = CGS_UCODE_ID_CP_MEC_JT2;
    315		break;
    316	case UCODE_ID_RLC_G:
    317		result = CGS_UCODE_ID_RLC_G;
    318		break;
    319	default:
    320		break;
    321	}
    322
    323	return result;
    324}
    325
    326static int smu8_smu_populate_single_scratch_task(
    327			struct pp_hwmgr *hwmgr,
    328			enum smu8_scratch_entry fw_enum,
    329			uint8_t type, bool is_last)
    330{
    331	uint8_t i;
    332	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    333	struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
    334	struct SMU_Task *task = &toc->tasks[smu8_smu->toc_entry_used_count++];
    335
    336	task->type = type;
    337	task->arg = smu8_translate_firmware_enum_to_arg(hwmgr, fw_enum);
    338	task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
    339
    340	for (i = 0; i < smu8_smu->scratch_buffer_length; i++)
    341		if (smu8_smu->scratch_buffer[i].firmware_ID == fw_enum)
    342			break;
    343
    344	if (i >= smu8_smu->scratch_buffer_length) {
    345		pr_err("Invalid Firmware Type\n");
    346		return -EINVAL;
    347	}
    348
    349	task->addr.low = lower_32_bits(smu8_smu->scratch_buffer[i].mc_addr);
    350	task->addr.high = upper_32_bits(smu8_smu->scratch_buffer[i].mc_addr);
    351	task->size_bytes = smu8_smu->scratch_buffer[i].data_size;
    352
    353	if (SMU8_SCRATCH_ENTRY_DATA_ID_IH_REGISTERS == fw_enum) {
    354		struct smu8_ih_meta_data *pIHReg_restore =
    355		     (struct smu8_ih_meta_data *)smu8_smu->scratch_buffer[i].kaddr;
    356		pIHReg_restore->command =
    357			METADATA_CMD_MODE0 | METADATA_PERFORM_ON_LOAD;
    358	}
    359
    360	return 0;
    361}
    362
    363static int smu8_smu_populate_single_ucode_load_task(
    364					struct pp_hwmgr *hwmgr,
    365					enum smu8_scratch_entry fw_enum,
    366					bool is_last)
    367{
    368	uint8_t i;
    369	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    370	struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
    371	struct SMU_Task *task = &toc->tasks[smu8_smu->toc_entry_used_count++];
    372
    373	task->type = TASK_TYPE_UCODE_LOAD;
    374	task->arg = smu8_translate_firmware_enum_to_arg(hwmgr, fw_enum);
    375	task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
    376
    377	for (i = 0; i < smu8_smu->driver_buffer_length; i++)
    378		if (smu8_smu->driver_buffer[i].firmware_ID == fw_enum)
    379			break;
    380
    381	if (i >= smu8_smu->driver_buffer_length) {
    382		pr_err("Invalid Firmware Type\n");
    383		return -EINVAL;
    384	}
    385
    386	task->addr.low = lower_32_bits(smu8_smu->driver_buffer[i].mc_addr);
    387	task->addr.high = upper_32_bits(smu8_smu->driver_buffer[i].mc_addr);
    388	task->size_bytes = smu8_smu->driver_buffer[i].data_size;
    389
    390	return 0;
    391}
    392
    393static int smu8_smu_construct_toc_for_rlc_aram_save(struct pp_hwmgr *hwmgr)
    394{
    395	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    396
    397	smu8_smu->toc_entry_aram = smu8_smu->toc_entry_used_count;
    398	smu8_smu_populate_single_scratch_task(hwmgr,
    399				SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM,
    400				TASK_TYPE_UCODE_SAVE, true);
    401
    402	return 0;
    403}
    404
    405static int smu8_smu_initialize_toc_empty_job_list(struct pp_hwmgr *hwmgr)
    406{
    407	int i;
    408	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    409	struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
    410
    411	for (i = 0; i < NUM_JOBLIST_ENTRIES; i++)
    412		toc->JobList[i] = (uint8_t)IGNORE_JOB;
    413
    414	return 0;
    415}
    416
    417static int smu8_smu_construct_toc_for_vddgfx_enter(struct pp_hwmgr *hwmgr)
    418{
    419	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    420	struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
    421
    422	toc->JobList[JOB_GFX_SAVE] = (uint8_t)smu8_smu->toc_entry_used_count;
    423	smu8_smu_populate_single_scratch_task(hwmgr,
    424				    SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH,
    425				    TASK_TYPE_UCODE_SAVE, false);
    426
    427	smu8_smu_populate_single_scratch_task(hwmgr,
    428				    SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM,
    429				    TASK_TYPE_UCODE_SAVE, true);
    430
    431	return 0;
    432}
    433
    434
    435static int smu8_smu_construct_toc_for_vddgfx_exit(struct pp_hwmgr *hwmgr)
    436{
    437	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    438	struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
    439
    440	toc->JobList[JOB_GFX_RESTORE] = (uint8_t)smu8_smu->toc_entry_used_count;
    441
    442	smu8_smu_populate_single_ucode_load_task(hwmgr,
    443				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE, false);
    444	smu8_smu_populate_single_ucode_load_task(hwmgr,
    445				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP, false);
    446	smu8_smu_populate_single_ucode_load_task(hwmgr,
    447				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME, false);
    448	smu8_smu_populate_single_ucode_load_task(hwmgr,
    449				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1, false);
    450
    451	if (hwmgr->chip_id == CHIP_STONEY)
    452		smu8_smu_populate_single_ucode_load_task(hwmgr,
    453				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1, false);
    454	else
    455		smu8_smu_populate_single_ucode_load_task(hwmgr,
    456				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2, false);
    457
    458	smu8_smu_populate_single_ucode_load_task(hwmgr,
    459				SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G, false);
    460
    461	/* populate scratch */
    462	smu8_smu_populate_single_scratch_task(hwmgr,
    463				SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH,
    464				TASK_TYPE_UCODE_LOAD, false);
    465
    466	smu8_smu_populate_single_scratch_task(hwmgr,
    467				SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM,
    468				TASK_TYPE_UCODE_LOAD, false);
    469
    470	smu8_smu_populate_single_scratch_task(hwmgr,
    471				SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM,
    472				TASK_TYPE_UCODE_LOAD, true);
    473
    474	return 0;
    475}
    476
    477static int smu8_smu_construct_toc_for_power_profiling(struct pp_hwmgr *hwmgr)
    478{
    479	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    480
    481	smu8_smu->toc_entry_power_profiling_index = smu8_smu->toc_entry_used_count;
    482
    483	smu8_smu_populate_single_scratch_task(hwmgr,
    484				SMU8_SCRATCH_ENTRY_UCODE_ID_POWER_PROFILING,
    485				TASK_TYPE_INITIALIZE, true);
    486	return 0;
    487}
    488
    489static int smu8_smu_construct_toc_for_bootup(struct pp_hwmgr *hwmgr)
    490{
    491	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    492
    493	smu8_smu->toc_entry_initialize_index = smu8_smu->toc_entry_used_count;
    494
    495	smu8_smu_populate_single_ucode_load_task(hwmgr,
    496				SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA0, false);
    497	if (hwmgr->chip_id != CHIP_STONEY)
    498		smu8_smu_populate_single_ucode_load_task(hwmgr,
    499				SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA1, false);
    500	smu8_smu_populate_single_ucode_load_task(hwmgr,
    501				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE, false);
    502	smu8_smu_populate_single_ucode_load_task(hwmgr,
    503				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP, false);
    504	smu8_smu_populate_single_ucode_load_task(hwmgr,
    505				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME, false);
    506	smu8_smu_populate_single_ucode_load_task(hwmgr,
    507				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1, false);
    508	if (hwmgr->chip_id != CHIP_STONEY)
    509		smu8_smu_populate_single_ucode_load_task(hwmgr,
    510				SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2, false);
    511	smu8_smu_populate_single_ucode_load_task(hwmgr,
    512				SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G, true);
    513
    514	return 0;
    515}
    516
    517static int smu8_smu_construct_toc_for_clock_table(struct pp_hwmgr *hwmgr)
    518{
    519	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    520
    521	smu8_smu->toc_entry_clock_table = smu8_smu->toc_entry_used_count;
    522
    523	smu8_smu_populate_single_scratch_task(hwmgr,
    524				SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE,
    525				TASK_TYPE_INITIALIZE, true);
    526
    527	return 0;
    528}
    529
    530static int smu8_smu_construct_toc(struct pp_hwmgr *hwmgr)
    531{
    532	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    533
    534	smu8_smu->toc_entry_used_count = 0;
    535	smu8_smu_initialize_toc_empty_job_list(hwmgr);
    536	smu8_smu_construct_toc_for_rlc_aram_save(hwmgr);
    537	smu8_smu_construct_toc_for_vddgfx_enter(hwmgr);
    538	smu8_smu_construct_toc_for_vddgfx_exit(hwmgr);
    539	smu8_smu_construct_toc_for_power_profiling(hwmgr);
    540	smu8_smu_construct_toc_for_bootup(hwmgr);
    541	smu8_smu_construct_toc_for_clock_table(hwmgr);
    542
    543	return 0;
    544}
    545
    546static int smu8_smu_populate_firmware_entries(struct pp_hwmgr *hwmgr)
    547{
    548	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    549	uint32_t firmware_type;
    550	uint32_t i;
    551	int ret;
    552	enum cgs_ucode_id ucode_id;
    553	struct cgs_firmware_info info = {0};
    554
    555	smu8_smu->driver_buffer_length = 0;
    556
    557	for (i = 0; i < ARRAY_SIZE(firmware_list); i++) {
    558
    559		firmware_type = smu8_translate_firmware_enum_to_arg(hwmgr,
    560					firmware_list[i]);
    561
    562		ucode_id = smu8_convert_fw_type_to_cgs(firmware_type);
    563
    564		ret = cgs_get_firmware_info(hwmgr->device,
    565							ucode_id, &info);
    566
    567		if (ret == 0) {
    568			smu8_smu->driver_buffer[i].mc_addr = info.mc_addr;
    569
    570			smu8_smu->driver_buffer[i].data_size = info.image_size;
    571
    572			smu8_smu->driver_buffer[i].firmware_ID = firmware_list[i];
    573			smu8_smu->driver_buffer_length++;
    574		}
    575	}
    576
    577	return 0;
    578}
    579
    580static int smu8_smu_populate_single_scratch_entry(
    581				struct pp_hwmgr *hwmgr,
    582				enum smu8_scratch_entry scratch_type,
    583				uint32_t ulsize_byte,
    584				struct smu8_buffer_entry *entry)
    585{
    586	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    587	uint32_t ulsize_aligned = SIZE_ALIGN_32(ulsize_byte);
    588
    589	entry->data_size = ulsize_byte;
    590	entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr +
    591				smu8_smu->smu_buffer_used_bytes;
    592	entry->mc_addr = smu8_smu->smu_buffer.mc_addr + smu8_smu->smu_buffer_used_bytes;
    593	entry->firmware_ID = scratch_type;
    594
    595	smu8_smu->smu_buffer_used_bytes += ulsize_aligned;
    596
    597	return 0;
    598}
    599
    600static int smu8_download_pptable_settings(struct pp_hwmgr *hwmgr, void **table)
    601{
    602	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    603	unsigned long i;
    604
    605	for (i = 0; i < smu8_smu->scratch_buffer_length; i++) {
    606		if (smu8_smu->scratch_buffer[i].firmware_ID
    607			== SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE)
    608			break;
    609	}
    610
    611	*table = (struct SMU8_Fusion_ClkTable *)smu8_smu->scratch_buffer[i].kaddr;
    612
    613	smum_send_msg_to_smc_with_parameter(hwmgr,
    614				PPSMC_MSG_SetClkTableAddrHi,
    615				upper_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
    616				NULL);
    617
    618	smum_send_msg_to_smc_with_parameter(hwmgr,
    619				PPSMC_MSG_SetClkTableAddrLo,
    620				lower_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
    621				NULL);
    622
    623	smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ExecuteJob,
    624				smu8_smu->toc_entry_clock_table,
    625				NULL);
    626
    627	smum_send_msg_to_smc(hwmgr, PPSMC_MSG_ClkTableXferToDram, NULL);
    628
    629	return 0;
    630}
    631
    632static int smu8_upload_pptable_settings(struct pp_hwmgr *hwmgr)
    633{
    634	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    635	unsigned long i;
    636
    637	for (i = 0; i < smu8_smu->scratch_buffer_length; i++) {
    638		if (smu8_smu->scratch_buffer[i].firmware_ID
    639				== SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE)
    640			break;
    641	}
    642
    643	smum_send_msg_to_smc_with_parameter(hwmgr,
    644				PPSMC_MSG_SetClkTableAddrHi,
    645				upper_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
    646				NULL);
    647
    648	smum_send_msg_to_smc_with_parameter(hwmgr,
    649				PPSMC_MSG_SetClkTableAddrLo,
    650				lower_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
    651				NULL);
    652
    653	smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ExecuteJob,
    654				smu8_smu->toc_entry_clock_table,
    655				NULL);
    656
    657	smum_send_msg_to_smc(hwmgr, PPSMC_MSG_ClkTableXferToSmu, NULL);
    658
    659	return 0;
    660}
    661
    662static int smu8_request_smu_load_fw(struct pp_hwmgr *hwmgr)
    663{
    664	struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
    665	uint32_t smc_address;
    666	uint32_t fw_to_check = 0;
    667	int ret;
    668
    669	amdgpu_ucode_init_bo(hwmgr->adev);
    670
    671	smu8_smu_populate_firmware_entries(hwmgr);
    672
    673	smu8_smu_construct_toc(hwmgr);
    674
    675	smc_address = SMU8_FIRMWARE_HEADER_LOCATION +
    676		offsetof(struct SMU8_Firmware_Header, UcodeLoadStatus);
    677
    678	smu8_write_smc_sram_dword(hwmgr, smc_address, 0, smc_address+4);
    679
    680	smum_send_msg_to_smc_with_parameter(hwmgr,
    681					PPSMC_MSG_DriverDramAddrHi,
    682					upper_32_bits(smu8_smu->toc_buffer.mc_addr),
    683					NULL);
    684
    685	smum_send_msg_to_smc_with_parameter(hwmgr,
    686					PPSMC_MSG_DriverDramAddrLo,
    687					lower_32_bits(smu8_smu->toc_buffer.mc_addr),
    688					NULL);
    689
    690	smum_send_msg_to_smc(hwmgr, PPSMC_MSG_InitJobs, NULL);
    691
    692	smum_send_msg_to_smc_with_parameter(hwmgr,
    693					PPSMC_MSG_ExecuteJob,
    694					smu8_smu->toc_entry_aram,
    695					NULL);
    696	smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ExecuteJob,
    697				smu8_smu->toc_entry_power_profiling_index,
    698				NULL);
    699
    700	smum_send_msg_to_smc_with_parameter(hwmgr,
    701					PPSMC_MSG_ExecuteJob,
    702					smu8_smu->toc_entry_initialize_index,
    703					NULL);
    704
    705	fw_to_check = UCODE_ID_RLC_G_MASK |
    706			UCODE_ID_SDMA0_MASK |
    707			UCODE_ID_SDMA1_MASK |
    708			UCODE_ID_CP_CE_MASK |
    709			UCODE_ID_CP_ME_MASK |
    710			UCODE_ID_CP_PFP_MASK |
    711			UCODE_ID_CP_MEC_JT1_MASK |
    712			UCODE_ID_CP_MEC_JT2_MASK;
    713
    714	if (hwmgr->chip_id == CHIP_STONEY)
    715		fw_to_check &= ~(UCODE_ID_SDMA1_MASK | UCODE_ID_CP_MEC_JT2_MASK);
    716
    717	ret = smu8_check_fw_load_finish(hwmgr, fw_to_check);
    718	if (ret) {
    719		pr_err("SMU firmware load failed\n");
    720		return ret;
    721	}
    722
    723	ret = smu8_load_mec_firmware(hwmgr);
    724	if (ret) {
    725		pr_err("Mec Firmware load failed\n");
    726		return ret;
    727	}
    728
    729	return 0;
    730}
    731
    732static int smu8_start_smu(struct pp_hwmgr *hwmgr)
    733{
    734	struct amdgpu_device *adev;
    735
    736	uint32_t index = SMN_MP1_SRAM_START_ADDR +
    737			 SMU8_FIRMWARE_HEADER_LOCATION +
    738			 offsetof(struct SMU8_Firmware_Header, Version);
    739
    740	if (hwmgr == NULL || hwmgr->device == NULL)
    741		return -EINVAL;
    742
    743	adev = hwmgr->adev;
    744
    745	cgs_write_register(hwmgr->device, mmMP0PUB_IND_INDEX, index);
    746	hwmgr->smu_version = cgs_read_register(hwmgr->device, mmMP0PUB_IND_DATA);
    747	pr_info("smu version %02d.%02d.%02d\n",
    748		((hwmgr->smu_version >> 16) & 0xFF),
    749		((hwmgr->smu_version >> 8) & 0xFF),
    750		(hwmgr->smu_version & 0xFF));
    751	adev->pm.fw_version = hwmgr->smu_version >> 8;
    752
    753	return smu8_request_smu_load_fw(hwmgr);
    754}
    755
    756static int smu8_smu_init(struct pp_hwmgr *hwmgr)
    757{
    758	int ret = 0;
    759	struct smu8_smumgr *smu8_smu;
    760
    761	smu8_smu = kzalloc(sizeof(struct smu8_smumgr), GFP_KERNEL);
    762	if (smu8_smu == NULL)
    763		return -ENOMEM;
    764
    765	hwmgr->smu_backend = smu8_smu;
    766
    767	smu8_smu->toc_buffer.data_size = 4096;
    768	smu8_smu->smu_buffer.data_size =
    769		ALIGN(UCODE_ID_RLC_SCRATCH_SIZE_BYTE, 32) +
    770		ALIGN(UCODE_ID_RLC_SRM_ARAM_SIZE_BYTE, 32) +
    771		ALIGN(UCODE_ID_RLC_SRM_DRAM_SIZE_BYTE, 32) +
    772		ALIGN(sizeof(struct SMU8_MultimediaPowerLogData), 32) +
    773		ALIGN(sizeof(struct SMU8_Fusion_ClkTable), 32);
    774
    775	ret = amdgpu_bo_create_kernel((struct amdgpu_device *)hwmgr->adev,
    776				smu8_smu->toc_buffer.data_size,
    777				PAGE_SIZE,
    778				AMDGPU_GEM_DOMAIN_VRAM,
    779				&smu8_smu->toc_buffer.handle,
    780				&smu8_smu->toc_buffer.mc_addr,
    781				&smu8_smu->toc_buffer.kaddr);
    782	if (ret)
    783		goto err2;
    784
    785	ret = amdgpu_bo_create_kernel((struct amdgpu_device *)hwmgr->adev,
    786				smu8_smu->smu_buffer.data_size,
    787				PAGE_SIZE,
    788				AMDGPU_GEM_DOMAIN_VRAM,
    789				&smu8_smu->smu_buffer.handle,
    790				&smu8_smu->smu_buffer.mc_addr,
    791				&smu8_smu->smu_buffer.kaddr);
    792	if (ret)
    793		goto err1;
    794
    795	if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
    796		SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH,
    797		UCODE_ID_RLC_SCRATCH_SIZE_BYTE,
    798		&smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
    799		pr_err("Error when Populate Firmware Entry.\n");
    800		goto err0;
    801	}
    802
    803	if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
    804		SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM,
    805		UCODE_ID_RLC_SRM_ARAM_SIZE_BYTE,
    806		&smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
    807		pr_err("Error when Populate Firmware Entry.\n");
    808		goto err0;
    809	}
    810	if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
    811		SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM,
    812		UCODE_ID_RLC_SRM_DRAM_SIZE_BYTE,
    813		&smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
    814		pr_err("Error when Populate Firmware Entry.\n");
    815		goto err0;
    816	}
    817
    818	if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
    819		SMU8_SCRATCH_ENTRY_UCODE_ID_POWER_PROFILING,
    820		sizeof(struct SMU8_MultimediaPowerLogData),
    821		&smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
    822		pr_err("Error when Populate Firmware Entry.\n");
    823		goto err0;
    824	}
    825
    826	if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
    827		SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE,
    828		sizeof(struct SMU8_Fusion_ClkTable),
    829		&smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
    830		pr_err("Error when Populate Firmware Entry.\n");
    831		goto err0;
    832	}
    833
    834	return 0;
    835
    836err0:
    837	amdgpu_bo_free_kernel(&smu8_smu->smu_buffer.handle,
    838				&smu8_smu->smu_buffer.mc_addr,
    839				&smu8_smu->smu_buffer.kaddr);
    840err1:
    841	amdgpu_bo_free_kernel(&smu8_smu->toc_buffer.handle,
    842				&smu8_smu->toc_buffer.mc_addr,
    843				&smu8_smu->toc_buffer.kaddr);
    844err2:
    845	kfree(smu8_smu);
    846	return -EINVAL;
    847}
    848
    849static int smu8_smu_fini(struct pp_hwmgr *hwmgr)
    850{
    851	struct smu8_smumgr *smu8_smu;
    852
    853	if (hwmgr == NULL || hwmgr->device == NULL)
    854		return -EINVAL;
    855
    856	smu8_smu = hwmgr->smu_backend;
    857	if (smu8_smu) {
    858		amdgpu_bo_free_kernel(&smu8_smu->toc_buffer.handle,
    859					&smu8_smu->toc_buffer.mc_addr,
    860					&smu8_smu->toc_buffer.kaddr);
    861		amdgpu_bo_free_kernel(&smu8_smu->smu_buffer.handle,
    862					&smu8_smu->smu_buffer.mc_addr,
    863					&smu8_smu->smu_buffer.kaddr);
    864		kfree(smu8_smu);
    865	}
    866
    867	return 0;
    868}
    869
    870static bool smu8_dpm_check_smu_features(struct pp_hwmgr *hwmgr,
    871				unsigned long check_feature)
    872{
    873	int result;
    874	uint32_t features;
    875
    876	result = smum_send_msg_to_smc_with_parameter(hwmgr,
    877				PPSMC_MSG_GetFeatureStatus,
    878				0,
    879				&features);
    880	if (result == 0) {
    881		if (features & check_feature)
    882			return true;
    883	}
    884
    885	return false;
    886}
    887
    888static bool smu8_is_dpm_running(struct pp_hwmgr *hwmgr)
    889{
    890	if (smu8_dpm_check_smu_features(hwmgr, SMU_EnabledFeatureScoreboard_SclkDpmOn))
    891		return true;
    892	return false;
    893}
    894
    895const struct pp_smumgr_func smu8_smu_funcs = {
    896	.name = "smu8_smu",
    897	.smu_init = smu8_smu_init,
    898	.smu_fini = smu8_smu_fini,
    899	.start_smu = smu8_start_smu,
    900	.check_fw_load_finish = smu8_check_fw_load_finish,
    901	.request_smu_load_fw = NULL,
    902	.request_smu_load_specific_fw = NULL,
    903	.get_argument = smu8_get_argument,
    904	.send_msg_to_smc = smu8_send_msg_to_smc,
    905	.send_msg_to_smc_with_parameter = smu8_send_msg_to_smc_with_parameter,
    906	.download_pptable_settings = smu8_download_pptable_settings,
    907	.upload_pptable_settings = smu8_upload_pptable_settings,
    908	.is_dpm_running = smu8_is_dpm_running,
    909};
    910