2017-03-06 14:49:53 -05:00
/*
* Copyright 2016 Advanced Micro Devices , Inc .
*
* Permission is hereby granted , free of charge , to any person obtaining a
* copy of this software and associated documentation files ( the " Software " ) ,
* to deal in the Software without restriction , including without limitation
* the rights to use , copy , modify , merge , publish , distribute , sublicense ,
* and / or sell copies of the Software , and to permit persons to whom the
* Software is furnished to do so , subject to the following conditions :
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software .
*
* THE SOFTWARE IS PROVIDED " AS IS " , WITHOUT WARRANTY OF ANY KIND , EXPRESS OR
* IMPLIED , INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY ,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT . IN NO EVENT SHALL
* THE COPYRIGHT HOLDER ( S ) OR AUTHOR ( S ) BE LIABLE FOR ANY CLAIM , DAMAGES OR
* OTHER LIABILITY , WHETHER IN AN ACTION OF CONTRACT , TORT OR OTHERWISE ,
* ARISING FROM , OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
* OTHER DEALINGS IN THE SOFTWARE .
*
*/
# include <linux/firmware.h>
# include <linux/slab.h>
# include <linux/module.h>
2017-04-24 13:50:21 +09:00
# include <drm/drmP.h>
2017-03-06 14:49:53 -05:00
# include "amdgpu.h"
2017-06-30 17:08:45 -04:00
# include "amdgpu_atombios.h"
2017-03-06 14:49:53 -05:00
# include "amdgpu_ih.h"
# include "amdgpu_uvd.h"
# include "amdgpu_vce.h"
# include "amdgpu_ucode.h"
# include "amdgpu_psp.h"
# include "atom.h"
# include "amd_pcie.h"
2017-11-23 11:09:07 +08:00
# include "uvd/uvd_7_0_offset.h"
2017-11-24 10:29:00 +08:00
# include "gc/gc_9_0_offset.h"
# include "gc/gc_9_0_sh_mask.h"
2017-11-15 16:01:30 +08:00
# include "sdma0/sdma0_4_0_offset.h"
# include "sdma1/sdma1_4_0_offset.h"
2017-11-15 18:09:33 +08:00
# include "hdp/hdp_4_0_offset.h"
# include "hdp/hdp_4_0_sh_mask.h"
2017-11-23 15:09:51 +08:00
# include "smuio/smuio_9_0_offset.h"
# include "smuio/smuio_9_0_sh_mask.h"
2017-03-06 14:49:53 -05:00
# include "soc15.h"
# include "soc15_common.h"
# include "gfx_v9_0.h"
# include "gmc_v9_0.h"
# include "gfxhub_v1_0.h"
# include "mmhub_v1_0.h"
2018-03-28 17:08:04 +08:00
# include "df_v1_7.h"
2018-04-04 14:32:10 +08:00
# include "df_v3_6.h"
2017-03-06 14:49:53 -05:00
# include "vega10_ih.h"
# include "sdma_v4_0.h"
# include "uvd_v7_0.h"
# include "vce_v4_0.h"
2016-12-28 13:36:00 -05:00
# include "vcn_v1_0.h"
2017-02-28 17:22:03 +08:00
# include "dce_virtual.h"
2017-03-08 15:06:47 +08:00
# include "mxgpu_ai.h"
2017-03-06 14:49:53 -05:00
# define mmMP0_MISC_CGTT_CTRL0 0x01b9
# define mmMP0_MISC_CGTT_CTRL0_BASE_IDX 0
# define mmMP0_MISC_LIGHT_SLEEP_CTRL 0x01ba
# define mmMP0_MISC_LIGHT_SLEEP_CTRL_BASE_IDX 0
/*
* Indirect registers accessor
*/
static u32 soc15_pcie_rreg ( struct amdgpu_device * adev , u32 reg )
{
unsigned long flags , address , data ;
u32 r ;
2017-11-28 17:01:21 -05:00
address = adev - > nbio_funcs - > get_pcie_index_offset ( adev ) ;
data = adev - > nbio_funcs - > get_pcie_data_offset ( adev ) ;
2017-03-06 14:49:53 -05:00
spin_lock_irqsave ( & adev - > pcie_idx_lock , flags ) ;
WREG32 ( address , reg ) ;
( void ) RREG32 ( address ) ;
r = RREG32 ( data ) ;
spin_unlock_irqrestore ( & adev - > pcie_idx_lock , flags ) ;
return r ;
}
static void soc15_pcie_wreg ( struct amdgpu_device * adev , u32 reg , u32 v )
{
unsigned long flags , address , data ;
2017-11-28 17:01:21 -05:00
address = adev - > nbio_funcs - > get_pcie_index_offset ( adev ) ;
data = adev - > nbio_funcs - > get_pcie_data_offset ( adev ) ;
2017-03-06 14:49:53 -05:00
spin_lock_irqsave ( & adev - > pcie_idx_lock , flags ) ;
WREG32 ( address , reg ) ;
( void ) RREG32 ( address ) ;
WREG32 ( data , v ) ;
( void ) RREG32 ( data ) ;
spin_unlock_irqrestore ( & adev - > pcie_idx_lock , flags ) ;
}
static u32 soc15_uvd_ctx_rreg ( struct amdgpu_device * adev , u32 reg )
{
unsigned long flags , address , data ;
u32 r ;
address = SOC15_REG_OFFSET ( UVD , 0 , mmUVD_CTX_INDEX ) ;
data = SOC15_REG_OFFSET ( UVD , 0 , mmUVD_CTX_DATA ) ;
spin_lock_irqsave ( & adev - > uvd_ctx_idx_lock , flags ) ;
WREG32 ( address , ( ( reg ) & 0x1ff ) ) ;
r = RREG32 ( data ) ;
spin_unlock_irqrestore ( & adev - > uvd_ctx_idx_lock , flags ) ;
return r ;
}
static void soc15_uvd_ctx_wreg ( struct amdgpu_device * adev , u32 reg , u32 v )
{
unsigned long flags , address , data ;
address = SOC15_REG_OFFSET ( UVD , 0 , mmUVD_CTX_INDEX ) ;
data = SOC15_REG_OFFSET ( UVD , 0 , mmUVD_CTX_DATA ) ;
spin_lock_irqsave ( & adev - > uvd_ctx_idx_lock , flags ) ;
WREG32 ( address , ( ( reg ) & 0x1ff ) ) ;
WREG32 ( data , ( v ) ) ;
spin_unlock_irqrestore ( & adev - > uvd_ctx_idx_lock , flags ) ;
}
static u32 soc15_didt_rreg ( struct amdgpu_device * adev , u32 reg )
{
unsigned long flags , address , data ;
u32 r ;
address = SOC15_REG_OFFSET ( GC , 0 , mmDIDT_IND_INDEX ) ;
data = SOC15_REG_OFFSET ( GC , 0 , mmDIDT_IND_DATA ) ;
spin_lock_irqsave ( & adev - > didt_idx_lock , flags ) ;
WREG32 ( address , ( reg ) ) ;
r = RREG32 ( data ) ;
spin_unlock_irqrestore ( & adev - > didt_idx_lock , flags ) ;
return r ;
}
static void soc15_didt_wreg ( struct amdgpu_device * adev , u32 reg , u32 v )
{
unsigned long flags , address , data ;
address = SOC15_REG_OFFSET ( GC , 0 , mmDIDT_IND_INDEX ) ;
data = SOC15_REG_OFFSET ( GC , 0 , mmDIDT_IND_DATA ) ;
spin_lock_irqsave ( & adev - > didt_idx_lock , flags ) ;
WREG32 ( address , ( reg ) ) ;
WREG32 ( data , ( v ) ) ;
spin_unlock_irqrestore ( & adev - > didt_idx_lock , flags ) ;
}
2017-07-03 22:37:44 +08:00
static u32 soc15_gc_cac_rreg ( struct amdgpu_device * adev , u32 reg )
{
unsigned long flags ;
u32 r ;
spin_lock_irqsave ( & adev - > gc_cac_idx_lock , flags ) ;
WREG32_SOC15 ( GC , 0 , mmGC_CAC_IND_INDEX , ( reg ) ) ;
r = RREG32_SOC15 ( GC , 0 , mmGC_CAC_IND_DATA ) ;
spin_unlock_irqrestore ( & adev - > gc_cac_idx_lock , flags ) ;
return r ;
}
static void soc15_gc_cac_wreg ( struct amdgpu_device * adev , u32 reg , u32 v )
{
unsigned long flags ;
spin_lock_irqsave ( & adev - > gc_cac_idx_lock , flags ) ;
WREG32_SOC15 ( GC , 0 , mmGC_CAC_IND_INDEX , ( reg ) ) ;
WREG32_SOC15 ( GC , 0 , mmGC_CAC_IND_DATA , ( v ) ) ;
spin_unlock_irqrestore ( & adev - > gc_cac_idx_lock , flags ) ;
}
2017-07-04 09:23:01 +08:00
static u32 soc15_se_cac_rreg ( struct amdgpu_device * adev , u32 reg )
{
unsigned long flags ;
u32 r ;
spin_lock_irqsave ( & adev - > se_cac_idx_lock , flags ) ;
WREG32_SOC15 ( GC , 0 , mmSE_CAC_IND_INDEX , ( reg ) ) ;
r = RREG32_SOC15 ( GC , 0 , mmSE_CAC_IND_DATA ) ;
spin_unlock_irqrestore ( & adev - > se_cac_idx_lock , flags ) ;
return r ;
}
static void soc15_se_cac_wreg ( struct amdgpu_device * adev , u32 reg , u32 v )
{
unsigned long flags ;
spin_lock_irqsave ( & adev - > se_cac_idx_lock , flags ) ;
WREG32_SOC15 ( GC , 0 , mmSE_CAC_IND_INDEX , ( reg ) ) ;
WREG32_SOC15 ( GC , 0 , mmSE_CAC_IND_DATA , ( v ) ) ;
spin_unlock_irqrestore ( & adev - > se_cac_idx_lock , flags ) ;
}
2017-03-06 14:49:53 -05:00
static u32 soc15_get_config_memsize ( struct amdgpu_device * adev )
{
2017-12-08 13:07:58 -05:00
return adev - > nbio_funcs - > get_memsize ( adev ) ;
2017-03-06 14:49:53 -05:00
}
static u32 soc15_get_xclk ( struct amdgpu_device * adev )
{
2017-09-29 15:41:43 +08:00
return adev - > clock . spll . reference_freq ;
2017-03-06 14:49:53 -05:00
}
void soc15_grbm_select ( struct amdgpu_device * adev ,
u32 me , u32 pipe , u32 queue , u32 vmid )
{
u32 grbm_gfx_cntl = 0 ;
grbm_gfx_cntl = REG_SET_FIELD ( grbm_gfx_cntl , GRBM_GFX_CNTL , PIPEID , pipe ) ;
grbm_gfx_cntl = REG_SET_FIELD ( grbm_gfx_cntl , GRBM_GFX_CNTL , MEID , me ) ;
grbm_gfx_cntl = REG_SET_FIELD ( grbm_gfx_cntl , GRBM_GFX_CNTL , VMID , vmid ) ;
grbm_gfx_cntl = REG_SET_FIELD ( grbm_gfx_cntl , GRBM_GFX_CNTL , QUEUEID , queue ) ;
WREG32 ( SOC15_REG_OFFSET ( GC , 0 , mmGRBM_GFX_CNTL ) , grbm_gfx_cntl ) ;
}
static void soc15_vga_set_state ( struct amdgpu_device * adev , bool state )
{
/* todo */
}
static bool soc15_read_disabled_bios ( struct amdgpu_device * adev )
{
/* todo */
return false ;
}
static bool soc15_read_bios_from_rom ( struct amdgpu_device * adev ,
u8 * bios , u32 length_bytes )
{
u32 * dw_ptr ;
u32 i , length_dw ;
if ( bios = = NULL )
return false ;
if ( length_bytes = = 0 )
return false ;
/* APU vbios image is part of sbios image */
if ( adev - > flags & AMD_IS_APU )
return false ;
dw_ptr = ( u32 * ) bios ;
length_dw = ALIGN ( length_bytes , 4 ) / 4 ;
/* set rom index to 0 */
WREG32 ( SOC15_REG_OFFSET ( SMUIO , 0 , mmROM_INDEX ) , 0 ) ;
/* read out the rom data */
for ( i = 0 ; i < length_dw ; i + + )
dw_ptr [ i ] = RREG32 ( SOC15_REG_OFFSET ( SMUIO , 0 , mmROM_DATA ) ) ;
return true ;
}
2017-11-28 17:01:21 -05:00
struct soc15_allowed_register_entry {
uint32_t hwip ;
uint32_t inst ;
uint32_t seg ;
uint32_t reg_offset ;
bool grbm_indexed ;
} ;
static struct soc15_allowed_register_entry soc15_allowed_read_registers [ ] = {
{ SOC15_REG_ENTRY ( GC , 0 , mmGRBM_STATUS ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmGRBM_STATUS2 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmGRBM_STATUS_SE0 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmGRBM_STATUS_SE1 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmGRBM_STATUS_SE2 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmGRBM_STATUS_SE3 ) } ,
{ SOC15_REG_ENTRY ( SDMA0 , 0 , mmSDMA0_STATUS_REG ) } ,
{ SOC15_REG_ENTRY ( SDMA1 , 0 , mmSDMA1_STATUS_REG ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_STAT ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_STALLED_STAT1 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_STALLED_STAT2 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_STALLED_STAT3 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_CPF_BUSY_STAT ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_CPF_STALLED_STAT1 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_CPF_STATUS ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_CPC_STALLED_STAT1 ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmCP_CPC_STATUS ) } ,
{ SOC15_REG_ENTRY ( GC , 0 , mmGB_ADDR_CONFIG ) } ,
2018-04-10 10:15:26 -05:00
{ SOC15_REG_ENTRY ( GC , 0 , mmDB_DEBUG2 ) } ,
2017-03-06 14:49:53 -05:00
} ;
static uint32_t soc15_read_indexed_register ( struct amdgpu_device * adev , u32 se_num ,
u32 sh_num , u32 reg_offset )
{
uint32_t val ;
mutex_lock ( & adev - > grbm_idx_mutex ) ;
if ( se_num ! = 0xffffffff | | sh_num ! = 0xffffffff )
amdgpu_gfx_select_se_sh ( adev , se_num , sh_num , 0xffffffff ) ;
val = RREG32 ( reg_offset ) ;
if ( se_num ! = 0xffffffff | | sh_num ! = 0xffffffff )
amdgpu_gfx_select_se_sh ( adev , 0xffffffff , 0xffffffff , 0xffffffff ) ;
mutex_unlock ( & adev - > grbm_idx_mutex ) ;
return val ;
}
2017-03-24 15:05:07 -04:00
static uint32_t soc15_get_register_value ( struct amdgpu_device * adev ,
bool indexed , u32 se_num ,
u32 sh_num , u32 reg_offset )
{
if ( indexed ) {
return soc15_read_indexed_register ( adev , se_num , sh_num , reg_offset ) ;
} else {
2017-11-29 13:51:32 -05:00
if ( reg_offset = = SOC15_REG_OFFSET ( GC , 0 , mmGB_ADDR_CONFIG ) )
2017-03-24 15:05:07 -04:00
return adev - > gfx . config . gb_addr_config ;
2018-04-10 10:15:26 -05:00
else if ( reg_offset = = SOC15_REG_OFFSET ( GC , 0 , mmDB_DEBUG2 ) )
return adev - > gfx . config . db_debug2 ;
2017-11-29 13:51:32 -05:00
return RREG32 ( reg_offset ) ;
2017-03-24 15:05:07 -04:00
}
}
2017-03-06 14:49:53 -05:00
static int soc15_read_register ( struct amdgpu_device * adev , u32 se_num ,
u32 sh_num , u32 reg_offset , u32 * value )
{
2017-04-12 12:53:18 +02:00
uint32_t i ;
2017-11-28 17:01:21 -05:00
struct soc15_allowed_register_entry * en ;
2017-03-06 14:49:53 -05:00
* value = 0 ;
for ( i = 0 ; i < ARRAY_SIZE ( soc15_allowed_read_registers ) ; i + + ) {
2017-11-28 17:01:21 -05:00
en = & soc15_allowed_read_registers [ i ] ;
if ( reg_offset ! = ( adev - > reg_offset [ en - > hwip ] [ en - > inst ] [ en - > seg ]
+ en - > reg_offset ) )
2017-03-06 14:49:53 -05:00
continue ;
2017-04-12 12:49:54 +02:00
* value = soc15_get_register_value ( adev ,
soc15_allowed_read_registers [ i ] . grbm_indexed ,
se_num , sh_num , reg_offset ) ;
2017-03-06 14:49:53 -05:00
return 0 ;
}
return - EINVAL ;
}
2017-11-28 17:01:21 -05:00
/**
* soc15_program_register_sequence - program an array of registers .
*
* @ adev : amdgpu_device pointer
* @ regs : pointer to the register array
* @ array_size : size of the register array
*
* Programs an array or registers with and and or masks .
* This is a helper for setting golden registers .
*/
void soc15_program_register_sequence ( struct amdgpu_device * adev ,
const struct soc15_reg_golden * regs ,
const u32 array_size )
{
const struct soc15_reg_golden * entry ;
u32 tmp , reg ;
int i ;
for ( i = 0 ; i < array_size ; + + i ) {
entry = & regs [ i ] ;
reg = adev - > reg_offset [ entry - > hwip ] [ entry - > instance ] [ entry - > segment ] + entry - > reg ;
if ( entry - > and_mask = = 0xffffffff ) {
tmp = entry - > or_mask ;
} else {
tmp = RREG32 ( reg ) ;
tmp & = ~ ( entry - > and_mask ) ;
tmp | = entry - > or_mask ;
}
WREG32 ( reg , tmp ) ;
}
}
2017-09-14 16:25:19 +08:00
static int soc15_asic_reset ( struct amdgpu_device * adev )
2017-03-06 14:49:53 -05:00
{
u32 i ;
2017-09-14 16:25:19 +08:00
amdgpu_atombios_scratch_regs_engine_hung ( adev , true ) ;
dev_info ( adev - > dev , " GPU reset \n " ) ;
2017-03-06 14:49:53 -05:00
/* disable BM */
pci_clear_master ( adev - > pdev ) ;
2017-09-14 16:25:19 +08:00
pci_save_state ( adev - > pdev ) ;
2018-01-23 16:27:31 -05:00
psp_gpu_reset ( adev ) ;
2017-09-14 16:25:19 +08:00
pci_restore_state ( adev - > pdev ) ;
2017-03-06 14:49:53 -05:00
/* wait for asic to come out of reset */
for ( i = 0 ; i < adev - > usec_timeout ; i + + ) {
2017-12-08 13:07:58 -05:00
u32 memsize = adev - > nbio_funcs - > get_memsize ( adev ) ;
2017-05-04 15:06:25 -04:00
if ( memsize ! = 0xffffffff )
2017-03-06 14:49:53 -05:00
break ;
udelay ( 1 ) ;
}
2017-06-30 17:08:45 -04:00
amdgpu_atombios_scratch_regs_engine_hung ( adev , false ) ;
2017-03-06 14:49:53 -05:00
return 0 ;
}
/*static int soc15_set_uvd_clock(struct amdgpu_device *adev, u32 clock,
u32 cntl_reg , u32 status_reg )
{
return 0 ;
} */
static int soc15_set_uvd_clocks ( struct amdgpu_device * adev , u32 vclk , u32 dclk )
{
/*int r;
r = soc15_set_uvd_clock ( adev , vclk , ixCG_VCLK_CNTL , ixCG_VCLK_STATUS ) ;
if ( r )
return r ;
r = soc15_set_uvd_clock ( adev , dclk , ixCG_DCLK_CNTL , ixCG_DCLK_STATUS ) ;
*/
return 0 ;
}
static int soc15_set_vce_clocks ( struct amdgpu_device * adev , u32 evclk , u32 ecclk )
{
/* todo */
return 0 ;
}
static void soc15_pcie_gen3_enable ( struct amdgpu_device * adev )
{
if ( pci_is_root_bus ( adev - > pdev - > bus ) )
return ;
if ( amdgpu_pcie_gen2 = = 0 )
return ;
if ( adev - > flags & AMD_IS_APU )
return ;
if ( ! ( adev - > pm . pcie_gen_mask & ( CAIL_PCIE_LINK_SPEED_SUPPORT_GEN2 |
CAIL_PCIE_LINK_SPEED_SUPPORT_GEN3 ) ) )
return ;
/* todo */
}
static void soc15_program_aspm ( struct amdgpu_device * adev )
{
if ( amdgpu_aspm = = 0 )
return ;
/* todo */
}
static void soc15_enable_doorbell_aperture ( struct amdgpu_device * adev ,
2017-12-08 13:07:58 -05:00
bool enable )
2017-03-06 14:49:53 -05:00
{
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > enable_doorbell_aperture ( adev , enable ) ;
adev - > nbio_funcs - > enable_doorbell_selfring_aperture ( adev , enable ) ;
2017-03-06 14:49:53 -05:00
}
static const struct amdgpu_ip_block_version vega10_common_ip_block =
{
. type = AMD_IP_BLOCK_TYPE_COMMON ,
. major = 2 ,
. minor = 0 ,
. rev = 0 ,
. funcs = & soc15_common_ip_funcs ,
} ;
2018-06-15 16:05:48 -05:00
static uint32_t soc15_get_rev_id ( struct amdgpu_device * adev )
{
return adev - > nbio_funcs - > get_rev_id ( adev ) ;
}
2017-03-06 14:49:53 -05:00
int soc15_set_ip_blocks ( struct amdgpu_device * adev )
{
2017-11-27 13:16:35 -05:00
/* Set IP register base before any HW register access */
switch ( adev - > asic_type ) {
case CHIP_VEGA10 :
2018-03-12 18:25:15 +08:00
case CHIP_VEGA12 :
2017-11-27 13:16:35 -05:00
case CHIP_RAVEN :
vega10_reg_base_init ( adev ) ;
break ;
2018-03-23 14:42:28 -05:00
case CHIP_VEGA20 :
vega20_reg_base_init ( adev ) ;
break ;
2017-11-27 13:16:35 -05:00
default :
return - EINVAL ;
}
2017-12-08 13:07:58 -05:00
if ( adev - > flags & AMD_IS_APU )
adev - > nbio_funcs = & nbio_v7_0_funcs ;
2018-03-23 14:44:28 -05:00
else if ( adev - > asic_type = = CHIP_VEGA20 )
adev - > nbio_funcs = & nbio_v7_4_funcs ;
2017-12-08 13:07:58 -05:00
else
adev - > nbio_funcs = & nbio_v6_1_funcs ;
2018-04-04 14:32:10 +08:00
if ( adev - > asic_type = = CHIP_VEGA20 )
adev - > df_funcs = & df_v3_6_funcs ;
else
adev - > df_funcs = & df_v1_7_funcs ;
2018-06-15 16:05:48 -05:00
adev - > rev_id = soc15_get_rev_id ( adev ) ;
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > detect_hw_virt ( adev ) ;
2017-03-08 15:00:48 +08:00
2017-03-08 15:06:47 +08:00
if ( amdgpu_sriov_vf ( adev ) )
adev - > virt . ops = & xgpu_ai_virt_ops ;
2017-03-06 14:49:53 -05:00
switch ( adev - > asic_type ) {
case CHIP_VEGA10 :
2018-03-06 22:35:19 -05:00
case CHIP_VEGA12 :
2018-04-20 18:35:42 +08:00
case CHIP_VEGA20 :
2017-12-15 16:18:00 -05:00
amdgpu_device_ip_block_add ( adev , & vega10_common_ip_block ) ;
amdgpu_device_ip_block_add ( adev , & gmc_v9_0_ip_block ) ;
amdgpu_device_ip_block_add ( adev , & vega10_ih_ip_block ) ;
2018-05-11 14:54:50 +08:00
if ( adev - > asic_type = = CHIP_VEGA20 )
amdgpu_device_ip_block_add ( adev , & psp_v11_0_ip_block ) ;
else
2018-04-24 11:20:16 +08:00
amdgpu_device_ip_block_add ( adev , & psp_v3_1_ip_block ) ;
2018-09-30 17:37:27 +08:00
amdgpu_device_ip_block_add ( adev , & gfx_v9_0_ip_block ) ;
amdgpu_device_ip_block_add ( adev , & sdma_v4_0_ip_block ) ;
2018-05-02 15:45:54 +08:00
if ( ! amdgpu_sriov_vf ( adev ) )
amdgpu_device_ip_block_add ( adev , & pp_smu_ip_block ) ;
2017-03-22 10:49:25 -04:00
if ( adev - > enable_virtual_display | | amdgpu_sriov_vf ( adev ) )
2017-12-15 16:18:00 -05:00
amdgpu_device_ip_block_add ( adev , & dce_virtual_ip_block ) ;
2017-04-19 17:28:47 -04:00
# if defined(CONFIG_DRM_AMD_DC)
else if ( amdgpu_device_has_dc_support ( adev ) )
2017-12-15 16:18:00 -05:00
amdgpu_device_ip_block_add ( adev , & dm_ip_block ) ;
2017-04-19 17:28:47 -04:00
# else
# warning "Enable CONFIG_DRM_AMD_DC for display support on SOC15."
# endif
2018-04-27 03:45:50 +08:00
if ( ! ( adev - > asic_type = = CHIP_VEGA20 & & amdgpu_sriov_vf ( adev ) ) ) {
amdgpu_device_ip_block_add ( adev , & uvd_v7_0_ip_block ) ;
amdgpu_device_ip_block_add ( adev , & vce_v4_0_ip_block ) ;
}
2017-03-06 14:49:53 -05:00
break ;
2016-12-08 10:09:13 +08:00
case CHIP_RAVEN :
2018-07-09 20:00:05 +08:00
amdgpu_device_ip_block_add ( adev , & vega10_common_ip_block ) ;
amdgpu_device_ip_block_add ( adev , & gmc_v9_0_ip_block ) ;
2017-12-15 16:18:00 -05:00
amdgpu_device_ip_block_add ( adev , & vega10_ih_ip_block ) ;
amdgpu_device_ip_block_add ( adev , & psp_v10_0_ip_block ) ;
2018-09-30 17:37:27 +08:00
amdgpu_device_ip_block_add ( adev , & gfx_v9_0_ip_block ) ;
amdgpu_device_ip_block_add ( adev , & sdma_v4_0_ip_block ) ;
2018-03-12 19:52:23 +08:00
amdgpu_device_ip_block_add ( adev , & pp_smu_ip_block ) ;
2017-06-02 14:52:18 -04:00
if ( adev - > enable_virtual_display | | amdgpu_sriov_vf ( adev ) )
2017-12-15 16:18:00 -05:00
amdgpu_device_ip_block_add ( adev , & dce_virtual_ip_block ) ;
2017-06-02 14:54:26 -04:00
# if defined(CONFIG_DRM_AMD_DC)
else if ( amdgpu_device_has_dc_support ( adev ) )
2017-12-15 16:18:00 -05:00
amdgpu_device_ip_block_add ( adev , & dm_ip_block ) ;
2017-06-02 14:54:26 -04:00
# else
# warning "Enable CONFIG_DRM_AMD_DC for display support on SOC15."
# endif
2017-12-15 16:18:00 -05:00
amdgpu_device_ip_block_add ( adev , & vcn_v1_0_ip_block ) ;
2016-12-08 10:09:13 +08:00
break ;
2017-03-06 14:49:53 -05:00
default :
return - EINVAL ;
}
return 0 ;
}
2018-01-19 14:17:40 +01:00
static void soc15_flush_hdp ( struct amdgpu_device * adev , struct amdgpu_ring * ring )
2017-09-06 18:06:45 -04:00
{
2018-01-19 14:17:40 +01:00
adev - > nbio_funcs - > hdp_flush ( adev , ring ) ;
2017-09-06 18:06:45 -04:00
}
2018-01-19 14:17:40 +01:00
static void soc15_invalidate_hdp ( struct amdgpu_device * adev ,
struct amdgpu_ring * ring )
2017-09-06 18:06:45 -04:00
{
2018-01-19 14:17:40 +01:00
if ( ! ring | | ! ring - > funcs - > emit_wreg )
WREG32_SOC15_NO_KIQ ( NBIO , 0 , mmHDP_READ_CACHE_INVALIDATE , 1 ) ;
else
amdgpu_ring_emit_wreg ( ring , SOC15_REG_OFFSET (
HDP , 0 , mmHDP_READ_CACHE_INVALIDATE ) , 1 ) ;
2017-09-06 18:06:45 -04:00
}
2018-03-29 14:39:46 -05:00
static bool soc15_need_full_reset ( struct amdgpu_device * adev )
{
/* change this when we implement soft reset */
return true ;
}
2017-03-06 14:49:53 -05:00
static const struct amdgpu_asic_funcs soc15_asic_funcs =
{
. read_disabled_bios = & soc15_read_disabled_bios ,
. read_bios_from_rom = & soc15_read_bios_from_rom ,
. read_register = & soc15_read_register ,
. reset = & soc15_asic_reset ,
. set_vga_state = & soc15_vga_set_state ,
. get_xclk = & soc15_get_xclk ,
. set_uvd_clocks = & soc15_set_uvd_clocks ,
. set_vce_clocks = & soc15_set_vce_clocks ,
. get_config_memsize = & soc15_get_config_memsize ,
2017-09-06 18:06:45 -04:00
. flush_hdp = & soc15_flush_hdp ,
. invalidate_hdp = & soc15_invalidate_hdp ,
2018-03-29 14:39:46 -05:00
. need_full_reset = & soc15_need_full_reset ,
2017-03-06 14:49:53 -05:00
} ;
static int soc15_common_early_init ( void * handle )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
adev - > smc_rreg = NULL ;
adev - > smc_wreg = NULL ;
adev - > pcie_rreg = & soc15_pcie_rreg ;
adev - > pcie_wreg = & soc15_pcie_wreg ;
adev - > uvd_ctx_rreg = & soc15_uvd_ctx_rreg ;
adev - > uvd_ctx_wreg = & soc15_uvd_ctx_wreg ;
adev - > didt_rreg = & soc15_didt_rreg ;
adev - > didt_wreg = & soc15_didt_wreg ;
2017-07-03 22:37:44 +08:00
adev - > gc_cac_rreg = & soc15_gc_cac_rreg ;
adev - > gc_cac_wreg = & soc15_gc_cac_wreg ;
2017-07-04 09:23:01 +08:00
adev - > se_cac_rreg = & soc15_se_cac_rreg ;
adev - > se_cac_wreg = & soc15_se_cac_wreg ;
2017-03-06 14:49:53 -05:00
adev - > asic_funcs = & soc15_asic_funcs ;
adev - > external_rev_id = 0xFF ;
switch ( adev - > asic_type ) {
case CHIP_VEGA10 :
adev - > cg_flags = AMD_CG_SUPPORT_GFX_MGCG |
AMD_CG_SUPPORT_GFX_MGLS |
AMD_CG_SUPPORT_GFX_RLC_LS |
AMD_CG_SUPPORT_GFX_CP_LS |
AMD_CG_SUPPORT_GFX_3D_CGCG |
AMD_CG_SUPPORT_GFX_3D_CGLS |
AMD_CG_SUPPORT_GFX_CGCG |
AMD_CG_SUPPORT_GFX_CGLS |
AMD_CG_SUPPORT_BIF_MGCG |
AMD_CG_SUPPORT_BIF_LS |
AMD_CG_SUPPORT_HDP_LS |
AMD_CG_SUPPORT_DRM_MGCG |
AMD_CG_SUPPORT_DRM_LS |
AMD_CG_SUPPORT_ROM_MGCG |
AMD_CG_SUPPORT_DF_MGCG |
AMD_CG_SUPPORT_SDMA_MGCG |
AMD_CG_SUPPORT_SDMA_LS |
AMD_CG_SUPPORT_MC_MGCG |
AMD_CG_SUPPORT_MC_LS ;
adev - > pg_flags = 0 ;
adev - > external_rev_id = 0x1 ;
break ;
2018-03-06 22:35:19 -05:00
case CHIP_VEGA12 :
2017-12-25 13:16:11 +08:00
adev - > cg_flags = AMD_CG_SUPPORT_GFX_MGCG |
AMD_CG_SUPPORT_GFX_MGLS |
AMD_CG_SUPPORT_GFX_CGCG |
AMD_CG_SUPPORT_GFX_CGLS |
AMD_CG_SUPPORT_GFX_3D_CGCG |
AMD_CG_SUPPORT_GFX_3D_CGLS |
AMD_CG_SUPPORT_GFX_CP_LS |
AMD_CG_SUPPORT_MC_LS |
AMD_CG_SUPPORT_MC_MGCG |
AMD_CG_SUPPORT_SDMA_MGCG |
AMD_CG_SUPPORT_SDMA_LS |
AMD_CG_SUPPORT_BIF_MGCG |
AMD_CG_SUPPORT_BIF_LS |
AMD_CG_SUPPORT_HDP_MGCG |
AMD_CG_SUPPORT_HDP_LS |
AMD_CG_SUPPORT_ROM_MGCG |
AMD_CG_SUPPORT_VCE_MGCG |
AMD_CG_SUPPORT_UVD_MGCG ;
2018-03-06 22:35:19 -05:00
adev - > pg_flags = 0 ;
2017-12-14 19:02:47 +08:00
adev - > external_rev_id = adev - > rev_id + 0x14 ;
2018-03-06 22:35:19 -05:00
break ;
2018-01-26 15:06:22 +08:00
case CHIP_VEGA20 :
2018-03-26 11:43:04 +08:00
adev - > cg_flags = AMD_CG_SUPPORT_GFX_MGCG |
AMD_CG_SUPPORT_GFX_MGLS |
AMD_CG_SUPPORT_GFX_CGCG |
AMD_CG_SUPPORT_GFX_CGLS |
AMD_CG_SUPPORT_GFX_3D_CGCG |
AMD_CG_SUPPORT_GFX_3D_CGLS |
AMD_CG_SUPPORT_GFX_CP_LS |
AMD_CG_SUPPORT_MC_LS |
AMD_CG_SUPPORT_MC_MGCG |
AMD_CG_SUPPORT_SDMA_MGCG |
AMD_CG_SUPPORT_SDMA_LS |
AMD_CG_SUPPORT_BIF_MGCG |
AMD_CG_SUPPORT_BIF_LS |
AMD_CG_SUPPORT_HDP_MGCG |
2018-05-28 09:22:09 +08:00
AMD_CG_SUPPORT_HDP_LS |
2018-03-26 11:43:04 +08:00
AMD_CG_SUPPORT_ROM_MGCG |
AMD_CG_SUPPORT_VCE_MGCG |
AMD_CG_SUPPORT_UVD_MGCG ;
2018-01-26 15:06:22 +08:00
adev - > pg_flags = 0 ;
adev - > external_rev_id = adev - > rev_id + 0x28 ;
break ;
2016-12-27 21:02:48 +08:00
case CHIP_RAVEN :
2018-06-19 10:46:42 -05:00
if ( adev - > rev_id > = 0x8 )
2018-09-13 15:41:57 -05:00
adev - > external_rev_id = adev - > rev_id + 0x81 ;
else if ( adev - > pdev - > device = = 0x15d8 )
adev - > external_rev_id = adev - > rev_id + 0x41 ;
else
adev - > external_rev_id = 0x1 ;
if ( adev - > rev_id > = 0x8 ) {
2018-06-19 10:46:42 -05:00
adev - > cg_flags = AMD_CG_SUPPORT_GFX_MGCG |
AMD_CG_SUPPORT_GFX_MGLS |
AMD_CG_SUPPORT_GFX_CP_LS |
AMD_CG_SUPPORT_GFX_3D_CGCG |
AMD_CG_SUPPORT_GFX_3D_CGLS |
AMD_CG_SUPPORT_GFX_CGCG |
AMD_CG_SUPPORT_GFX_CGLS |
AMD_CG_SUPPORT_BIF_LS |
AMD_CG_SUPPORT_HDP_LS |
AMD_CG_SUPPORT_ROM_MGCG |
AMD_CG_SUPPORT_MC_MGCG |
AMD_CG_SUPPORT_MC_LS |
AMD_CG_SUPPORT_SDMA_MGCG |
AMD_CG_SUPPORT_SDMA_LS |
AMD_CG_SUPPORT_VCN_MGCG ;
2018-09-13 15:41:57 -05:00
adev - > pg_flags = AMD_PG_SUPPORT_SDMA | AMD_PG_SUPPORT_VCN ;
} else if ( adev - > pdev - > device = = 0x15d8 ) {
adev - > cg_flags = AMD_CG_SUPPORT_GFX_MGLS |
AMD_CG_SUPPORT_GFX_CP_LS |
AMD_CG_SUPPORT_GFX_3D_CGCG |
AMD_CG_SUPPORT_GFX_3D_CGLS |
AMD_CG_SUPPORT_GFX_CGCG |
AMD_CG_SUPPORT_GFX_CGLS |
AMD_CG_SUPPORT_BIF_LS |
AMD_CG_SUPPORT_HDP_LS |
AMD_CG_SUPPORT_ROM_MGCG |
AMD_CG_SUPPORT_MC_MGCG |
AMD_CG_SUPPORT_MC_LS |
AMD_CG_SUPPORT_SDMA_MGCG |
AMD_CG_SUPPORT_SDMA_LS ;
adev - > pg_flags = AMD_PG_SUPPORT_SDMA |
AMD_PG_SUPPORT_MMHUB |
2018-09-21 14:47:45 -04:00
AMD_PG_SUPPORT_VCN |
AMD_PG_SUPPORT_VCN_DPG ;
2018-09-13 15:41:57 -05:00
} else {
2018-06-19 10:46:42 -05:00
adev - > cg_flags = AMD_CG_SUPPORT_GFX_MGCG |
AMD_CG_SUPPORT_GFX_MGLS |
AMD_CG_SUPPORT_GFX_RLC_LS |
AMD_CG_SUPPORT_GFX_CP_LS |
AMD_CG_SUPPORT_GFX_3D_CGCG |
AMD_CG_SUPPORT_GFX_3D_CGLS |
AMD_CG_SUPPORT_GFX_CGCG |
AMD_CG_SUPPORT_GFX_CGLS |
AMD_CG_SUPPORT_BIF_MGCG |
AMD_CG_SUPPORT_BIF_LS |
AMD_CG_SUPPORT_HDP_MGCG |
AMD_CG_SUPPORT_HDP_LS |
AMD_CG_SUPPORT_DRM_MGCG |
AMD_CG_SUPPORT_DRM_LS |
AMD_CG_SUPPORT_ROM_MGCG |
AMD_CG_SUPPORT_MC_MGCG |
AMD_CG_SUPPORT_MC_LS |
AMD_CG_SUPPORT_SDMA_MGCG |
AMD_CG_SUPPORT_SDMA_LS |
AMD_CG_SUPPORT_VCN_MGCG ;
2018-05-17 16:03:47 +08:00
2018-09-13 15:41:57 -05:00
adev - > pg_flags = AMD_PG_SUPPORT_SDMA | AMD_PG_SUPPORT_VCN ;
}
2018-08-10 16:22:26 +08:00
2018-08-07 17:05:22 +08:00
if ( adev - > powerplay . pp_feature & PP_GFXOFF_MASK )
adev - > pg_flags | = AMD_PG_SUPPORT_GFX_PG |
AMD_PG_SUPPORT_CP |
AMD_PG_SUPPORT_RLC_SMU_HS ;
2018-07-10 20:22:36 +08:00
break ;
2017-03-06 14:49:53 -05:00
default :
/* FIXME: not supported yet */
return - EINVAL ;
}
2017-04-21 14:06:09 +08:00
if ( amdgpu_sriov_vf ( adev ) ) {
amdgpu_virt_init_setting ( adev ) ;
xgpu_ai_mailbox_set_irq_funcs ( adev ) ;
}
2017-03-06 14:49:53 -05:00
return 0 ;
}
2017-04-05 13:04:50 +08:00
static int soc15_common_late_init ( void * handle )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
if ( amdgpu_sriov_vf ( adev ) )
xgpu_ai_mailbox_get_irq ( adev ) ;
return 0 ;
}
2017-03-06 14:49:53 -05:00
static int soc15_common_sw_init ( void * handle )
{
2017-04-05 13:04:50 +08:00
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
if ( amdgpu_sriov_vf ( adev ) )
xgpu_ai_mailbox_add_irq_id ( adev ) ;
2017-03-06 14:49:53 -05:00
return 0 ;
}
static int soc15_common_sw_fini ( void * handle )
{
return 0 ;
}
static int soc15_common_hw_init ( void * handle )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
/* enable pcie gen2/3 link */
soc15_pcie_gen3_enable ( adev ) ;
/* enable aspm */
soc15_program_aspm ( adev ) ;
2017-07-06 13:43:55 -04:00
/* setup nbio registers */
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > init_registers ( adev ) ;
2017-03-06 14:49:53 -05:00
/* enable the doorbell aperture */
soc15_enable_doorbell_aperture ( adev , true ) ;
return 0 ;
}
static int soc15_common_hw_fini ( void * handle )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
/* disable the doorbell aperture */
soc15_enable_doorbell_aperture ( adev , false ) ;
2017-04-05 13:04:50 +08:00
if ( amdgpu_sriov_vf ( adev ) )
xgpu_ai_mailbox_put_irq ( adev ) ;
2017-03-06 14:49:53 -05:00
return 0 ;
}
static int soc15_common_suspend ( void * handle )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
return soc15_common_hw_fini ( adev ) ;
}
static int soc15_common_resume ( void * handle )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
return soc15_common_hw_init ( adev ) ;
}
static bool soc15_common_is_idle ( void * handle )
{
return true ;
}
static int soc15_common_wait_for_idle ( void * handle )
{
return 0 ;
}
static int soc15_common_soft_reset ( void * handle )
{
return 0 ;
}
static void soc15_update_hdp_light_sleep ( struct amdgpu_device * adev , bool enable )
{
uint32_t def , data ;
def = data = RREG32 ( SOC15_REG_OFFSET ( HDP , 0 , mmHDP_MEM_POWER_LS ) ) ;
if ( enable & & ( adev - > cg_flags & AMD_CG_SUPPORT_HDP_LS ) )
data | = HDP_MEM_POWER_LS__LS_ENABLE_MASK ;
else
data & = ~ HDP_MEM_POWER_LS__LS_ENABLE_MASK ;
if ( def ! = data )
WREG32 ( SOC15_REG_OFFSET ( HDP , 0 , mmHDP_MEM_POWER_LS ) , data ) ;
}
static void soc15_update_drm_clock_gating ( struct amdgpu_device * adev , bool enable )
{
uint32_t def , data ;
def = data = RREG32 ( SOC15_REG_OFFSET ( MP0 , 0 , mmMP0_MISC_CGTT_CTRL0 ) ) ;
if ( enable & & ( adev - > cg_flags & AMD_CG_SUPPORT_DRM_MGCG ) )
data & = ~ ( 0x01000000 |
0x02000000 |
0x04000000 |
0x08000000 |
0x10000000 |
0x20000000 |
0x40000000 |
0x80000000 ) ;
else
data | = ( 0x01000000 |
0x02000000 |
0x04000000 |
0x08000000 |
0x10000000 |
0x20000000 |
0x40000000 |
0x80000000 ) ;
if ( def ! = data )
WREG32 ( SOC15_REG_OFFSET ( MP0 , 0 , mmMP0_MISC_CGTT_CTRL0 ) , data ) ;
}
static void soc15_update_drm_light_sleep ( struct amdgpu_device * adev , bool enable )
{
uint32_t def , data ;
def = data = RREG32 ( SOC15_REG_OFFSET ( MP0 , 0 , mmMP0_MISC_LIGHT_SLEEP_CTRL ) ) ;
if ( enable & & ( adev - > cg_flags & AMD_CG_SUPPORT_DRM_LS ) )
data | = 1 ;
else
data & = ~ 1 ;
if ( def ! = data )
WREG32 ( SOC15_REG_OFFSET ( MP0 , 0 , mmMP0_MISC_LIGHT_SLEEP_CTRL ) , data ) ;
}
static void soc15_update_rom_medium_grain_clock_gating ( struct amdgpu_device * adev ,
bool enable )
{
uint32_t def , data ;
def = data = RREG32 ( SOC15_REG_OFFSET ( SMUIO , 0 , mmCGTT_ROM_CLK_CTRL0 ) ) ;
if ( enable & & ( adev - > cg_flags & AMD_CG_SUPPORT_ROM_MGCG ) )
data & = ~ ( CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE0_MASK |
CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE1_MASK ) ;
else
data | = CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE0_MASK |
CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE1_MASK ;
if ( def ! = data )
WREG32 ( SOC15_REG_OFFSET ( SMUIO , 0 , mmCGTT_ROM_CLK_CTRL0 ) , data ) ;
}
static int soc15_common_set_clockgating_state ( void * handle ,
enum amd_clockgating_state state )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
2017-03-22 18:02:40 +08:00
if ( amdgpu_sriov_vf ( adev ) )
return 0 ;
2017-03-06 14:49:53 -05:00
switch ( adev - > asic_type ) {
case CHIP_VEGA10 :
2018-03-06 22:35:19 -05:00
case CHIP_VEGA12 :
2018-01-26 15:10:55 +08:00
case CHIP_VEGA20 :
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > update_medium_grain_clock_gating ( adev ,
2017-03-06 14:49:53 -05:00
state = = AMD_CG_STATE_GATE ? true : false ) ;
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > update_medium_grain_light_sleep ( adev ,
2017-03-06 14:49:53 -05:00
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_hdp_light_sleep ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_drm_clock_gating ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_drm_light_sleep ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_rom_medium_grain_clock_gating ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
2018-03-28 17:08:04 +08:00
adev - > df_funcs - > update_medium_grain_clock_gating ( adev ,
2017-03-06 14:49:53 -05:00
state = = AMD_CG_STATE_GATE ? true : false ) ;
break ;
2017-01-18 18:12:59 +08:00
case CHIP_RAVEN :
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > update_medium_grain_clock_gating ( adev ,
2017-01-18 18:12:59 +08:00
state = = AMD_CG_STATE_GATE ? true : false ) ;
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > update_medium_grain_light_sleep ( adev ,
2017-01-18 18:12:59 +08:00
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_hdp_light_sleep ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_drm_clock_gating ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_drm_light_sleep ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
soc15_update_rom_medium_grain_clock_gating ( adev ,
state = = AMD_CG_STATE_GATE ? true : false ) ;
break ;
2017-03-06 14:49:53 -05:00
default :
break ;
}
return 0 ;
}
2017-03-24 10:46:16 +08:00
static void soc15_common_get_clockgating_state ( void * handle , u32 * flags )
{
struct amdgpu_device * adev = ( struct amdgpu_device * ) handle ;
int data ;
if ( amdgpu_sriov_vf ( adev ) )
* flags = 0 ;
2017-12-08 13:07:58 -05:00
adev - > nbio_funcs - > get_clockgating_state ( adev , flags ) ;
2017-03-24 10:46:16 +08:00
/* AMD_CG_SUPPORT_HDP_LS */
data = RREG32 ( SOC15_REG_OFFSET ( HDP , 0 , mmHDP_MEM_POWER_LS ) ) ;
if ( data & HDP_MEM_POWER_LS__LS_ENABLE_MASK )
* flags | = AMD_CG_SUPPORT_HDP_LS ;
/* AMD_CG_SUPPORT_DRM_MGCG */
data = RREG32 ( SOC15_REG_OFFSET ( MP0 , 0 , mmMP0_MISC_CGTT_CTRL0 ) ) ;
if ( ! ( data & 0x01000000 ) )
* flags | = AMD_CG_SUPPORT_DRM_MGCG ;
/* AMD_CG_SUPPORT_DRM_LS */
data = RREG32 ( SOC15_REG_OFFSET ( MP0 , 0 , mmMP0_MISC_LIGHT_SLEEP_CTRL ) ) ;
if ( data & 0x1 )
* flags | = AMD_CG_SUPPORT_DRM_LS ;
/* AMD_CG_SUPPORT_ROM_MGCG */
data = RREG32 ( SOC15_REG_OFFSET ( SMUIO , 0 , mmCGTT_ROM_CLK_CTRL0 ) ) ;
if ( ! ( data & CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE0_MASK ) )
* flags | = AMD_CG_SUPPORT_ROM_MGCG ;
2018-03-28 17:08:04 +08:00
adev - > df_funcs - > get_clockgating_state ( adev , flags ) ;
2017-03-24 10:46:16 +08:00
}
2017-03-06 14:49:53 -05:00
static int soc15_common_set_powergating_state ( void * handle ,
enum amd_powergating_state state )
{
/* todo */
return 0 ;
}
const struct amd_ip_funcs soc15_common_ip_funcs = {
. name = " soc15_common " ,
. early_init = soc15_common_early_init ,
2017-04-05 13:04:50 +08:00
. late_init = soc15_common_late_init ,
2017-03-06 14:49:53 -05:00
. sw_init = soc15_common_sw_init ,
. sw_fini = soc15_common_sw_fini ,
. hw_init = soc15_common_hw_init ,
. hw_fini = soc15_common_hw_fini ,
. suspend = soc15_common_suspend ,
. resume = soc15_common_resume ,
. is_idle = soc15_common_is_idle ,
. wait_for_idle = soc15_common_wait_for_idle ,
. soft_reset = soc15_common_soft_reset ,
. set_clockgating_state = soc15_common_set_clockgating_state ,
. set_powergating_state = soc15_common_set_powergating_state ,
2017-03-24 10:46:16 +08:00
. get_clockgating_state = soc15_common_get_clockgating_state ,
2017-03-06 14:49:53 -05:00
} ;