2015-04-20 16:55:21 -04:00
/*
* Copyright 2014 Advanced Micro Devices , Inc .
* All Rights Reserved .
*
* Permission is hereby granted , free of charge , to any person obtaining a
* copy of this software and associated documentation files ( the
* " Software " ) , to deal in the Software without restriction , including
* without limitation the rights to use , copy , modify , merge , publish ,
* distribute , sub license , and / or sell copies of the Software , and to
* permit persons to whom the Software is furnished to do so , subject to
* the following conditions :
*
* THE SOFTWARE IS PROVIDED " AS IS " , WITHOUT WARRANTY OF ANY KIND , EXPRESS OR
* IMPLIED , INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY ,
* FITNESS FOR A PARTICULAR PURPOSE AND NON - INFRINGEMENT . IN NO EVENT SHALL
* THE COPYRIGHT HOLDERS , AUTHORS AND / OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM ,
* DAMAGES OR OTHER LIABILITY , WHETHER IN AN ACTION OF CONTRACT , TORT OR
* OTHERWISE , ARISING FROM , OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
* USE OR OTHER DEALINGS IN THE SOFTWARE .
*
* The above copyright notice and this permission notice ( including the
* next paragraph ) shall be included in all copies or substantial portions
* of the Software .
*
*/
/*
* Authors :
* Christian König < christian . koenig @ amd . com >
*/
# include <drm/drmP.h>
# include "amdgpu.h"
# include "amdgpu_trace.h"
2018-02-06 20:32:35 -05:00
# include "amdgpu_amdkfd.h"
2015-04-20 16:55:21 -04:00
2015-08-20 14:47:40 +08:00
struct amdgpu_sync_entry {
struct hlist_node node ;
2016-10-25 13:00:45 +01:00
struct dma_fence * fence ;
2017-11-13 14:47:52 -05:00
bool explicit ;
2015-08-20 14:47:40 +08:00
} ;
2016-02-16 11:24:58 +01:00
static struct kmem_cache * amdgpu_sync_slab ;
2015-04-20 16:55:21 -04:00
/**
* amdgpu_sync_create - zero init sync object
*
* @ sync : sync object to initialize
*
* Just clear the sync object for now .
*/
void amdgpu_sync_create ( struct amdgpu_sync * sync )
{
2015-08-20 14:47:40 +08:00
hash_init ( sync - > fences ) ;
2015-04-20 16:55:21 -04:00
sync - > last_vm_update = NULL ;
}
2016-02-16 16:23:02 +01:00
/**
* amdgpu_sync_same_dev - test if fence belong to us
*
* @ adev : amdgpu device to use for the test
* @ f : fence to test
*
* Test if the fence was issued by us .
*/
2016-10-25 13:00:45 +01:00
static bool amdgpu_sync_same_dev ( struct amdgpu_device * adev ,
struct dma_fence * f )
2015-08-20 18:33:59 +08:00
{
2017-12-06 17:49:39 +01:00
struct drm_sched_fence * s_fence = to_drm_sched_fence ( f ) ;
2015-08-20 18:33:59 +08:00
2015-09-08 20:22:31 +02:00
if ( s_fence ) {
struct amdgpu_ring * ring ;
ring = container_of ( s_fence - > sched , struct amdgpu_ring , sched ) ;
return ring - > adev = = adev ;
}
2015-08-20 18:33:59 +08:00
return false ;
}
2016-02-16 16:23:02 +01:00
/**
* amdgpu_sync_get_owner - extract the owner of a fence
*
* @ fence : fence get the owner from
*
* Extract who originally created the fence .
*/
2016-10-25 13:00:45 +01:00
static void * amdgpu_sync_get_owner ( struct dma_fence * f )
2015-08-20 18:33:59 +08:00
{
2018-02-06 20:32:35 -05:00
struct drm_sched_fence * s_fence ;
struct amdgpu_amdkfd_fence * kfd_fence ;
if ( ! f )
return AMDGPU_FENCE_OWNER_UNDEFINED ;
2016-02-16 16:23:02 +01:00
2018-02-06 20:32:35 -05:00
s_fence = to_drm_sched_fence ( f ) ;
2015-08-20 18:33:59 +08:00
if ( s_fence )
2016-02-16 16:23:02 +01:00
return s_fence - > owner ;
2016-02-16 10:57:10 +01:00
2018-02-06 20:32:35 -05:00
kfd_fence = to_amdgpu_amdkfd_fence ( f ) ;
if ( kfd_fence )
return AMDGPU_FENCE_OWNER_KFD ;
2016-02-16 16:23:02 +01:00
return AMDGPU_FENCE_OWNER_UNDEFINED ;
2015-08-20 18:33:59 +08:00
}
2016-02-16 16:23:02 +01:00
/**
* amdgpu_sync_keep_later - Keep the later fence
*
* @ keep : existing fence to test
* @ fence : new fence
*
* Either keep the existing fence or the new one , depending which one is later .
*/
2016-10-25 13:00:45 +01:00
static void amdgpu_sync_keep_later ( struct dma_fence * * keep ,
struct dma_fence * fence )
2015-10-22 10:53:16 +02:00
{
2016-10-25 13:00:45 +01:00
if ( * keep & & dma_fence_is_later ( * keep , fence ) )
2015-10-22 10:53:16 +02:00
return ;
2016-10-25 13:00:45 +01:00
dma_fence_put ( * keep ) ;
* keep = dma_fence_get ( fence ) ;
2015-10-22 10:53:16 +02:00
}
2016-02-15 12:33:02 +01:00
/**
* amdgpu_sync_add_later - add the fence to the hash
*
* @ sync : sync object to add the fence to
* @ f : fence to add
*
* Tries to add the fence to an existing hash entry . Returns true when an entry
* was found , false otherwise .
*/
2017-12-05 08:03:13 -05:00
static bool amdgpu_sync_add_later ( struct amdgpu_sync * sync , struct dma_fence * f , bool explicit )
2016-02-15 12:33:02 +01:00
{
struct amdgpu_sync_entry * e ;
hash_for_each_possible ( sync - > fences , e , node , f - > context ) {
if ( unlikely ( e - > fence - > context ! = f - > context ) )
continue ;
amdgpu_sync_keep_later ( & e - > fence , f ) ;
2017-12-05 08:03:13 -05:00
/* Preserve eplicit flag to not loose pipe line sync */
e - > explicit | = explicit ;
2016-02-15 12:33:02 +01:00
return true ;
}
return false ;
}
2015-04-20 16:55:21 -04:00
/**
2015-07-06 22:06:40 +02:00
* amdgpu_sync_fence - remember to sync to this fence
2015-04-20 16:55:21 -04:00
*
* @ sync : sync object to add fence to
* @ fence : fence to sync to
*
*/
2015-07-06 22:06:40 +02:00
int amdgpu_sync_fence ( struct amdgpu_device * adev , struct amdgpu_sync * sync ,
2017-11-13 14:47:52 -05:00
struct dma_fence * f , bool explicit )
2015-04-20 16:55:21 -04:00
{
2015-08-20 14:47:40 +08:00
struct amdgpu_sync_entry * e ;
2015-04-20 16:55:21 -04:00
2015-07-06 22:06:40 +02:00
if ( ! f )
return 0 ;
2015-08-20 18:33:59 +08:00
if ( amdgpu_sync_same_dev ( adev , f ) & &
2016-02-16 16:23:02 +01:00
amdgpu_sync_get_owner ( f ) = = AMDGPU_FENCE_OWNER_VM )
2015-10-22 10:53:16 +02:00
amdgpu_sync_keep_later ( & sync - > last_vm_update , f ) ;
2015-08-20 18:33:59 +08:00
2017-12-05 08:03:13 -05:00
if ( amdgpu_sync_add_later ( sync , f , explicit ) )
2015-08-20 14:47:40 +08:00
return 0 ;
2015-04-20 16:55:21 -04:00
2016-02-16 11:24:58 +01:00
e = kmem_cache_alloc ( amdgpu_sync_slab , GFP_KERNEL ) ;
2016-01-18 14:49:45 +01:00
if ( ! e )
return - ENOMEM ;
2015-04-20 16:55:21 -04:00
2017-11-13 14:47:52 -05:00
e - > explicit = explicit ;
2016-01-18 14:49:45 +01:00
hash_add ( sync - > fences , & e - > node , f - > context ) ;
2016-10-25 13:00:45 +01:00
e - > fence = dma_fence_get ( f ) ;
2015-07-06 22:06:40 +02:00
return 0 ;
2015-04-20 16:55:21 -04:00
}
/**
2016-01-15 11:05:21 +08:00
* amdgpu_sync_resv - sync to a reservation object
2015-04-20 16:55:21 -04:00
*
* @ sync : sync object to add fences from reservation object to
* @ resv : reservation object with embedded fence
2017-09-15 20:44:06 -04:00
* @ explicit_sync : true if we should only sync to the exclusive fence
2015-04-20 16:55:21 -04:00
*
2016-01-15 11:05:21 +08:00
* Sync to the fence
2015-04-20 16:55:21 -04:00
*/
int amdgpu_sync_resv ( struct amdgpu_device * adev ,
struct amdgpu_sync * sync ,
struct reservation_object * resv ,
2017-09-15 20:44:06 -04:00
void * owner , bool explicit_sync )
2015-04-20 16:55:21 -04:00
{
struct reservation_object_list * flist ;
2016-10-25 13:00:45 +01:00
struct dma_fence * f ;
2015-08-24 16:59:54 +08:00
void * fence_owner ;
2015-04-20 16:55:21 -04:00
unsigned i ;
int r = 0 ;
2015-05-12 23:17:19 +08:00
if ( resv = = NULL )
return - EINVAL ;
2015-04-20 16:55:21 -04:00
/* always sync to the exclusive fence */
f = reservation_object_get_excl ( resv ) ;
2017-11-13 14:47:52 -05:00
r = amdgpu_sync_fence ( adev , sync , f , false ) ;
2015-04-20 16:55:21 -04:00
flist = reservation_object_get_list ( resv ) ;
if ( ! flist | | r )
return r ;
for ( i = 0 ; i < flist - > shared_count ; + + i ) {
f = rcu_dereference_protected ( flist - > shared [ i ] ,
reservation_object_held ( resv ) ) ;
2018-02-06 20:32:35 -05:00
/* We only want to trigger KFD eviction fences on
* evict or move jobs . Skip KFD fences otherwise .
*/
fence_owner = amdgpu_sync_get_owner ( f ) ;
if ( fence_owner = = AMDGPU_FENCE_OWNER_KFD & &
owner ! = AMDGPU_FENCE_OWNER_UNDEFINED )
continue ;
2015-08-24 16:59:54 +08:00
if ( amdgpu_sync_same_dev ( adev , f ) ) {
2015-07-27 15:40:35 +02:00
/* VM updates are only interesting
* for other VM updates and moves .
*/
2015-10-27 17:28:24 +01:00
if ( ( owner ! = AMDGPU_FENCE_OWNER_UNDEFINED ) & &
( fence_owner ! = AMDGPU_FENCE_OWNER_UNDEFINED ) & &
2015-07-27 15:40:35 +02:00
( ( owner = = AMDGPU_FENCE_OWNER_VM ) ! =
2015-08-24 16:59:54 +08:00
( fence_owner = = AMDGPU_FENCE_OWNER_VM ) ) )
2015-07-06 22:06:40 +02:00
continue ;
2017-11-24 11:41:52 +01:00
/* Ignore fence from the same owner and explicit one as
2015-07-27 15:40:35 +02:00
* long as it isn ' t undefined .
*/
if ( owner ! = AMDGPU_FENCE_OWNER_UNDEFINED & &
2017-11-24 11:41:52 +01:00
( fence_owner = = owner | | explicit_sync ) )
2015-07-27 15:40:35 +02:00
continue ;
}
2017-11-13 14:47:52 -05:00
r = amdgpu_sync_fence ( adev , sync , f , false ) ;
2015-07-06 22:06:40 +02:00
if ( r )
break ;
2015-04-20 16:55:21 -04:00
}
return r ;
}
2016-02-15 12:33:02 +01:00
/**
2016-06-01 10:47:36 +02:00
* amdgpu_sync_peek_fence - get the next fence not signaled yet
2016-02-15 12:33:02 +01:00
*
* @ sync : the sync object
2016-05-23 14:26:39 +02:00
* @ ring : optional ring to use for test
2016-02-15 12:33:02 +01:00
*
2016-06-01 10:47:36 +02:00
* Returns the next fence not signaled yet without removing it from the sync
* object .
2016-02-15 12:33:02 +01:00
*/
2016-10-25 13:00:45 +01:00
struct dma_fence * amdgpu_sync_peek_fence ( struct amdgpu_sync * sync ,
struct amdgpu_ring * ring )
2016-02-15 12:33:02 +01:00
{
struct amdgpu_sync_entry * e ;
struct hlist_node * tmp ;
int i ;
hash_for_each_safe ( sync - > fences , i , tmp , e , node ) {
2016-10-25 13:00:45 +01:00
struct dma_fence * f = e - > fence ;
2017-12-06 17:49:39 +01:00
struct drm_sched_fence * s_fence = to_drm_sched_fence ( f ) ;
2016-05-23 14:26:39 +02:00
2017-08-11 09:34:33 +08:00
if ( dma_fence_is_signaled ( f ) ) {
hash_del ( & e - > node ) ;
dma_fence_put ( f ) ;
kmem_cache_free ( amdgpu_sync_slab , e ) ;
continue ;
}
2016-05-23 14:26:39 +02:00
if ( ring & & s_fence ) {
/* For fences from the same ring it is sufficient
* when they are scheduled .
*/
2016-06-01 10:47:36 +02:00
if ( s_fence - > sched = = & ring - > sched ) {
2016-10-25 13:00:45 +01:00
if ( dma_fence_is_signaled ( & s_fence - > scheduled ) )
2016-06-01 10:47:36 +02:00
continue ;
2016-02-15 12:33:02 +01:00
2016-06-01 10:47:36 +02:00
return & s_fence - > scheduled ;
}
2016-02-15 12:33:02 +01:00
}
2016-06-01 10:47:36 +02:00
return f ;
2016-02-15 12:33:02 +01:00
}
2016-06-01 10:47:36 +02:00
return NULL ;
2016-02-15 12:33:02 +01:00
}
2016-05-23 16:19:44 +02:00
/**
* amdgpu_sync_get_fence - get the next fence from the sync object
*
* @ sync : sync object to use
2017-11-13 14:47:52 -05:00
* @ explicit : true if the next fence is explicit
2016-05-23 16:19:44 +02:00
*
* Get and removes the next fence from the sync object not signaled yet .
*/
2017-11-13 14:47:52 -05:00
struct dma_fence * amdgpu_sync_get_fence ( struct amdgpu_sync * sync , bool * explicit )
2015-08-25 11:05:36 +02:00
{
struct amdgpu_sync_entry * e ;
struct hlist_node * tmp ;
2016-10-25 13:00:45 +01:00
struct dma_fence * f ;
2015-08-25 11:05:36 +02:00
int i ;
hash_for_each_safe ( sync - > fences , i , tmp , e , node ) {
f = e - > fence ;
2017-11-13 14:47:52 -05:00
if ( explicit )
* explicit = e - > explicit ;
2015-08-25 11:05:36 +02:00
hash_del ( & e - > node ) ;
2016-02-16 11:24:58 +01:00
kmem_cache_free ( amdgpu_sync_slab , e ) ;
2015-08-25 11:05:36 +02:00
2016-10-25 13:00:45 +01:00
if ( ! dma_fence_is_signaled ( f ) )
2015-08-25 11:05:36 +02:00
return f ;
2016-10-25 13:00:45 +01:00
dma_fence_put ( f ) ;
2015-08-25 11:05:36 +02:00
}
return NULL ;
}
2018-02-06 20:32:37 -05:00
/**
* amdgpu_sync_clone - clone a sync object
*
* @ source : sync object to clone
* @ clone : pointer to destination sync object
*
* Adds references to all unsignaled fences in @ source to @ clone . Also
* removes signaled fences from @ source while at it .
*/
int amdgpu_sync_clone ( struct amdgpu_sync * source , struct amdgpu_sync * clone )
{
struct amdgpu_sync_entry * e ;
struct hlist_node * tmp ;
struct dma_fence * f ;
int i , r ;
hash_for_each_safe ( source - > fences , i , tmp , e , node ) {
f = e - > fence ;
if ( ! dma_fence_is_signaled ( f ) ) {
r = amdgpu_sync_fence ( NULL , clone , f , e - > explicit ) ;
if ( r )
return r ;
} else {
hash_del ( & e - > node ) ;
dma_fence_put ( f ) ;
kmem_cache_free ( amdgpu_sync_slab , e ) ;
}
}
dma_fence_put ( clone - > last_vm_update ) ;
clone - > last_vm_update = dma_fence_get ( source - > last_vm_update ) ;
return 0 ;
}
2017-05-15 15:09:15 -04:00
int amdgpu_sync_wait ( struct amdgpu_sync * sync , bool intr )
{
struct amdgpu_sync_entry * e ;
struct hlist_node * tmp ;
int i , r ;
hash_for_each_safe ( sync - > fences , i , tmp , e , node ) {
r = dma_fence_wait ( e - > fence , intr ) ;
if ( r )
return r ;
hash_del ( & e - > node ) ;
dma_fence_put ( e - > fence ) ;
kmem_cache_free ( amdgpu_sync_slab , e ) ;
}
return 0 ;
}
2015-04-20 16:55:21 -04:00
/**
* amdgpu_sync_free - free the sync object
*
* @ sync : sync object to use
*
2016-01-15 11:05:21 +08:00
* Free the sync object .
2015-04-20 16:55:21 -04:00
*/
2016-02-03 15:11:39 +01:00
void amdgpu_sync_free ( struct amdgpu_sync * sync )
2015-04-20 16:55:21 -04:00
{
2015-08-20 14:47:40 +08:00
struct amdgpu_sync_entry * e ;
struct hlist_node * tmp ;
2015-04-20 16:55:21 -04:00
unsigned i ;
2015-08-20 14:47:40 +08:00
hash_for_each_safe ( sync - > fences , i , tmp , e , node ) {
hash_del ( & e - > node ) ;
2016-10-25 13:00:45 +01:00
dma_fence_put ( e - > fence ) ;
2016-02-16 11:24:58 +01:00
kmem_cache_free ( amdgpu_sync_slab , e ) ;
2015-08-20 14:47:40 +08:00
}
2016-10-25 13:00:45 +01:00
dma_fence_put ( sync - > last_vm_update ) ;
2015-04-20 16:55:21 -04:00
}
2016-02-16 11:24:58 +01:00
/**
* amdgpu_sync_init - init sync object subsystem
*
* Allocate the slab allocator .
*/
int amdgpu_sync_init ( void )
{
amdgpu_sync_slab = kmem_cache_create (
" amdgpu_sync " , sizeof ( struct amdgpu_sync_entry ) , 0 ,
SLAB_HWCACHE_ALIGN , NULL ) ;
if ( ! amdgpu_sync_slab )
return - ENOMEM ;
return 0 ;
}
/**
* amdgpu_sync_fini - fini sync object subsystem
*
* Free the slab allocator .
*/
void amdgpu_sync_fini ( void )
{
kmem_cache_destroy ( amdgpu_sync_slab ) ;
}