2013-01-18 13:42:19 +04:00
/*
* Copyright ( C ) 2004 , 2007 - 2010 , 2011 - 2012 Synopsys , Inc . ( www . synopsys . com )
*
* This program is free software ; you can redistribute it and / or modify
* it under the terms of the GNU General Public License version 2 as
* published by the Free Software Foundation .
*
* vineetg : May 2011
* - Refactored get_new_mmu_context ( ) to only handle live - mm .
* retiring - mm handled in other hooks
*
* Vineetg : March 25 th , 2008 : Bug # 92690
* - Major rewrite of Core ASID allocation routine get_new_mmu_context
*
* Amit Bhor , Sameer Dhavale : Codito Technologies 2004
*/
# ifndef _ASM_ARC_MMU_CONTEXT_H
# define _ASM_ARC_MMU_CONTEXT_H
# include <asm/arcregs.h>
# include <asm/tlb.h>
# include <asm-generic/mm_hooks.h>
/* ARC700 ASID Management
*
* ARC MMU provides 8 - bit ASID ( 0. .255 ) to TAG TLB entries , allowing entries
* with same vaddr ( different tasks ) to co - exit . This provides for
* " Fast Context Switch " i . e . no TLB flush on ctxt - switch
*
* Linux assigns each task a unique ASID . A simple round - robin allocation
2013-08-23 17:46:34 +04:00
* of H / w ASID is done using software tracker @ asid_cpu .
2013-01-18 13:42:19 +04:00
* When it reaches max 255 , the allocation cycle starts afresh by flushing
* the entire TLB and wrapping ASID back to zero .
*
ARC: [ASID] Track ASID allocation cycles/generations
This helps remove asid-to-mm reverse map
While mm->context.id contains the ASID assigned to a process, our ASID
allocator also used asid_mm_map[] reverse map. In a new allocation
cycle (mm->ASID >= @asid_cache), the Round Robin ASID allocator used this
to check if new @asid_cache belonged to some mm2 (from prev cycle).
If so, it could locate that mm using the ASID reverse map, and mark that
mm as unallocated ASID, to force it to refresh at the time of switch_mm()
However, for SMP, the reverse map has to be maintained per CPU, so
becomes 2 dimensional, hence got rid of it.
With reverse map gone, it is NOT possible to reach out to current
assignee. So we track the ASID allocation generation/cycle and
on every switch_mm(), check if the current generation of CPU ASID is
same as mm's ASID; If not it is refreshed.
(Based loosely on arch/sh implementation)
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
2013-07-26 02:45:50 +04:00
* A new allocation cycle , post rollover , could potentially reassign an ASID
* to a different task . Thus the rule is to refresh the ASID in a new cycle .
2013-08-23 17:46:34 +04:00
* The 32 bit @ asid_cpu ( and mm - > asid ) have 8 bits MMU PID and rest 24 bits
ARC: [ASID] Track ASID allocation cycles/generations
This helps remove asid-to-mm reverse map
While mm->context.id contains the ASID assigned to a process, our ASID
allocator also used asid_mm_map[] reverse map. In a new allocation
cycle (mm->ASID >= @asid_cache), the Round Robin ASID allocator used this
to check if new @asid_cache belonged to some mm2 (from prev cycle).
If so, it could locate that mm using the ASID reverse map, and mark that
mm as unallocated ASID, to force it to refresh at the time of switch_mm()
However, for SMP, the reverse map has to be maintained per CPU, so
becomes 2 dimensional, hence got rid of it.
With reverse map gone, it is NOT possible to reach out to current
assignee. So we track the ASID allocation generation/cycle and
on every switch_mm(), check if the current generation of CPU ASID is
same as mm's ASID; If not it is refreshed.
(Based loosely on arch/sh implementation)
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
2013-07-26 02:45:50 +04:00
* serve as cycle / generation indicator and natural 32 bit unsigned math
* automagically increments the generation when lower 8 bits rollover .
2013-01-18 13:42:19 +04:00
*/
ARC: [ASID] Track ASID allocation cycles/generations
This helps remove asid-to-mm reverse map
While mm->context.id contains the ASID assigned to a process, our ASID
allocator also used asid_mm_map[] reverse map. In a new allocation
cycle (mm->ASID >= @asid_cache), the Round Robin ASID allocator used this
to check if new @asid_cache belonged to some mm2 (from prev cycle).
If so, it could locate that mm using the ASID reverse map, and mark that
mm as unallocated ASID, to force it to refresh at the time of switch_mm()
However, for SMP, the reverse map has to be maintained per CPU, so
becomes 2 dimensional, hence got rid of it.
With reverse map gone, it is NOT possible to reach out to current
assignee. So we track the ASID allocation generation/cycle and
on every switch_mm(), check if the current generation of CPU ASID is
same as mm's ASID; If not it is refreshed.
(Based loosely on arch/sh implementation)
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
2013-07-26 02:45:50 +04:00
# define MM_CTXT_ASID_MASK 0x000000ff /* MMU PID reg :8 bit PID */
# define MM_CTXT_CYCLE_MASK (~MM_CTXT_ASID_MASK)
# define MM_CTXT_FIRST_CYCLE (MM_CTXT_ASID_MASK + 1)
# define MM_CTXT_NO_ASID 0UL
2013-01-18 13:42:19 +04:00
2013-08-23 17:46:34 +04:00
# define asid_mm(mm, cpu) mm->context.asid[cpu]
# define hw_pid(mm, cpu) (asid_mm(mm, cpu) & MM_CTXT_ASID_MASK)
2013-01-18 13:42:19 +04:00
2013-08-23 17:46:34 +04:00
DECLARE_PER_CPU ( unsigned int , asid_cache ) ;
# define asid_cpu(cpu) per_cpu(asid_cache, cpu)
2013-01-18 13:42:19 +04:00
/*
2013-07-25 00:53:45 +04:00
* Get a new ASID if task doesn ' t have a valid one ( unalloc or from prev cycle )
* Also set the MMU PID register to existing / updated ASID
2013-01-18 13:42:19 +04:00
*/
static inline void get_new_mmu_context ( struct mm_struct * mm )
{
2013-08-23 17:46:34 +04:00
const unsigned int cpu = smp_processor_id ( ) ;
2013-01-18 13:42:19 +04:00
unsigned long flags ;
local_irq_save ( flags ) ;
2013-07-25 00:53:45 +04:00
/*
* Move to new ASID if it was not from current alloc - cycle / generation .
ARC: [ASID] Track ASID allocation cycles/generations
This helps remove asid-to-mm reverse map
While mm->context.id contains the ASID assigned to a process, our ASID
allocator also used asid_mm_map[] reverse map. In a new allocation
cycle (mm->ASID >= @asid_cache), the Round Robin ASID allocator used this
to check if new @asid_cache belonged to some mm2 (from prev cycle).
If so, it could locate that mm using the ASID reverse map, and mark that
mm as unallocated ASID, to force it to refresh at the time of switch_mm()
However, for SMP, the reverse map has to be maintained per CPU, so
becomes 2 dimensional, hence got rid of it.
With reverse map gone, it is NOT possible to reach out to current
assignee. So we track the ASID allocation generation/cycle and
on every switch_mm(), check if the current generation of CPU ASID is
same as mm's ASID; If not it is refreshed.
(Based loosely on arch/sh implementation)
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
2013-07-26 02:45:50 +04:00
* This is done by ensuring that the generation bits in both mm - > ASID
* and cpu ' s ASID counter are exactly same .
2013-07-25 00:53:45 +04:00
*
* Note : Callers needing new ASID unconditionally , independent of
* generation , e . g . local_flush_tlb_mm ( ) for forking parent ,
* first need to destroy the context , setting it to invalid
* value .
*/
2013-08-23 17:46:34 +04:00
if ( ! ( ( asid_mm ( mm , cpu ) ^ asid_cpu ( cpu ) ) & MM_CTXT_CYCLE_MASK ) )
2013-07-25 00:53:45 +04:00
goto set_hw ;
ARC: [ASID] Track ASID allocation cycles/generations
This helps remove asid-to-mm reverse map
While mm->context.id contains the ASID assigned to a process, our ASID
allocator also used asid_mm_map[] reverse map. In a new allocation
cycle (mm->ASID >= @asid_cache), the Round Robin ASID allocator used this
to check if new @asid_cache belonged to some mm2 (from prev cycle).
If so, it could locate that mm using the ASID reverse map, and mark that
mm as unallocated ASID, to force it to refresh at the time of switch_mm()
However, for SMP, the reverse map has to be maintained per CPU, so
becomes 2 dimensional, hence got rid of it.
With reverse map gone, it is NOT possible to reach out to current
assignee. So we track the ASID allocation generation/cycle and
on every switch_mm(), check if the current generation of CPU ASID is
same as mm's ASID; If not it is refreshed.
(Based loosely on arch/sh implementation)
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
2013-07-26 02:45:50 +04:00
/* move to new ASID and handle rollover */
2013-08-23 17:46:34 +04:00
if ( unlikely ( ! ( + + asid_cpu ( cpu ) & MM_CTXT_ASID_MASK ) ) ) {
2013-01-18 13:42:19 +04:00
2013-10-27 13:19:02 +04:00
local_flush_tlb_all ( ) ;
2013-01-18 13:42:19 +04:00
ARC: [ASID] Track ASID allocation cycles/generations
This helps remove asid-to-mm reverse map
While mm->context.id contains the ASID assigned to a process, our ASID
allocator also used asid_mm_map[] reverse map. In a new allocation
cycle (mm->ASID >= @asid_cache), the Round Robin ASID allocator used this
to check if new @asid_cache belonged to some mm2 (from prev cycle).
If so, it could locate that mm using the ASID reverse map, and mark that
mm as unallocated ASID, to force it to refresh at the time of switch_mm()
However, for SMP, the reverse map has to be maintained per CPU, so
becomes 2 dimensional, hence got rid of it.
With reverse map gone, it is NOT possible to reach out to current
assignee. So we track the ASID allocation generation/cycle and
on every switch_mm(), check if the current generation of CPU ASID is
same as mm's ASID; If not it is refreshed.
(Based loosely on arch/sh implementation)
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
2013-07-26 02:45:50 +04:00
/*
* Above checke for rollover of 8 bit ASID in 32 bit container .
* If the container itself wrapped around , set it to a non zero
* " generation " to distinguish from no context
*/
2013-08-23 17:46:34 +04:00
if ( ! asid_cpu ( cpu ) )
asid_cpu ( cpu ) = MM_CTXT_FIRST_CYCLE ;
ARC: [ASID] Track ASID allocation cycles/generations
This helps remove asid-to-mm reverse map
While mm->context.id contains the ASID assigned to a process, our ASID
allocator also used asid_mm_map[] reverse map. In a new allocation
cycle (mm->ASID >= @asid_cache), the Round Robin ASID allocator used this
to check if new @asid_cache belonged to some mm2 (from prev cycle).
If so, it could locate that mm using the ASID reverse map, and mark that
mm as unallocated ASID, to force it to refresh at the time of switch_mm()
However, for SMP, the reverse map has to be maintained per CPU, so
becomes 2 dimensional, hence got rid of it.
With reverse map gone, it is NOT possible to reach out to current
assignee. So we track the ASID allocation generation/cycle and
on every switch_mm(), check if the current generation of CPU ASID is
same as mm's ASID; If not it is refreshed.
(Based loosely on arch/sh implementation)
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
2013-07-26 02:45:50 +04:00
}
2013-01-18 13:42:19 +04:00
/* Assign new ASID to tsk */
2013-08-23 17:46:34 +04:00
asid_mm ( mm , cpu ) = asid_cpu ( cpu ) ;
2013-01-18 13:42:19 +04:00
2013-07-25 00:53:45 +04:00
set_hw :
2013-08-23 17:46:34 +04:00
write_aux_reg ( ARC_REG_PID , hw_pid ( mm , cpu ) | MMU_ENABLE ) ;
2013-01-18 13:42:19 +04:00
local_irq_restore ( flags ) ;
}
/*
* Initialize the context related info for a new mm_struct
* instance .
*/
static inline int
init_new_context ( struct task_struct * tsk , struct mm_struct * mm )
{
2013-08-23 17:46:34 +04:00
int i ;
for_each_possible_cpu ( i )
asid_mm ( mm , i ) = MM_CTXT_NO_ASID ;
2013-01-18 13:42:19 +04:00
return 0 ;
}
2013-08-23 17:46:34 +04:00
static inline void destroy_context ( struct mm_struct * mm )
{
unsigned long flags ;
/* Needed to elide CONFIG_DEBUG_PREEMPT warning */
local_irq_save ( flags ) ;
asid_mm ( mm , smp_processor_id ( ) ) = MM_CTXT_NO_ASID ;
local_irq_restore ( flags ) ;
}
2013-01-18 13:42:19 +04:00
/* Prepare the MMU for task: setup PID reg with allocated ASID
If task doesn ' t have an ASID ( never alloc or stolen , get a new ASID )
*/
static inline void switch_mm ( struct mm_struct * prev , struct mm_struct * next ,
struct task_struct * tsk )
{
2013-10-27 13:19:02 +04:00
const int cpu = smp_processor_id ( ) ;
/*
* Note that the mm_cpumask is " aggregating " only , we don ' t clear it
* for the switched - out task , unlike some other arches .
* It is used to enlist cpus for sending TLB flush IPIs and not sending
* it to CPUs where a task once ran - on , could cause stale TLB entry
* re - use , specially for a multi - threaded task .
* e . g . T1 runs on C1 , migrates to C3 . T2 running on C2 munmaps .
* For a non - aggregating mm_cpumask , IPI not sent C1 , and if T1
* were to re - migrate to C1 , it could access the unmapped region
* via any existing stale TLB entries .
*/
cpumask_set_cpu ( cpu , mm_cpumask ( next ) ) ;
2013-01-18 13:42:23 +04:00
# ifndef CONFIG_SMP
2013-01-18 13:42:19 +04:00
/* PGD cached in MMU reg to avoid 3 mem lookups: task->mm->pgd */
write_aux_reg ( ARC_REG_SCRATCH_DATA0 , next - > pgd ) ;
2013-01-18 13:42:23 +04:00
# endif
2013-01-18 13:42:19 +04:00
2013-07-25 00:53:45 +04:00
get_new_mmu_context ( next ) ;
2013-01-18 13:42:19 +04:00
}
2013-07-25 04:31:08 +04:00
/*
* Called at the time of execve ( ) to get a new ASID
* Note the subtlety here : get_new_mmu_context ( ) behaves differently here
* vs . in switch_mm ( ) . Here it always returns a new ASID , because mm has
* an unallocated " initial " value , while in latter , it moves to a new ASID ,
* only if it was unallocated
*/
# define activate_mm(prev, next) switch_mm(prev, next, NULL)
2013-01-18 13:42:19 +04:00
/* it seemed that deactivate_mm( ) is a reasonable place to do book-keeping
* for retiring - mm . However destroy_context ( ) still needs to do that because
* between mm_release ( ) = > deactive_mm ( ) and
* mmput = > . . = > __mmdrop ( ) = > destroy_context ( )
* there is a good chance that task gets sched - out / in , making it ' s ASID valid
* again ( this teased me for a whole day ) .
*/
# define deactivate_mm(tsk, mm) do { } while (0)
# define enter_lazy_tlb(mm, tsk)
# endif /* __ASM_ARC_MMU_CONTEXT_H */