2016-04-18 23:56:29 +03:00
#!/usr/bin/env python
# Copyright (c) 2016 Ansible, Inc.
# All Rights Reserved
import os
import sys
# Python
from collections import defaultdict
from optparse import make_option , OptionParser
2017-01-25 17:16:07 +03:00
from datetime import datetime
2016-12-20 17:55:08 +03:00
import logging
2016-04-18 23:56:29 +03:00
# Django
2016-07-26 21:50:48 +03:00
import django
2016-04-18 23:56:29 +03:00
from django . utils . timezone import now
2016-12-22 20:03:56 +03:00
base_dir = os . path . abspath ( # Convert into absolute path string
os . path . join ( # Current file's grandparent directory
os . path . join ( # Current file's parent directory
os . path . dirname ( # Current file's directory
os . path . abspath ( __file__ ) # Current file path
) ,
os . pardir
) ,
os . pardir
)
)
if base_dir not in sys . path :
sys . path . insert ( 1 , base_dir )
2016-04-18 23:56:29 +03:00
2016-07-26 21:50:48 +03:00
os . environ . setdefault ( " DJANGO_SETTINGS_MODULE " , " awx.settings.development " ) # noqa
django . setup ( ) # noqa
2016-04-18 23:56:29 +03:00
2016-12-22 20:03:56 +03:00
from django . db import transaction # noqa
# awx
from awx . main . models import * # noqa
2017-01-30 17:17:46 +03:00
from awx . main . signals import ( # noqa
2017-02-16 21:55:18 +03:00
disable_activity_stream ,
disable_computed_fields
2017-01-30 17:17:46 +03:00
)
2016-12-22 20:03:56 +03:00
2016-04-18 23:56:29 +03:00
option_list = [
make_option ( ' --organizations ' , action = ' store ' , type = ' int ' , default = 3 ,
help = ' Number of organizations to create ' ) ,
make_option ( ' --users ' , action = ' store ' , type = ' int ' , default = 10 ,
help = ' Number of users to create ' ) ,
make_option ( ' --teams ' , action = ' store ' , type = ' int ' , default = 5 ,
help = ' Number of teams to create ' ) ,
make_option ( ' --projects ' , action = ' store ' , type = ' int ' , default = 10 ,
help = ' Number of projects to create ' ) ,
make_option ( ' --job-templates ' , action = ' store ' , type = ' int ' , default = 20 ,
help = ' Number of job templates to create ' ) ,
make_option ( ' --credentials ' , action = ' store ' , type = ' int ' , default = 5 ,
help = ' Number of credentials to create ' ) ,
make_option ( ' --inventories ' , action = ' store ' , type = ' int ' , default = 5 ,
help = ' Number of credentials to create ' ) ,
make_option ( ' --inventory-groups ' , action = ' store ' , type = ' int ' , default = 10 ,
help = ' Number of credentials to create ' ) ,
make_option ( ' --inventory-hosts ' , action = ' store ' , type = ' int ' , default = 40 ,
help = ' number of credentials to create ' ) ,
2016-12-22 22:37:02 +03:00
make_option ( ' --wfjts ' , action = ' store ' , type = ' int ' , default = 15 ,
help = ' number of workflow job templates to create ' ) ,
make_option ( ' --nodes ' , action = ' store ' , type = ' int ' , default = 200 ,
help = ' number of workflow job template nodes to create ' ) ,
2016-12-27 06:40:32 +03:00
make_option ( ' --labels ' , action = ' store ' , type = ' int ' , default = 100 ,
2016-12-26 03:42:56 +03:00
help = ' labels to create, will associate 10x as many ' ) ,
2016-04-18 23:56:29 +03:00
make_option ( ' --jobs ' , action = ' store ' , type = ' int ' , default = 200 ,
help = ' number of job entries to create ' ) ,
make_option ( ' --job-events ' , action = ' store ' , type = ' int ' , default = 500 ,
help = ' number of job event entries to create ' ) ,
make_option ( ' --pretend ' , action = ' store_true ' ,
help = " Don ' t commit the data to the database " ) ,
2016-12-27 06:40:32 +03:00
make_option ( ' --preset ' , action = ' store ' , type = ' string ' , default = ' ' ,
help = " Preset data set to use " ) ,
2016-04-18 23:56:29 +03:00
make_option ( ' --prefix ' , action = ' store ' , type = ' string ' , default = ' ' ,
help = " Prefix generated names with this string " ) ,
#make_option('--spread-bias', action='store', type='string', default='exponential',
# help='"exponential" to bias associations exponentially front loaded for - for ex'),
]
parser = OptionParser ( option_list = option_list )
options , remainder = parser . parse_args ( )
options = vars ( options )
2016-12-27 06:40:32 +03:00
if options [ ' preset ' ] :
2016-12-20 17:55:08 +03:00
print ' Using preset data numbers set ' + str ( options [ ' preset ' ] )
2017-01-03 17:09:43 +03:00
# Read the numbers of resources from presets file, if provided
presets_filename = os . path . abspath ( os . path . join (
os . path . dirname ( os . path . abspath ( __file__ ) ) , ' presets.tsv ' ) )
with open ( presets_filename ) as f :
text = f . read ( )
split_lines = [ line . split ( ' \t ' ) for line in text . split ( ' \n ' ) ]
keys = split_lines [ 0 ] [ 1 : ]
try :
col = keys . index ( options [ ' preset ' ] )
except ValueError :
raise Exception ( ' Preset " %s " dataset not found, options are %s ' % ( options [ ' preset ' ] , keys ) )
options . update ( { cols [ 0 ] : cols [ col + 1 ] for cols in split_lines } )
2016-12-27 06:40:32 +03:00
2016-04-18 23:56:29 +03:00
n_organizations = int ( options [ ' organizations ' ] )
n_users = int ( options [ ' users ' ] )
n_teams = int ( options [ ' teams ' ] )
n_projects = int ( options [ ' projects ' ] )
n_job_templates = int ( options [ ' job_templates ' ] )
n_credentials = int ( options [ ' credentials ' ] )
n_inventories = int ( options [ ' inventories ' ] )
n_inventory_groups = int ( options [ ' inventory_groups ' ] )
n_inventory_hosts = int ( options [ ' inventory_hosts ' ] )
2016-12-22 22:37:02 +03:00
n_wfjts = int ( options [ ' wfjts ' ] )
n_nodes = int ( options [ ' nodes ' ] )
2016-12-26 03:42:56 +03:00
n_labels = int ( options [ ' labels ' ] )
2016-04-18 23:56:29 +03:00
n_jobs = int ( options [ ' jobs ' ] )
n_job_events = int ( options [ ' job_events ' ] )
prefix = options [ ' prefix ' ]
organizations = [ ]
users = [ ]
teams = [ ]
projects = [ ]
job_templates = [ ]
credentials = [ ]
inventories = [ ]
inventory_groups = [ ]
inventory_hosts = [ ]
2016-12-22 22:37:02 +03:00
wfjts = [ ]
nodes = [ ]
2016-12-26 03:42:56 +03:00
labels = [ ]
2016-04-18 23:56:29 +03:00
jobs = [ ]
#job_events = []
2016-11-16 04:59:39 +03:00
2016-04-18 23:56:29 +03:00
def spread ( n , m ) :
ret = [ ]
# At least one in each slot, split up the rest exponentially so the first
# buckets contain a lot of entries
for i in xrange ( m ) :
if n > 0 :
ret . append ( 1 )
n - = 1
else :
ret . append ( 0 )
for i in xrange ( m ) :
n_in_this_slot = n / / 2
n - = n_in_this_slot
ret [ i ] + = n_in_this_slot
if n > 0 and len ( ret ) :
ret [ 0 ] + = n
return ret
2016-11-16 00:32:27 +03:00
2016-04-18 23:56:29 +03:00
ids = defaultdict ( lambda : 0 )
2016-12-25 07:17:09 +03:00
bulk_data_description = ' From Tower bulk-data script '
2016-12-22 22:37:02 +03:00
# function to cycle through a list
def yield_choice ( alist ) :
ix = 0
while True :
2016-12-25 07:17:09 +03:00
yield alist [ ix ]
2016-12-22 22:37:02 +03:00
ix + = 1
if ix > = len ( alist ) :
ix = 0
2016-04-18 23:56:29 +03:00
class Rollback ( Exception ) :
pass
2016-12-26 02:45:58 +03:00
# Normally the modified_by field is populated by the crum library automatically,
# but since this is ran outside the request-response cycle that won't work.
# It is disaled here.
def mock_save ( self , * args , * * kwargs ) :
return super ( PrimordialModel , self ) . save ( * args , * * kwargs )
2017-01-03 17:09:43 +03:00
2017-01-30 17:17:46 +03:00
def mock_computed_fields ( self , * * kwargs ) :
pass
2016-12-26 02:45:58 +03:00
PrimordialModel . save = mock_save
2017-01-25 17:16:07 +03:00
startTime = datetime . now ( )
2016-04-18 23:56:29 +03:00
try :
2016-04-22 00:04:09 +03:00
with transaction . atomic ( ) :
2017-02-16 21:55:18 +03:00
with batch_role_ancestor_rebuilding ( ) , disable_computed_fields ( ) :
2016-12-22 20:03:56 +03:00
admin , created = User . objects . get_or_create ( username = ' admin ' , is_superuser = True )
if created :
admin . is_superuser = True
admin . save ( )
admin . set_password ( ' test ' )
admin . save ( )
org_admin , created = User . objects . get_or_create ( username = ' org_admin ' )
if created :
org_admin . set_password ( ' test ' )
org_admin . save ( )
org_member , created = User . objects . get_or_create ( username = ' org_member ' )
if created :
org_member . set_password ( ' test ' )
org_member . save ( )
prj_admin , created = User . objects . get_or_create ( username = ' prj_admin ' )
if created :
prj_admin . set_password ( ' test ' )
prj_admin . save ( )
jt_admin , created = User . objects . get_or_create ( username = ' jt_admin ' )
if created :
jt_admin . set_password ( ' test ' )
jt_admin . save ( )
inv_admin , created = User . objects . get_or_create ( username = ' inv_admin ' )
if created :
inv_admin . set_password ( ' test ' )
inv_admin . save ( )
2016-04-18 23:56:29 +03:00
print ( ' # Creating %d organizations ' % n_organizations )
for i in xrange ( n_organizations ) :
sys . stdout . write ( ' \r %d ' % ( i + 1 ) )
sys . stdout . flush ( )
2016-12-22 21:53:19 +03:00
org , _ = Organization . objects . get_or_create ( name = ' %s Organization %d ' % ( prefix , i ) )
2016-04-18 23:56:29 +03:00
organizations . append ( org )
if i == 0 :
org . admin_role . members . add ( org_admin )
org . member_role . members . add ( org_admin )
org . member_role . members . add ( org_member )
org . member_role . members . add ( prj_admin )
org . member_role . members . add ( jt_admin )
org . member_role . members . add ( inv_admin )
2016-12-26 03:42:56 +03:00
organization_gen = yield_choice ( organizations )
2016-04-18 23:56:29 +03:00
print ( ' ' )
print ( ' # Creating %d users ' % n_users )
org_idx = 0
for n in spread ( n_users , n_organizations ) :
for i in range ( n ) :
ids [ ' user ' ] + = 1
user_id = ids [ ' user ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , organizations [ org_idx ] . name , i + 1 ) )
sys . stdout . flush ( )
2016-12-22 21:53:19 +03:00
user , _ = User . objects . get_or_create ( username = ' %s user- %d ' % ( prefix , user_id ) )
2016-04-18 23:56:29 +03:00
organizations [ org_idx ] . member_role . members . add ( user )
users . append ( user )
org_idx + = 1
print ( ' ' )
2016-12-22 22:37:02 +03:00
creator_gen = yield_choice ( users )
for i in range ( 6 ) :
next ( creator_gen )
modifier_gen = yield_choice ( users )
2016-04-18 23:56:29 +03:00
print ( ' # Creating %d teams ' % n_teams )
org_idx = 0
for n in spread ( n_teams , n_organizations ) :
org = organizations [ org_idx ]
for i in range ( n ) :
ids [ ' team ' ] + = 1
team_id = ids [ ' team ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , org . name , i + 1 ) )
sys . stdout . flush ( )
2016-12-22 22:37:02 +03:00
team , _ = Team . objects . get_or_create (
name = ' %s Team %d Org %d ' % ( prefix , team_id , org_idx ) , organization = org ,
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) )
)
2016-04-18 23:56:29 +03:00
teams . append ( team )
org_idx + = 1
print ( ' ' )
print ( ' # Adding users to teams ' )
for org in organizations :
org_teams = [ t for t in org . teams . all ( ) ]
org_users = [ u for u in org . member_role . members . all ( ) ]
print ( ' Spreading %d users accross %d teams for %s ' % ( len ( org_users ) , len ( org_teams ) , org . name ) )
# Our normal spread for most users
cur_user_idx = 0
cur_team_idx = 0
for n in spread ( len ( org_users ) , len ( org_teams ) ) :
team = org_teams [ cur_team_idx ]
for i in range ( n ) :
if cur_user_idx < len ( org_users ) :
user = org_users [ cur_user_idx ]
team . member_role . members . add ( user )
cur_user_idx + = 1
cur_team_idx + = 1
# First user gets added to all teams
for team in org_teams :
team . member_role . members . add ( org_users [ 0 ] )
print ( ' # Creating %d credentials for users ' % ( n_credentials - n_credentials / / 2 ) )
user_idx = 0
for n in spread ( n_credentials - n_credentials / / 2 , n_users ) :
user = users [ user_idx ]
for i in range ( n ) :
ids [ ' credential ' ] + = 1
sys . stdout . write ( ' \r %d ' % ( ids [ ' credential ' ] ) )
sys . stdout . flush ( )
credential_id = ids [ ' credential ' ]
2016-12-26 02:45:58 +03:00
credential , _ = Credential . objects . get_or_create (
name = ' %s Credential %d User %d ' % ( prefix , credential_id , user_idx ) ,
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) ) ,
2017-03-30 21:47:48 +03:00
credential_type = CredentialType . from_v1_kind ( ' ssh ' )
2016-12-26 02:45:58 +03:00
)
2016-06-24 23:55:07 +03:00
credential . admin_role . members . add ( user )
2016-04-18 23:56:29 +03:00
credentials . append ( credential )
user_idx + = 1
print ( ' ' )
2016-12-25 07:17:09 +03:00
credential_gen = yield_choice ( credentials )
2016-04-18 23:56:29 +03:00
print ( ' # Creating %d credentials for teams ' % ( n_credentials / / 2 ) )
team_idx = 0
starting_credential_id = ids [ ' credential ' ]
for n in spread ( n_credentials - n_credentials / / 2 , n_teams ) :
team = teams [ team_idx ]
for i in range ( n ) :
ids [ ' credential ' ] + = 1
sys . stdout . write ( ' \r %d ' % ( ids [ ' credential ' ] - starting_credential_id ) )
sys . stdout . flush ( )
credential_id = ids [ ' credential ' ]
2016-12-26 02:45:58 +03:00
credential , _ = Credential . objects . get_or_create (
name = ' %s Credential %d team %d ' % ( prefix , credential_id , team_idx ) ,
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) ) ,
2017-03-30 21:47:48 +03:00
credential_type = CredentialType . from_v1_kind ( ' ssh ' )
2016-12-26 02:45:58 +03:00
)
2016-06-24 23:55:07 +03:00
credential . admin_role . parents . add ( team . member_role )
2016-04-18 23:56:29 +03:00
credentials . append ( credential )
team_idx + = 1
print ( ' ' )
print ( ' # Creating %d projects ' % n_projects )
org_idx = 0
for n in spread ( n_projects , n_organizations ) :
org = organizations [ org_idx ]
for i in range ( n ) :
ids [ ' project ' ] + = 1
project_id = ids [ ' project ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , org . name , i + 1 ) )
sys . stdout . flush ( )
2016-12-26 02:45:58 +03:00
project , _ = Project . objects . get_or_create (
name = ' %s Project %d Org %d ' % ( prefix , project_id , org_idx ) ,
organization = org ,
2017-01-03 17:09:43 +03:00
defaults = dict (
created_by = next ( creator_gen ) , modified_by = next ( modifier_gen ) ,
scm_url = ' https://github.com/jlaska/ansible-playbooks.git ' ,
scm_type = ' git ' ,
playbook_files = [
" check.yml " , " debug-50.yml " , " debug.yml " , " debug2.yml " ,
" debug_extra_vars.yml " , " dynamic_inventory.yml " ,
" environ_test.yml " , " fail_unless.yml " , " pass_unless.yml " ,
" pause.yml " , " ping-20.yml " , " ping.yml " ,
" setfact_50.yml " , " vault.yml "
] )
2016-12-26 02:45:58 +03:00
)
2016-04-18 23:56:29 +03:00
projects . append ( project )
if org_idx == 0 and i == 0 :
project . admin_role . members . add ( prj_admin )
org_idx + = 1
print ( ' ' )
print ( ' # Creating %d inventories ' % n_inventories )
org_idx = 0
for n in spread ( n_inventories , min ( n_inventories / / 4 + 1 , n_organizations ) ) :
org = organizations [ org_idx ]
for i in range ( n ) :
ids [ ' inventory ' ] + = 1
inventory_id = ids [ ' inventory ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , org . name , i + 1 ) )
sys . stdout . flush ( )
2016-12-26 02:45:58 +03:00
inventory , _ = Inventory . objects . get_or_create (
name = ' %s Inventory %d Org %d ' % ( prefix , inventory_id , org_idx ) ,
organization = org ,
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) ) ,
variables = ' { " ansible_connection " : " local " } '
)
2016-04-18 23:56:29 +03:00
inventories . append ( inventory )
if org_idx == 0 and i == 0 :
inventory . admin_role . members . add ( inv_admin )
org_idx + = 1
print ( ' ' )
print ( ' # Creating %d inventory_groups ' % n_inventory_groups )
inv_idx = 0
for n in spread ( n_inventory_groups , n_inventories ) :
inventory = inventories [ inv_idx ]
parent_list = [ None ] * 3
for i in range ( n ) :
ids [ ' group ' ] + = 1
group_id = ids [ ' group ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , inventory . name , i + 1 ) )
sys . stdout . flush ( )
2016-12-22 21:53:19 +03:00
group , _ = Group . objects . get_or_create (
2016-04-18 23:56:29 +03:00
name = ' %s Group %d Inventory %d ' % ( prefix , group_id , inv_idx ) ,
inventory = inventory ,
2016-12-26 02:45:58 +03:00
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) )
2016-04-18 23:56:29 +03:00
)
# Have each group have up to 3 parent groups
for parent_n in range ( 3 ) :
if i / / 4 + parent_n < len ( parent_list ) and parent_list [ i / / 4 + parent_n ] :
group . parents . add ( parent_list [ i / / 4 + parent_n ] )
if parent_list [ i / / 4 ] is None :
parent_list [ i / / 4 ] = group
else :
parent_list . append ( group )
inventory_groups . append ( group )
inv_idx + = 1
print ( ' ' )
print ( ' # Creating %d inventory_hosts ' % n_inventory_hosts )
group_idx = 0
for n in spread ( n_inventory_hosts , n_inventory_groups ) :
group = inventory_groups [ group_idx ]
for i in range ( n ) :
ids [ ' host ' ] + = 1
host_id = ids [ ' host ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , group . name , i + 1 ) )
sys . stdout . flush ( )
2016-12-26 02:45:58 +03:00
host , _ = Host . objects . get_or_create (
name = ' %s .host- %06d .group- %05d .dummy ' % ( prefix , host_id , group_idx ) ,
inventory = group . inventory ,
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) )
)
2016-04-18 23:56:29 +03:00
# Add the host to up to 3 groups
host . groups . add ( group )
for m in range ( 2 ) :
if group_idx + m < len ( inventory_groups ) and group . inventory . id == inventory_groups [ group_idx + m ] . inventory . id :
host . groups . add ( inventory_groups [ group_idx + m ] )
inventory_hosts . append ( host )
group_idx + = 1
print ( ' ' )
print ( ' # Creating %d job_templates ' % n_job_templates )
project_idx = 0
inv_idx = 0
for n in spread ( n_job_templates , n_projects ) :
project = projects [ project_idx ]
for i in range ( n ) :
ids [ ' job_template ' ] + = 1
job_template_id = ids [ ' job_template ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , project . name , i + 1 ) )
sys . stdout . flush ( )
inventory = None
org_inv_count = project . organization . inventories . count ( )
if org_inv_count > 0 :
inventory = project . organization . inventories . all ( ) [ inv_idx % org_inv_count ]
2016-12-26 02:45:58 +03:00
extra_kwargs = { }
if ids [ ' job_template ' ] % 5 == 0 :
extra_kwargs [ ' cloud_credential ' ] = next ( credential_gen )
if ids [ ' job_template ' ] % 7 == 0 :
extra_kwargs [ ' network_credential ' ] = next ( credential_gen )
2016-04-18 23:56:29 +03:00
2016-12-22 21:53:19 +03:00
job_template , _ = JobTemplate . objects . get_or_create (
2016-04-18 23:56:29 +03:00
name = ' %s Job Template %d Project %d ' % ( prefix , job_template_id , project_idx ) ,
2016-12-27 06:40:32 +03:00
defaults = dict (
2017-01-24 18:15:33 +03:00
inventory = inventory ,
project = project ,
credential = next ( credential_gen ) ,
2016-12-27 06:40:32 +03:00
created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) ,
2017-01-24 18:15:33 +03:00
playbook = " debug.yml " ,
* * extra_kwargs )
2016-04-18 23:56:29 +03:00
)
2016-12-27 06:40:32 +03:00
job_template . _is_new = _
2016-04-18 23:56:29 +03:00
job_templates . append ( job_template )
inv_idx + = 1
if project_idx == 0 and i == 0 :
job_template . admin_role . members . add ( jt_admin )
project_idx + = 1
2017-01-21 20:45:40 +03:00
if n > 0 :
print ( ' ' )
2016-04-18 23:56:29 +03:00
2016-12-22 22:37:02 +03:00
print ( ' # Creating %d Workflow Job Templates ' % n_wfjts )
org_idx = 0
for n in spread ( n_wfjts , n_organizations ) :
org = organizations [ org_idx ]
for i in range ( n ) :
ids [ ' wfjts ' ] + = 1
wfjt_id = ids [ ' wfjts ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , org . name , i + 1 ) )
sys . stdout . flush ( )
wfjt , _ = WorkflowJobTemplate . objects . get_or_create (
name = ' %s WFJT %d Org %d ' % ( prefix , wfjt_id , org_idx ) ,
description = bulk_data_description ,
2016-12-26 02:45:58 +03:00
organization = org ,
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) )
2016-12-22 22:37:02 +03:00
)
2016-12-26 03:42:56 +03:00
wfjt . _is_new = _
2016-12-22 22:37:02 +03:00
wfjts . append ( wfjt )
org_idx + = 1
2017-01-21 20:45:40 +03:00
if n :
print ( ' ' )
2016-12-22 22:37:02 +03:00
print ( ' # Creating %d Workflow Job Template nodes ' % n_nodes )
wfjt_idx = 0
for n in spread ( n_nodes , n_wfjts ) :
wfjt = wfjts [ wfjt_idx ]
2016-12-26 03:42:56 +03:00
if not wfjt . _is_new :
continue
2016-12-22 22:37:02 +03:00
jt_gen = yield_choice ( job_templates )
inv_gen = yield_choice ( inventories )
cred_gen = yield_choice ( credentials )
2016-12-25 07:17:09 +03:00
parent_idx = 0
wfjt_nodes = [ ]
2016-12-22 22:37:02 +03:00
for i in range ( n ) :
ids [ ' nodes ' ] + = 1
node_id = ids [ ' nodes ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , wfjt . name , i + 1 ) )
sys . stdout . flush ( )
kwargs = dict (
workflow_job_template = wfjt ,
2016-12-26 03:42:56 +03:00
unified_job_template = next ( jt_gen ) ,
modified = now ( )
2016-12-22 22:37:02 +03:00
)
if i % 2 == 0 :
# only apply inventories for every other node
kwargs [ ' inventory ' ] = next ( inv_gen )
if i % 3 == 0 :
# only apply prompted credential every 3rd node
kwargs [ ' credential ' ] = next ( cred_gen )
node , _ = WorkflowJobTemplateNode . objects . get_or_create (
2016-12-26 03:42:56 +03:00
* * kwargs
2016-12-22 22:37:02 +03:00
)
2016-12-25 07:17:09 +03:00
# nodes.append(node)
wfjt_nodes . append ( node )
if i < = 3 :
continue
parent_node = wfjt_nodes [ parent_idx ]
if parent_node . workflow_job_template != node . workflow_job_template :
raise Exception ( " Programming error, associating nodes in different workflows " )
elif parent_node == node :
raise Exception ( " error, self association " )
if parent_idx % 2 == 0 :
parent_node . always_nodes . add ( node )
else :
if ( i + 1 ) % 3 == 0 :
parent_node . failure_nodes . add ( node )
else :
parent_node . success_nodes . add ( node )
parent_idx = ( parent_idx + 7 ) % len ( wfjt_nodes )
2016-12-22 22:37:02 +03:00
wfjt_idx + = 1
2017-01-21 20:45:40 +03:00
if n :
print ( ' ' )
2016-12-22 22:37:02 +03:00
2016-12-26 03:42:56 +03:00
print ( ' # Creating %d Labels ' % n_labels )
org_idx = 0
for n in spread ( n_labels , n_organizations ) :
org = organizations [ org_idx ]
for i in range ( n ) :
ids [ ' labels ' ] + = 1
label_id = ids [ ' labels ' ]
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , org . name , i + 1 ) )
sys . stdout . flush ( )
label , _ = Label . objects . get_or_create (
2016-12-27 06:40:32 +03:00
name = ' %s L_ %d o %d ' % ( prefix , label_id , org_idx ) ,
2016-12-26 03:42:56 +03:00
organization = org ,
defaults = dict ( created_by = next ( creator_gen ) ,
modified_by = next ( modifier_gen ) )
)
labels . append ( label )
org_idx + = 1
2017-01-21 20:45:40 +03:00
if n :
print ( ' ' )
2016-12-26 03:42:56 +03:00
label_gen = yield_choice ( labels )
print ( ' # Adding labels to job templates ' )
jt_idx = 0
2017-01-03 17:09:43 +03:00
for n in spread ( n_labels * 7 , n_job_templates ) :
2017-01-30 17:17:46 +03:00
if n == 0 :
continue
2016-12-26 03:42:56 +03:00
jt = job_templates [ jt_idx ]
2016-12-27 06:40:32 +03:00
if not jt . _is_new :
continue
2016-12-26 03:42:56 +03:00
print ( ' Giving %d labels to %s JT ' % ( n , jt . name ) )
for i in range ( n ) :
jt . labels . add ( next ( label_gen ) )
jt_idx + = 1
2016-12-27 06:40:32 +03:00
print ( ' # Adding labels to workflow job templates ' )
wfjt_idx = 0
2017-01-03 17:09:43 +03:00
for n in spread ( n_labels * 3 , n_wfjts ) :
2016-12-27 06:40:32 +03:00
wfjt = wfjts [ wfjt_idx ]
if not jt . _is_new :
continue
print ( ' Giving %d labels to %s WFJT ' % ( n , wfjt . name ) )
for i in range ( n ) :
wfjt . labels . add ( next ( label_gen ) )
wfjt_idx + = 1
2016-12-20 17:55:08 +03:00
# Disable logging here, because it will mess up output format
logger = logging . getLogger ( ' awx.main ' )
logger . propagate = False
2016-04-18 23:56:29 +03:00
print ( ' # Creating %d jobs ' % n_jobs )
group_idx = 0
job_template_idx = 0
2016-12-20 17:55:08 +03:00
job_i = 0
2016-04-18 23:56:29 +03:00
for n in spread ( n_jobs , n_job_templates ) :
job_template = job_templates [ job_template_idx ]
for i in range ( n ) :
sys . stdout . write ( ' \r Assigning %d to %s : %d ' % ( n , job_template . name , i + 1 ) )
sys . stdout . flush ( )
2016-12-26 03:42:56 +03:00
if len ( jobs ) % 4 == 0 :
2016-12-25 07:17:09 +03:00
job_stat = ' failed '
2016-12-26 03:42:56 +03:00
elif len ( jobs ) % 11 == 0 :
2016-12-25 07:17:09 +03:00
job_stat = ' canceled '
2016-12-20 17:55:08 +03:00
else :
job_stat = ' successful '
2017-01-18 21:58:15 +03:00
job , _ = Job . objects . get_or_create (
2016-12-27 06:40:32 +03:00
job_template = job_template ,
2016-12-20 17:55:08 +03:00
status = job_stat , name = " %s - %d " % ( job_template . name , job_i ) ,
2016-12-27 06:40:32 +03:00
project = job_template . project , inventory = job_template . inventory ,
credential = job_template . credential ,
cloud_credential = job_template . cloud_credential ,
network_credential = job_template . network_credential
)
2017-01-18 21:58:15 +03:00
job . _is_new = _
2016-04-18 23:56:29 +03:00
jobs . append ( job )
2016-12-20 17:55:08 +03:00
job_i + = 1
2017-01-18 21:58:15 +03:00
if not job . _is_new :
2016-12-20 17:55:08 +03:00
group_idx + = 1
2017-01-18 21:58:15 +03:00
continue
2017-01-21 20:45:40 +03:00
if i + 1 == n :
2016-12-27 06:40:32 +03:00
job_template . last_job = job
if job_template . pk % 5 == 0 :
job_template . current_job = job
job_template . save ( )
2016-04-18 23:56:29 +03:00
2016-12-20 17:55:08 +03:00
if job . _is_new :
with transaction . atomic ( ) :
if job_template . inventory :
inv_groups = [ g for g in job_template . inventory . groups . all ( ) ]
if len ( inv_groups ) :
JobHostSummary . objects . bulk_create ( [
JobHostSummary (
job = job , host = h , host_name = h . name , processed = 1 ,
created = now ( ) , modified = now ( )
)
for h in inv_groups [ group_idx % len ( inv_groups ) ] . hosts . all ( ) [ : 100 ]
] )
2016-04-18 23:56:29 +03:00
group_idx + = 1
job_template_idx + = 1
if n :
print ( ' ' )
print ( ' # Creating %d job events ' % n_job_events )
job_idx = 0
for n in spread ( n_job_events , n_jobs ) :
job = jobs [ job_idx ]
2017-01-23 01:09:29 +03:00
# Check if job already has events, for idempotence
2016-12-27 06:40:32 +03:00
if not job . _is_new :
continue
2017-01-30 17:17:46 +03:00
# Bulk create in chunks with maximum chunk size
MAX_BULK_CREATE = 100
for j in range ( ( n / MAX_BULK_CREATE ) + 1 ) :
n_subgroup = MAX_BULK_CREATE
if j == n / MAX_BULK_CREATE :
# on final pass, create the remainder
n_subgroup = n % MAX_BULK_CREATE
2017-02-21 20:08:03 +03:00
sys . stdout . write ( ' \r Creating %d job events for job %d , subgroup: %d ' % ( n , job . id , j + 1 ) )
2017-01-30 17:17:46 +03:00
sys . stdout . flush ( )
JobEvent . objects . bulk_create ( [
JobEvent (
created = now ( ) ,
modified = now ( ) ,
job = job ,
event = ' runner_on_ok '
)
for i in range ( n_subgroup )
] )
2016-04-18 23:56:29 +03:00
job_idx + = 1
if n :
print ( ' ' )
if options [ ' pretend ' ] :
raise Rollback ( )
except Rollback :
print ( ' Rolled back changes ' )
pass
2017-01-21 20:45:40 +03:00
print ( ' ' )
2017-01-25 17:16:07 +03:00
print ( ' script execution time: {} ' . format ( datetime . now ( ) - startTime ) )