mirror of
https://github.com/ansible/awx.git
synced 2024-10-31 23:51:09 +03:00
437d9843d1
- Fixed issue #5528 Signed-off-by: Hideki Saito <saito@fgrep.org>
727 lines
25 KiB
Makefile
727 lines
25 KiB
Makefile
PYTHON ?= python3
|
|
PYTHON_VERSION = $(shell $(PYTHON) -c "from distutils.sysconfig import get_python_version; print(get_python_version())")
|
|
SITELIB=$(shell $(PYTHON) -c "from distutils.sysconfig import get_python_lib; print(get_python_lib())")
|
|
OFFICIAL ?= no
|
|
PACKER ?= packer
|
|
PACKER_BUILD_OPTS ?= -var 'official=$(OFFICIAL)' -var 'aw_repo_url=$(AW_REPO_URL)'
|
|
NODE ?= node
|
|
NPM_BIN ?= npm
|
|
DEPS_SCRIPT ?= packaging/bundle/deps.py
|
|
GIT_BRANCH ?= $(shell git rev-parse --abbrev-ref HEAD)
|
|
MANAGEMENT_COMMAND ?= awx-manage
|
|
IMAGE_REPOSITORY_AUTH ?=
|
|
IMAGE_REPOSITORY_BASE ?= https://gcr.io
|
|
VERSION := $(shell cat VERSION)
|
|
|
|
# NOTE: This defaults the container image version to the branch that's active
|
|
COMPOSE_TAG ?= $(GIT_BRANCH)
|
|
COMPOSE_HOST ?= $(shell hostname)
|
|
|
|
VENV_BASE ?= /venv
|
|
COLLECTION_VENV ?= /awx_devel/awx_collection_test_venv
|
|
SCL_PREFIX ?=
|
|
CELERY_SCHEDULE_FILE ?= /var/lib/awx/beat.db
|
|
|
|
DEV_DOCKER_TAG_BASE ?= gcr.io/ansible-tower-engineering
|
|
# Python packages to install only from source (not from binary wheels)
|
|
# Comma separated list
|
|
SRC_ONLY_PKGS ?= cffi,pycparser,psycopg2,twilio
|
|
# These should be upgraded in the AWX and Ansible venv before attempting
|
|
# to install the actual requirements
|
|
VENV_BOOTSTRAP ?= pip==19.3.1 setuptools==41.6.0
|
|
|
|
# Determine appropriate shasum command
|
|
UNAME_S := $(shell uname -s)
|
|
ifeq ($(UNAME_S),Linux)
|
|
SHASUM_BIN ?= sha256sum
|
|
endif
|
|
ifeq ($(UNAME_S),Darwin)
|
|
SHASUM_BIN ?= shasum -a 256
|
|
endif
|
|
|
|
# Get the branch information from git
|
|
GIT_DATE := $(shell git log -n 1 --format="%ai")
|
|
DATE := $(shell date -u +%Y%m%d%H%M)
|
|
|
|
NAME ?= awx
|
|
GIT_REMOTE_URL = $(shell git config --get remote.origin.url)
|
|
|
|
# TAR build parameters
|
|
SDIST_TAR_NAME=$(NAME)-$(VERSION)
|
|
WHEEL_NAME=$(NAME)-$(VERSION)
|
|
|
|
SDIST_COMMAND ?= sdist
|
|
WHEEL_COMMAND ?= bdist_wheel
|
|
SDIST_TAR_FILE ?= $(SDIST_TAR_NAME).tar.gz
|
|
WHEEL_FILE ?= $(WHEEL_NAME)-py2-none-any.whl
|
|
|
|
# UI flag files
|
|
UI_DEPS_FLAG_FILE = awx/ui/.deps_built
|
|
UI_RELEASE_DEPS_FLAG_FILE = awx/ui/.release_deps_built
|
|
UI_RELEASE_FLAG_FILE = awx/ui/.release_built
|
|
|
|
I18N_FLAG_FILE = .i18n_built
|
|
|
|
.PHONY: awx-link clean clean-tmp clean-venv requirements requirements_dev \
|
|
develop refresh adduser migrate dbchange runserver \
|
|
receiver test test_unit test_coverage coverage_html \
|
|
dev_build release_build release_clean sdist \
|
|
ui-docker-machine ui-docker ui-release ui-devel \
|
|
ui-test ui-deps ui-test-ci VERSION
|
|
|
|
# remove ui build artifacts
|
|
clean-ui: clean-languages
|
|
rm -rf awx/ui/static/
|
|
rm -rf awx/ui/node_modules/
|
|
rm -rf awx/ui/test/unit/reports/
|
|
rm -rf awx/ui/test/spec/reports/
|
|
rm -rf awx/ui/test/e2e/reports/
|
|
rm -rf awx/ui/client/languages/
|
|
rm -rf awx/ui_next/node_modules/
|
|
rm -rf awx/ui_next/coverage/
|
|
rm -rf awx/ui_next/build/locales/_build/
|
|
rm -f $(UI_DEPS_FLAG_FILE)
|
|
rm -f $(UI_RELEASE_DEPS_FLAG_FILE)
|
|
rm -f $(UI_RELEASE_FLAG_FILE)
|
|
|
|
clean-tmp:
|
|
rm -rf tmp/
|
|
|
|
clean-venv:
|
|
rm -rf venv/
|
|
|
|
clean-dist:
|
|
rm -rf dist
|
|
|
|
clean-schema:
|
|
rm -rf swagger.json
|
|
rm -rf schema.json
|
|
rm -rf reference-schema.json
|
|
|
|
clean-languages:
|
|
rm -f $(I18N_FLAG_FILE)
|
|
find . -type f -regex ".*\.mo$$" -delete
|
|
|
|
# Remove temporary build files, compiled Python files.
|
|
clean: clean-ui clean-api clean-awxkit clean-dist
|
|
rm -rf awx/public
|
|
rm -rf awx/lib/site-packages
|
|
rm -rf awx/job_status
|
|
rm -rf awx/job_output
|
|
rm -rf reports
|
|
rm -rf tmp
|
|
rm -rf $(I18N_FLAG_FILE)
|
|
mkdir tmp
|
|
|
|
clean-api:
|
|
rm -rf build $(NAME)-$(VERSION) *.egg-info
|
|
find . -type f -regex ".*\.py[co]$$" -delete
|
|
find . -type d -name "__pycache__" -delete
|
|
rm -f awx/awx_test.sqlite3*
|
|
rm -rf requirements/vendor
|
|
rm -rf awx/projects
|
|
|
|
clean-awxkit:
|
|
rm -rf awxkit/*.egg-info awxkit/.tox awxkit/build/*
|
|
|
|
# convenience target to assert environment variables are defined
|
|
guard-%:
|
|
@if [ "$${$*}" = "" ]; then \
|
|
echo "The required environment variable '$*' is not set"; \
|
|
exit 1; \
|
|
fi
|
|
|
|
virtualenv: virtualenv_ansible virtualenv_awx
|
|
|
|
# virtualenv_* targets do not use --system-site-packages to prevent bugs installing packages
|
|
# but Ansible venvs are expected to have this, so that must be done after venv creation
|
|
virtualenv_ansible:
|
|
if [ "$(VENV_BASE)" ]; then \
|
|
if [ ! -d "$(VENV_BASE)" ]; then \
|
|
mkdir $(VENV_BASE); \
|
|
fi; \
|
|
if [ ! -d "$(VENV_BASE)/ansible" ]; then \
|
|
virtualenv -p python $(VENV_BASE)/ansible && \
|
|
$(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) $(VENV_BOOTSTRAP); \
|
|
fi; \
|
|
fi
|
|
|
|
virtualenv_ansible_py3:
|
|
if [ "$(VENV_BASE)" ]; then \
|
|
if [ ! -d "$(VENV_BASE)" ]; then \
|
|
mkdir $(VENV_BASE); \
|
|
fi; \
|
|
if [ ! -d "$(VENV_BASE)/ansible" ]; then \
|
|
virtualenv -p $(PYTHON) $(VENV_BASE)/ansible; \
|
|
$(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) $(VENV_BOOTSTRAP); \
|
|
fi; \
|
|
fi
|
|
|
|
# flit is needed for offline install of certain packages, specifically ptyprocess
|
|
# it is needed for setup, but not always recognized as a setup dependency
|
|
# similar to pip, setuptools, and wheel, these are all needed here as a bootstrapping issues
|
|
virtualenv_awx:
|
|
if [ "$(VENV_BASE)" ]; then \
|
|
if [ ! -d "$(VENV_BASE)" ]; then \
|
|
mkdir $(VENV_BASE); \
|
|
fi; \
|
|
if [ ! -d "$(VENV_BASE)/awx" ]; then \
|
|
virtualenv -p $(PYTHON) $(VENV_BASE)/awx; \
|
|
$(VENV_BASE)/awx/bin/pip install $(PIP_OPTIONS) $(VENV_BOOTSTRAP) && \
|
|
$(VENV_BASE)/awx/bin/pip install $(PIP_OPTIONS) flit; \
|
|
fi; \
|
|
fi
|
|
|
|
# --ignore-install flag is not used because *.txt files should specify exact versions
|
|
requirements_ansible: virtualenv_ansible
|
|
if [[ "$(PIP_OPTIONS)" == *"--no-index"* ]]; then \
|
|
cat requirements/requirements_ansible.txt requirements/requirements_ansible_local.txt | $(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) -r /dev/stdin ; \
|
|
else \
|
|
cat requirements/requirements_ansible.txt requirements/requirements_ansible_git.txt | $(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) --no-binary $(SRC_ONLY_PKGS) -r /dev/stdin ; \
|
|
fi
|
|
$(VENV_BASE)/ansible/bin/pip uninstall --yes -r requirements/requirements_ansible_uninstall.txt
|
|
# Same effect as using --system-site-packages flag on venv creation
|
|
rm $(shell ls -d $(VENV_BASE)/ansible/lib/python* | head -n 1)/no-global-site-packages.txt
|
|
|
|
requirements_ansible_py3: virtualenv_ansible_py3
|
|
if [[ "$(PIP_OPTIONS)" == *"--no-index"* ]]; then \
|
|
cat requirements/requirements_ansible.txt requirements/requirements_ansible_local.txt | $(VENV_BASE)/ansible/bin/pip3 install $(PIP_OPTIONS) -r /dev/stdin ; \
|
|
else \
|
|
cat requirements/requirements_ansible.txt requirements/requirements_ansible_git.txt | $(VENV_BASE)/ansible/bin/pip3 install $(PIP_OPTIONS) --no-binary $(SRC_ONLY_PKGS) -r /dev/stdin ; \
|
|
fi
|
|
$(VENV_BASE)/ansible/bin/pip3 uninstall --yes -r requirements/requirements_ansible_uninstall.txt
|
|
# Same effect as using --system-site-packages flag on venv creation
|
|
rm $(shell ls -d $(VENV_BASE)/ansible/lib/python* | head -n 1)/no-global-site-packages.txt
|
|
|
|
requirements_ansible_dev:
|
|
if [ "$(VENV_BASE)" ]; then \
|
|
$(VENV_BASE)/ansible/bin/pip install pytest mock; \
|
|
fi
|
|
|
|
# Install third-party requirements needed for AWX's environment.
|
|
# this does not use system site packages intentionally
|
|
requirements_awx: virtualenv_awx
|
|
if [[ "$(PIP_OPTIONS)" == *"--no-index"* ]]; then \
|
|
cat requirements/requirements.txt requirements/requirements_local.txt | $(VENV_BASE)/awx/bin/pip install $(PIP_OPTIONS) -r /dev/stdin ; \
|
|
else \
|
|
cat requirements/requirements.txt requirements/requirements_git.txt | $(VENV_BASE)/awx/bin/pip install $(PIP_OPTIONS) --no-binary $(SRC_ONLY_PKGS) -r /dev/stdin ; \
|
|
fi
|
|
$(VENV_BASE)/awx/bin/pip uninstall --yes -r requirements/requirements_tower_uninstall.txt
|
|
|
|
requirements_awx_dev:
|
|
$(VENV_BASE)/awx/bin/pip install -r requirements/requirements_dev.txt
|
|
|
|
requirements: requirements_ansible requirements_awx
|
|
|
|
requirements_dev: requirements_awx requirements_ansible_py3 requirements_awx_dev requirements_ansible_dev
|
|
|
|
requirements_test: requirements
|
|
|
|
# "Install" awx package in development mode.
|
|
develop:
|
|
@if [ "$(VIRTUAL_ENV)" ]; then \
|
|
pip uninstall -y awx; \
|
|
$(PYTHON) setup.py develop; \
|
|
else \
|
|
pip uninstall -y awx; \
|
|
$(PYTHON) setup.py develop; \
|
|
fi
|
|
|
|
version_file:
|
|
mkdir -p /var/lib/awx/; \
|
|
if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
python -c "import awx; print(awx.__version__)" > /var/lib/awx/.awx_version; \
|
|
|
|
# Do any one-time init tasks.
|
|
comma := ,
|
|
init:
|
|
if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(MANAGEMENT_COMMAND) provision_instance --hostname=$(COMPOSE_HOST); \
|
|
$(MANAGEMENT_COMMAND) register_queue --queuename=tower --instance_percent=100;\
|
|
if [ "$(AWX_GROUP_QUEUES)" == "tower,thepentagon" ]; then \
|
|
$(MANAGEMENT_COMMAND) provision_instance --hostname=isolated; \
|
|
$(MANAGEMENT_COMMAND) register_queue --queuename='thepentagon' --hostnames=isolated --controller=tower; \
|
|
$(MANAGEMENT_COMMAND) generate_isolated_key > /awx_devel/awx/main/isolated/authorized_keys; \
|
|
fi;
|
|
|
|
# Refresh development environment after pulling new code.
|
|
refresh: clean requirements_dev version_file develop migrate
|
|
|
|
# Create Django superuser.
|
|
adduser:
|
|
$(MANAGEMENT_COMMAND) createsuperuser
|
|
|
|
# Create database tables and apply any new migrations.
|
|
migrate:
|
|
if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(MANAGEMENT_COMMAND) migrate --noinput
|
|
|
|
# Run after making changes to the models to create a new migration.
|
|
dbchange:
|
|
$(MANAGEMENT_COMMAND) makemigrations
|
|
|
|
server_noattach:
|
|
tmux new-session -d -s awx 'exec make uwsgi'
|
|
tmux rename-window 'AWX'
|
|
tmux select-window -t awx:0
|
|
tmux split-window -v 'exec make dispatcher'
|
|
tmux new-window 'exec make daphne'
|
|
tmux select-window -t awx:1
|
|
tmux rename-window 'WebSockets'
|
|
tmux split-window -h 'exec make runworker'
|
|
tmux split-window -v 'exec make nginx'
|
|
tmux new-window 'exec make receiver'
|
|
tmux select-window -t awx:2
|
|
tmux rename-window 'Extra Services'
|
|
tmux select-window -t awx:0
|
|
|
|
server: server_noattach
|
|
tmux -2 attach-session -t awx
|
|
|
|
# Use with iterm2's native tmux protocol support
|
|
servercc: server_noattach
|
|
tmux -2 -CC attach-session -t awx
|
|
|
|
supervisor:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
supervisord --pidfile=/tmp/supervisor_pid -n
|
|
|
|
collectstatic:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
mkdir -p awx/public/static && $(PYTHON) manage.py collectstatic --clear --noinput > /dev/null 2>&1
|
|
|
|
uwsgi: collectstatic
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
uwsgi -b 32768 --socket 127.0.0.1:8050 --module=awx.wsgi:application --home=/venv/awx --chdir=/awx_devel/ --vacuum --processes=5 --harakiri=120 --master --no-orphans --py-autoreload 1 --max-requests=1000 --stats /tmp/stats.socket --lazy-apps --logformat "%(addr) %(method) %(uri) - %(proto) %(status)" --hook-accepting1="exec:supervisorctl restart tower-processes:awx-dispatcher tower-processes:awx-receiver"
|
|
|
|
daphne:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
daphne -b 127.0.0.1 -p 8051 awx.asgi:channel_layer
|
|
|
|
runworker:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(PYTHON) manage.py runworker --only-channels websocket.*
|
|
|
|
# Run the built-in development webserver (by default on http://localhost:8013).
|
|
runserver:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(PYTHON) manage.py runserver
|
|
|
|
# Run to start the background task dispatcher for development.
|
|
dispatcher:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(PYTHON) manage.py run_dispatcher
|
|
|
|
|
|
# Run to start the zeromq callback receiver
|
|
receiver:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(PYTHON) manage.py run_callback_receiver
|
|
|
|
nginx:
|
|
nginx -g "daemon off;"
|
|
|
|
jupyter:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(MANAGEMENT_COMMAND) shell_plus --notebook
|
|
|
|
reports:
|
|
mkdir -p $@
|
|
|
|
pep8: reports
|
|
@(set -o pipefail && $@ | tee reports/$@.report)
|
|
|
|
flake8: reports
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
(set -o pipefail && $@ | tee reports/$@.report)
|
|
|
|
pyflakes: reports
|
|
@(set -o pipefail && $@ | tee reports/$@.report)
|
|
|
|
pylint: reports
|
|
@(set -o pipefail && $@ | reports/$@.report)
|
|
|
|
genschema: reports
|
|
$(MAKE) swagger PYTEST_ARGS="--genschema --create-db "
|
|
mv swagger.json schema.json
|
|
|
|
swagger: reports
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
(set -o pipefail && py.test $(PYTEST_ARGS) awx/conf/tests/functional awx/main/tests/functional/api awx/main/tests/docs --release=$(VERSION_TARGET) | tee reports/$@.report)
|
|
|
|
check: flake8 pep8 # pyflakes pylint
|
|
|
|
awx-link:
|
|
cp -R /tmp/awx.egg-info /awx_devel/ || true
|
|
sed -i "s/placeholder/$(shell cat VERSION)/" /awx_devel/awx.egg-info/PKG-INFO
|
|
cp -f /tmp/awx.egg-link /venv/awx/lib/python$(PYTHON_VERSION)/site-packages/awx.egg-link
|
|
|
|
TEST_DIRS ?= awx/main/tests/unit awx/main/tests/functional awx/conf/tests awx/sso/tests
|
|
|
|
# Run all API unit tests.
|
|
test:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
PYTHONDONTWRITEBYTECODE=1 py.test -p no:cacheprovider -n auto $(TEST_DIRS)
|
|
cd awxkit && $(VENV_BASE)/awx/bin/tox -re py2,py3
|
|
awx-manage check_migrations --dry-run --check -n 'vNNN_missing_migration_file'
|
|
|
|
prepare_collection_venv:
|
|
rm -rf $(COLLECTION_VENV)
|
|
mkdir $(COLLECTION_VENV)
|
|
$(VENV_BASE)/awx/bin/pip install --target=$(COLLECTION_VENV) git+https://github.com/ansible/tower-cli.git
|
|
|
|
COLLECTION_TEST_DIRS ?= awx_collection/test/awx
|
|
COLLECTION_PACKAGE ?= awx
|
|
COLLECTION_NAMESPACE ?= awx
|
|
|
|
test_collection:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
PYTHONPATH=$(COLLECTION_VENV):/awx_devel/awx_collection:$PYTHONPATH:/usr/lib/python3.6/site-packages py.test $(COLLECTION_TEST_DIRS)
|
|
|
|
flake8_collection:
|
|
flake8 awx_collection/ # Different settings, in main exclude list
|
|
|
|
test_collection_all: prepare_collection_venv test_collection flake8_collection
|
|
|
|
test_collection_sanity:
|
|
rm -rf sanity
|
|
mkdir -p sanity/ansible_collections/awx
|
|
cp -Ra awx_collection sanity/ansible_collections/awx/awx # symlinks do not work
|
|
cd sanity/ansible_collections/awx/awx && git init && git add . # requires both this file structure and a git repo, so there you go
|
|
cd sanity/ansible_collections/awx/awx && ansible-test sanity
|
|
|
|
build_collection:
|
|
ansible-playbook -i localhost, awx_collection/template_galaxy.yml -e collection_package=$(COLLECTION_PACKAGE) -e collection_namespace=$(COLLECTION_NAMESPACE) -e collection_version=$(VERSION)
|
|
ansible-galaxy collection build awx_collection --force --output-path=awx_collection
|
|
|
|
install_collection: build_collection
|
|
rm -rf ~/.ansible/collections/ansible_collections/awx/awx
|
|
ansible-galaxy collection install awx_collection/awx-awx-$(VERSION).tar.gz
|
|
|
|
test_unit:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
py.test awx/main/tests/unit awx/conf/tests/unit awx/sso/tests/unit
|
|
|
|
# Run all API unit tests with coverage enabled.
|
|
test_coverage:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
py.test --create-db --cov=awx --cov-report=xml --junitxml=./reports/junit.xml $(TEST_DIRS)
|
|
|
|
# Output test coverage as HTML (into htmlcov directory).
|
|
coverage_html:
|
|
coverage html
|
|
|
|
# Run API unit tests across multiple Python/Django versions with Tox.
|
|
test_tox:
|
|
tox -v
|
|
|
|
# Make fake data
|
|
DATA_GEN_PRESET = ""
|
|
bulk_data:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(PYTHON) tools/data_generators/rbac_dummy_data_generator.py --preset=$(DATA_GEN_PRESET)
|
|
|
|
# l10n TASKS
|
|
# --------------------------------------
|
|
|
|
# check for UI po files
|
|
HAVE_PO := $(shell ls awx/ui/po/*.po 2>/dev/null)
|
|
check-po:
|
|
ifdef HAVE_PO
|
|
# Should be 'Language: zh-CN' but not 'Language: zh_CN' in zh_CN.po
|
|
for po in awx/ui/po/*.po ; do \
|
|
echo $$po; \
|
|
mo="awx/ui/po/`basename $$po .po`.mo"; \
|
|
msgfmt --check --verbose $$po -o $$mo; \
|
|
if test "$$?" -ne 0 ; then \
|
|
exit -1; \
|
|
fi; \
|
|
rm $$mo; \
|
|
name=`echo "$$po" | grep '-'`; \
|
|
if test "x$$name" != x ; then \
|
|
right_name=`echo $$language | sed -e 's/-/_/'`; \
|
|
echo "ERROR: WRONG $$name CORRECTION: $$right_name"; \
|
|
exit -1; \
|
|
fi; \
|
|
language=`grep '^"Language:' "$$po" | grep '_'`; \
|
|
if test "x$$language" != x ; then \
|
|
right_language=`echo $$language | sed -e 's/_/-/'`; \
|
|
echo "ERROR: WRONG $$language CORRECTION: $$right_language in $$po"; \
|
|
exit -1; \
|
|
fi; \
|
|
done;
|
|
else
|
|
@echo No PO files
|
|
endif
|
|
|
|
# generate UI .pot
|
|
pot: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run pot
|
|
|
|
# generate django .pot .po
|
|
LANG = "en-us"
|
|
messages:
|
|
@if [ "$(VENV_BASE)" ]; then \
|
|
. $(VENV_BASE)/awx/bin/activate; \
|
|
fi; \
|
|
$(PYTHON) manage.py makemessages -l $(LANG) --keep-pot
|
|
|
|
# generate l10n .json .mo
|
|
languages: $(I18N_FLAG_FILE)
|
|
|
|
$(I18N_FLAG_FILE): $(UI_RELEASE_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run languages
|
|
$(PYTHON) tools/scripts/compilemessages.py
|
|
touch $(I18N_FLAG_FILE)
|
|
|
|
# End l10n TASKS
|
|
# --------------------------------------
|
|
|
|
# UI RELEASE TASKS
|
|
# --------------------------------------
|
|
ui-release: $(UI_RELEASE_FLAG_FILE)
|
|
|
|
$(UI_RELEASE_FLAG_FILE): $(I18N_FLAG_FILE) $(UI_RELEASE_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run build-release
|
|
touch $(UI_RELEASE_FLAG_FILE)
|
|
|
|
$(UI_RELEASE_DEPS_FLAG_FILE):
|
|
PUPPETEER_SKIP_CHROMIUM_DOWNLOAD=1 $(NPM_BIN) --unsafe-perm --prefix awx/ui ci --no-save awx/ui
|
|
touch $(UI_RELEASE_DEPS_FLAG_FILE)
|
|
|
|
# END UI RELEASE TASKS
|
|
# --------------------------------------
|
|
|
|
# UI TASKS
|
|
# --------------------------------------
|
|
ui-deps: $(UI_DEPS_FLAG_FILE)
|
|
|
|
$(UI_DEPS_FLAG_FILE):
|
|
@if [ -f ${UI_RELEASE_DEPS_FLAG_FILE} ]; then \
|
|
rm -rf awx/ui/node_modules; \
|
|
rm -f ${UI_RELEASE_DEPS_FLAG_FILE}; \
|
|
fi; \
|
|
$(NPM_BIN) --unsafe-perm --prefix awx/ui ci --no-save awx/ui
|
|
touch $(UI_DEPS_FLAG_FILE)
|
|
|
|
ui-docker-machine: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run ui-docker-machine -- $(MAKEFLAGS)
|
|
|
|
# Native docker. Builds UI and raises BrowserSync & filesystem polling.
|
|
ui-docker: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run ui-docker -- $(MAKEFLAGS)
|
|
|
|
# Builds UI with development UI without raising browser-sync or filesystem polling.
|
|
ui-devel: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run build-devel -- $(MAKEFLAGS)
|
|
|
|
ui-test: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run test
|
|
|
|
ui-lint: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) run --prefix awx/ui jshint
|
|
$(NPM_BIN) run --prefix awx/ui lint
|
|
|
|
# A standard go-to target for API developers to use building the frontend
|
|
ui: clean-ui ui-devel
|
|
|
|
ui-test-ci: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) --prefix awx/ui run test:ci
|
|
$(NPM_BIN) --prefix awx/ui run unit
|
|
|
|
jshint: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) run --prefix awx/ui jshint
|
|
$(NPM_BIN) run --prefix awx/ui lint
|
|
|
|
ui-zuul-lint-and-test: $(UI_DEPS_FLAG_FILE)
|
|
$(NPM_BIN) run --prefix awx/ui jshint
|
|
$(NPM_BIN) run --prefix awx/ui lint
|
|
$(NPM_BIN) --prefix awx/ui run test:ci
|
|
$(NPM_BIN) --prefix awx/ui run unit
|
|
|
|
# END UI TASKS
|
|
# --------------------------------------
|
|
|
|
# UI NEXT TASKS
|
|
# --------------------------------------
|
|
|
|
ui-next-lint:
|
|
$(NPM_BIN) --prefix awx/ui_next install
|
|
$(NPM_BIN) run --prefix awx/ui_next lint
|
|
$(NPM_BIN) run --prefix awx/ui_next prettier-check
|
|
|
|
ui-next-test:
|
|
$(NPM_BIN) --prefix awx/ui_next install
|
|
$(NPM_BIN) run --prefix awx/ui_next test
|
|
|
|
ui-next-zuul-lint-and-test:
|
|
$(NPM_BIN) --prefix awx/ui_next install
|
|
$(NPM_BIN) run --prefix awx/ui_next lint
|
|
$(NPM_BIN) run --prefix awx/ui_next prettier-check
|
|
$(NPM_BIN) run --prefix awx/ui_next test
|
|
|
|
# END UI NEXT TASKS
|
|
# --------------------------------------
|
|
|
|
# Build a pip-installable package into dist/ with a timestamped version number.
|
|
dev_build:
|
|
$(PYTHON) setup.py dev_build
|
|
|
|
# Build a pip-installable package into dist/ with the release version number.
|
|
release_build:
|
|
$(PYTHON) setup.py release_build
|
|
|
|
dist/$(SDIST_TAR_FILE): ui-release VERSION
|
|
$(PYTHON) setup.py $(SDIST_COMMAND)
|
|
|
|
dist/$(WHEEL_FILE): ui-release
|
|
$(PYTHON) setup.py $(WHEEL_COMMAND)
|
|
|
|
sdist: dist/$(SDIST_TAR_FILE)
|
|
@echo "#############################################"
|
|
@echo "Artifacts:"
|
|
@echo dist/$(SDIST_TAR_FILE)
|
|
@echo "#############################################"
|
|
|
|
wheel: dist/$(WHEEL_FILE)
|
|
@echo "#############################################"
|
|
@echo "Artifacts:"
|
|
@echo dist/$(WHEEL_FILE)
|
|
@echo "#############################################"
|
|
|
|
# Build setup bundle tarball
|
|
setup-bundle-build:
|
|
mkdir -p $@
|
|
|
|
docker-auth:
|
|
@if [ "$(IMAGE_REPOSITORY_AUTH)" ]; then \
|
|
echo "$(IMAGE_REPOSITORY_AUTH)" | docker login -u oauth2accesstoken --password-stdin $(IMAGE_REPOSITORY_BASE); \
|
|
fi;
|
|
|
|
# This directory is bind-mounted inside of the development container and
|
|
# needs to be pre-created for permissions to be set correctly. Otherwise,
|
|
# Docker will create this directory as root.
|
|
awx/projects:
|
|
@mkdir -p $@
|
|
|
|
# Docker isolated rampart
|
|
docker-compose-isolated: awx/projects
|
|
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml -f tools/docker-isolated-override.yml up
|
|
|
|
# Docker Compose Development environment
|
|
docker-compose: docker-auth awx/projects
|
|
CURRENT_UID=$(shell id -u) OS="$(shell docker info | grep 'Operating System')" TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml up --no-recreate awx
|
|
|
|
docker-compose-cluster: docker-auth awx/projects
|
|
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose-cluster.yml up
|
|
|
|
docker-compose-credential-plugins: docker-auth awx/projects
|
|
echo -e "\033[0;31mTo generate a CyberArk Conjur API key: docker exec -it tools_conjur_1 conjurctl account create quick-start\033[0m"
|
|
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml -f tools/docker-credential-plugins-override.yml up --no-recreate awx
|
|
|
|
docker-compose-test: docker-auth awx/projects
|
|
cd tools && CURRENT_UID=$(shell id -u) OS="$(shell docker info | grep 'Operating System')" TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /bin/bash
|
|
|
|
docker-compose-runtest: awx/projects
|
|
cd tools && CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /start_tests.sh
|
|
|
|
docker-compose-build-swagger: awx/projects
|
|
cd tools && CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /start_tests.sh swagger
|
|
|
|
detect-schema-change: genschema
|
|
curl https://s3.amazonaws.com/awx-public-ci-files/schema.json -o reference-schema.json
|
|
# Ignore differences in whitespace with -b
|
|
diff -u -b reference-schema.json schema.json
|
|
|
|
docker-compose-clean: awx/projects
|
|
cd tools && CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm -w /awx_devel --service-ports awx make clean
|
|
cd tools && TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose rm -sf
|
|
|
|
docker-compose-build: awx-devel-build
|
|
|
|
# Base development image build
|
|
awx-devel-build:
|
|
docker build -t ansible/awx_devel -f tools/docker-compose/Dockerfile \
|
|
--cache-from=$(DEV_DOCKER_TAG_BASE)/awx_devel:$(COMPOSE_TAG) .
|
|
docker tag ansible/awx_devel $(DEV_DOCKER_TAG_BASE)/awx_devel:$(COMPOSE_TAG)
|
|
#docker push $(DEV_DOCKER_TAG_BASE)/awx_devel:$(COMPOSE_TAG)
|
|
|
|
# For use when developing on "isolated" AWX deployments
|
|
docker-compose-isolated-build: awx-devel-build
|
|
docker build -t ansible/awx_isolated -f tools/docker-isolated/Dockerfile .
|
|
docker tag ansible/awx_isolated $(DEV_DOCKER_TAG_BASE)/awx_isolated:$(COMPOSE_TAG)
|
|
#docker push $(DEV_DOCKER_TAG_BASE)/awx_isolated:$(COMPOSE_TAG)
|
|
|
|
MACHINE?=default
|
|
docker-clean:
|
|
eval $$(docker-machine env $(MACHINE))
|
|
$(foreach container_id,$(shell docker ps -f name=tools_awx -aq),docker stop $(container_id); docker rm -f $(container_id);)
|
|
-docker images | grep "awx_devel" | awk '{print $$1 ":" $$2}' | xargs docker rmi
|
|
|
|
docker-refresh: docker-clean docker-compose
|
|
|
|
# Docker Development Environment with Elastic Stack Connected
|
|
docker-compose-elk: docker-auth awx/projects
|
|
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml -f tools/elastic/docker-compose.logstash-link.yml -f tools/elastic/docker-compose.elastic-override.yml up --no-recreate
|
|
|
|
docker-compose-cluster-elk: docker-auth awx/projects
|
|
TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose-cluster.yml -f tools/elastic/docker-compose.logstash-link-cluster.yml -f tools/elastic/docker-compose.elastic-override.yml up --no-recreate
|
|
|
|
prometheus:
|
|
docker run -u0 --net=tools_default --link=`docker ps | egrep -o "tools_awx(_run)?_([^ ]+)?"`:awxweb --volume `pwd`/tools/prometheus:/prometheus --name prometheus -d -p 0.0.0.0:9090:9090 prom/prometheus --web.enable-lifecycle --config.file=/prometheus/prometheus.yml
|
|
|
|
minishift-dev:
|
|
ansible-playbook -i localhost, -e devtree_directory=$(CURDIR) tools/clusterdevel/start_minishift_dev.yml
|
|
|
|
clean-elk:
|
|
docker stop tools_kibana_1
|
|
docker stop tools_logstash_1
|
|
docker stop tools_elasticsearch_1
|
|
docker rm tools_logstash_1
|
|
docker rm tools_elasticsearch_1
|
|
docker rm tools_kibana_1
|
|
|
|
psql-container:
|
|
docker run -it --net tools_default --rm postgres:10 sh -c 'exec psql -h "postgres" -p "5432" -U postgres'
|
|
|
|
VERSION:
|
|
@echo "awx: $(VERSION)"
|