1
0
mirror of https://github.com/1C-Company/docker_fresh.git synced 2024-12-04 10:24:48 +02:00
docker_fresh/start.py

485 lines
15 KiB
Python

import subprocess
import os
import modules.helper as helper
import sys
import json
import threading
import time
import codecs
from datetime import datetime
host_name = '.1cfresh.dev'
sup_password = '123Qwer'
new_server = False
global_debug = False
info_base_list = []
configurations = {}
docker_run_str = 'docker run --rm -v {}:/out_files alpine'.format(helper.this_path)
docker_compose_str = 'docker-compose -f workdir/docker-compose.yml '
work_dir = '/out_files/workdir/'
work_dir_other = work_dir + 'mnt/other-files/'
local_work_dir = helper.replace_sep(helper.this_path + 'workdir/')
path_to_1c = ''
class colors:
GREEN = '\033[92m'
WHITE = '\033[97m'
RED = '\033[91m'
class ib_prop:
a_name = 'ИмяВнешнейПубликации'
a_desc = 'ИмяФайлаШаблонаВнешненийПубликации'
int_name = 'ИмяВнутреннейПубликации'
int_desc = 'ИмяФайлаШаблонаВнутреннейПубликации'
conf_file = 'ИмяФайлаКонфигурации'
name = 'ИмяВКластере'
job = 'БлокироватьРаботуРегЗаданийПриСоздании'
adm = 'Администратор'
class ProgressThread(threading.Thread):
desc = ''
def __init__(self, desc):
self.desc = desc
super().__init__(target=self._spin)
self._stopevent = threading.Event()
def stop(self):
self._stopevent.set()
def _spin(self):
while not self._stopevent.isSet():
for t in [' ', '. ', '.. ', '...']:
print(self.desc, t, end='\r')
time.sleep(0.5)
class DoThread(threading.Thread):
is_good = False
def run(self):
try:
threading.Thread.run(self)
except:
self.is_good = False
pass
else:
self.is_good = True
def print_description(function_to_decorate):
def wrapper(*args, **kwargs):
if 'desc' in kwargs:
desc = kwargs['desc']
else:
desc = ''
all_desc = function_to_decorate.__doc__ + desc
task = DoThread(target=function_to_decorate, args=args, kwargs=kwargs)
task.start()
progress_thread = ProgressThread(all_desc)
progress_thread.start()
task.join()
progress_thread.stop()
while progress_thread.is_alive():
time.sleep(0.2)
if not task.is_good: exit(1)
print(all_desc, '...', '{}done'.format(colors.GREEN), colors.WHITE)
return wrapper
def call(command, remote=True, debug=False, action='', measure_duration=False, silent=True):
commands = []
if remote:
commands.append(docker_run_str)
commands.append(command)
if action != '' and (debug or global_debug):
print(action, end='\r')
if debug or global_debug:
print(' '.join(commands))
if not silent or global_debug:
stdout = None
stderr = None
else:
stdout = subprocess.PIPE
stderr = subprocess.PIPE
start_time = datetime.now()
result = subprocess.call(' '.join(commands), shell=True, stdout=stdout, stderr=stderr)
end_time = datetime.now() - start_time
if action != '' and (debug or global_debug):
print(action, 'is fihish.', 'Duration:{}'.format(
end_time) if measure_duration else '')
return result
@print_description
def get_configurations_data():
"""Get configuration data"""
is_fail = False
with codecs.open('other_files/params.json', 'r', 'utf-8') as json_file:
data = json.load(json_file)
for ib_data in data['ИнформационныеБазы']:
if not os.path.isfile('distr/{}'.format(ib_data['ИмяФайлаКонфигурации'])):
is_fail = True
print(colors.RED, 'File {} not found'.format(ib_data['ИмяФайлаКонфигурации']), colors.WHITE)
else:
info_base_list.append(ib_data)
if is_fail: exit(1)
def check_call_work(result, action, ib_name):
if result == 0: return
print(action, ib_name, 'is failed')
f = open(local_work_dir + 'mnt/{}_{}.out'.format(action, ib_name))
print(f.read())
f.close()
exit(1)
def prepare_new_ib(ib_name, int_name, conf_file_name, job_block):
job_dn_str = 'Y' if job_block else 'N'
action = 'create_ib'
result = call(' '.join(helper.create_ib_command(host_name, ib_name, conf_file_name, job_dn_str, action)),
remote=False,
action='Creating ' + ib_name,
measure_duration=True)
check_call_work(result, action, ib_name)
action = 'install_control_ext'
result = call(' '.join(helper.install_control_ext_command(host_name, ib_name, action)),
remote=False,
action='Installing control extension',
measure_duration=True)
check_call_work(result, action, ib_name)
ext_name = helper.replace_sep(local_work_dir + 'mnt/' + ib_name + '.cfe')
if os.path.isfile(ext_name):
action = 'install_ext'
result = call(' '.join(helper.install_ext_command(host_name, ib_name, action)),
remote=False,
action='Installing extension',
measure_duration=True)
check_call_work(result, action, ib_name)
if ib_name == 'sm':
action = 'install_sm_ext'
post_data = '/mnt/other-files/params.json'
result = call(' '.join(helper.install_sm_ext_command(host_name, ib_name, action)),
remote=False,
action='Installing gate control extension',
measure_duration=True)
check_call_work(result, action, ib_name)
else:
post_data = ''
action = 'disable_safe_mode'
result = call(' '.join(helper.disable_safe_mode(host_name, ib_name, action)),
remote=False,
action='Disabling safe mode for extensions',
measure_duration=True)
check_call_work(result, action, ib_name)
str_post = '-d @{}'.format(post_data) if post_data != '' else ''
call('docker exec web.{} curl {} -X POST http://localhost/int/{}/hs/api.1cfresh/init'.format(host_name, str_post, int_name),
remote=False,
action='Initialization',
measure_duration=True)
@print_description
def delete_volumes():
"""Delete volumes"""
call('docker volume rm workdir_1c_pg_data', remote=False)
call('docker volume rm workdir_1c_pg_socket', remote=False)
call('docker volume rm workdir_1c_server_data', remote=False)
@print_description
def prepare_bases():
"""Prepare all bases"""
sm_ib = None
for ib_data in info_base_list:
if ib_data[ib_prop.name] == 'sm':
sm_ib = ib_data
continue
prepare_new_ib(
ib_name=ib_data[ib_prop.name],
int_name=ib_data[ib_prop.int_name],
conf_file_name=ib_data[ib_prop.conf_file],
job_block=ib_data[ib_prop.job]
)
if ib_data[ib_prop.job]:
enable_job(ib_data[ib_prop.name], ib_data[ib_prop.adm])
# prepare sm base
prepare_new_ib(
ib_name=sm_ib[ib_prop.name],
int_name=sm_ib[ib_prop.int_name],
conf_file_name=sm_ib[ib_prop.conf_file],
job_block=sm_ib[ib_prop.job]
)
if ib_prop.job:
enable_job(ib_data[ib_prop.name], ib_data[ib_prop.adm])
@print_description
def renew_nginx_files():
"""Renew nginx files"""
conf_catalog = work_dir + 'artifacts/nginx/conf/'
call('mkdir -p {}'.format(conf_catalog))
call('sh -c "cp -r /out_files/conf/nginx/* {}"'.format(conf_catalog))
call('sh -c "sed -i \'s/hosthosthost/{}/g\' {}*.conf"'.format(host_name, conf_catalog))
call('sh -c "sed -i \'s/sitesitesite/site.{}/g\' {}*.conf"'.format(host_name, conf_catalog))
call('sh -c "sed -i \'s/webwebweb/web.{}/g\' {}*.conf"'.format(host_name, conf_catalog))
call('sh -c "sed -i \'s/gategategate/gate.{}/g\' {}*.conf"'.format(host_name, conf_catalog))
call('sh -c "sed -i \'s/sitesitesite/site.{}/g\' {}conf.d/*.conf"'.format(host_name, conf_catalog))
call('sh -c "sed -i \'s/hosthosthost/{}/g\' {}conf.d/*.conf"'.format(host_name, conf_catalog))
@print_description
def renew_workdir():
"""Renew wordir"""
call('rm -rf /out_files/workdir')
call('mkdir -p {}mnt'.format(work_dir))
call('mkdir -p {}artifacts/web/conf'.format(work_dir))
call('sh -c "cp /out_files/conf/web/httpd.conf {}artifacts/web/conf/httpd.conf"'.format(work_dir))
call('sh -c "cp /out_files/distr/*.cf {}mnt/"'.format(work_dir))
call('sh -c "cp /out_files/distr/*.cfe {}mnt/"'.format(work_dir))
@print_description
def renew_docker_compose():
"""Renew docker-compose file"""
call('cp /out_files/docker-compose_pattern.yml /out_files/workdir/docker-compose.yml')
call('sh -c "sed -i \'s/HOSTNAMEREPLACE/{}/\' {}/*.yml"'.format(host_name, work_dir))
call('sh -c "sed -i \'s/PATH_TO_1C_REPLACE/{}/\' {}/*.yml"'.format(path_to_1c.replace('/','\/'), work_dir))
@print_description
def renew_other_files():
"""Renew other-files"""
call('rm -rf ' + work_dir_other)
call('cp -r /out_files/other_files/ ' + work_dir_other)
call('sh -c "sed -i \'s/HOSTNAMEREPLACE/{}/\' {}vrd/*.vrd"'.format(host_name, work_dir_other))
call('sh -c "sed -i \'s/HOSTNAMEREPLACE/{}/\' {}params.json"'.format(host_name, work_dir_other))
@print_description
def create_bucket():
"""Create new bucket to 1C"""
call('mkdir /out_files/workdir/artifacts/s3/files')
@print_description
def publish_sevises():
"""Publish services"""
for ib_data in info_base_list:
if ib_data[ib_prop.a_name] != '':
call(' '.join(helper.web_publish_command(
host_name=host_name,
conf_name=ib_data[ib_prop.a_name],
internal=False,
descriptor=ib_data[ib_prop.a_desc],
base_name=ib_data[ib_prop.name])), remote=False)
call(' '.join(helper.web_publish_command(
host_name=host_name,
conf_name=ib_data[ib_prop.int_name],
internal=True,
descriptor=ib_data[ib_prop.int_desc],
base_name=ib_data[ib_prop.name])), remote=False)
# publish special services
call(' '.join(helper.web_publish_command(
host_name, 'openid', False, 'openid', 'sm')), remote=False)
call(' '.join(helper.web_publish_command(host_name, 'sc', True,
'sessioncontrol', 'sm;Usr=SessionControl;Pwd=' + sup_password)), remote=False)
call(' '.join(helper.web_publish_command(host_name, 'extreg', True,
'extreg', 'sm;Usr=ExtReg;Pwd=' + sup_password)), remote=False)
# restart Apache
call('docker exec web.{} chown -R usr1cv8:grp1cv8 /var/www'.format(host_name), remote=False)
call('docker exec web.{} httpd -k graceful'.format(host_name), remote=False)
@print_description
def set_full_host_name(is_new):
"""Set full hostname"""
global host_name
if is_new:
part_host_name = helper.get_host_name(sys.argv)
f = open('.hostname', 'w+')
f.write(part_host_name)
f.close()
else:
f = open('.hostname')
part_host_name = f.read()
f.close()
host_name = part_host_name + host_name
print('host name is', host_name)
@print_description
def create_db_site():
"""Create db for site"""
call('docker exec -t db.{} sh -c "/usr/bin/psql -U postgres -f {}'.format(host_name, '/create_db_site.psql"'),
remote=False)
@print_description
def create_db_forum():
"""Create db for forum"""
call('docker exec -t db.{} sh -c "/usr/bin/psql -U postgres -f {}'.format(host_name, '/create_db_forum.psql"'),
remote=False)
@print_description
def delete_control_extension(ib_name, user, desc):
"""Delete control extension"""
call(' '.join(helper.delete_control_extension(ib_name, host_name, user)), remote=False)
def delete_all_control_ext():
for ib_data in info_base_list:
delete_control_extension(
ib_name=ib_data[ib_prop.name],
user= None if ib_data[ib_prop.adm] == '' else ib_data[ib_prop.adm],
desc=ib_data[ib_prop.name]
)
@print_description
def configurate_site():
"""Configurate site settings"""
call(' '.join(helper.edit_site_settings(host_name, sup_password)), remote=False)
call('docker exec -t web.{0} curl https://{0}/settings/finish_configuration'.format(host_name), remote=False)
call(' '.join(helper.enable_manual_registration(host_name)), remote=False)
call(' '.join(helper.enable_openid(host_name)), remote=False)
call(' '.join(helper.add_solution(
host_name=host_name,
brief_desc='"БТС"',
full_desc='"БТС"',
display_order=0,
id='smtl',
possibilities='"БТС"',
title='"Библиотека технологии сервиса"'
)), remote=False)
@print_description
def init_gate():
"""Initialization gate"""
call('docker exec -t web.{0} curl --user Администратор: https://{0}/a/adm/hs/docker_control/update_appgate'.format(host_name),
remote=False)
@print_description
def wait_postgres():
"""Waiting for postgres"""
call('docker exec -t db.{} /wait_postgres.sh'.format(host_name), remote=False)
@print_description
def wait_site():
"""Waiting for site"""
call('docker exec -t site.{} /wait_site.sh'.format(host_name), remote=False)
def enable_job(base_name, user):
call('docker exec -t ras.{} deployka scheduledjobs unlock -db {} -db-user "{}"'.format(host_name, base_name, user),
remote=False)
@print_description
def down_containers():
"""Down all conteiners"""
call(docker_compose_str + 'down', remote=False)
@print_description
def get_path_to_1c():
"""Getting path to 1C"""
global path_to_1c
cmd = "docker run --rm fresh/core sh -c \"find / -name '1cv8c' | sed 's/1cv8c//g'\""
pipe = subprocess.PIPE
p = subprocess.Popen(cmd, shell=True, stdin=pipe, stdout=pipe, stderr=pipe, close_fds=True)
path_to_1c = p.stdout.read().decode('utf-8').strip()
print('path to 1C: ' + path_to_1c)
global_start_time = datetime.now()
print('{}Fresh is starting{} at {}'.format(colors.GREEN, colors.WHITE, global_start_time))
down_containers()
new_server = '-new' in sys.argv
global_debug = '-debug' in sys.argv
set_full_host_name(new_server)
get_path_to_1c()
helper.init(path_to_1c)
if new_server:
renew_workdir()
get_configurations_data()
renew_nginx_files()
renew_docker_compose()
renew_other_files()
delete_volumes()
# start db srv ras web gate conteiners
call(docker_compose_str + 'up -d db srv ras web gate s3', remote=False, silent=False)
wait_postgres()
if new_server:
create_bucket()
publish_sevises()
prepare_bases()
create_db_site()
create_db_forum()
# start site forum nginx conteiners
call(docker_compose_str + 'up -d nginx site', remote=False, silent=False)
wait_site()
if new_server:
delete_all_control_ext()
configurate_site()
init_gate()
global_end_time = datetime.now() - global_start_time
print('{}Fresh started{}'.format(colors.GREEN, colors.WHITE), 'Duration:', global_end_time)