def generate(backup_type, storages, path_to_file): """ The function generate a configuration file job. """ backup_type = backup_type[0] path_to_file = path_to_file[0] template_path = f'{TEMPLATES_DIR}/backup_type/{backup_type}.conf' if path_to_file.startswith('/'): general_function.create_dirs( job_name=backup_type, dirs_pairs={os.path.dirname(path_to_file): ''}) general_function.copy_ofs(template_path, path_to_file) try: fd = open(path_to_file, 'a') except (OSError, PermissionError, FileNotFoundError) as e: messange_info = f"Couldn't open file {path_to_file}:{e}!" general_function.print_info(messange_info) sys.exit(1) if backup_type in config.supported_db_backup_type: job_type = 'databases' elif backup_type in config.supported_file_backup_type: job_type = 'files' else: job_type = 'external' for storage in storages: storage_template_path = f'{TEMPLATES_DIR}/storages/{storage}.conf' with open(storage_template_path, 'r', encoding='utf-8') as f: str_storage = f.read() str_storage = str_storage.replace('backup_type', backup_type) str_storage = str_storage.replace('job_type', job_type) if backup_type == 'inc_files': str_storage = str_storage.replace('inc_files/dump', 'inc') str_storage = re.sub( r"[ ]*store:[\s]*days: ''[\s]*weeks: ''[\s]*month: ''[\s]*", '', str_storage) if backup_type == 'desc_files': str_storage = str_storage.replace('desc_files/dump', 'desc/dump') if backup_type == 'external': str_storage = str_storage.replace('external/dump', 'dump') fd.write(str_storage) fd.close() os.chmod(path_to_file, 0o600) general_function.print_info( f"Successfully generated '{path_to_file}' configuration file!")
def create_links_and_copies(link_dict, copy_dict, job_name): if link_dict: for dst, src in link_dict.items(): try: general_function.create_symlink(src, dst) except general_function.MyError as err: log_and_mail.writelog('ERROR', f"Can't create symlink {src} -> {dst}: {err}", config.filelog_fd, job_name) if copy_dict: for dst, src in copy_dict.items(): try: general_function.copy_ofs(src, dst) except general_function.MyError as err: log_and_mail.writelog('ERROR', f"Can't copy {src} -> {dst}: {err}", config.filelog_fd, job_name)
def periodic_backup(full_tmp_path, general_local_dst_path, remote_dir, storage, subdir_name, days_count, weeks_count, job_name, host, share): daily_subdir_name = "daily" weekly_subdir_name = "weekly" monthly_subdir_name = "monthly" link_dict = {} dow = general_function.get_time_now("dow") backup_file_name = os.path.basename(full_tmp_path) full_dst_path = os.path.join(general_local_dst_path, subdir_name) dst_dirs = [] daily_dir = os.path.join(general_local_dst_path, daily_subdir_name) weekly_dir = os.path.join(general_local_dst_path, weekly_subdir_name) monthly_dir = os.path.join(general_local_dst_path, monthly_subdir_name) if storage == 'local': if subdir_name == monthly_subdir_name: dst_dirs.append(monthly_dir) if dow == config.dow_backup and int(weeks_count): src_link = os.path.join(general_local_dst_path, monthly_subdir_name, backup_file_name) dst_link = os.path.join(general_local_dst_path, weekly_subdir_name, backup_file_name) dst_dirs.append(weekly_dir) link_dict[dst_link] = src_link if int(days_count): src_link = os.path.join(general_local_dst_path, monthly_subdir_name, backup_file_name) dst_link = os.path.join(general_local_dst_path, daily_subdir_name, backup_file_name) dst_dirs.append(daily_dir) link_dict[dst_link] = src_link elif subdir_name == weekly_subdir_name and storage == 'local': dst_dirs.append(weekly_dir) if int(days_count): src_link = os.path.join(general_local_dst_path, weekly_subdir_name, backup_file_name) dst_link = os.path.join(general_local_dst_path, daily_subdir_name, backup_file_name) dst_dirs.append(daily_dir) link_dict[dst_link] = src_link else: dst_dirs.append(daily_dir) else: dst_dirs.append(full_dst_path) for dst_dir in set(dst_dirs): dirs_for_log = general_function.get_dirs_for_log( dst_dir, remote_dir, storage) general_function.create_dirs(job_name='', dirs_pairs={dst_dir: dirs_for_log}) if storage == 'local': try: general_function.move_ofs(full_tmp_path, full_dst_path) except general_function.MyError as err: log_and_mail.writelog( 'ERROR', "Can't move '%s' file '%s' -> '%s' on '%s' storage: %s" % (subdir_name, full_tmp_path, full_dst_path, storage, err), config.filelog_fd, job_name) else: log_and_mail.writelog( 'INFO', "Successfully moved '%s' file '%s' -> '%s' on '%s' storage." % (subdir_name, full_tmp_path, full_dst_path, storage), config.filelog_fd, job_name) if link_dict: for key in link_dict.keys(): src = link_dict[key] dst = key try: general_function.create_symlink(src, dst) except general_function.MyError as err: log_and_mail.writelog( 'ERROR', "Can't create symlink '%s' -> '%s' on 'local' storage: %s" % (src, dst, err), config.filelog_fd, job_name) else: try: general_function.copy_ofs(full_tmp_path, full_dst_path) dirs_for_log = general_function.get_dirs_for_log( full_dst_path, remote_dir, storage) except general_function.MyError as err: if storage != 'smb': log_and_mail.writelog( 'ERROR', "Can't copy '%s' file '%s' -> '%s' directory on '%s' storage(%s): %s" % (subdir_name, full_tmp_path, dirs_for_log, storage, host, err), config.filelog_fd, job_name) else: log_and_mail.writelog( 'ERROR', "Can't copy '%s' file '%s' -> '%s' directory in '%s' share on '%s' storage(%s): %s" % (subdir_name, full_tmp_path, dirs_for_log, share, storage, host, err), config.filelog_fd, job_name) else: if storage != 'smb': log_and_mail.writelog( 'INFO', "Successfully copied '%s' file '%s' -> '%s' directory on '%s' storage(%s)." % (subdir_name, full_tmp_path, dirs_for_log, storage, host), config.filelog_fd, job_name) else: log_and_mail.writelog( 'INFO', "Successfully copied '%s' file '%s' -> '%s' directory in '%s' share on '%s' storage(%s)." % (subdir_name, full_tmp_path, dirs_for_log, share, storage, host), config.filelog_fd, job_name)
def create_inc_file(local_dst_dirname, remote_dir, part_of_dir_path, backup_file_name, target, exclude_list, gzip, job_name, storage, host, share): ''' The function determines whether to collect a full backup or incremental, prepares all the necessary information. ''' date_year = general_function.get_time_now('year') date_month = general_function.get_time_now('moy') date_day = general_function.get_time_now('dom') if int(date_day) < 11: daily_prefix = 'day_01' elif int(date_day) < 21: daily_prefix = 'day_11' else: daily_prefix = 'day_21' year_dir = os.path.join(local_dst_dirname, part_of_dir_path, date_year) initial_dir = os.path.join(year_dir, 'year') # Path to full backup month_dir = os.path.join(year_dir, 'month_%s' % (date_month), 'monthly') daily_dir = os.path.join(year_dir, 'month_%s' % (date_month), 'daily', daily_prefix) year_inc_file = os.path.join(initial_dir, 'year.inc') month_inc_file = os.path.join(month_dir, 'month.inc') daily_inc_file = os.path.join(daily_dir, 'daily.inc') link_dict = {} # dict for symlink with pairs like dst: src copy_dict = {} # dict for copy with pairs like dst: src # Before we proceed to collect a copy, we need to delete the copies for the same month last year # if they are to not save extra archives old_year = int(date_year) - 1 old_year_dir = os.path.join(local_dst_dirname, part_of_dir_path, str(old_year)) if os.path.isdir(old_year_dir): old_month_dir = os.path.join(old_year_dir, 'month_%s' % (date_month)) del_old_inc_file(old_year_dir, old_month_dir) if not os.path.isfile(year_inc_file): # There is no original index file, so we need to check the existence of an year directory if os.path.isdir(year_dir): # There is a directory, but there is no file itself, then something went wrong, so # we delete this directory with all the data inside, because even if they are there # continue to collect incremental copies it will not be able to general_function.del_file_objects(job_name, year_dir) dirs_for_log = general_function.get_dirs_for_log( year_dir, remote_dir, storage) file_for_log = os.path.join(dirs_for_log, os.path.basename(year_inc_file)) log_and_mail.writelog('ERROR', "The file %s not found, so the directory %s is cleared." +\ "Incremental backup will be reinitialized " %(file_for_log, dirs_for_log), config.filelog_fd, job_name) # Initialize the incremental backup, i.e. collect a full copy dirs_for_log = general_function.get_dirs_for_log( initial_dir, remote_dir, storage) general_function.create_dirs(job_name=job_name, dirs_pairs={initial_dir: dirs_for_log}) # Get the current list of files and write to the year inc file meta_info = get_index(target, exclude_list) with open(year_inc_file, "w") as index_file: json.dump(meta_info, index_file) full_backup_path = general_function.get_full_path( initial_dir, backup_file_name, 'tar', gzip) general_files_func.create_tar('files', full_backup_path, target, gzip, 'inc_files', job_name, remote_dir, storage, host, share) # After creating the full copy, you need to make the symlinks for the inc.file and # the most collected copy in the month directory of the current month # as well as in the decade directory if it's local, scp the repository and # copy inc.file for other types of repositories that do not support symlynk. month_dirs_for_log = general_function.get_dirs_for_log( month_dir, remote_dir, storage) daily_dirs_for_log = general_function.get_dirs_for_log( daily_dir, remote_dir, storage) general_function.create_dirs(job_name=job_name, dirs_pairs={ month_dir: month_dirs_for_log, daily_dir: daily_dirs_for_log }) if storage in 'local, scp': link_dict[month_inc_file] = year_inc_file link_dict[os.path.join( month_dir, os.path.basename(full_backup_path))] = full_backup_path link_dict[daily_inc_file] = year_inc_file link_dict[os.path.join( daily_dir, os.path.basename(full_backup_path))] = full_backup_path else: copy_dict[month_inc_file] = year_inc_file copy_dict[daily_inc_file] = year_inc_file else: symlink_dir = '' if int(date_day) == 1: # It is necessary to collect monthly incremental backup relative to the year copy old_meta_info = specific_function.parser_json(year_inc_file) new_meta_info = get_index(target, exclude_list) general_inc_backup_dir = month_dir # It is also necessary to make a symlink for inc files and backups to the directory with the first decade symlink_dir = daily_dir general_dirs_for_log = general_function.get_dirs_for_log( general_inc_backup_dir, remote_dir, storage) symlink_dirs_for_log = general_function.get_dirs_for_log( symlink_dir, remote_dir, storage) general_function.create_dirs(job_name=job_name, dirs_pairs={ general_inc_backup_dir: general_dirs_for_log, symlink_dir: symlink_dirs_for_log }) with open(month_inc_file, "w") as index_file: json.dump(new_meta_info, index_file) elif int(date_day) == 11 or int(date_day) == 21: # It is necessary to collect a ten-day incremental backup relative to a monthly copy try: old_meta_info = specific_function.parser_json(month_inc_file) except general_function.MyError as e: log_and_mail.writelog( 'ERROR', "Couldn't open old month meta info file '%s': %s!" % (month_inc_file, e), config.filelog_fd, job_name) return 2 new_meta_info = get_index(target, exclude_list) general_inc_backup_dir = daily_dir general_dirs_for_log = general_function.get_dirs_for_log( general_inc_backup_dir, remote_dir, storage) general_function.create_dirs( job_name=job_name, dirs_pairs={general_inc_backup_dir: general_dirs_for_log}) with open(daily_inc_file, "w") as index_file: json.dump(new_meta_info, index_file) else: # It is necessary to collect a normal daily incremental backup relative to a ten-day copy try: old_meta_info = specific_function.parser_json(daily_inc_file) except general_function.MyError as e: log_and_mail.writelog( 'ERROR', "Couldn't open old decade meta info file '%s': %s!" % (daily_inc_file, e), config.filelog_fd, job_name) return 2 new_meta_info = get_index(target, exclude_list) general_inc_backup_dir = daily_dir general_dirs_for_log = general_function.get_dirs_for_log( general_inc_backup_dir, remote_dir, storage) general_function.create_dirs( job_name=job_name, dirs_pairs={general_inc_backup_dir: general_dirs_for_log}) # Calculate the difference between the old and new file states diff_json = compute_diff(new_meta_info, old_meta_info) inc_backup_path = general_function.get_full_path( general_inc_backup_dir, backup_file_name, 'tar', gzip) # Define the list of files that need to be included in the archive target_change_list = diff_json['modify'] # Form GNU.dumpdir headers dict_directory = {} # Dict to store pairs like dir:GNU.dumpdir excludes = r'|'.join([ fnmatch.translate(x)[:-7] for x in general_files_func.EXCLUDE_FILES ]) or r'$.' for dir_name, dirs, files in os.walk(target): first_level_files = [] if re.match(excludes, dir_name): continue for file in files: if re.match(excludes, os.path.join(dir_name, file)): continue first_level_files.append(file) first_level_subdirs = dirs dict_directory[dir_name] = get_gnu_dumpdir_format( diff_json, dir_name, target, excludes, first_level_subdirs, first_level_files) create_inc_tar(inc_backup_path, remote_dir, dict_directory, target_change_list, gzip, job_name, storage, host, share) if symlink_dir: if storage in 'local, scp': link_dict[daily_inc_file] = month_inc_file else: copy_dict[daily_inc_file] = month_inc_file if link_dict: for key in link_dict.keys(): src = link_dict[key] dst = key try: general_function.create_symlink(src, dst) except general_function.MyError as err: log_and_mail.writelog( 'ERROR', "Can't create symlink %s -> %s: %s" % (src, dst, err), config.filelog_fd, job_name) if copy_dict: for key in copy_dict.keys(): src = copy_dict[key] dst = key try: general_function.copy_ofs(src, dst) except general_function.MyError as err: log_and_mail.writelog( 'ERROR', "Can't copy %s -> %s: %s" % (src, dst, err), config.filelog_fd, job_name)