422 lines
11 KiB
Nix
422 lines
11 KiB
Nix
{ ... }: {
|
|
text = ''
|
|
export _archive_pattern="_[0-9]{12}-[[:alnum:]]{40}"
|
|
export _archive_pattern_compressed="_[0-9]{12}-[[:alnum:]]{40}\.t(ar|gz|xz)$"
|
|
export _archive_pattern_fast="_[0-9]{12}-[[:alnum:]]{40}\.tar$"
|
|
|
|
# Archive using multiple threads. Uses 75% of free RAM.
|
|
# All directories by default.
|
|
# Supports .archiveignore exclude file.
|
|
# Usage: archive [DIRS]
|
|
function archive() {
|
|
local IFS=$'\n'
|
|
local targets=(''${@})
|
|
[[ "''${targets}" = "" ]] && targets=($(_ls_dir))
|
|
|
|
process() {
|
|
# Skip if already an archive.
|
|
_is_archive "''${target}" && {
|
|
_iterate_skip "Already an archive."
|
|
return 0
|
|
};
|
|
|
|
# Cut full paths.
|
|
[[ "''${target##*/}" = "" ]] || target="''${target##*/}"
|
|
|
|
local date=$(_archive_date)
|
|
|
|
# Parse name.
|
|
local name=$(parse_pascal ''${target})
|
|
|
|
# Exclude support.
|
|
local exclude=""
|
|
[[ -f ".archiveignore" ]] && exclude="--exclude-from=.archiveignore"
|
|
[[ -f "''${target}/.archiveignore" ]] && exclude="--exclude-from=''${target}/.archiveignore"
|
|
|
|
# Create archive.
|
|
local hash=$(tar ''${exclude} -c ''${target} | pv -s $(/usr/bin/env du -sb ''${target} | awk '{print $1}') | xz -9e --threads=0 --memlimit=$(_archive_memlimit) | tee ''${name}.txz | sha1sum | cut -d\ -f1)
|
|
|
|
# Append hash to target name.
|
|
local new_name="''${name}_''${date}-''${hash}.txz"
|
|
mv -- ''${name}.txz ''${new_name} && echo ''${new_name}
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Creates a simple archive.
|
|
# If it is a file, it just reformats file name to match archive name.
|
|
# For dirs, it first creates a tar archive.
|
|
# All files by default.
|
|
# Usage: archive_fast [DIRS]
|
|
function archive_fast() {
|
|
local IFS=$'\n'
|
|
local targets=("''${@}")
|
|
[[ "''${targets}" = "" ]] && targets=($(_ls_file))
|
|
|
|
process() {
|
|
# Skip if already an archive.
|
|
_is_archive "''${target}" && {
|
|
_iterate_skip "Already an archive."
|
|
return 0
|
|
};
|
|
|
|
# Cut full paths.
|
|
[[ "''${target##*/}" = "" ]] || target="''${target##*/}"
|
|
|
|
# Start timestamp.
|
|
local date=$(_archive_date)
|
|
|
|
# Exclude support.
|
|
local exclude
|
|
[[ -f ".archiveignore" ]] && exclude="--exclude-from=.archiveignore"
|
|
[[ -f "''${target}/.archiveignore" ]] && exclude="--exclude-from=''${target}/.archiveignore"
|
|
|
|
local name
|
|
local extension
|
|
|
|
if [[ -d "''${target}" ]]; then
|
|
name=$(parse_pascal "''${target}")
|
|
|
|
# Create archive.
|
|
local hash=$(tar ''${exclude} -c "''${target}" | pv -s $(/usr/bin/env du -sb "''${target}" | awk '{print $1}') | tee "''${name}".tar | sha1sum | cut -d\ -f1)
|
|
|
|
# Append hash to target name.
|
|
local new_name="''${name}_''${date}-''${hash}.tar"
|
|
mv -- "''${name}".tar ''${new_name} && echo ''${new_name}
|
|
else
|
|
name=$(parse_pascal "''${target%.*}")
|
|
extension=".''${target##*.}"
|
|
|
|
# Check if extension matches name, then drop it.
|
|
[[ "''${extension}" = ".''${target%.*}" ]] && extension=""
|
|
|
|
local hash=$(pv "''${target}" | sha1sum | cut -d\ -f1)
|
|
local new_name="''${name}_''${date}-''${hash}''${extension}"
|
|
mv -- "''${target}" "''${new_name}" && echo ''${new_name}
|
|
fi
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Check archives integrity.
|
|
# Checks all archives by default.
|
|
# Usage: archive_check [FILES]
|
|
function archive_check() {
|
|
local IFS=$'\n'
|
|
local targets=(''${@})
|
|
[[ "''${targets}" = "" ]] && targets=($(_ls_archive))
|
|
|
|
process() {
|
|
_archive_check "''${target}"
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Delete old versions of an archive.
|
|
# All archives with 1 version by default.
|
|
# Usage: archive_prune [NAME] [VERSIONS]
|
|
function archive_prune() {
|
|
local IFS=$'\n'
|
|
local targets=(''${1})
|
|
local versions=''${2}
|
|
|
|
[[ "''${targets}" = "" ]] && targets=($(_archive_names))
|
|
[[ "''${versions}" = "" ]] && versions=1
|
|
|
|
if [[ ''${#} -gt 2 ]]; then
|
|
help archive_prune
|
|
return 2
|
|
fi
|
|
|
|
process() {
|
|
local prune=($(ls | grep -E "^''${target}''${_archive_pattern}" | sort -r | sed -e "1,''${versions}d"))
|
|
|
|
for archive in ''${prune[@]}; do
|
|
rm -- "''${archive}" && echo "''${archive}"
|
|
done
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Delete specified or all archive files.
|
|
# Usage: archive_rm [FILES]
|
|
function archive_rm() {
|
|
local IFS=$'\n'
|
|
local targets=(''${@})
|
|
[[ "''${targets}" = "" ]] && targets=($(_ls_archive))
|
|
|
|
process() {
|
|
rm -- "''${target}" && echo "''${target}"
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Recompress previously created archive_fast with better compression.
|
|
# Usage: archive_compress [FILES]
|
|
function archive_compress() {
|
|
local IFS=$'\n'
|
|
local targets=(''${@})
|
|
[[ "''${targets}" = "" ]] && targets=$(_ls_archive_fast)
|
|
|
|
process() {
|
|
local data=($(_archive_parse "''${target}"))
|
|
local tmp="''${data[0]}.txz"
|
|
|
|
# Check that old format.
|
|
if [[ "''${data[3]}" != "tar" ]]; then
|
|
_iterate_skip "Not in .tar format!"
|
|
return 0
|
|
fi
|
|
|
|
# Check integrity.
|
|
_archive_check "''${target}" || return 1
|
|
|
|
# Recompress.
|
|
local hash=$(pv "''${target}" | xz -9e --threads=0 --memlimit=$(_archive_memlimit) | tee "''${tmp}" | sha1sum | cut -d\ -f1)
|
|
|
|
# Rename.
|
|
local new_name="''${data[0]}_''${data[1]}-''${hash}.txz"
|
|
mv -- ''${tmp} ''${new_name} && rm ''${target} && echo ''${new_name}
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Rename archives.
|
|
# If no name specified, it simplifies archive's name.
|
|
# If no archives specified, apply to all archives.
|
|
# Usage: archive_name [ARCHIVE] [NAME]
|
|
function archive_name() {
|
|
local IFS=$'\n'
|
|
local targets="''${1}"
|
|
local name="''${2}"
|
|
[[ "''${targets}" = "" ]] && targets=($(_ls_archive))
|
|
|
|
process() {
|
|
# Simplify name by default.
|
|
if [[ "''${name}" = "" || ''${count} -gt 1 ]]; then
|
|
name="''${target%_*}"
|
|
name="$(parse_pascal ''${name})"
|
|
fi
|
|
|
|
# Remove old name.
|
|
local data="''${target##*_}"
|
|
local new_name="''${name}_''${data}"
|
|
|
|
# Check for the same name.
|
|
[[ "''${target}" = "''${new_name}" ]] && return 0
|
|
|
|
# Check for existing target.
|
|
if [[ -f "''${new_name}" ]]; then
|
|
_error "''${new_name}: Already exists!"
|
|
return 1
|
|
fi
|
|
|
|
# Rename.
|
|
mv -- ''${target} ''${new_name} && echo ''${new_name}
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Extract previously created archive with checksum validation.
|
|
# Supports unarchiving exact paths from the remote machines (rsync syntax).
|
|
# Usage: unarchive [HOST:FILES]
|
|
function unarchive() {
|
|
local IFS=$'\n'
|
|
local targets=(''${@})
|
|
[[ "''${targets}" = "" ]] && targets=$(_ls_archive_compressed)
|
|
|
|
process() {
|
|
# Validate.
|
|
# _archive_check "''${target}" || return 1
|
|
if ! _is_compressed_archive "''${target}"; then
|
|
_iterate_skip "Not a compressed archive."
|
|
return 0
|
|
fi
|
|
|
|
# Remote archives.
|
|
local remote
|
|
local file="''${target}"
|
|
|
|
if [[ "''${target//\\:/}" == *:* ]]; then
|
|
local host="''${target%%:*}"
|
|
file="''${target#*:}"
|
|
remote=(trysudo ssh ''${host})
|
|
fi
|
|
|
|
# Extract.
|
|
case "''${file##*.}" in
|
|
"txz")
|
|
''${remote[@]} pv -f ''${file} | xz -d | tar -xf -
|
|
;;
|
|
"tar")
|
|
''${remote[@]} pv -f ''${file} | tar -xf -
|
|
;;
|
|
"tgz")
|
|
''${remote[@]} pv -f ''${file} | gzip -d | tar -xf -
|
|
;;
|
|
esac
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Change archive's filesystem time to match creation date.
|
|
# Usage: archive_touch [FILES]
|
|
function archive_touch() {
|
|
local IFS=$'\n'
|
|
local targets=(''${@})
|
|
[[ "''${targets}" = "" ]] && targets=$(_ls_archive)
|
|
|
|
process() {
|
|
local data=($(_archive_parse "''${target}"))
|
|
local date=''${data[1]}
|
|
touch -t ''${date} -- ''${target}
|
|
}
|
|
|
|
_iterate_targets process ''${targets[@]}
|
|
}
|
|
|
|
# Parse archive file name to get: name, date, hash and format.
|
|
# Usage: _archive_parse <FILENAME>
|
|
function _archive_parse() {
|
|
local input="''${1}"
|
|
local name="''${input%_*}"
|
|
local format="''${input##*.}"
|
|
local data="''${input##*_}"; data="''${data%.*}"
|
|
local date="''${data%%-*}"
|
|
local hash="''${data##*-}"
|
|
|
|
# No extension if no dots.
|
|
[[ "''${input}" = *\.* ]] || format=""
|
|
|
|
echo "''${name}"
|
|
echo "''${date}"
|
|
echo "''${hash}"
|
|
echo "''${format}"
|
|
}
|
|
|
|
# Autocomplete for archive_name function.
|
|
# First arg is the archives list, second one is selected archive's current name.
|
|
function _comp_archive_name() {
|
|
local IFS=$'\n'
|
|
COMPREPLY=()
|
|
|
|
local cur="''${COMP_WORDS[COMP_CWORD]}"
|
|
local prev="''${COMP_WORDS[COMP_CWORD-1]}"
|
|
local command="''${COMP_WORDS[0]}"
|
|
|
|
if [[ "''${prev}" = "''${command}" ]]; then
|
|
COMPREPLY=( $(compgen -W "$(ls | grep -E ''${_archive_pattern})" -- ''${cur}) )
|
|
return 0
|
|
else
|
|
local data=($(_archive_parse ''${prev}))
|
|
local name="''${data[0]}"
|
|
COMPREPLY=( $(compgen -W "''${name}" -- ''${cur}) )
|
|
return 0
|
|
fi
|
|
}
|
|
|
|
# Autocomplete with archives in current dir.
|
|
function _comp_archive_grep() {
|
|
_autocomplete_grep ''${_archive_pattern}
|
|
}
|
|
|
|
# Autocomplete with fast archives in current dir.
|
|
function _comp_archive_grep_fast() {
|
|
_autocomplete_grep ''${_archive_pattern_fast}
|
|
}
|
|
|
|
# Get date for a new archive.
|
|
function _archive_date() {
|
|
date +%Y%m%d%H%M
|
|
}
|
|
|
|
# Get names of all archives.
|
|
function _archive_names() {
|
|
local IFS=$'\n'
|
|
local archives=($(_ls_archive))
|
|
local names=()
|
|
|
|
for archive in ''${archives[@]}; do
|
|
local data=($(_archive_parse ''${archive}))
|
|
names+=(''${data[0]})
|
|
done
|
|
|
|
# Remove copies.
|
|
names=($(printf '%s\n' "''${names[@]}" | sort -u))
|
|
|
|
printf '%s\n' "''${names[@]}"
|
|
}
|
|
|
|
# Autocomplete with names of all archives.
|
|
function _comp_archive_names() {
|
|
_autocomplete_first $(_archive_names)
|
|
}
|
|
|
|
# Check if file is an archive.
|
|
function _is_archive() {
|
|
local out=$(echo "''${*}" | grep -E ''${_archive_pattern})
|
|
|
|
[[ "''${out}" != "" ]]
|
|
}
|
|
|
|
# Check if file is a compressed archive.
|
|
function _is_compressed_archive() {
|
|
local out=$(echo "''${*}" | grep -E ''${_archive_pattern_compressed})
|
|
|
|
[[ "''${out}" != "" ]]
|
|
}
|
|
|
|
# List all archives.
|
|
function _ls_archive() {
|
|
ls | grep -E ''${_archive_pattern}
|
|
}
|
|
|
|
# List fast archives.
|
|
function _ls_archive_fast() {
|
|
ls | grep -E ''${_archive_pattern_fast}
|
|
}
|
|
|
|
# List fast archives.
|
|
function _ls_archive_compressed() {
|
|
ls | grep -E ''${_archive_pattern_compressed}
|
|
}
|
|
|
|
# Filter input for archives only.
|
|
function _filter_archive() {
|
|
grep -E ''${_archive_pattern}
|
|
}
|
|
|
|
function _archive_memlimit() {
|
|
local mem_free=$(_mem_free)
|
|
local mem_limit=$((mem_free*3/4))
|
|
|
|
echo "''${mem_limit}MiB"
|
|
}
|
|
|
|
function _archive_check() {
|
|
# Extract hash from name.
|
|
local data=($(_archive_parse ''${target}))
|
|
local saved=''${data[2]}
|
|
|
|
# Calculate actual hash.
|
|
local actual=$(pv ''${target} | sha1sum | cut -d\ -f1)
|
|
|
|
# Compare hashes.
|
|
[[ "''${actual}" = "''${saved}" ]] || _error "Archive check failed."
|
|
}
|
|
|
|
# complete -o filenames -F _comp_archive_grep archive_check unarchive archive_rm archive_touch
|
|
# complete -o filenames -F _comp_archive_grep_fast archive_xz
|
|
complete -o filenames -F _comp_archive_name archive_name
|
|
complete -o filenames -F _comp_archive_names archive_prune
|
|
'';
|
|
}
|