first commit
30
bin/.applyTemplate-completion.bash
Executable file
@ -0,0 +1,30 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_applyTemplate_completions () {
|
||||
declare -a options
|
||||
OPTIONS=("-help" "-timestamp")
|
||||
COMPREPLY=()
|
||||
|
||||
local CUR OPTIONS_COUNT=0
|
||||
CUR=${COMP_WORDS[COMP_CWORD]}
|
||||
|
||||
# compte les options déjà utilisé et les retire de la liste des possible
|
||||
for ITEM in ${COMP_WORDS[@]:1}
|
||||
do
|
||||
if [[ " ${OPTIONS[*]} " =~ " ${ITEM} " ]] ; then
|
||||
let "OPTIONS_COUNT++"
|
||||
OPTIONS=(${OPTIONS[@]/${ITEM}})
|
||||
else
|
||||
break
|
||||
fi
|
||||
done
|
||||
|
||||
# si la position est dans des options "ou juste après"
|
||||
((COMP_CWORD <= OPTIONS_COUNT+1)) && [[ "${CUR}" =~ ^- ]] && COMPREPLY=( $(compgen -W "${OPTIONS[*]}" -- "${CUR}" ) ) && return 0
|
||||
|
||||
# si la position est après des options ou que l'on ne commence pas par "-" on cherche des fichiers
|
||||
((COMP_CWORD <= OPTIONS_COUNT+2)) && COMPREPLY=($(compgen -f -- "${CUR}")) && return 0
|
||||
|
||||
return 0
|
||||
}
|
||||
complete -F _applyTemplate_completions applyTemplate.sh
|
382
bin/.commonFunctions.sh
Executable file
@ -0,0 +1,382 @@
|
||||
# commun fonctions for KAZ
|
||||
|
||||
#TODO; toutes les fonctions ci-dessous devraient être commentées
|
||||
|
||||
#KI : françois
|
||||
#KOI : tout un tas de trucs utiles pour la gestion de l'infra kaz (à mettre dans chaque script)
|
||||
#KAN :
|
||||
# maj le 27/01/2024 by FAB: recherche de tous les srv kaz dispo (via le DNS)
|
||||
# maj le 15/04/2024 by FAB: correction getPahekoOrgaList
|
||||
|
||||
# https://wiki.bash-hackers.org/scripting/terminalcodes
|
||||
BOLD='[1m'
|
||||
RED='[0;31m'
|
||||
GREEN='[0;32m'
|
||||
YELLOW='[0;33m'
|
||||
BLUE='[0;34m'
|
||||
MAGENTA='[0;35m'
|
||||
CYAN='[0;36m'
|
||||
NC='[0m' # No Color
|
||||
NL='
|
||||
'
|
||||
|
||||
########################################
|
||||
setKazVars () {
|
||||
# KAZ_ROOT must be set
|
||||
if [ -z "${KAZ_ROOT}" ]; then
|
||||
printKazError "\n\n *** KAZ_ROOT not defined! ***\n"
|
||||
exit
|
||||
fi
|
||||
export KAZ_KEY_DIR="${KAZ_ROOT}/secret"
|
||||
export KAZ_BIN_DIR="${KAZ_ROOT}/bin"
|
||||
export KAZ_CONF_DIR="${KAZ_ROOT}/config"
|
||||
export KAZ_CONF_PROXY_DIR="${KAZ_CONF_DIR}/proxy"
|
||||
export KAZ_COMP_DIR="${KAZ_ROOT}/dockers"
|
||||
export KAZ_STATE_DIR="${KAZ_ROOT}/state"
|
||||
|
||||
export KAZ_GIT_DIR="${KAZ_ROOT}/git"
|
||||
export KAZ_DNLD_DIR="${KAZ_ROOT}/download"
|
||||
export KAZ_DNLD_PAHEKO_DIR="${KAZ_DNLD_DIR}/paheko"
|
||||
|
||||
export APPLY_TMPL=${KAZ_BIN_DIR}/applyTemplate.sh
|
||||
export DOCKERS_ENV="${KAZ_CONF_DIR}/dockers.env"
|
||||
|
||||
export DOCK_LIB="/var/lib/docker"
|
||||
export DOCK_VOL="${DOCK_LIB}/volumes"
|
||||
export DOCK_VOL_PAHEKO_ORGA="${DOCK_LIB}/volumes/paheko_assoUsers/_data/"
|
||||
|
||||
export NAS_VOL="/mnt/disk-nas1/docker/volumes/"
|
||||
}
|
||||
|
||||
########################################
|
||||
|
||||
printKazMsg () {
|
||||
# $1 msg
|
||||
echo -e "${CYAN}${BOLD}$1${NC}"
|
||||
}
|
||||
|
||||
printKazError () {
|
||||
# $1 msb
|
||||
echo -e "${RED}${BOLD}$1${NC}"
|
||||
}
|
||||
|
||||
########################################
|
||||
checkContinue () {
|
||||
local rep
|
||||
while : ; do
|
||||
read -p "Do you want to continue? [yes]" rep
|
||||
case "${rep}" in
|
||||
""|[yYoO]* )
|
||||
break
|
||||
;;
|
||||
[Nn]* )
|
||||
exit
|
||||
;;
|
||||
* )
|
||||
echo "Please answer yes or no."
|
||||
;;
|
||||
esac
|
||||
done
|
||||
}
|
||||
|
||||
checkDockerRunning () {
|
||||
# $1 docker name
|
||||
# $2 service name
|
||||
if ! [[ "$(docker ps -f "name=$1" | grep -w "$1")" ]]; then
|
||||
printKazError "$2 not running... abort"
|
||||
return 1
|
||||
fi
|
||||
return 0
|
||||
}
|
||||
|
||||
########################################
|
||||
testValidIp () {
|
||||
# $1 ip
|
||||
local ip=$1
|
||||
local stat=1
|
||||
|
||||
if [[ $ip =~ ^[0-9]{1,3}\.[0-9]{1,3}\.[0-9]{1,3}\.[0-9]{1,3}$ ]]; then
|
||||
OIFS=$IFS
|
||||
IFS='.'
|
||||
ip=($ip)
|
||||
IFS=$OIFS
|
||||
[[ ${ip[0]} -le 255 && ${ip[1]} -le 255 && ${ip[2]} -le 255 && ${ip[3]} -le 255 ]]
|
||||
stat=$?
|
||||
fi
|
||||
return $stat
|
||||
}
|
||||
|
||||
########################################
|
||||
getValInFile () {
|
||||
# $1 filename
|
||||
# $2 varname
|
||||
grep "^\s*$2\s*=" $1 2>/dev/null | head -1 | sed "s%^\s*$2\s*=\(.*\)$%\1%"
|
||||
}
|
||||
|
||||
getList () {
|
||||
# $1 filename
|
||||
(cat "$1"|sort; echo) | sed -e "s/\(.*\)[ \t]*#.*$/\1/" -e "s/^[ \t]*\(.*\)$/\1/" -e "/^$/d"
|
||||
}
|
||||
|
||||
getPahekoPluginList () {
|
||||
ls "${KAZ_DNLD_PAHEKO_DIR}" | grep -v "paheko-"
|
||||
}
|
||||
|
||||
getPahekoOrgaList () {
|
||||
# ls "${DOCK_VOL_PAHEKO_ORGA}"
|
||||
find ${DOCK_VOL_PAHEKO_ORGA} -mindepth 1 -maxdepth 1 -type d -printf '%f\n' | sort
|
||||
}
|
||||
|
||||
getAvailableComposes () {
|
||||
ls "${KAZ_COMP_DIR}" | grep -v -- "^.*-orga$"
|
||||
}
|
||||
|
||||
getAvailableOrgas () {
|
||||
|
||||
#KI : Fab
|
||||
#KOI : donne la liste de toutes les orgas pour un serveur donné sinon serveur courant
|
||||
#KAN : 27/01/2024
|
||||
|
||||
#en entrée
|
||||
SITE_DST="$1"
|
||||
|
||||
if [ -n "${SITE_DST}" ];then
|
||||
ssh -p 2201 root@${SITE_DST}.${domain} "ls \"${KAZ_COMP_DIR}\" | grep -- \"^.*-orga$\""
|
||||
else
|
||||
ls "${KAZ_COMP_DIR}" | grep -- "^.*-orga$"
|
||||
fi
|
||||
|
||||
}
|
||||
|
||||
getAvailableServices () {
|
||||
local service
|
||||
for service in paheko cloud collabora agora wiki wp; do
|
||||
echo "${service}"
|
||||
done
|
||||
}
|
||||
|
||||
########################################
|
||||
filterInList () {
|
||||
# $* ref list filter
|
||||
# stdin candidats
|
||||
local compose
|
||||
while read compose ; do
|
||||
if [[ " $* " =~ " ${compose} " ]]; then
|
||||
echo ${compose}
|
||||
fi
|
||||
done | sort -u
|
||||
}
|
||||
|
||||
filterNotInList () {
|
||||
# $* ref list filter
|
||||
# stdin candidats
|
||||
local compose
|
||||
while read compose ; do
|
||||
if [[ ! " $* " =~ " ${compose} " ]]; then
|
||||
echo ${compose}
|
||||
fi
|
||||
done | sort -u
|
||||
}
|
||||
|
||||
filterAvailableComposes () {
|
||||
# $* candidats
|
||||
local AVAILABLE_COMPOSES=$(getAvailableComposes;getAvailableOrgas)
|
||||
if [ $# -eq 0 ] ; then
|
||||
echo ${AVAILABLE_COMPOSES}
|
||||
fi
|
||||
local compose
|
||||
for compose in $*
|
||||
do
|
||||
compose=${compose%/}
|
||||
if [[ ! "${NL}${AVAILABLE_COMPOSES}${NL}" =~ "${NL}${compose}${NL}" ]]; then
|
||||
local subst=""
|
||||
for item in ${AVAILABLE_COMPOSES}; do
|
||||
[[ "${item}" =~ "${compose}" ]] && echo ${item} && subst="${subst} ${item}"
|
||||
done
|
||||
if [ -z "${subst}" ] ; then
|
||||
echo "${RED}${BOLD}Unknown compose: ${compose} not in "${AVAILABLE_COMPOSES}"${NC}" >&2
|
||||
#exit 1
|
||||
else
|
||||
echo "${BLUE}${BOLD}substitute compose: ${compose} => "${subst}"${NC}" >&2
|
||||
fi
|
||||
else
|
||||
echo "${compose}"
|
||||
fi
|
||||
done | sort -u
|
||||
}
|
||||
|
||||
########################################
|
||||
serviceOnInOrga () {
|
||||
# $1 orga name
|
||||
# $2 service name
|
||||
# default value
|
||||
local composeFile="${KAZ_COMP_DIR}/$1-orga/docker-compose.yml"
|
||||
if [[ ! -f "${composeFile}" ]]
|
||||
then
|
||||
echo "$3"
|
||||
else
|
||||
grep -q "$2" "${composeFile}" 2>/dev/null && echo on || echo off
|
||||
fi
|
||||
}
|
||||
|
||||
########################################
|
||||
waitUrl () {
|
||||
# $1 URL to waitfor
|
||||
# $2 timeout en secondes (optional)
|
||||
starttime=$(date +%s)
|
||||
if [[ $(curl --connect-timeout 2 -s -D - "$1" -o /dev/null 2>/dev/null | head -n1) != *[23]0[0-9]* ]]; then
|
||||
printKazMsg "service not available ($1). Please wait..."
|
||||
echo curl --connect-timeout 2 -s -D - "$1" -o /dev/null \| head -n1
|
||||
while [[ $(curl --connect-timeout 2 -s -D - "$1" -o /dev/null 2>/dev/null | head -n1) != *[23]0[0-9]* ]]
|
||||
do
|
||||
sleep 5
|
||||
if [ $# -gt 1 ]; then
|
||||
actualtime=$(date +%s)
|
||||
delta=$(($actualtime-$starttime))
|
||||
[[ $2 -lt $delta ]] && return 1
|
||||
fi
|
||||
done
|
||||
fi
|
||||
return 0
|
||||
}
|
||||
|
||||
########################################
|
||||
waitContainerHealthy () {
|
||||
# $1 ContainerName
|
||||
# $2 timeout en secondes (optional)
|
||||
|
||||
healthy="false"
|
||||
starttime=$(date +%s)
|
||||
running="false"
|
||||
[[ $(docker ps -f name="$1" | grep -w "$1") ]] && running="true"
|
||||
[[ $running == "true" && $(docker inspect -f {{.State.Health.Status}} "$1") == "healthy" ]] && healthy="true"
|
||||
if [[ ! $running == "true" || ! $healthy == "true" ]]; then
|
||||
printKazMsg "Docker not healthy ($1). Please wait..."
|
||||
while [[ ! $running == "true" || ! $healthy == "true" ]]
|
||||
do
|
||||
sleep 5
|
||||
if [ $# -gt 1 ]; then
|
||||
actualtime=$(date +%s)
|
||||
delta=$(($actualtime-$starttime))
|
||||
[[ $2 -lt $delta ]] && printKazMsg "Docker not healthy ($1)... abort..." && return 1
|
||||
fi
|
||||
[[ ! $running == "true" ]] && [[ $(docker ps -f name="$1" | grep -w "$1") ]] && running="true"
|
||||
[[ $running == "true" && $(docker inspect -f {{.State.Health.Status}} "$1") == "healthy" ]] && healthy="true"
|
||||
done
|
||||
fi
|
||||
return 0
|
||||
}
|
||||
########################################
|
||||
waitContainerRunning () {
|
||||
# $1 ContainerName
|
||||
# $2 timeout en secondes (optional)
|
||||
|
||||
starttime=$(date +%s)
|
||||
running="false"
|
||||
[[ $(docker ps -f name="$1" | grep -w "$1") ]] && running="true"
|
||||
if [[ ! $running == "true" ]]; then
|
||||
printKazMsg "Docker not running ($1). Please wait..."
|
||||
while [[ ! $running == "true" ]]
|
||||
do
|
||||
sleep 5
|
||||
if [ $# -gt 1 ]; then
|
||||
actualtime=$(date +%s)
|
||||
delta=$(($actualtime-$starttime))
|
||||
[[ $2 -lt $delta ]] && printKazMsg "Docker did not start ($1)... abort..." && return 1
|
||||
fi
|
||||
[[ ! $running == "true" ]] && [[ $(docker ps -f name="$1" | grep -w "$1") ]] && running="true"
|
||||
done
|
||||
fi
|
||||
return 0
|
||||
}
|
||||
|
||||
########################################
|
||||
downloadFile () {
|
||||
# $1 URL to download
|
||||
# $2 new filename (optional)
|
||||
if [ $# -lt 1 ] || [ $# -gt 2 ]; then
|
||||
printKazError "downloadFile: bad arg number"
|
||||
return
|
||||
fi
|
||||
URL=$1
|
||||
if [ -z "$2" ]; then
|
||||
FILENAME="$(basename $1)"
|
||||
else
|
||||
FILENAME="$2"
|
||||
fi
|
||||
|
||||
if [ ! -f "${FILENAME}" ]; then
|
||||
printKazMsg " - load ${URL}"
|
||||
curl -L -o "${FILENAME}" "${URL}"
|
||||
else
|
||||
TMP="${FILENAME}.tmp"
|
||||
rm -f "${TMP}"
|
||||
curl -L -o "${TMP}" "${URL}"
|
||||
if ! cmp -s "${TMP}" "${FILENAME}" 2>/dev/null; then
|
||||
mv "${TMP}" "${FILENAME}"
|
||||
else
|
||||
rm -f "${TMP}"
|
||||
fi
|
||||
fi
|
||||
}
|
||||
|
||||
unzipInDir () {
|
||||
# $1 zipfile
|
||||
# $2 destDir
|
||||
|
||||
if [ $# -ne 2 ]; then
|
||||
printKazError "unzipInDir: bad arg number"
|
||||
return
|
||||
fi
|
||||
if ! [[ $1 == *.zip ]]; then
|
||||
printKazError "unzipInDir: $1 is not a zip file"
|
||||
return
|
||||
fi
|
||||
if ! [[ -d $2 ]]; then
|
||||
printKazError "$2 is not destination dir"
|
||||
return
|
||||
fi
|
||||
|
||||
destName="$2/$(basename "${1%.zip}")"
|
||||
if [[ -d "${destName}" ]]; then
|
||||
printKazError "${destName} already exist"
|
||||
return
|
||||
fi
|
||||
|
||||
tmpDir=$2/tmp-$$
|
||||
trap 'rm -rf "${tmpDir}"' EXIT
|
||||
unzip "$1" -d "${tmpDir}"
|
||||
srcDir=$(ls -1 "${tmpDir}")
|
||||
case $(wc -l <<< $srcDir) in
|
||||
0)
|
||||
printKazError "empty zip file : $1"
|
||||
rmdir "${tmpDir}"
|
||||
return
|
||||
;;
|
||||
1)
|
||||
mv "${tmpDir}/${srcDir}" "${destName}"
|
||||
rmdir "${tmpDir}"
|
||||
;;
|
||||
*)
|
||||
printKazError "zip file $1 is not a tree (${srcDir})"
|
||||
return
|
||||
;;
|
||||
esac
|
||||
}
|
||||
########################################
|
||||
|
||||
get_Serveurs_Kaz () {
|
||||
|
||||
#KI : Fab
|
||||
#KOI : donne la liste de tous les serveurs kaz sous le format srv1;srv2;srv3;.... en intérogeant le DNS
|
||||
#KAN : 27/01/2024
|
||||
|
||||
liste=`dig -t TXT srv.kaz.bzh +short`
|
||||
#on nettoie
|
||||
liste=$(echo "$liste" | sed 's/\;/ /g')
|
||||
liste=$(echo "$liste" | sed 's/\"//g')
|
||||
#renvoi srv1 srv2 srv3 ....
|
||||
echo ${liste}
|
||||
}
|
||||
########################################
|
||||
|
66
bin/.container-completion.bash
Executable file
@ -0,0 +1,66 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_container_completions () {
|
||||
KAZ_ROOT=$(cd "$(dirname ${COMP_WORDS[0]})"/..; pwd)
|
||||
COMPREPLY=()
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]} card=${#COMP_WORDS[@]} i w skip=0
|
||||
for ((i=1 ; i<cword; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
[[ "${w}" == -* ]] && ((skip++))
|
||||
done
|
||||
local arg_pos w i cmd= names=
|
||||
((arg_pos = cword - skip))
|
||||
for ((i=1 ; i<card; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
if [ -z "${cmd}" ] ; then
|
||||
[[ "${w}" == -* ]] || cmd="${w}"
|
||||
continue
|
||||
fi
|
||||
names="${names} ${w}"
|
||||
done
|
||||
|
||||
case "$cur" in
|
||||
-*)
|
||||
COMPREPLY=( $(compgen -W "-h -n" -- "${cur}" ) ) ;;
|
||||
*)
|
||||
local cmd_available="status start stop save"
|
||||
case "${arg_pos}" in
|
||||
1)
|
||||
# $1 of container.sh
|
||||
COMPREPLY=($(compgen -W "${cmd_available}" -- "${cur}"))
|
||||
;;
|
||||
*)
|
||||
# $2-* of container.sh
|
||||
[[ " ${cmd_available} " =~ " ${cmd} " ]] || return 0
|
||||
# select set of names
|
||||
local names_set="available"
|
||||
case "${cmd}" in
|
||||
status)
|
||||
names_set="available"
|
||||
;;
|
||||
start)
|
||||
names_set="disable"
|
||||
;;
|
||||
stop)
|
||||
names_set="enable"
|
||||
;;
|
||||
save)
|
||||
names_set="validate"
|
||||
;;
|
||||
esac
|
||||
local available_args=$("${KAZ_ROOT}/bin/kazList.sh" "compose" "${names_set}")
|
||||
# remove previous selected target
|
||||
local proposal item
|
||||
for item in ${available_args} ; do
|
||||
[[ " ${names} " =~ " ${item} " ]] || proposal="${proposal} ${item}"
|
||||
done
|
||||
COMPREPLY=($(compgen -W "${proposal}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
complete -F _container_completions container.sh
|
19
bin/.dns-completion.bash
Executable file
@ -0,0 +1,19 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_dns_completions () {
|
||||
local cur find
|
||||
COMPREPLY=()
|
||||
cur=${COMP_WORDS[COMP_CWORD]}
|
||||
case "$cur" in
|
||||
-*)
|
||||
COMPREPLY=( $(compgen -W "-h -n -f" -- "${cur}" ) ) ;;
|
||||
*)
|
||||
find=""
|
||||
for arg in ${COMP_WORDS[@]} ; do
|
||||
[[ " list add del " =~ " ${arg} " ]] && find="arg"
|
||||
done
|
||||
[ -z "${find}" ] && COMPREPLY=($(compgen -W "init list add del" -- "${cur}")) ;;
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
complete -F _dns_completions dns.sh
|
79
bin/.foreign-domain-completion.bash
Executable file
@ -0,0 +1,79 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_foreign_domain_completions () {
|
||||
KAZ_ROOT=$(cd "$(dirname ${COMP_WORDS[0]})"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
COMPREPLY=()
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]} card=${#COMP_WORDS[@]} i w skip=0
|
||||
for ((i=1 ; i<cword; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
[[ "${w}" == -* ]] && ((skip++))
|
||||
done
|
||||
local arg_pos w i cmd= opt1= opt2=
|
||||
((arg_pos = cword - skip))
|
||||
for ((i=1 ; i<card; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
if [ -z "${cmd}" ] ; then
|
||||
[[ "${w}" == -* ]] || cmd="${w}"
|
||||
continue
|
||||
fi
|
||||
if [ -z "${opt1}" ] ; then
|
||||
[[ "${w}" == -* ]] || opt1="${w}"
|
||||
continue
|
||||
fi
|
||||
if [ -z "${opt2}" ] ; then
|
||||
[[ "${w}" == -* ]] || opt2="${w}"
|
||||
break
|
||||
fi
|
||||
done
|
||||
|
||||
case "$cur" in
|
||||
-*)
|
||||
COMPREPLY=( $(compgen -W "-h -n" -- "${cur}" ) ) ;;
|
||||
*)
|
||||
local cmd_available="list add del"
|
||||
if [ "${arg_pos}" == 1 ]; then
|
||||
# $1 of foreign-domain.sh .sh
|
||||
COMPREPLY=($(compgen -W "${cmd_available}" -- "${cur}"))
|
||||
else
|
||||
. "${KAZ_CONF_DIR}/dockers.env"
|
||||
case "${cmd}" in
|
||||
|
||||
"list")
|
||||
;;
|
||||
"add")
|
||||
case "${arg_pos}" in
|
||||
2)
|
||||
declare -a availableOrga
|
||||
availableOrga=($(sed -e "s/\(.*\)[ \t]*#.*$/\1/" -e "s/^[ \t]*\(.*\)-orga$/\1/" -e "/^$/d" "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
COMPREPLY=($(compgen -W "${availableOrga[*]}" -- "${cur}"))
|
||||
;;
|
||||
3)
|
||||
local availableComposes=$(${KAZ_COMP_DIR}/${opt1}-orga/orga-gen.sh -l)
|
||||
COMPREPLY=($(compgen -W "${availableComposes[*]}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
;;
|
||||
"del")
|
||||
case "${arg_pos}" in
|
||||
1)
|
||||
;;
|
||||
*)
|
||||
local availableComposes=$(${KAZ_BIN_DIR}/kazList.sh service validate|sed -e "s/\bcollabora\b//" -e "s/ / /")
|
||||
declare -a availableDomaine
|
||||
availableDomaine=($(for compose in ${availableComposes[@]} ; do
|
||||
sed -e "s/[ \t]*\([^#]*\)#.*/\1/g" -e "/^$/d" -e "s/.*server_name[ \t]\([^ ;]*\).*/\1/" "${KAZ_CONF_PROXY_DIR}/${compose}_kaz_name"
|
||||
done))
|
||||
COMPREPLY=($(compgen -W "${availableDomaine[*]}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
complete -F _foreign_domain_completions foreign-domain.sh
|
22
bin/.gestContainers-completion.bash
Normal file
@ -0,0 +1,22 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_gestContainers_completion () {
|
||||
KAZ_ROOT=$(cd "$(dirname ${COMP_WORDS[0]})"/..; pwd)
|
||||
COMPREPLY=()
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]}
|
||||
case "$cur" in
|
||||
-*)
|
||||
local proposal="-h --help -n --simu -q --quiet -m --main -M --nas --local -v --version -l --list -cloud -agora -wp -wiki -office -I --install -r -t -exec --optim -occ -u -i -a -U--upgrade -p --post -mmctl"
|
||||
COMPREPLY=( $(compgen -W "${proposal}" -- "${cur}" ) )
|
||||
;;
|
||||
*)
|
||||
# orga name
|
||||
local available_orga=$("${KAZ_BIN_DIR}/kazList.sh" "service" "enable" 2>/dev/null)
|
||||
COMPREPLY=($(compgen -W "${available_orga}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
complete -F _gestContainers_completion gestContainers.sh
|
51
bin/.kazDockerNet-completion.bash
Executable file
@ -0,0 +1,51 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_kazDockerNet_completion () {
|
||||
KAZ_ROOT=$(cd "$(dirname ${COMP_WORDS[0]})"/..; pwd)
|
||||
COMPREPLY=()
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]} card=${#COMP_WORDS[@]} i w skip=0
|
||||
for ((i=1 ; i<cword; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
[[ "${w}" == -* ]] && ((skip++))
|
||||
done
|
||||
local arg_pos w i cmd= names=
|
||||
((arg_pos = cword - skip))
|
||||
for ((i=1 ; i<card; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
if [ -z "${cmd}" ] ; then
|
||||
[[ "${w}" == -* ]] || cmd="${w}"
|
||||
continue
|
||||
fi
|
||||
names="${names} ${w}"
|
||||
done
|
||||
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]}
|
||||
case "$cur" in
|
||||
-*)
|
||||
COMPREPLY=( $(compgen -W "-h -n" -- "${cur}" ) )
|
||||
;;
|
||||
*)
|
||||
local cmd_available="list add"
|
||||
case "${cword}" in
|
||||
1)
|
||||
COMPREPLY=($(compgen -W "${cmd_available}" -- "${cur}"))
|
||||
;;
|
||||
*)
|
||||
[[ "${cmd}" = "add" ]] || return 0
|
||||
local available_args=$("${KAZ_BIN_DIR}/kazList.sh" "compose" "available" 2>/dev/null)
|
||||
local used=$("${KAZ_BIN_DIR}/kazDockerNet.sh" "list" | grep "name:" | sed -e "s%\bname:\s*%%" -e "s%\bbridge\b\s*%%" -e "s%Net\b%%g")
|
||||
local proposal item
|
||||
for item in ${available_args} ; do
|
||||
[[ " ${names} " =~ " ${item} " ]] || [[ " ${used} " =~ " ${item} " ]] || proposal="${proposal} ${item}"
|
||||
done
|
||||
COMPREPLY=($(compgen -W "${proposal}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
;;
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
complete -F _kazDockerNet_completion kazDockerNet.sh
|
83
bin/.kazList-completion.bash
Executable file
@ -0,0 +1,83 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_kazList_completions () {
|
||||
#KAZ_ROOT=$(cd "$(dirname ${COMP_WORDS[0]})"/..; pwd)
|
||||
COMPREPLY=()
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]} card=${#COMP_WORDS[@]} i w skip=0
|
||||
for ((i=1 ; i<cword; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
[[ "${w}" == -* ]] && ((skip++))
|
||||
done
|
||||
local arg_pos w i cmd= opt= names=
|
||||
((arg_pos = cword - skip))
|
||||
for ((i=1 ; i<card; i++)); do
|
||||
w="${COMP_WORDS[i]}"
|
||||
if [ -z "${cmd}" ]; then
|
||||
[[ "${w}" == -* ]] || cmd="${w}"
|
||||
continue
|
||||
fi
|
||||
if [ -z "${opt}" ]; then
|
||||
[[ "${w}" == -* ]] || opt="${w}"
|
||||
continue
|
||||
fi
|
||||
names="${names} ${w}"
|
||||
done
|
||||
#(echo "A cword:${cword} / arg_pos:${arg_pos} / card:${card} / cur:${cur} / cmd:${cmd} / opt:${opt} / names:${names} " >> /dev/pts/1)
|
||||
|
||||
case "${cur}" in
|
||||
-*)
|
||||
COMPREPLY=($(compgen -W "-h --help" -- "${cur}"))
|
||||
;;
|
||||
*)
|
||||
local cmd_available="compose service"
|
||||
local opt_available="available validate enable disable status"
|
||||
case "${arg_pos}" in
|
||||
1)
|
||||
# $1 of kazList.sh
|
||||
COMPREPLY=($(compgen -W "${cmd_available}" -- "${cur}"))
|
||||
;;
|
||||
2)
|
||||
# $2 of kazList.sh
|
||||
COMPREPLY=($(compgen -W "${opt_available}" -- "${cur}"))
|
||||
;;
|
||||
*)
|
||||
# $3-* of kazList.sh
|
||||
[[ " ${cmd_available} " =~ " ${cmd} " ]] || return 0
|
||||
# select set of names
|
||||
local names_set="${opt}"
|
||||
local available_args
|
||||
case "${cmd}" in
|
||||
service)
|
||||
case "${names_set}" in
|
||||
available|validate)
|
||||
return 0
|
||||
;;
|
||||
*)
|
||||
available_args=$("${COMP_WORDS[0]}" "compose" "enable" "orga" 2>/dev/null)
|
||||
;;
|
||||
esac
|
||||
;;
|
||||
compose)
|
||||
case "${names_set}" in
|
||||
validate|enable|disable)
|
||||
;;
|
||||
*)
|
||||
names_set="available"
|
||||
;;
|
||||
esac
|
||||
available_args=$("${COMP_WORDS[0]}" "${cmd}" "${names_set}")
|
||||
;;
|
||||
esac
|
||||
# remove previous selected target
|
||||
local proposal item
|
||||
for item in ${available_args} ; do
|
||||
[[ " ${names} " =~ " ${item} " ]] || proposal="${proposal} ${item}"
|
||||
done
|
||||
COMPREPLY=($(compgen -W "${proposal}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
|
||||
complete -F _kazList_completions kazList.sh
|
39
bin/.mvOrga2Nas-completion.bash
Executable file
@ -0,0 +1,39 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_mv_orga_nas_completion () {
|
||||
KAZ_ROOT=$(cd "$(dirname ${COMP_WORDS[0]})"/..; pwd)
|
||||
COMPREPLY=()
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]} card=${#COMP_WORDS[@]} i w skip=0
|
||||
for ((i=1 ; i<cword; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
[[ "${w}" == -* ]] && ((skip++))
|
||||
done
|
||||
local arg_pos w i names=
|
||||
((arg_pos = cword - skip))
|
||||
for ((i=1 ; i<card; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
if [[ "${w}" == -* ]]; then
|
||||
continue
|
||||
fi
|
||||
names="${names} ${w}"
|
||||
done
|
||||
local KAZ_LIST="${KAZ_BIN_DIR}/kazList.sh"
|
||||
case "$cur" in
|
||||
-*)
|
||||
local proposal="-h -n"
|
||||
COMPREPLY=( $(compgen -W "${proposal}" -- "${cur}" ) )
|
||||
;;
|
||||
*)
|
||||
local available_orga=$("${KAZ_LIST}" "compose" "enable" "orga" 2>/dev/null | sed "s/-orga\b//g")
|
||||
local proposal= item
|
||||
for item in ${available_orga} ; do
|
||||
[[ " ${names} " =~ " ${item} " ]] || proposal="${proposal} ${item}"
|
||||
done
|
||||
COMPREPLY=($(compgen -W "${proposal}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
complete -F _mv_orga_nas_completion mvOrga2Nas.sh
|
63
bin/.orga-gen-completion.bash
Executable file
@ -0,0 +1,63 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_orga_gen_completion () {
|
||||
KAZ_ROOT=$(cd "$(dirname ${COMP_WORDS[0]})"/../..; pwd)
|
||||
ORGA_DIR=$(cd "$(dirname ${COMP_WORDS[0]})"; basename $(pwd))
|
||||
COMPREPLY=()
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
local cword=${COMP_CWORD} cur=${COMP_WORDS[COMP_CWORD]} card=${#COMP_WORDS[@]} i w skip=0
|
||||
for ((i=1 ; i<cword; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
[[ "${w}" == -* ]] && ((skip++))
|
||||
[[ "${w}" == +* ]] && ((skip++))
|
||||
done
|
||||
local arg_pos w i addOpt= rmOpt= names=
|
||||
((arg_pos = cword - skip))
|
||||
for ((i=1 ; i<card; i++)) ; do
|
||||
w="${COMP_WORDS[i]}"
|
||||
if [[ "${w}" == -* ]]; then
|
||||
rmOpt="${rmOpt} ${w}"
|
||||
continue
|
||||
fi
|
||||
if [[ "${w}" == '+'* ]]; then
|
||||
addOpt="${addOpt} ${w}"
|
||||
continue
|
||||
fi
|
||||
names="${names} ${w}"
|
||||
done
|
||||
local KAZ_LIST="${KAZ_BIN_DIR}/kazList.sh"
|
||||
case "$cur" in
|
||||
-*)
|
||||
local available_services item proposal="-h -l" listOpt="available"
|
||||
[ -n "${names}" ] && listOpt="enable ${names}"
|
||||
[[ "${ORGA_DIR}" = "orgaTmpl" ]] || listOpt="enable ${ORGA_DIR%-orga}"
|
||||
available_services=$("${KAZ_LIST}" service ${listOpt} 2>/dev/null | tr ' ' '\n' | sed "s/\(..*\)/-\1/")
|
||||
for item in ${available_services} ; do
|
||||
[[ " ${rmOpt} " =~ " ${item} " ]] || proposal="${proposal} ${item}"
|
||||
done
|
||||
COMPREPLY=( $(compgen -W "${proposal}" -- "${cur}" ) )
|
||||
;;
|
||||
'+'*)
|
||||
local available_services item proposal= listOpt="available"
|
||||
[ -n "${names}" ] && listOpt="disable ${names}"
|
||||
[[ "${ORGA_DIR}" = "orgaTmpl" ]] || listOpt="disable ${ORGA_DIR%-orga}"
|
||||
available_services=$("${KAZ_LIST}" service ${listOpt} 2>/dev/null | tr ' ' '\n' | sed "s/\(..*\)/+\1/")
|
||||
for item in ${available_services} ; do
|
||||
[[ " ${addOpt} " =~ " ${item} " ]] || proposal="${proposal} ${item}"
|
||||
done
|
||||
COMPREPLY=( $(compgen -W "${proposal}" -- "${cur}" ) )
|
||||
;;
|
||||
*)
|
||||
[[ "${ORGA_DIR}" = "orgaTmpl" ]] || return 0;
|
||||
local available_orga=$("${KAZ_LIST}" "compose" "enable" "orga" 2>/dev/null | sed "s/-orga\b//g")
|
||||
local proposal= item
|
||||
for item in ${available_orga} ; do
|
||||
[[ " ${names} " =~ " ${item} " ]] || proposal="${proposal} ${item}"
|
||||
done
|
||||
COMPREPLY=($(compgen -W "${proposal}" -- "${cur}"))
|
||||
;;
|
||||
esac
|
||||
return 0
|
||||
}
|
||||
complete -F _orga_gen_completion orga-gen.sh
|
11
bin/.updateLook-completion.bash
Executable file
@ -0,0 +1,11 @@
|
||||
#/usr/bin/env bash
|
||||
|
||||
_update_look_nas_completion () {
|
||||
COMPREPLY=()
|
||||
local cur=${COMP_WORDS[COMP_CWORD]}
|
||||
local THEMES=$(cd "$(dirname ${COMP_WORDS[0]})"/look ; ls -F -- '.' | grep '/$' | sed 's%/%%' | tr '\n' ' ' | sed 's% $%%')
|
||||
|
||||
COMPREPLY=($(compgen -W "${THEMES}" -- "${cur}"))
|
||||
return 0
|
||||
}
|
||||
complete -F _update_look_nas_completion updateLook.sh
|
102
bin/applyTemplate.sh
Executable file
@ -0,0 +1,102 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Met à jour la configuration de ${CONF} en fonction du modèle ${TMPL}
|
||||
# Viariables misent à jour :
|
||||
# - __DOMAIN__
|
||||
# Il est possible de prendre en considération ou d'occulter des blocks.
|
||||
# Le début du block est repéré par une ligne contenant {{XXX
|
||||
# La fin du block est repéré par une ligne contenant }}
|
||||
# L'affiche est fonction de XXX
|
||||
# XXX = on => affichage systématique
|
||||
# XXX = off => masquage systématique
|
||||
# XXX = compose => affichage si la variable d'environnement proxy_compose à la valeur on
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
. "${DOCKERS_ENV}"
|
||||
. "${KAZ_KEY_DIR}/SetAllPass.sh"
|
||||
|
||||
usage () {
|
||||
echo $(basename "$0") " [-h] [-help] [-timestamp] template dst"
|
||||
echo " -h"
|
||||
echo " -help Display this help."
|
||||
echo " -timestamp produce timestamp comment."
|
||||
}
|
||||
|
||||
TIMESTAMP=""
|
||||
|
||||
case "$1" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
shift
|
||||
exit;;
|
||||
'-time' | '-timestamp' )
|
||||
TIMESTAMP=YES
|
||||
shift;;
|
||||
esac
|
||||
|
||||
# no more export in .env
|
||||
PROXY_VARS=$(set | grep "proxy_.*=")
|
||||
for var in ${PROXY_VARS}
|
||||
do
|
||||
export ${var}
|
||||
done
|
||||
|
||||
(
|
||||
# $1 = template
|
||||
# $2 = target
|
||||
if [ "${TIMESTAMP}" == "YES" ]; then
|
||||
echo "# Generated by $(pwd)$(basename $0)"
|
||||
echo "# à partir du modèle $1"
|
||||
echo "#" $(date "+%x %X")
|
||||
echo
|
||||
fi
|
||||
|
||||
sed \
|
||||
-e "/^[ \t]*$/d"\
|
||||
-e "/^[ ]*#.*$/d"\
|
||||
-e "s|__CACHET_HOST__|${cachetHost}|g"\
|
||||
-e "s|__CALC_HOST__|${calcHost}|g"\
|
||||
-e "s|__CLOUD_HOST__|${cloudHost}|g"\
|
||||
-e "s|__DATE_HOST__|${dateHost}|g"\
|
||||
-e "s|__DOKUWIKI_HOST__|${dokuwikiHost}|g"\
|
||||
-e "s|__DOMAIN__|${domain}|g"\
|
||||
-e "s|__FILE_HOST__|${fileHost}|g"\
|
||||
-e "s|__PAHEKO_API_PASSWORD__|${paheko_API_PASSWORD}|g"\
|
||||
-e "s|__PAHEKO_API_USER__|${paheko_API_USER}|g"\
|
||||
-e "s|__PAHEKO_HOST__|${pahekoHost}|g"\
|
||||
-e "s|__GIT_HOST__|${gitHost}|g"\
|
||||
-e "s|__GRAV_HOST__|${gravHost}|g"\
|
||||
-e "s|__HTTP_PROTO__|${httpProto}|g"\
|
||||
-e "s|__LDAP_HOST__|${ldapHost}|g"\
|
||||
-e "s|__LDAPUI_HOST__|${ldapUIHost}|g"\
|
||||
-e "s|__MATTER_HOST__|${matterHost}|g"\
|
||||
-e "s|__OFFICE_HOST__|${officeHost}|g"\
|
||||
-e "s|__PAD_HOST__|${padHost}|g"\
|
||||
-e "s|__QUOTAS_HOST__|${quotasHost}|g"\
|
||||
-e "s|__SMTP_HOST__|${smtpHost}|g"\
|
||||
-e "s|__SYMPADB__|${sympaDBName}|g"\
|
||||
-e "s|__SYMPA_HOST__|${sympaHost}|g"\
|
||||
-e "s|__SYMPA_MYSQL_DATABASE__|${sympa_MYSQL_DATABASE}|g"\
|
||||
-e "s|__SYMPA_MYSQL_PASSWORD__|${sympa_MYSQL_PASSWORD}|g"\
|
||||
-e "s|__SYMPA_MYSQL_USER__|${sympa_MYSQL_USER}|g"\
|
||||
-e "s|__VIGILO_HOST__|${vigiloHost}|g"\
|
||||
-e "s|__WEBMAIL_HOST__|${webmailHost}|g"\
|
||||
-e "s|__CASTOPOD_HOST__|${castopodHost}|g"\
|
||||
-e "s|__IMAPSYNC_HOST__|${imapsyncHost}|g"\
|
||||
-e "s|__YAKFORMS_HOST__|${yakformsHost}|g"\
|
||||
-e "s|__WORDPRESS_HOST__|${wordpressHost}|g"\
|
||||
-e "s|__MOBILIZON_HOST__|${mobilizonHost}|g"\
|
||||
-e "s|__API_HOST__|${apiHost}|g"\
|
||||
-e "s|__VAULTWARDEN_HOST__|${vaultwardenHost}|g"\
|
||||
-e "s|__DOMAIN_SYMPA__|${domain_sympa}|g"\
|
||||
$1 | awk '
|
||||
BEGIN {cp=1}
|
||||
/}}/ {cp=1 ; next};
|
||||
/{{on/ {cp=1; next};
|
||||
/{{off/ {cp=0; next};
|
||||
match($0, /{{[a-zA-Z0-9_]+/) {cp=(ENVIRON["proxy_" substr($0,RSTART+2,RLENGTH)] == "on"); next};
|
||||
{if (cp) print $0};'
|
||||
) > $2
|
315
bin/checkEnvFiles.sh
Executable file
@ -0,0 +1,315 @@
|
||||
#!/bin/bash
|
||||
|
||||
export KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
RUN_PASS_DIR="secret"
|
||||
TMPL_PASS_DIR="secret.tmpl"
|
||||
RUN_PASS_FILE="${RUN_PASS_DIR}/SetAllPass.sh"
|
||||
TMPL_PASS_FILE="${TMPL_PASS_DIR}/SetAllPass.sh"
|
||||
NEED_GEN=
|
||||
|
||||
########################################
|
||||
|
||||
usage () {
|
||||
echo "Usage: $0 [-n] [-h]"
|
||||
echo " -h help"
|
||||
exit 1
|
||||
}
|
||||
|
||||
case "$1" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
|
||||
[ "$#" -eq 0 ] || usage
|
||||
|
||||
########################################
|
||||
# check system
|
||||
|
||||
for prg in kompare; do
|
||||
if ! type "${prg}" > /dev/null; then
|
||||
printKazError "$0 need ${prg}"
|
||||
echo "please run \"apt-get install ${prg}\""
|
||||
exit
|
||||
fi
|
||||
done
|
||||
|
||||
cd "${KAZ_ROOT}"
|
||||
########################################
|
||||
# get lvalues in script
|
||||
getVars () {
|
||||
# $1 : filename
|
||||
grep "^[^#]*=" $1 | sed 's/\([^=]*\).*/\1/' | sort -u
|
||||
}
|
||||
|
||||
# get lvalues in script
|
||||
getSettedVars () {
|
||||
# $1 : filename
|
||||
grep "^[^#]*=..*" $1 | grep -v '^[^#]*=".*--clean_val--.*"' | grep -v '^[^#]*="${' | sort -u
|
||||
}
|
||||
|
||||
getVarFormVal () {
|
||||
# $1 searched value
|
||||
# $2 filename
|
||||
grep "^[^#]*=$1" $2 | sed 's/\s*\([^=]*\).*/\1/'
|
||||
}
|
||||
|
||||
########################################
|
||||
# synchronized SetAllPass.sh (find missing lvalues)
|
||||
updatePassFile () {
|
||||
# $1 : ref filename
|
||||
# $2 : target filename
|
||||
|
||||
REF_FILE="$1"
|
||||
TARGET_FILE="$2"
|
||||
NEED_UPDATE=
|
||||
while : ; do
|
||||
declare -a listRef listTarget missing
|
||||
listRef=($(getVars "${REF_FILE}"))
|
||||
listTarget=($(getVars "${TARGET_FILE}"))
|
||||
missing=($(comm -23 <(printf "%s\n" ${listRef[@]}) <(printf "%s\n" ${listTarget[@]})))
|
||||
if [ -n "${missing}" ]; then
|
||||
echo "missing vars in ${YELLOW}${BOLD}${TARGET_FILE}${NC}:${RED}${BOLD}" ${missing[@]} "${NC}"
|
||||
read -p "Do you want to add them? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
emacs "${REF_FILE}" "${TARGET_FILE}"
|
||||
NEED_UPDATE=true
|
||||
break
|
||||
;;
|
||||
[Nn]*)
|
||||
break
|
||||
;;
|
||||
esac
|
||||
else
|
||||
break
|
||||
fi
|
||||
done
|
||||
}
|
||||
|
||||
updatePassFile "${TMPL_PASS_FILE}" "${RUN_PASS_FILE}"
|
||||
[ -n "${NEED_UPDATE}" ] && NEED_GEN=true
|
||||
updatePassFile "${RUN_PASS_FILE}" "${TMPL_PASS_FILE}"
|
||||
|
||||
########################################
|
||||
# check empty pass in TMPL_PASS_FILE
|
||||
declare -a settedVars
|
||||
settedVars=($(getSettedVars "${TMPL_PASS_FILE}"))
|
||||
if [ -n "${settedVars}" ]; then
|
||||
echo "unclear password in ${YELLOW}${BOLD}${TMPL_PASS_FILE}${NC}:${BLUE}${BOLD}"
|
||||
for var in ${settedVars[@]}; do
|
||||
echo -e "\t${var}"
|
||||
done
|
||||
echo "${NC}"
|
||||
read -p "Do you want to clear them? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
emacs "${TMPL_PASS_FILE}"
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
|
||||
########################################
|
||||
# check new files env-*
|
||||
createMissingEnv () {
|
||||
# $1 : ref dir
|
||||
# $2 : target dir
|
||||
REF_DIR="$1"
|
||||
TARGET_DIR="$2"
|
||||
NEED_UPDATE=
|
||||
|
||||
declare -a listRef listTarget missing
|
||||
listRef=($(cd "${REF_DIR}"; ls -1 env-* | grep -v '~$'))
|
||||
listTarget=($(cd "${TARGET_DIR}"; ls -1 env-* | grep -v '~$'))
|
||||
missing=($(comm -23 <(printf "%s\n" ${listRef[@]}) <(printf "%s\n" ${listTarget[@]})))
|
||||
for envFile in ${missing[@]}; do
|
||||
read -p "Do you want to create ${GREEN}${BOLD}${TARGET_DIR}/${envFile}${NC}? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
cp "${REF_DIR}/${envFile}" "${TARGET_DIR}/${envFile}"
|
||||
NEED_UPDATE=true
|
||||
;;
|
||||
esac
|
||||
done
|
||||
}
|
||||
|
||||
createMissingEnv "${RUN_PASS_DIR}" "${TMPL_PASS_DIR}"
|
||||
[ -n "${NEED_UPDATE}" ] && NEED_GEN=true
|
||||
createMissingEnv "${TMPL_PASS_DIR}" "${RUN_PASS_DIR}"
|
||||
[ -n "${NEED_UPDATE}" ] && NEED_GEN=true
|
||||
|
||||
########################################
|
||||
# check missing values in env-* between RUN and TMPL
|
||||
declare -a listTmpl listRun listCommonFiles
|
||||
listTmplFiles=($(cd "${TMPL_PASS_DIR}"; ls -1 env-* | grep -v '~$'))
|
||||
listRunFiles=($(cd "${RUN_PASS_DIR}"; ls -1 env-* | grep -v '~$'))
|
||||
listCommonFiles=($(comm -3 <(printf "%s\n" ${listTmplFiles[@]}) <(printf "%s\n" ${listRunFiles[@]})))
|
||||
for envFile in ${listCommonFiles[@]}; do
|
||||
while : ; do
|
||||
TMPL_FILE="${TMPL_PASS_DIR}/${envFile}"
|
||||
RUN_FILE="${RUN_PASS_DIR}/${envFile}"
|
||||
declare -a listRef list2Target missingInRun missingInTmpl
|
||||
listTmplVars=($(getVars "${TMPL_FILE}"))
|
||||
listRunVars=($(getVars "${RUN_FILE}"))
|
||||
missingInTmpl=($(comm -23 <(printf "%s\n" ${listTmplVars[@]}) <(printf "%s\n" ${listRunVars[@]})))
|
||||
missingInRun=($(comm -13 <(printf "%s\n" ${listTmplVars[@]}) <(printf "%s\n" ${listRunVars[@]})))
|
||||
if [ -n "${missingInTmpl}" ] || [ -n "${missingInRun}" ]; then
|
||||
[ -n "${missingInTmpl}" ] &&
|
||||
echo "missing vars in ${YELLOW}${BOLD}${TMPL_FILE}${NC}:${RED}${BOLD}" ${missingInTmpl[@]} "${NC}"
|
||||
[ -n "${missingInRun}" ] &&
|
||||
echo "missing vars in ${YELLOW}${BOLD}${RUN_FILE}${NC}:${RED}${BOLD}" ${missingInRun[@]} "${NC}"
|
||||
read -p "Do you want to add them? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
emacs "${TMPL_FILE}" "${RUN_FILE}"
|
||||
[ -n "${missingInTmpl}" ] && NEED_GEN=true
|
||||
break
|
||||
;;
|
||||
[Nn]*)
|
||||
break
|
||||
;;
|
||||
esac
|
||||
else
|
||||
break
|
||||
fi
|
||||
done
|
||||
done
|
||||
|
||||
########################################
|
||||
# check empty pass in env-*
|
||||
for envFile in $(ls -1 "${TMPL_PASS_DIR}/"env-* | grep -v '~$'); do
|
||||
settedVars=($(getSettedVars "${envFile}"))
|
||||
if [ -n "${settedVars}" ]; then
|
||||
echo "unclear password in ${GREEN}${BOLD}${envFile}${NC}:${BLUE}${BOLD}"
|
||||
for var in ${settedVars[@]}; do
|
||||
echo -e "\t${var}"
|
||||
done
|
||||
echo "${NC}"
|
||||
read -p "Do you want to clear them? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
emacs "${envFile}"
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
done
|
||||
|
||||
########################################
|
||||
# check extention in dockers.env
|
||||
declare -a missing
|
||||
missing=($(for DIR in "${RUN_PASS_DIR}" "${TMPL_PASS_DIR}"; do
|
||||
for envFile in $(ls -1 "${DIR}/"env-* | grep -v '~$'); do
|
||||
val="${envFile#*env-}"
|
||||
varName=$(getVarFormVal "${val}" "${DOCKERS_ENV}")
|
||||
if [ -z "${varName}" ]; then
|
||||
echo "${val}"
|
||||
fi
|
||||
done
|
||||
done | sort -u))
|
||||
if [ -n "${missing}" ]; then
|
||||
echo "missing def in ${GREEN}${BOLD}${DOCKERS_ENV}${NC}:${BLUE}${BOLD}"
|
||||
for var in ${missing[@]}; do
|
||||
echo -e "\t${var}"
|
||||
done
|
||||
echo "${NC}"
|
||||
read -p "Do you want to add them? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
emacs "${DOCKERS_ENV}"
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
|
||||
########################################
|
||||
# check env-* in updateDockerPassword.sh
|
||||
missing=($(for DIR in "${RUN_PASS_DIR}" "${TMPL_PASS_DIR}"; do
|
||||
for envFile in $(ls -1 "${DIR}/"env-* | grep -v '~$'); do
|
||||
val="${envFile#*env-}"
|
||||
varName=$(getVarFormVal "${val}" "${DOCKERS_ENV}")
|
||||
[ -z "${varName}" ] && continue
|
||||
prefixe=$(grep "^\s*updateEnv.*${varName}" "${KAZ_BIN_DIR}/updateDockerPassword.sh" |
|
||||
sed 's/\s*updateEnv[^"]*"\([^"]*\)".*/\1/' | sort -u)
|
||||
if [ -z "${prefixe}" ]; then
|
||||
echo "${envFile#*/}_(\${KAZ_KEY_DIR}/env-\${"${varName}"})"
|
||||
fi
|
||||
done
|
||||
done | sort -u))
|
||||
if [ -n "${missing}" ]; then
|
||||
echo "missing update in ${GREEN}${BOLD}${KAZ_BIN_DIR}/updateDockerPassword.sh${NC}:${BLUE}${BOLD}"
|
||||
for var in ${missing[@]}; do
|
||||
echo -e "\t${var}"
|
||||
done
|
||||
echo "${NC}"
|
||||
read -p "Do you want to add them? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
emacs "${KAZ_BIN_DIR}/updateDockerPassword.sh"
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
|
||||
########################################
|
||||
# synchronized SetAllPass.sh and env-*
|
||||
updateEnvFiles () {
|
||||
# $1 secret dir
|
||||
DIR=$1
|
||||
listRef=($(getVars "${DIR}/SetAllPass.sh"))
|
||||
missing=($(for envFile in $(ls -1 "${DIR}/"env-* | grep -v '~$'); do
|
||||
val="${envFile#*env-}"
|
||||
varName=$(getVarFormVal "${val}" "${DOCKERS_ENV}")
|
||||
[ -z "${varName}" ] && continue
|
||||
prefixe=$(grep "^\s*updateEnv.*${varName}" "${KAZ_BIN_DIR}/updateDockerPassword.sh" |
|
||||
sed 's/\s*updateEnv[^"]*"\([^"]*\)".*/\1/' | sort -u)
|
||||
[ -z "${prefixe}" ] && continue
|
||||
listVarsInEnv=($(getVars "${envFile}"))
|
||||
for var in ${listVarsInEnv[@]}; do
|
||||
[[ ! " ${listRef[@]} " =~ " ${prefixe}_${var} " ]] && echo "${prefixe}_${var}"
|
||||
done
|
||||
# XXX doit exister dans SetAllPass.sh avec le prefixe
|
||||
done))
|
||||
if [ -n "${missing}" ]; then
|
||||
echo "missing update in ${GREEN}${BOLD}${DIR}/SetAllPass.sh${NC}:${BLUE}${BOLD}"
|
||||
for var in ${missing[@]}; do
|
||||
echo -e "\t${var}"
|
||||
done
|
||||
echo "${NC}"
|
||||
read -p "Do you want to add them? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
emacs "${DIR}/SetAllPass.sh"
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
}
|
||||
|
||||
updateEnvFiles "${RUN_PASS_DIR}"
|
||||
updateEnvFiles "${TMPL_PASS_DIR}"
|
||||
|
||||
# XXX chercher les variables non utilisées dans les SetAllPass.sh
|
||||
|
||||
if [ -n "${NEED_GEN}" ]; then
|
||||
while : ; do
|
||||
read -p "Do you want to generate blank values? [y/n]: " yn
|
||||
case $yn in
|
||||
""|[Yy]*)
|
||||
"${KAZ_BIN_DIR}/secretGen.sh"
|
||||
break
|
||||
;;
|
||||
[Nn]*)
|
||||
break
|
||||
;;
|
||||
esac
|
||||
done
|
||||
fi
|
||||
|
||||
|
||||
# XXX config/dockers.tmpl.env
|
||||
|
||||
|
||||
# XXX ! vérifier init pour dockers.env
|
||||
|
||||
|
||||
|
11
bin/checkEnvPassword.sh
Executable file
@ -0,0 +1,11 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd $(dirname $0)/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
for filename in "${KAZ_KEY_DIR}/"env-*Serv "${KAZ_KEY_DIR}/"env-*DB; do
|
||||
if grep -q "^[^#=]*=\s*$" "${filename}" 2>/dev/null; then
|
||||
echo "${filename}"
|
||||
fi
|
||||
done
|
24
bin/configKaz.sh
Executable file
@ -0,0 +1,24 @@
|
||||
#!/bin/sh -e
|
||||
|
||||
. /usr/share/debconf/confmodule
|
||||
db_version 2.0
|
||||
|
||||
if [ "$1" = "fset" ]; then
|
||||
db_fset kaz/mode seen false
|
||||
db_fset kaz/domain seen false
|
||||
db_go
|
||||
fi
|
||||
if [ "$1" = "reset" ]; then
|
||||
db_reset kaz/mode
|
||||
db_reset kaz/domain
|
||||
db_go
|
||||
fi
|
||||
|
||||
#db_set kaz/domain test
|
||||
|
||||
|
||||
db_title "a b c"
|
||||
|
||||
db_input critical kaz/mode
|
||||
db_input critical kaz/domain
|
||||
db_go
|
11
bin/configKaz.sh.templates
Executable file
@ -0,0 +1,11 @@
|
||||
Template: kaz/mode
|
||||
Type: select
|
||||
Choices: prod, dev, local
|
||||
Default: local
|
||||
Description: Mode
|
||||
|
||||
Template: kaz/domain
|
||||
Type: string
|
||||
Description: domain name
|
||||
Default: kaz.bzh
|
||||
|
319
bin/container.sh
Executable file
@ -0,0 +1,319 @@
|
||||
#!/bin/bash
|
||||
|
||||
# En cas d'absence de postfix, il faut lancer :
|
||||
# docker network create postfix_mailNet
|
||||
|
||||
# démare/arrête un compose
|
||||
# sauvegarde la base de données d'un compose
|
||||
# met à jours les paramètres de configuration du mandataire (proxy)
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
cd "${KAZ_BIN_DIR}"
|
||||
PATH_SAUVE="/home/sauve/"
|
||||
export SIMU=""
|
||||
|
||||
declare -a availableComposesNoNeedMail availableMailComposes availableComposesNeedMail availableProxyComposes availableOrga
|
||||
availableComposesNoNeedMail=($(getList "${KAZ_CONF_DIR}/container-withoutMail.list"))
|
||||
availableMailComposes=($(getList "${KAZ_CONF_DIR}/container-mail.list"))
|
||||
availableComposesNeedMail=($(getList "${KAZ_CONF_DIR}/container-withMail.list"))
|
||||
availableProxyComposes=($(getList "${KAZ_CONF_DIR}/container-proxy.list"))
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
availableComposesNeedMail+=( "${availableOrga[@]}" )
|
||||
|
||||
knownedComposes+=( ${availableMailComposes[@]} )
|
||||
knownedComposes+=( ${availableProxyComposes[@]} )
|
||||
knownedComposes+=( ${availableComposesNoNeedMail[@]} )
|
||||
knownedComposes+=( ${availableComposesNeedMail[@]} )
|
||||
|
||||
usage () {
|
||||
echo "Usage: $0 [-n] {status|start|stop|save} [compose]..."
|
||||
echo " -n : simulation"
|
||||
echo " status : docker-compose status (default all compose available)"
|
||||
echo " start : start composes (default all compose validate)"
|
||||
echo " stop : stop composes (default all compose enable)"
|
||||
echo " save : save all known database"
|
||||
echo " [compose] : in ${knownedComposes[@]}"
|
||||
exit 1
|
||||
}
|
||||
|
||||
doCompose () {
|
||||
# $1 dans ("up -d" "down")
|
||||
# $2 nom du répertoire du compose
|
||||
echo "compose: $1 $2"
|
||||
${SIMU} cd "${KAZ_COMP_DIR}/$2"
|
||||
if [ ! -h .env ] ; then
|
||||
echo "create .env in $2"
|
||||
${SIMU} ln -fs ../../config/dockers.env .env
|
||||
fi
|
||||
${SIMU} docker-compose $1
|
||||
|
||||
if [ "$2" = "cachet" ] && [ "$1" != "down" ]; then
|
||||
NEW_KEY=$(cd "${KAZ_COMP_DIR}/$2" ; docker-compose logs | grep APP_KEY=base64: | sed "s/^.*'APP_KEY=\(base64:[^']*\)'.*$/\1/" | tail -1)
|
||||
if [ -n "${NEW_KEY}" ]; then
|
||||
printKazMsg "cachet key change"
|
||||
# change key
|
||||
${SIMU} sed -i \
|
||||
-e 's%^\(\s*cachet_APP_KEY=\).*$%\1"'"${NEW_KEY}"'"%' \
|
||||
"${KAZ_KEY_DIR}/SetAllPass.sh"
|
||||
${SIMU} "${KAZ_BIN_DIR}/secretGen.sh"
|
||||
# restart
|
||||
${SIMU} docker-compose $1
|
||||
fi
|
||||
fi
|
||||
}
|
||||
|
||||
doComposes () {
|
||||
# $1 dans ("up -d" "down")
|
||||
# $2+ nom des répertoires des composes
|
||||
cmd=$1
|
||||
shift
|
||||
for compose in $@ ; do
|
||||
doCompose "${cmd}" ${compose}
|
||||
done
|
||||
}
|
||||
|
||||
updateProxy () {
|
||||
# $1 dans ("on" "off")
|
||||
# $2 nom des répertoires des composes
|
||||
cmd=$1
|
||||
shift
|
||||
echo "update proxy ${cmd}: $@"
|
||||
date=$(date "+%x %X")
|
||||
for compose in $@ ; do
|
||||
composeFlag=${compose//-/_}
|
||||
entry="proxy_${composeFlag}="
|
||||
newline="${entry}${cmd} # update by $(basename $0) at ${date}"
|
||||
if ! grep -q "proxy_${composeFlag}=" "${DOCKERS_ENV}" 2> /dev/null ; then
|
||||
if [[ -n "${SIMU}" ]] ; then
|
||||
echo "${newline} >> ${DOCKERS_ENV}"
|
||||
else
|
||||
echo "${newline}" >> "${DOCKERS_ENV}"
|
||||
fi
|
||||
else
|
||||
${SIMU} sed -i \
|
||||
-e "s|${entry}.*|${newline}|g" \
|
||||
"${DOCKERS_ENV}"
|
||||
fi
|
||||
done
|
||||
for item in "${availableProxyComposes[@]}"; do
|
||||
${SIMU} ${KAZ_COMP_DIR}/${item}/proxy-gen.sh
|
||||
done
|
||||
}
|
||||
|
||||
saveDB () {
|
||||
#attention, soucis avec l'option "-ti" qui ne semble pas rendre la main avec docker exec
|
||||
|
||||
containerName=$1
|
||||
userName=$2
|
||||
userPass=$3
|
||||
dbName=$4
|
||||
backName=$5
|
||||
if [[ -n "${SIMU}" ]] ; then
|
||||
${SIMU} "docker exec ${containerName} mysqldump --user=${userName} --password=${userPass} ${dbName} | gzip > $PATH_SAUVE${backName}.sql.gz"
|
||||
else
|
||||
docker exec ${containerName} mysqldump --user=${userName} --password=${userPass} ${dbName} | gzip > $PATH_SAUVE${backName}.sql.gz
|
||||
fi
|
||||
}
|
||||
|
||||
declare -a enableComposesNoNeedMail enableMailComposes enableComposesNeedMail enableProxyComposes
|
||||
|
||||
enableComposesNoNeedMail=()
|
||||
enableMailComposes=()
|
||||
enableComposesNeedMail=()
|
||||
enableProxyComposes=()
|
||||
|
||||
startComposes () {
|
||||
./kazDockerNet.sh add ${enableComposesNoNeedMail[@]} ${enableProxyComposes[@]} ${enableMailComposes[@]} ${enableComposesNeedMail[@]}
|
||||
[ ${#enableComposesNeedMail[@]} -ne 0 ] && [[ ! "${enableMailComposes[@]}" =~ "postfix" ]] && ./kazDockerNet.sh add postfix
|
||||
[[ "${enableComposesNeedMail[@]}" =~ "paheko" ]] && ${SIMU} ${KAZ_COMP_DIR}/paheko/paheko-gen.sh
|
||||
doComposes "up -d" ${enableComposesNoNeedMail[@]}
|
||||
doComposes "up -d" ${enableMailComposes[@]}
|
||||
doComposes "up -d" ${enableComposesNeedMail[@]}
|
||||
updateProxy "on" ${enableComposesNoNeedMail[@]} ${enableComposesNeedMail[@]}
|
||||
doComposes "up -d" ${enableProxyComposes[@]}
|
||||
for item in "${enableProxyComposes[@]}"; do
|
||||
${SIMU} ${KAZ_COMP_DIR}/${item}/reload.sh
|
||||
done
|
||||
if grep -q "^.s*proxy_web.s*=.s*on" "${DOCKERS_ENV}" 2> /dev/null ; then
|
||||
${SIMU} ${KAZ_COMP_DIR}/web/web-gen.sh
|
||||
fi
|
||||
}
|
||||
|
||||
stopComposes () {
|
||||
updateProxy "off" ${enableComposesNoNeedMail[@]} ${enableComposesNeedMail[@]}
|
||||
doComposes "down" ${enableProxyComposes[@]}
|
||||
doComposes "down" ${enableComposesNeedMail[@]}
|
||||
doComposes "down" ${enableMailComposes[@]}
|
||||
doComposes "down" ${enableComposesNoNeedMail[@]}
|
||||
if grep -q "^.s*proxy_web.s*=.s*on" "${DOCKERS_ENV}" 2> /dev/null ; then
|
||||
${SIMU} ${KAZ_COMP_DIR}/web/web-gen.sh
|
||||
fi
|
||||
}
|
||||
|
||||
statusComposes () {
|
||||
${KAZ_ROOT}/bin/kazList.sh compose status ${enableMailComposes[@]} ${enableProxyComposes[@]} ${enableComposesNoNeedMail[@]} ${enableComposesNeedMail[@]}
|
||||
}
|
||||
|
||||
saveComposes () {
|
||||
. "${DOCKERS_ENV}"
|
||||
. "${KAZ_ROOT}/secret/SetAllPass.sh"
|
||||
|
||||
savedComposes+=( ${enableMailComposes[@]} )
|
||||
savedComposes+=( ${enableProxyComposes[@]} )
|
||||
savedComposes+=( ${enableComposesNoNeedMail[@]} )
|
||||
savedComposes+=( ${enableComposesNeedMail[@]} )
|
||||
|
||||
for compose in ${savedComposes[@]}
|
||||
do
|
||||
case "${compose}" in
|
||||
jirafeau)
|
||||
# rien à faire (fichiers)
|
||||
;;
|
||||
ethercalc)
|
||||
#inutile car le backup de /var/lib/docker/volumes/ethercalc_calcDB/_data/dump.rdb est suffisant
|
||||
;;
|
||||
#grav)
|
||||
# ???
|
||||
#;;
|
||||
#postfix)
|
||||
sympa)
|
||||
echo "save sympa"
|
||||
saveDB ${sympaDBName} "${sympa_MYSQL_USER}" "${sympa_MYSQL_PASSWORD}" "${sympa_MYSQL_DATABASE}" sympa
|
||||
;;
|
||||
web)
|
||||
# rien à faire (fichiers)
|
||||
;;
|
||||
etherpad)
|
||||
echo "save pad"
|
||||
saveDB ${etherpadDBName} "${etherpad_MYSQL_USER}" "${etherpad_MYSQL_PASSWORD}" "${etherpad_MYSQL_DATABASE}" etherpad
|
||||
;;
|
||||
framadate)
|
||||
echo "save date"
|
||||
saveDB ${framadateDBName} "${framadate_MYSQL_USER}" "${framadate_MYSQL_PASSWORD}" "${framadate_MYSQL_DATABASE}" framadate
|
||||
;;
|
||||
cloud)
|
||||
echo "save cloud"
|
||||
saveDB ${nextcloudDBName} "${nextcloud_MYSQL_USER}" "${nextcloud_MYSQL_PASSWORD}" "${nextcloud_MYSQL_DATABASE}" nextcloud
|
||||
;;
|
||||
paheko)
|
||||
# rien à faire (fichiers)
|
||||
;;
|
||||
mattermost)
|
||||
echo "save mattermost"
|
||||
saveDB ${mattermostDBName} "${mattermost_MYSQL_USER}" "${mattermost_MYSQL_PASSWORD}" "${mattermost_MYSQL_DATABASE}" mattermost
|
||||
;;
|
||||
dokuwiki)
|
||||
# rien à faire (fichiers)
|
||||
;;
|
||||
*-orga)
|
||||
ORGA=${compose%-orga}
|
||||
echo "save ${ORGA}"
|
||||
if grep -q "cloud:" "${KAZ_COMP_DIR}/${compose}/docker-compose.yml" 2> /dev/null ; then
|
||||
echo " => cloud"
|
||||
saveDB "${ORGA}-DB" "${nextcloud_MYSQL_USER}" "${nextcloud_MYSQL_PASSWORD}" "${nextcloud_MYSQL_DATABASE}" "${ORGA}-cloud"
|
||||
fi
|
||||
if grep -q "agora:" "${KAZ_COMP_DIR}/${compose}/docker-compose.yml" 2> /dev/null ; then
|
||||
echo " => mattermost"
|
||||
saveDB "${ORGA}-DB" "${mattermost_MYSQL_USER}" "${mattermost_MYSQL_PASSWORD}" "${mattermost_MYSQL_DATABASE}" "${ORGA}-mattermost"
|
||||
fi
|
||||
if grep -q "wordpress:" "${KAZ_COMP_DIR}/${compose}/docker-compose.yml" 2> /dev/null ; then
|
||||
echo " => wordpress"
|
||||
saveDB "${ORGA}-DB" "${wp_MYSQL_USER}" "${wp_MYSQL_PASSWORD}" "${wp_MYSQL_DATABASE}" "${ORGA}-wordpress"
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
done
|
||||
}
|
||||
|
||||
if [ "$#" -eq 0 ] ; then
|
||||
usage
|
||||
fi
|
||||
|
||||
if [ "$1" == "-h" ] ; then
|
||||
usage
|
||||
shift
|
||||
fi
|
||||
|
||||
if [ "$1" == "-n" ] ; then
|
||||
export SIMU=echo
|
||||
shift
|
||||
fi
|
||||
|
||||
DCK_CMD=""
|
||||
SAVE_CMD=""
|
||||
case "$1" in
|
||||
start)
|
||||
DCK_CMD="startComposes"
|
||||
shift
|
||||
;;
|
||||
|
||||
stop)
|
||||
DCK_CMD="stopComposes"
|
||||
shift
|
||||
;;
|
||||
|
||||
save)
|
||||
SAVE_CMD="saveComposes"
|
||||
shift
|
||||
;;
|
||||
|
||||
status)
|
||||
DCK_CMD="statusComposes"
|
||||
shift
|
||||
;;
|
||||
*)
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
|
||||
if [ $# -eq 0 ] ; then
|
||||
enableComposesNoNeedMail=("${availableComposesNoNeedMail[@]}")
|
||||
enableMailComposes=("${availableMailComposes[@]}")
|
||||
enableComposesNeedMail=("${availableComposesNeedMail[@]}")
|
||||
enableProxyComposes=("${availableProxyComposes[@]}")
|
||||
else
|
||||
if [ "${DCK_CMD}" = "startComposes" ] ; then
|
||||
enableProxyComposes=("${availableProxyComposes[@]}")
|
||||
fi
|
||||
fi
|
||||
|
||||
for compose in $*
|
||||
do
|
||||
compose=${compose%/}
|
||||
if [[ ! " ${knownedComposes[@]} " =~ " ${compose} " ]]; then
|
||||
declare -a subst
|
||||
subst=()
|
||||
for item in "${knownedComposes[@]}"; do
|
||||
[[ "${item}" =~ "${compose}" ]] && subst+=(${item})
|
||||
done
|
||||
if [ "${subst}" = "" ] ; then
|
||||
echo
|
||||
echo "Unknown compose: ${RED}${BOLD}${compose}${NC} not in ${YELLOW}${BOLD}${knownedComposes[*]}${NC}"
|
||||
echo
|
||||
exit 1
|
||||
else
|
||||
echo "substitute compose: ${YELLOW}${BOLD}${compose} => ${subst[@]}${NC}"
|
||||
fi
|
||||
fi
|
||||
for item in "${availableMailComposes[@]}"; do
|
||||
[[ "${item}" =~ "${compose}" ]] && enableMailComposes+=("${item}")
|
||||
done
|
||||
for item in "${availableProxyComposes[@]}"; do
|
||||
[[ "${item}" =~ "${compose}" ]] && enableProxyComposes=("${item}")
|
||||
done
|
||||
for item in "${availableComposesNoNeedMail[@]}"; do
|
||||
[[ "${item}" =~ "${compose}" ]] && enableComposesNoNeedMail+=("${item}")
|
||||
done
|
||||
for item in "${availableComposesNeedMail[@]}"; do
|
||||
[[ "${item}" =~ "${compose}" ]] && enableComposesNeedMail+=("${item}")
|
||||
done
|
||||
done
|
||||
|
||||
[[ ! -z "${DCK_CMD}" ]] && "${DCK_CMD}" && exit 0
|
||||
|
||||
[[ ! -z "${SAVE_CMD}" ]] && "${SAVE_CMD}" && exit 0
|
||||
|
||||
exit 1
|
104
bin/createEmptyPasswd.sh
Executable file
@ -0,0 +1,104 @@
|
||||
#!/bin/bash
|
||||
|
||||
cd $(dirname $0)/..
|
||||
|
||||
mkdir -p emptySecret
|
||||
rsync -aHAX --info=progress2 --delete secret/ emptySecret/
|
||||
|
||||
cd emptySecret/
|
||||
|
||||
. ../config/dockers.env
|
||||
. ./SetAllPass.sh
|
||||
|
||||
# pour mise au point
|
||||
# SIMU=echo
|
||||
|
||||
cleanEnvDB(){
|
||||
# $1 = prefix
|
||||
# $2 = envName
|
||||
# $3 = containerName of DB
|
||||
rootPass="--root_password--"
|
||||
dbName="--database_name--"
|
||||
userName="--user_name--"
|
||||
userPass="--user_password--"
|
||||
|
||||
${SIMU} sed -i \
|
||||
-e "s/MYSQL_ROOT_PASSWORD=.*/MYSQL_ROOT_PASSWORD=${rootPass}/g" \
|
||||
-e "s/MYSQL_DATABASE=.*/MYSQL_DATABASE=${dbName}/g" \
|
||||
-e "s/MYSQL_USER=.*/MYSQL_USER=${userName}/g" \
|
||||
-e "s/MYSQL_PASSWORD=.*/MYSQL_PASSWORD=${userPass}/g" \
|
||||
"$2"
|
||||
}
|
||||
|
||||
cleanEnv(){
|
||||
# $1 = prefix
|
||||
# $2 = envName
|
||||
for varName in $(grep "^[a-zA-Z_]*=" $2 | sed "s/^\([^=]*\)=.*/\1/g")
|
||||
do
|
||||
srcName="$1_${varName}"
|
||||
srcVal="--clean_val--"
|
||||
${SIMU} sed -i \
|
||||
-e "s~^[ ]*${varName}=.*$~${varName}=${srcVal}~" \
|
||||
"$2"
|
||||
done
|
||||
}
|
||||
|
||||
cleanPasswd(){
|
||||
${SIMU} sed -i \
|
||||
-e 's/^\([# ]*[^#= ]*\)=".[^{][^"]*"/\1="--clean_val--"/g' \
|
||||
./SetAllPass.sh
|
||||
}
|
||||
|
||||
####################
|
||||
# main
|
||||
|
||||
# read -r -p "Do you want to remove all password? [Y/n] " input
|
||||
|
||||
# case $input in
|
||||
# [yY][eE][sS]|[yY])
|
||||
# echo "Remove all password"
|
||||
# ;;
|
||||
# [nN][oO]|[nN])
|
||||
# echo "Abort"
|
||||
# ;;
|
||||
# *)
|
||||
# echo "Invalid input..."
|
||||
# exit 1
|
||||
# ;;
|
||||
# esac
|
||||
|
||||
cleanPasswd
|
||||
|
||||
cleanEnvDB "etherpad" "./env-${etherpadDBName}" "${etherpadDBName}"
|
||||
cleanEnvDB "framadate" "./env-${framadateDBName}" "${framadateDBName}"
|
||||
cleanEnvDB "git" "./env-${gitDBName}" "${gitDBName}"
|
||||
cleanEnvDB "mattermost" "./env-${mattermostDBName}" "${mattermostDBName}"
|
||||
cleanEnvDB "nextcloud" "./env-${nextcloudDBName}" "${nextcloudDBName}"
|
||||
cleanEnvDB "roundcube" "./env-${roundcubeDBName}" "${roundcubeDBName}"
|
||||
cleanEnvDB "sso" "./env-${ssoDBName}" "${ssoDBName}"
|
||||
cleanEnvDB "sympa" "./env-${sympaDBName}" "${sympaDBName}"
|
||||
cleanEnvDB "vigilo" "./env-${vigiloDBName}" "${vigiloDBName}"
|
||||
cleanEnvDB "wp" "./env-${wordpressDBName}" "${wordpressDBName}"
|
||||
|
||||
cleanEnv "etherpad" "./env-${etherpadServName}"
|
||||
cleanEnv "gandi" "./env-gandi"
|
||||
cleanEnv "jirafeau" "./env-${jirafeauServName}"
|
||||
cleanEnv "mattermost" "./env-${mattermostServName}"
|
||||
cleanEnv "nextcloud" "./env-${nextcloudServName}"
|
||||
cleanEnv "office" "./env-${officeServName}"
|
||||
cleanEnv "roundcube" "./env-${roundcubeServName}"
|
||||
cleanEnv "sso" "./env-${ssoServName}"
|
||||
cleanEnv "vigilo" "./env-${vigiloServName}"
|
||||
cleanEnv "wp" "./env-${wordpressServName}"
|
||||
|
||||
cat > allow_admin_ip <<EOF
|
||||
# ip for admin access only
|
||||
|
||||
# local test
|
||||
allow 127.0.0.0/8;
|
||||
allow 192.168.0.0/16;
|
||||
|
||||
EOF
|
||||
|
||||
chmod -R go= .
|
||||
chmod -R +X .
|
16
bin/createSrcDocker.sh
Executable file
@ -0,0 +1,16 @@
|
||||
#!/bin/bash
|
||||
|
||||
cd $(dirname $0)
|
||||
|
||||
./setOwner.sh
|
||||
./createEmptyPasswd.sh
|
||||
|
||||
cd ../..
|
||||
|
||||
FILE_NAME="/tmp/$(date +'%Y%M%d')-KAZ.tar.bz2"
|
||||
|
||||
tar -cjf "${FILE_NAME}" --transform s/emptySecret/secret/ \
|
||||
./kaz/emptySecret/ ./kaz/bin ./kaz/config ./kaz/dockers
|
||||
|
||||
ls -l "${FILE_NAME}"
|
||||
|
796
bin/createUser.sh
Executable file
@ -0,0 +1,796 @@
|
||||
#!/bin/bash
|
||||
|
||||
# kan: 30/03/2021
|
||||
# koi: créer les users dans le système KAZ, le KazWorld, to become a kaznaute, a kaaaaaaaznaute!
|
||||
# ki : fab
|
||||
# test git du 02/10/2023 depuis snster
|
||||
|
||||
# !!! need by htpasswd
|
||||
# apt-get install apache2-utils dos2unix
|
||||
|
||||
# rechercher tous les TODO du script pour le reste à faire
|
||||
##########################################################
|
||||
|
||||
# fonctionnement :
|
||||
# vérification de l'existence du fichier des demandes et création si absent
|
||||
# on garnit les variables
|
||||
# on vérifie les variables
|
||||
# on créé un mdp utilisable par tous les services (identifiant : email kaz)
|
||||
# pour chacun des services KAZ (NC / WP / DOKUWIKI)
|
||||
# * on vérifie si le sous-domaine existe, on le créé sinon
|
||||
# * on créé le user et le met admin si nécessaire
|
||||
# * s'il existe déjà, rollback (y compris sur les autres services)
|
||||
# pour paheko, on vérifie si le sous-domaine existe, on le créé sinon
|
||||
# pour mattermost, on créé le user et l'équipe si nécé=essaire, sur l'agora de base
|
||||
# tout est ok, on créé l'email
|
||||
# on créé le mail d'inscription avec tout le détail des services créés (url/user)
|
||||
# on inscrit le user dans la liste infos@${domain_sympa}
|
||||
# on avertit contact@kaz.bzh et on post dans l'agora/creation_compte
|
||||
|
||||
# TODO : utilisez la req sql pour attaquer paheko et créer createUser.txt en auto et modifier le champ dans paheko ACTION de "à créer" à "aucune"
|
||||
|
||||
# on récupère toutes les variables et mdp
|
||||
# on prend comme source des repertoire le dossier du dessus ( /kaz dans notre cas )
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
cd "${KAZ_ROOT}"
|
||||
. "${DOCKERS_ENV}"
|
||||
. "${KAZ_KEY_DIR}/SetAllPass.sh"
|
||||
|
||||
# DOCK_DIR="${KAZ_COMP_DIR}" # ???
|
||||
|
||||
SETUP_MAIL="docker exec -ti mailServ setup"
|
||||
|
||||
# on détermine le script appelant, le fichier log et le fichier source, tous issus de la même racine
|
||||
PRG=$(basename $0)
|
||||
RACINE=${PRG%.sh}
|
||||
|
||||
CREATE_ORGA_CMD="${KAZ_CONF_DIR}/orgaTmpl/orga-gen.sh"
|
||||
|
||||
mkdir -p "${KAZ_ROOT}/tmp" "${KAZ_ROOT}/log"
|
||||
|
||||
# fichier source dans lequel se trouve les infos sur les utilisateurs à créer
|
||||
FILE="${KAZ_ROOT}/tmp/${RACINE}.txt"
|
||||
# fichier de log pour
|
||||
LOG="${KAZ_ROOT}/log/${RACINE}.log"
|
||||
# TODO : risque si 2 admins lance en même temps
|
||||
|
||||
CMD_LOGIN="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-1-LOGIN.sh"
|
||||
CMD_SYMPA="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-2-SYMPA.sh"
|
||||
CMD_ORGA="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-3-ORGA.sh"
|
||||
CMD_PROXY="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-4-PROXY.sh"
|
||||
CMD_FIRST="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-5-FIRST.sh"
|
||||
CMD_INIT="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-6-INIT.sh"
|
||||
CMD_PAHEKO="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-7-PAHEKO.sh"
|
||||
CMD_MSG="${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-8-MSG.sh"
|
||||
TEMP_PAHEKO="${KAZ_ROOT}/tmp/${RACINE}.TEMP_PHAEKO.cvs"
|
||||
|
||||
URL_SITE="${domain}"
|
||||
URL_WEBMAIL="${webmailHost}.${domain}"
|
||||
URL_LISTE="${sympaHost}.${domain}"
|
||||
URL_AGORA="${matterHost}.${domain}"
|
||||
URL_MDP="${ldapUIHost}.${domain}"
|
||||
# URL_PAHEKO="kaz-${pahekoHost}.${domain}"
|
||||
URL_PAHEKO="${httpProto}://${paheko_API_USER}:${paheko_API_PASSWORD}@kaz-paheko.${domain}"
|
||||
|
||||
availableProxyComposes=($(getList "${KAZ_CONF_DIR}/container-proxy.list"))
|
||||
|
||||
NL_LIST=infos@${domain_sympa}
|
||||
|
||||
# indiqué dans le mail d'inscription
|
||||
# (mail+cloud base+agora : max=3, min=2)
|
||||
NB_SERVICES_BASE=0
|
||||
# max : 5, min : 0
|
||||
NB_SERVICES_DEDIES=0
|
||||
# note qu'on rajoute dans le mail pour les orgas
|
||||
MESSAGE_MAIL_ORGA_1=""
|
||||
MESSAGE_MAIL_ORGA_2=""
|
||||
MESSAGE_MAIL_ORGA_3=""
|
||||
|
||||
############################
|
||||
# Traitement des arguments #
|
||||
############################
|
||||
|
||||
CREATE_ORGA="true"
|
||||
SIMULATION=YES
|
||||
|
||||
usage () {
|
||||
echo "${PRG} [-h] [-s] [-e] [-v] [-u]"
|
||||
echo " version 1.0"
|
||||
echo " Create users in kaz world using ${FILE} as source file. All logs in ${LOG}"
|
||||
echo " -h Display this help."
|
||||
echo " -s Simulate. none user created but you can see the result in ${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-*.sh"
|
||||
echo " -e Execute commands. user or orga will be created !!!"
|
||||
echo " -u create only user (don't create orga)"
|
||||
echo " -v or -V display site informations"
|
||||
}
|
||||
|
||||
for ARG in $*; do
|
||||
case "${ARG}" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
shift
|
||||
exit;;
|
||||
-s)
|
||||
shift;;
|
||||
-e)
|
||||
SIMULATION=NO
|
||||
shift;;
|
||||
-u)
|
||||
# only user => no orga
|
||||
CREATE_ORGA=
|
||||
shift;;
|
||||
'-v' | '-V' )
|
||||
echo "${PRG}, root : ${KAZ_ROOT}, on domain : ${URL_SITE}"
|
||||
exit
|
||||
;;
|
||||
*)
|
||||
usage
|
||||
echo "${PRG}: ${RED}unknown parameter${NC}"
|
||||
shift
|
||||
exit;;
|
||||
esac
|
||||
done
|
||||
|
||||
##################################
|
||||
# Inventaire des comptes à créer #
|
||||
##################################
|
||||
|
||||
# la recherche des comptes à créé avec la commande :
|
||||
# bin/interoPaheko.sh
|
||||
|
||||
# création d'un fichier vide
|
||||
# TODO : même code ici et dans interoPaheko.sh => risque de divergence
|
||||
if [ ! -s "${FILE}" ];then
|
||||
echo "${RED}"
|
||||
echo "ERREUR : le fichier ${FILE} n'existait pas"
|
||||
echo "Il vient d'être créé. Vous pouvez le compléter."
|
||||
echo "${NC}"
|
||||
cat > "${FILE}" <<EOF
|
||||
# -- fichier de création des comptes KAZ
|
||||
# --
|
||||
# -- 1 ligne par compte
|
||||
# -- champs séparés par ";". les espaces en début et en fin sont enlevés
|
||||
# -- laisser vide si pas de donnée
|
||||
# -- pas d'espace dans les variables
|
||||
# --
|
||||
# -- ORGA : nom de l'organisation (max 23 car), vide sinon
|
||||
# -- ADMIN_ORGA : O/N indique si le user est admin de l'orga (va le créer comme admin du NC de l'orga et admin de l'équipe agora)
|
||||
# -- NC_ORGA : O/N indique si l'orga a demandé un NC
|
||||
# -- PAHEKO_ORGA : O/N indique si l'orga a demandé un paheko
|
||||
# -- WP_ORGA : O/N indique si l'orga a demandé un wp
|
||||
# -- AGORA_ORGA : O/N indique si l'orga a demandé un mattermost
|
||||
# -- WIKI_ORGA : O/N indique si l'orga a demandé un wiki
|
||||
# -- NC_BASE : O/N indique si le user doit être inscrit dans le NC de base
|
||||
# -- GROUPE_NC_BASE : soit null soit le groupe dans le NC de base
|
||||
# -- EQUIPE_AGORA : soit null soit equipe agora (max 23 car)
|
||||
# -- QUOTA = (1/10/20/...) en GB
|
||||
# --
|
||||
# NOM ; PRENOM ; EMAIL_SOUHAITE ; EMAIL_SECOURS ; ORGA ; ADMIN_ORGA ; NC_ORGA ; PAHEKO_ORGA ; WP_ORGA ; AGORA_ORGA ; WIKI_ORGA ; NC_BASE ; GROUPE_NC_BASE ; EQUIPE_AGORA ; QUOTA
|
||||
|
||||
# exemple pour un compte découverte :
|
||||
# loufoque ; le_mec; loufoque.le-mec@kaz.bzh ; gregomondo@kaz.bzh; ; N; N; N; N; N; N;N;;; 1
|
||||
|
||||
# exemple pour un compte asso de l'orga gogol avec le service dédié NC uniquement + une équipe dans l'agora
|
||||
# loufoque ; le_mec; loufoque.le-mec@kaz.bzh ; gregomondo@kaz.bzh; gogol ; O; O; N; N; N; N;N;;gogol_team; 10
|
||||
|
||||
EOF
|
||||
exit
|
||||
fi
|
||||
ALL_LINES=$(sed -e "/^[ \t]*#.*$/d" -e "/^[ \t]*$/d" "${FILE}")
|
||||
if [ -z "${ALL_LINES}" ];then
|
||||
usage
|
||||
echo "${PRG}: ${RED}nothing to do in ${FILE}${NC}"
|
||||
exit
|
||||
fi
|
||||
|
||||
###################
|
||||
# Initialisations #
|
||||
###################
|
||||
|
||||
# emails et les alias KAZ déjà créés
|
||||
TFILE_EMAIL="$(mktemp /tmp/${RACINE}.XXXXXXXXX.TFILE_EMAIL)"
|
||||
# comptes mattermost
|
||||
TFILE_MM="$(mktemp /tmp/${RACINE}.XXXXXXXXX.TFILE_MM)"
|
||||
# l'ident NextCloud
|
||||
TEMP_USER_NC="$(mktemp /tmp/${RACINE}.XXXXXXXXX.TEMP_USER_NC)"
|
||||
# le groupe NextCloud
|
||||
TEMP_GROUP_NC="$(mktemp /tmp/${RACINE}.XXXXXXXXX.TEMP_GROUP_NC)"
|
||||
# l'ident WP
|
||||
TEMP_USER_WP="$(mktemp /tmp/${RACINE}.XXXXXXXXX.TEMP_USER_WP)"
|
||||
|
||||
trap "rm -f '${TFILE_EMAIL}' '${TFILE_MM}' '${TEMP_USER_NC}' '${TEMP_GROUP_NC}' '${TEMP_USER_WP}'" 0 1 2 3 15
|
||||
|
||||
for i in "${CMD_LOGIN}" "${CMD_SYMPA}" "${CMD_ORGA}" "${CMD_PROXY}" "${CMD_FIRST}" "${CMD_INIT}" "${CMD_PAHEKO}" "${CMD_MSG}"; do
|
||||
echo "#!/bin/bash" > "${i}" && chmod +x "${i}"
|
||||
done
|
||||
|
||||
echo "numero,nom,quota_disque,action_auto" > "${TEMP_PAHEKO}"
|
||||
echo "curl \"https://${paheko_API_USER}:${paheko_API_PASSWORD}@kaz-paheko.kaz.bzh/api/user/import\" -T \"${TEMP_PAHEKO}\"" >> "${CMD_PAHEKO}"
|
||||
|
||||
#echo "récupération des login postfix... "
|
||||
## on stocke les emails et les alias KAZ déjà créés
|
||||
#(
|
||||
# ${SETUP_MAIL} email list
|
||||
# ${SETUP_MAIL} alias list
|
||||
#) | cut -d ' ' -f 2 | grep @ | sort > "${TFILE_EMAIL}"
|
||||
# did on supprime le ^M en fin de fichier pour pas faire planter les grep
|
||||
#dos2unix "${TFILE_EMAIL}"
|
||||
|
||||
echo "on récupère tous les emails (secours/alias/kaz) sur le ldap"
|
||||
FILE_LDIF=/home/sauve/ldap.ldif
|
||||
/kaz/bin/ldap/ldap_sauve.sh
|
||||
gunzip ${FILE_LDIF}.gz -f
|
||||
grep -aEiorh '([[:alnum:]]+([._-][[:alnum:]]+)*@[[:alnum:]]+([._-][[:alnum:]]+)*\.[[:alpha:]]{2,6})' ${FILE_LDIF} | sort -u > ${TFILE_EMAIL}
|
||||
|
||||
echo "récupération des login mattermost... "
|
||||
docker exec -ti mattermostServ bin/mmctl user list --all | grep ":.*(" | cut -d ':' -f 2 | cut -d ' ' -f 2 | sort > "${TFILE_MM}"
|
||||
|
||||
dos2unix "${TFILE_MM}"
|
||||
echo "done"
|
||||
|
||||
# se connecter à l'agora pour ensuite pouvoir passer toutes les commandes mmctl
|
||||
echo "docker exec -ti mattermostServ bin/mmctl auth login ${httpProto}://${URL_AGORA} --name local-server --username ${mattermost_user} --password ${mattermost_pass}" | tee -a "${CMD_INIT}"
|
||||
|
||||
# vérif des emails
|
||||
regex="^(([A-Za-z0-9]+((\.|\-|\_|\+)?[A-Za-z0-9]?)*[A-Za-z0-9]+)|[A-Za-z0-9]+)@(([A-Za-z0-9]+)+((\.|\-|\_)?([A-Za-z0-9]+)+)*)+\.([A-Za-z]{2,})+$"
|
||||
function validator {
|
||||
if ! [[ "$1" =~ ${regex} ]]; then
|
||||
# printf "* %-48s \e[1;31m[fail]\e[m\n" "${1}"
|
||||
(
|
||||
echo
|
||||
echo "ERREUR : le paramètre ${RED}${BOLD}$1 n'est pas un email valide${NC} - on stoppe tout - aucun utilisateur de créé"
|
||||
echo
|
||||
) | tee -a "${LOG}"
|
||||
exit 1
|
||||
fi
|
||||
}
|
||||
|
||||
######################################
|
||||
# Boucle lecture des comptes à créer #
|
||||
######################################
|
||||
echo -e "$(date '+%Y-%m-%d %H:%M:%S') : ${PRG} - sauvegarde des utilisateurs à créer" | tee "${LOG}"
|
||||
cat "${FILE}" >> "${LOG}"
|
||||
|
||||
LDAP_IP=$(docker inspect -f '{{.NetworkSettings.Networks.ldapNet.IPAddress}}' ldapServ)
|
||||
ALL_ORGA=
|
||||
|
||||
while read ligne; do
|
||||
|
||||
# | xargs permet de faire un trim
|
||||
NOM=$(awk -F ";" '{print $1}' <<< "${ligne}" | xargs)
|
||||
PRENOM=$(awk -F ";" '{print $2}' <<< "${ligne}" | xargs)
|
||||
|
||||
declare -A tab_email
|
||||
tab_email[EMAIL_SOUHAITE]=$(awk -F ";" '{print $3}' <<< "${ligne}" | xargs)
|
||||
tab_email[EMAIL_SECOURS]=$(awk -F ";" '{print $4}' <<< "${ligne}" | xargs)
|
||||
|
||||
ORGA=$(awk -F ";" '{print $5}' <<< "${ligne}" | xargs)
|
||||
ORGA=${ORGA,,}
|
||||
|
||||
declare -A service
|
||||
service[ADMIN_ORGA]=$(awk -F ";" '{print $6}' <<< "${ligne}" | xargs)
|
||||
service[NC_ORGA]=$(awk -F ";" '{print $7}' <<< "${ligne}" | xargs)
|
||||
service[PAHEKO_ORGA]=$(awk -F ";" '{print $8}' <<< "${ligne}" | xargs)
|
||||
service[WP_ORGA]=$(awk -F ";" '{print $9}' <<< "${ligne}" | xargs)
|
||||
service[AGORA_ORGA]=$(awk -F ";" '{print $10}' <<< "${ligne}" | xargs)
|
||||
service[WIKI_ORGA]=$(awk -F ";" '{print $11}' <<< "${ligne}" | xargs)
|
||||
service[NC_BASE]=$(awk -F ";" '{print $12}' <<< "${ligne}" | xargs)
|
||||
|
||||
GROUPE_NC_BASE=$(awk -F ";" '{print $13}' <<< "${ligne}" | xargs)
|
||||
GROUPE_NC_BASE="${GROUPE_NC_BASE,,}"
|
||||
EQUIPE_AGORA=$(awk -F ";" '{print $14}' <<< "${ligne}" | xargs)
|
||||
EQUIPE_AGORA=${EQUIPE_AGORA,,}
|
||||
QUOTA=$(awk -F ";" '{print $15}' <<< "${ligne}" | xargs)
|
||||
PASSWORD=$(awk -F ";" '{print $16}' <<< "${ligne}" | xargs)
|
||||
|
||||
IDENT_KAZ=$(unaccent utf8 "${PRENOM,,}.${NOM,,}")
|
||||
EMAIL_SOUHAITE=${tab_email[EMAIL_SOUHAITE]}
|
||||
EMAIL_SECOURS=${tab_email[EMAIL_SECOURS]}
|
||||
|
||||
echo -e "${NL}***************************** traitement de ${ligne}" | tee -a "${LOG}"
|
||||
|
||||
###########################
|
||||
# Vérification des champs #
|
||||
###########################
|
||||
|
||||
for k in "${!tab_email[@]}"; do
|
||||
validator "${tab_email[${k}]}"
|
||||
done
|
||||
|
||||
# vérif des champs O/N
|
||||
for k in "${!service[@]}"; do
|
||||
if [ "${service[${k}]}" != "O" -a "${service[${k}]}" != "N" ]; then
|
||||
(
|
||||
echo "${RED}"
|
||||
echo "${k} : ${service[${k}]}"
|
||||
echo "ERREUR : le paramètre ${k} accepte O ou N - on stoppe tout - aucun utilisateur de créé"
|
||||
echo "${NC}"
|
||||
) | tee -a "${LOG}"
|
||||
exit 1
|
||||
fi
|
||||
done
|
||||
|
||||
# taille ORGA et EQUIPE_AGORA
|
||||
TAILLE_MAX="23"
|
||||
if [ "${#ORGA}" -gt "${TAILLE_MAX}" ]; then
|
||||
(
|
||||
echo "${RED}"
|
||||
echo "ERREUR : le paramètre ORGA est trop grand : ${ORGA} , taille max : ${TAILLE_MAX} - on stoppe tout - aucun utilisateur de créé"
|
||||
echo "${NC}"
|
||||
) | tee -a "${LOG}"
|
||||
exit 1
|
||||
fi
|
||||
if [ "${#ORGA}" -gt "0" ]; then
|
||||
if [[ "${ORGA}" =~ ^[[:alnum:]-]+$ ]]; then
|
||||
echo "ok"
|
||||
else
|
||||
(
|
||||
echo "${RED}"
|
||||
echo "ERREUR : le paramètre ORGA ne contient pas les caractères autorisés : ${ORGA} - on stoppe tout - aucun utilisateur de créé"
|
||||
echo "${NC}"
|
||||
) | tee -a "${LOG}"
|
||||
exit 1
|
||||
fi
|
||||
fi
|
||||
|
||||
if [ "${#EQUIPE_AGORA}" -gt "${TAILLE_MAX}" ]; then
|
||||
(
|
||||
echo "${RED}"
|
||||
echo "ERREUR : le paramètre EQUIPE_AGORA est trop grand : ${EQUIPE_AGORA} , taille max : ${TAILLE_MAX} - on stoppe tout - aucun utilisateur de créé"
|
||||
echo "${NC}"
|
||||
) | tee -a "${LOG}"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# vérif quota est entier
|
||||
if ! [[ "${QUOTA}" =~ ^[[:digit:]]+$ ]]; then
|
||||
(
|
||||
echo
|
||||
echo "ERREUR : ${RED}${BOLD}QUOTA n'est pas numérique : ${QUOTA}${NC} - on stoppe tout - aucun utilisateur de créé"
|
||||
) | tee -a "${LOG}"
|
||||
fi
|
||||
|
||||
####################################################
|
||||
# cree un mdp acceptable par postfix/nc/mattermost #
|
||||
####################################################
|
||||
if [ -z ${PASSWORD} ]; then
|
||||
PASSWORD=_`apg -n 1 -m 10 -M NCL -d`_
|
||||
fi
|
||||
SEND_MSG_CREATE=
|
||||
if [ -n "${ORGA}" -a -z "${CREATE_ORGA}" ]; then
|
||||
# skeep orga
|
||||
continue
|
||||
fi
|
||||
|
||||
####################################################################
|
||||
# TODO: Test de l'identKAZ du ldap, il faut l'unicité. si KO, STOP #
|
||||
####################################################################
|
||||
|
||||
###################################
|
||||
# Création du compe de messagerie #
|
||||
###################################
|
||||
# le mail existe t-il déjà ?
|
||||
if grep -q "^${EMAIL_SOUHAITE}$" "${TFILE_EMAIL}"; then
|
||||
echo "${EMAIL_SOUHAITE} existe déjà" | tee -a "${LOG}"
|
||||
else
|
||||
SEND_MSG_CREATE=true
|
||||
echo "${EMAIL_SOUHAITE} n'existe pas" | tee -a "${LOG}"
|
||||
echo "${SETUP_MAIL} email add ${EMAIL_SOUHAITE} ${PASSWORD}" | tee -a "${CMD_LOGIN}"
|
||||
echo "${SETUP_MAIL} quota set ${EMAIL_SOUHAITE} ${QUOTA}G" | tee -a "${CMD_LOGIN}"
|
||||
# LDAP, à tester
|
||||
user=$(echo ${EMAIL_SOUHAITE} | awk -F '@' '{print $1}')
|
||||
domain=$(echo ${EMAIL_SOUHAITE} | awk -F '@' '{print $2}')
|
||||
pass=$(mkpasswd -m sha512crypt ${PASSWORD})
|
||||
echo "echo -e '\n\ndn: cn=${EMAIL_SOUHAITE},ou=users,${ldap_root}\n\
|
||||
changeType: add\n\
|
||||
objectclass: inetOrgPerson\n\
|
||||
objectClass: PostfixBookMailAccount\n\
|
||||
objectClass: nextcloudAccount\n\
|
||||
objectClass: kaznaute\n\
|
||||
sn: ${PRENOM} ${NOM}\n\
|
||||
mail: ${EMAIL_SOUHAITE}\n\
|
||||
mailEnabled: TRUE\n\
|
||||
mailGidNumber: 5000\n\
|
||||
mailHomeDirectory: /var/mail/${domain}/${user}/\n\
|
||||
mailQuota: ${QUOTA}G\n\
|
||||
mailStorageDirectory: maildir:/var/mail/${domain}/${user}/\n\
|
||||
mailUidNumber: 5000\n\
|
||||
mailDeSecours: ${EMAIL_SECOURS}\n\
|
||||
identifiantKaz: ${IDENT_KAZ}\n\
|
||||
quota: ${QUOTA}\n\
|
||||
nextcloudEnabled: TRUE\n\
|
||||
nextcloudQuota: ${QUOTA} GB\n\
|
||||
mobilizonEnabled: TRUE\n\
|
||||
agoraEnabled: TRUE\n\
|
||||
userPassword: {CRYPT}${pass}\n\n' | ldapmodify -c -H ldap://${LDAP_IP} -D \"cn=${ldap_LDAP_ADMIN_USERNAME},${ldap_root}\" -x -w ${ldap_LDAP_ADMIN_PASSWORD}" | tee -a "${CMD_LOGIN}"
|
||||
fi
|
||||
#userPassword: {CRYPT}\$6\$${pass}\n\n\" | ldapmodify -c -H ldap://${LDAP_IP} -D \"cn=${ldap_LDAP_CONFIG_ADMIN_USERNAME},${ldap_root}\" -x -w ${ldap_LDAP_CONFIG_ADMIN_PASSWORD}" | tee -a "${CMD_LOGIN}"
|
||||
|
||||
CREATE_ORGA_SERVICES=""
|
||||
|
||||
#############
|
||||
# NEXTCLOUD #
|
||||
#############
|
||||
# on recalcul l'url de NC
|
||||
if [ "${ORGA}" != "" -a "${service[NC_ORGA]}" == "O" ]; then
|
||||
URL_NC="${ORGA}-${cloudHost}.${domain}"
|
||||
|
||||
# si le cloud de l'orga n'est pas up alors on le créé
|
||||
nb=$(docker ps | grep "${ORGA}-${cloudHost}" | wc -l)
|
||||
if [ "${nb}" == "0" ];then
|
||||
echo " * +cloud +collabora ${ORGA}"
|
||||
CREATE_ORGA_SERVICES="${CREATE_ORGA_SERVICES} +cloud +collabora"
|
||||
# installe les plugins initiaux dans "/kaz/bin/gestClouds.sh"
|
||||
fi
|
||||
|
||||
NB_SERVICES_DEDIES=$((NB_SERVICES_DEDIES+1))
|
||||
|
||||
else
|
||||
URL_NC="${cloudHost}.${domain}"
|
||||
NB_SERVICES_BASE=$((NB_SERVICES_BASE+1))
|
||||
fi
|
||||
|
||||
MESSAGE_MAIL_ORGA_1="${MESSAGE_MAIL_ORGA_1}${NL}* un bureau virtuel pour stocker des fichiers/calendriers/contacts et partager avec vos connaissances : ${httpProto}://${URL_NC}"
|
||||
|
||||
# le user existe t-il déjà sur NC ?
|
||||
curl -o "${TEMP_USER_NC}" -X GET -H 'OCS-APIRequest:true' "${httpProto}://admin:${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}@${URL_NC}/ocs/v1.php/cloud/users?search=${IDENT_KAZ}"
|
||||
if grep -q "<element>${IDENT_KAZ}</element>" "${TEMP_USER_NC}"; then
|
||||
echo "${IDENT_KAZ} existe déjà sur ${URL_NC}" | tee -a "${LOG}"
|
||||
else
|
||||
|
||||
# on créé l'utilisateur sur NC sauf si c'est le NC général, on ne créé jamais l'utilisateur7
|
||||
if [ ${URL_NC} != "${cloudHost}.${domain}" ]; then
|
||||
|
||||
echo "curl -X POST -H 'OCS-APIRequest:true' ${httpProto}://admin:${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}@${URL_NC}/ocs/v1.php/cloud/users \
|
||||
-d userid='${IDENT_KAZ}' \
|
||||
-d displayName='${PRENOM} ${NOM}' \
|
||||
-d password='${PASSWORD}' \
|
||||
-d email='${EMAIL_SOUHAITE}' \
|
||||
-d quota='${QUOTA}GB' \
|
||||
-d language='fr' \
|
||||
" | tee -a "${CMD_INIT}"
|
||||
|
||||
fi
|
||||
|
||||
# s'il est admin de son orga, on le met admin
|
||||
if [ "${service[ADMIN_ORGA]}" == "O" -a "${ORGA}" != "" -a "${service[NC_ORGA]}" == "O" ]; then
|
||||
echo "curl -X POST -H 'OCS-APIRequest:true' ${httpProto}://${nextcloud_NEXTCLOUD_ADMIN_USER}:${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}@${URL_NC}/ocs/v1.php/cloud/users/${IDENT_KAZ}/groups -d groupid='admin'" | tee -a "${CMD_INIT}"
|
||||
fi
|
||||
|
||||
# faut-il mettre le user NC dans un groupe particulier sur le NC de base ?
|
||||
if [ "${GROUPE_NC_BASE}" != "" -a "${service[NC_BASE]}" == "O" ]; then
|
||||
# le groupe existe t-il déjà ?
|
||||
curl -o "${TEMP_GROUP_NC}" -X GET -H 'OCS-APIRequest:true' "${httpProto}://admin:${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}@${URL_NC}/ocs/v1.php/cloud/groups?search=${GROUPE_NC_BASE}"
|
||||
nb=$(grep "<element>${GROUPE_NC_BASE}</element>" "${TEMP_GROUP_NC}" | wc -l)
|
||||
if [ "${nb}" == "0" ];then
|
||||
echo "curl -X POST -H 'OCS-APIRequest:true' ${httpProto}://admin:${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}@${URL_NC}/ocs/v1.php/cloud/groups -d groupid=${GROUPE_NC_BASE}" | tee -a "${CMD_INIT}"
|
||||
fi
|
||||
# puis attacher le user au groupe
|
||||
echo "curl -X POST -H 'OCS-APIRequest:true' ${httpProto}://admin:${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}@${URL_NC}/ocs/v1.php/cloud/users/${IDENT_KAZ}/groups -d groupid=${GROUPE_NC_BASE}" | tee -a "${CMD_INIT}"
|
||||
fi
|
||||
fi
|
||||
|
||||
#############
|
||||
# WORDPRESS #
|
||||
#############
|
||||
|
||||
# TODO : pour l'utilisation de l'api : https://www.hostinger.com/tutorials/wordpress-rest-api
|
||||
|
||||
if [ "${ORGA}" != "" -a "${service[WP_ORGA]}" == "O" ]; then
|
||||
|
||||
URL_WP_ORGA="${ORGA}-${wordpressHost}.${domain}"
|
||||
|
||||
# si le wp de l'orga n'est pas up alors on le créé
|
||||
nb=$(docker ps | grep "${ORGA}-${wordpressHost}" | wc -l)
|
||||
if [ "${nb}" == "0" ];then
|
||||
echo " * +wp ${ORGA}"
|
||||
CREATE_ORGA_SERVICES="${CREATE_ORGA_SERVICES} +wp"
|
||||
fi
|
||||
|
||||
NB_SERVICES_DEDIES=$((NB_SERVICES_DEDIES+1))
|
||||
MESSAGE_MAIL_ORGA_1="${MESSAGE_MAIL_ORGA_1}${NL}* un site web de type wordpress : ${httpProto}://${URL_WP_ORGA}/wp-admin/"
|
||||
|
||||
# TODO : vérif existance user
|
||||
# # le user existe t-il déjà sur le wp ?
|
||||
# curl -o "${TEMP_USER_WP}" -X GET "${httpProto}://${wp_WORDPRESS_ADMIN_USER}:${wp_WORDPRESS_ADMIN_PASSWORD}@${URL_WP_ORGA}/ocs/v1.php/cloud/users?search=${IDENT_KAZ}"
|
||||
# nb_user_wp_orga=$(grep "<element>${IDENT_KAZ}</element>" "${TEMP_USER_WP}" | wc -l)
|
||||
# if [ "${nb_user_wp_orga}" != "0" ];then
|
||||
# (
|
||||
# echo "${RED}"
|
||||
# echo "ERREUR : ${IDENT_KAZ} existe déjà sur ${URL_WP_ORGA} - on stoppe tout - aucun utilisateur de créé"
|
||||
# echo "${NC}"
|
||||
# ) | tee -a "${LOG}"
|
||||
#
|
||||
# # ROLLBACK - on vire le user de NC
|
||||
# if [ "${nb_user_nc_orga}" != "0" ];then
|
||||
# (
|
||||
# echo "${RED}"
|
||||
# echo "ERREUR : ${IDENT_KAZ} existe déjà sur ${URL_NC} - on stoppe tout - aucun utilisateur de créé"
|
||||
# echo "${NC}"
|
||||
# ) | tee -a "${LOG}"
|
||||
#
|
||||
# # on supprime l'utilisateur sur NC.
|
||||
# echo "curl -X DELETE -H 'OCS-APIRequest:true' ${httpProto}://admin:${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}@${URL_NC}/ocs/v1.php/cloud/users \
|
||||
# -d userid='${IDENT_KAZ}' \
|
||||
# " | tee -a "${CMD_INIT}"
|
||||
# fi
|
||||
#
|
||||
# exit 1
|
||||
# fi
|
||||
|
||||
# TODO : créer le user et le mettre admin si nécessaire
|
||||
# if [ "${service[ADMIN_ORGA]}" == "O" ]; then
|
||||
# :
|
||||
# else
|
||||
# :
|
||||
# fi
|
||||
fi
|
||||
|
||||
############
|
||||
# PAHEKO #
|
||||
############
|
||||
|
||||
|
||||
if [ "${ORGA}" != "" -a "${service[PAHEKO_ORGA]}" == "O" ]; then
|
||||
|
||||
URL_PAHEKO_ORGA="${ORGA}-${pahekoHost}.${domain}"
|
||||
|
||||
# il n'y a pas de docker spécifique paheko (je cree toujours paheko)
|
||||
echo " * +paheko ${ORGA}"
|
||||
CREATE_ORGA_SERVICES="${CREATE_ORGA_SERVICES} +paheko"
|
||||
|
||||
NB_SERVICES_DEDIES=$((NB_SERVICES_DEDIES+1))
|
||||
MESSAGE_MAIL_ORGA_1="${MESSAGE_MAIL_ORGA_1}${NL}* un service de gestion adhérents/clients : ${httpProto}://${URL_PAHEKO_ORGA}"
|
||||
|
||||
if [ "${service[ADMIN_ORGA]}" == "O" ]; then
|
||||
MESSAGE_MAIL_ORGA_1="${MESSAGE_MAIL_ORGA_1} (l'installation est à terminer en vous rendant sur le site)"
|
||||
fi
|
||||
fi
|
||||
|
||||
|
||||
############
|
||||
# DOKUWIKI #
|
||||
############
|
||||
|
||||
if [ "${ORGA}" != "" -a "${service[WIKI_ORGA]}" == "O" ]; then
|
||||
|
||||
URL_WIKI_ORGA="${ORGA}-${dokuwikiHost}.${domain}"
|
||||
|
||||
# si le wiki de l'orga n'est pas up alors on le créé
|
||||
nb=$(docker ps | grep "${ORGA}-${dokuwikiHost}" | wc -l)
|
||||
if [ "${nb}" == "0" ];then
|
||||
echo " * +wiki ${ORGA}"
|
||||
CREATE_ORGA_SERVICES="${CREATE_ORGA_SERVICES} +wiki"
|
||||
fi
|
||||
|
||||
NB_SERVICES_DEDIES=$((NB_SERVICES_DEDIES+1))
|
||||
MESSAGE_MAIL_ORGA_1="${MESSAGE_MAIL_ORGA_1}${NL}* un wiki dédié pour votre documentation : ${httpProto}://${URL_WIKI_ORGA}"
|
||||
|
||||
|
||||
# TODO : ??? à voir https://www.dokuwiki.org/devel:xmlrpc:clients
|
||||
|
||||
if grep -q "^${IDENT_KAZ}:" "${DOCK_VOL}/orga_${ORGA}-wikiConf/_data/users.auth.php" 2>/dev/null; then
|
||||
echo "${IDENT_KAZ} existe déjà sur ${URL_WIKI_ORGA}" | tee -a "${LOG}"
|
||||
else
|
||||
echo "echo \"${IDENT_KAZ}:$(htpasswd -bnBC 10 "" ${PASSWORD}):${PRENOM} ${NOM}:${EMAIL_SOUHAITE}:admin,user\" >> \"${DOCK_VOL}/orga_${ORGA}-wikiConf/_data/users.auth.php\"" | tee -a "${CMD_INIT}"
|
||||
fi
|
||||
fi
|
||||
|
||||
##############
|
||||
# MATTERMOST #
|
||||
##############
|
||||
|
||||
# on ne gère pas la création du docker dédié mattermost
|
||||
if [ "${ORGA}" != "" -a "${service[AGORA_ORGA]}" == "O" ]; then
|
||||
echo "# ******************************************************************************" | tee -a "${CMD_INIT}"
|
||||
echo "# **************************** ATTENTION ***************************************" | tee -a "${CMD_INIT}"
|
||||
echo "# ******************************************************************************" | tee -a "${CMD_INIT}"
|
||||
echo "# Mattermost dédié : on ne fait rien." | tee -a "${CMD_INIT}"
|
||||
echo "# ******************************************************************************" | tee -a "${CMD_INIT}"
|
||||
fi
|
||||
|
||||
if grep -q "^${IDENT_KAZ}$" "${TFILE_MM}" 2>/dev/null; then
|
||||
echo "${IDENT_KAZ} existe déjà sur mattermost" | tee -a "${LOG}"
|
||||
else
|
||||
# on créé le compte mattermost
|
||||
echo "docker exec -ti mattermostServ bin/mmctl user create --email ${EMAIL_SOUHAITE} --username ${IDENT_KAZ} --password ${PASSWORD}" | tee -a "${CMD_LOGIN}"
|
||||
# et enfin on ajoute toujours le user à l'équipe KAZ et aux 2 channels publiques
|
||||
echo "docker exec -ti mattermostServ bin/mmctl team users add kaz ${EMAIL_SOUHAITE}" | tee -a "${CMD_LOGIN}"
|
||||
echo "docker exec -ti mattermostServ bin/mmctl channel users add kaz:une-question--un-soucis ${EMAIL_SOUHAITE}" | tee -a "${CMD_LOGIN}"
|
||||
echo "docker exec -ti mattermostServ bin/mmctl channel users add kaz:cafe-du-commerce--ouvert-2424h ${EMAIL_SOUHAITE}" | tee -a "${CMD_LOGIN}"
|
||||
NB_SERVICES_BASE=$((NB_SERVICES_BASE+1))
|
||||
fi
|
||||
|
||||
if [ "${EQUIPE_AGORA}" != "" -a "${EQUIPE_AGORA}" != "kaz" ]; then
|
||||
# l'équipe existe t-elle déjà ?
|
||||
nb=$(docker exec mattermostServ bin/mmctl team list | grep -w "${EQUIPE_AGORA}" | wc -l)
|
||||
if [ "${nb}" == "0" ];then # non, on la créé en mettant le user en admin de l'équipe
|
||||
echo "docker exec -ti mattermostServ bin/mmctl team create --name ${EQUIPE_AGORA} --display_name ${EQUIPE_AGORA} --email ${EMAIL_SOUHAITE}" --private | tee -a "${CMD_INIT}"
|
||||
fi
|
||||
# puis ajouter le user à l'équipe
|
||||
echo "docker exec -ti mattermostServ bin/mmctl team users add ${EQUIPE_AGORA} ${EMAIL_SOUHAITE}" | tee -a "${CMD_INIT}"
|
||||
fi
|
||||
|
||||
if [ -n "${CREATE_ORGA_SERVICES}" ]; then
|
||||
SEND_MSG_CREATE=true
|
||||
echo "${CREATE_ORGA_CMD}" --create ${CREATE_ORGA_SERVICES} "${ORGA}" | tee -a "${CMD_ORGA}"
|
||||
echo "${CREATE_ORGA_CMD}" --init ${CREATE_ORGA_SERVICES} "${ORGA}" | tee -a "${CMD_FIRST}"
|
||||
ALL_ORGA="${ALL_ORGA} ${ORGA}"
|
||||
fi
|
||||
|
||||
##########################
|
||||
# Inscription newsletter #
|
||||
##########################
|
||||
|
||||
# TODO : utiliser liste sur dev également
|
||||
|
||||
# on inscrit le user sur sympa, à la liste infos@${domain_sympa}
|
||||
# docker exec -ti sympaServ /usr/lib/sympa/bin/sympa_soap_client.pl --soap_url=https://listes.kaz.sns/sympasoap --trusted_application=SOAP_USER --trusted_application_password=SOAP_PASSWORD --proxy_vars="USER_EMAIL=contact1@kaz.sns" --service=which
|
||||
if [[ "${mode}" = "dev" ]]; then
|
||||
echo "# DEV, on teste l'inscription à sympa"| tee -a "${CMD_SYMPA}"
|
||||
LISTMASTER=$(echo ${sympa_LISTMASTERS} | cut -d',' -f1)
|
||||
echo "docker exec -ti sympaServ /usr/lib/sympa/bin/sympa_soap_client.pl --soap_url=${httpProto}://${URL_LISTE}/sympasoap --trusted_application=${sympa_SOAP_USER} --trusted_application_password=${sympa_SOAP_PASSWORD} --proxy_vars=\"USER_EMAIL=${LISTMASTER}\" --service=add --service_parameters=\"${NL_LIST},${EMAIL_SOUHAITE}\"" | tee -a "${CMD_SYMPA}"
|
||||
else
|
||||
echo "# PROD, on inscrit à sympa"| tee -a "${CMD_SYMPA}"
|
||||
LISTMASTER=$(echo ${sympa_LISTMASTERS} | cut -d',' -f1)
|
||||
echo "docker exec -ti sympaServ /usr/lib/sympa/bin/sympa_soap_client.pl --soap_url=${httpProto}://${URL_LISTE}/sympasoap --trusted_application=${sympa_SOAP_USER} --trusted_application_password=${sympa_SOAP_PASSWORD} --proxy_vars=\"USER_EMAIL=${LISTMASTER}\" --service=add --service_parameters=\"${NL_LIST},${EMAIL_SOUHAITE}\"" | tee -a "${CMD_SYMPA}"
|
||||
echo "docker exec -ti sympaServ /usr/lib/sympa/bin/sympa_soap_client.pl --soap_url=${httpProto}://${URL_LISTE}/sympasoap --trusted_application=${sympa_SOAP_USER} --trusted_application_password=${sympa_SOAP_PASSWORD} --proxy_vars=\"USER_EMAIL=${LISTMASTER}\" --service=add --service_parameters=\"${NL_LIST},${EMAIL_SECOURS}\"" | tee -a "${CMD_SYMPA}"
|
||||
fi
|
||||
|
||||
if [ "${service[ADMIN_ORGA]}" == "O" ]; then
|
||||
MESSAGE_MAIL_ORGA_2="${MESSAGE_MAIL_ORGA_2}Comme administrateur de votre organisation, vous pouvez créer des listes de diffusion en vous rendant sur ${httpProto}://${URL_LISTE}"
|
||||
fi
|
||||
|
||||
###################
|
||||
# update paheko #
|
||||
###################
|
||||
|
||||
# TODO : problème si 2 comptes partagent le même email souhaité (cela ne devrait pas arriver)
|
||||
curl -s "https://${paheko_API_USER}:${paheko_API_PASSWORD}@kaz-paheko.kaz.bzh/api/sql" -d "SELECT numero,nom,quota_disque from users WHERE email='${EMAIL_SOUHAITE}'" | jq '.results[] | .numero,.nom,.quota_disque ' | tr \\n ',' | sed 's/,$/,Aucune\n/' >> "${TEMP_PAHEKO}"
|
||||
|
||||
####################
|
||||
# Inscription MAIL #
|
||||
####################
|
||||
|
||||
if [ "${NB_SERVICES_DEDIES}" != "0" ];then
|
||||
MESSAGE_MAIL_ORGA_1="${NL} dont ${NB_SERVICES_DEDIES} service(s) dédié(s) pour votre organisation:${NL} ${MESSAGE_MAIL_ORGA_1}"
|
||||
fi
|
||||
|
||||
if [ -z "${SEND_MSG_CREATE}" ]; then
|
||||
# rien de créé => pas de message
|
||||
continue
|
||||
fi
|
||||
|
||||
#si admin alors msg pour indiquer qu'il faut nous renvoyer ce qu'il souhaite comme service.
|
||||
if [ "${service[ADMIN_ORGA]}" == "O" ]; then
|
||||
MESSAGE_MAIL_ORGA_3="${MESSAGE_MAIL_ORGA_3}En tant qu'association/famille/société. Vous avez la possibilité d'ouvrir, quand vous le voulez, des services kaz, il vous suffit de nous le demander.
|
||||
|
||||
Pourquoi n'ouvrons-nous pas tous les services tout de suite ? parce que nous aimons la sobriété et que nous préservons notre espace disque ;)
|
||||
A quoi sert d'avoir un site web si on ne l'utilise pas, n'est-ce pas ?
|
||||
|
||||
Par retour de mail, dites-nous de quoi vous avez besoin tout de suite entre:
|
||||
* une comptabilité : un service de gestion adhérents/clients
|
||||
* un site web de type WordPress
|
||||
* un cloud : bureau virtuel pour stocker des fichiers/calendriers/contacts et partager avec vos connaissances
|
||||
|
||||
Une fois que vous aurez répondu à ce mail, votre demande sera traitée manuellement.
|
||||
"
|
||||
fi
|
||||
|
||||
# on envoie le mail de bienvenue
|
||||
MAIL_KAZ="Bonjour,
|
||||
|
||||
Bienvenue chez KAZ!
|
||||
|
||||
Vous disposez de $((${NB_SERVICES_BASE} + ${NB_SERVICES_DEDIES})) services kaz avec authentification :
|
||||
|
||||
* une messagerie classique : ${httpProto}://${URL_WEBMAIL}
|
||||
* une messagerie instantanée pour discuter au sein d'équipes : ${httpProto}://${URL_AGORA}
|
||||
|
||||
Votre email et identifiant pour tous ces services : ${EMAIL_SOUHAITE}
|
||||
Le mot de passe : ${PASSWORD}
|
||||
|
||||
Pour changer votre mot de passe de messagerie, c'est ici: ${httpProto}://${URL_MDP}
|
||||
Si vous avez perdu votre mot de passe, c'est ici: ${httpProto}://${URL_MDP}/?action=sendtoken
|
||||
|
||||
Vous pouvez accéder à votre messagerie :
|
||||
* soit depuis votre webmail : ${httpProto}://${URL_WEBMAIL}
|
||||
* soit depuis votre bureau virtuel : ${httpProto}://${URL_NC}
|
||||
* soit depuis un client de messagerie comme thunderbird
|
||||
|
||||
${MESSAGE_MAIL_ORGA_3}
|
||||
|
||||
Vous avez quelques docs intéressantes sur le wiki de kaz:
|
||||
|
||||
* Migrer son site internet wordpress vers kaz
|
||||
https://wiki.kaz.bzh/wordpress/start#migrer_son_site_wordpress_vers_kaz
|
||||
|
||||
* Migrer sa messagerie vers kaz
|
||||
https://wiki.kaz.bzh/messagerie/gmail/start
|
||||
|
||||
* Démarrer simplement avec son cloud
|
||||
https://wiki.kaz.bzh/nextcloud/start
|
||||
|
||||
Votre quota est de ${QUOTA}GB. Si vous souhaitez plus de place pour vos fichiers ou la messagerie, faites-nous signe !
|
||||
|
||||
Pour accéder à la messagerie instantanée et communiquer avec les membres de votre équipe ou ceux de kaz : ${httpProto}://${URL_AGORA}/login
|
||||
|
||||
${MESSAGE_MAIL_ORGA_2}
|
||||
|
||||
Enfin, vous disposez de tous les autres services KAZ où l'authentification n'est pas nécessaire : ${httpProto}://${URL_SITE}
|
||||
|
||||
En cas de soucis, n'hésitez pas à poser vos questions sur le canal 'Une question ? un soucis' de l'agora dispo ici : ${httpProto}://${URL_AGORA}
|
||||
|
||||
Si vous avez besoin d'accompagnement pour votre site, votre cloud, votre compta, votre migration de messagerie,... nous proposons des formations mensuelles gratuites. Si vous souhaitez être accompagné par un professionnel, nous pouvons vous donner une liste de pros, référencés par KAZ.
|
||||
|
||||
À bientôt ;)
|
||||
|
||||
La collégiale de KAZ. "
|
||||
|
||||
echo "docker exec -i mailServ mailx -a 'Content-Type: text/plain; charset=\"UTF-8\"' -r contact@kaz.bzh -s \"KAZ: confirmation d'inscription\" ${EMAIL_SOUHAITE} ${EMAIL_SECOURS} << EOF
|
||||
${MAIL_KAZ}
|
||||
EOF" | tee -a "${CMD_MSG}"
|
||||
|
||||
# on envoie le mail de confirmation d'inscription à contact
|
||||
MAIL_KAZ="*****POST AUTOMATIQUE******
|
||||
Hello,
|
||||
${NOM} ${PRENOM} vient d'être inscrit avec l'email ${EMAIL_SOUHAITE}
|
||||
quota : ${QUOTA}GB
|
||||
|
||||
NC_BASE : ${service[NC_BASE]}
|
||||
groupe NC base : ${GROUPE_NC_BASE}
|
||||
équipe agora base : ${EQUIPE_AGORA}
|
||||
email de secours : ${EMAIL_SECOURS}
|
||||
|
||||
ORGA : ${ORGA}
|
||||
ADMIN_ORGA : ${service[ADMIN_ORGA]}
|
||||
NC_ORGA : ${service[NC_ORGA]}
|
||||
PAHEKO_ORGA : ${service[PAHEKO_ORGA]}
|
||||
WP_ORGA : ${service[WP_ORGA]}
|
||||
AGORA_ORGA : ${service[AGORA_ORGA]}
|
||||
WIKI_ORGA : ${service[WIKI_ORGA]}
|
||||
|
||||
bisou!"
|
||||
echo "docker exec -i mailServ mailx -a 'Content-Type: text/plain; charset=\"UTF-8\"' -r contact@kaz.bzh -s \"KAZ: confirmation d'inscription\" ${EMAIL_CONTACT} << EOF
|
||||
${MAIL_KAZ}
|
||||
EOF" | tee -a "${CMD_MSG}"
|
||||
|
||||
echo " # on envoie la confirmation d'inscription sur l'agora " | tee -a "${CMD_MSG}"
|
||||
echo "docker exec -ti mattermostServ bin/mmctl post create kaz:Creation-Comptes --message \"${MAIL_KAZ}\"" | tee -a "${CMD_MSG}"
|
||||
|
||||
# fin des inscriptions
|
||||
done <<< "${ALL_LINES}"
|
||||
|
||||
if [[ -n "${ALL_ORGA}" ]]; then
|
||||
echo "sleep 2" | tee -a "${CMD_PROXY}"
|
||||
echo "${KAZ_BIN_DIR}/container.sh start ${ALL_ORGA}" | tee -a "${CMD_PROXY}"
|
||||
for item in "${availableProxyComposes[@]}"; do
|
||||
echo "cd \"${KAZ_COMP_DIR}/${item}/\"; ./proxy-gen.sh; docker-compose up -d; ./reload.sh " | tee -a "${CMD_PROXY}"
|
||||
done
|
||||
fi
|
||||
|
||||
###########################
|
||||
# Lancement des commandes #
|
||||
###########################
|
||||
if [ "${SIMULATION}" == "NO" ]; then
|
||||
echo "on exécute"
|
||||
"${CMD_LOGIN}"
|
||||
# on attend qques secondes que le mail soit bien créé avant de continuer (prob de lecture de la BAL : à investiguer)
|
||||
# je rallonge à 20s car je vois que le docker sympa ne connait pas toujours l'email kaz créé
|
||||
echo "on attend 20s pour que la création des emails soit certaine"
|
||||
sleep 20
|
||||
"${CMD_SYMPA}"
|
||||
"${CMD_ORGA}"
|
||||
"${CMD_PROXY}"
|
||||
"${CMD_FIRST}"
|
||||
"${CMD_INIT}"
|
||||
"${CMD_PAHEKO}"
|
||||
"${CMD_MSG}"
|
||||
else
|
||||
echo "Aucune commande n'a été lancée : Possibilité de le faire à la main. cf ${KAZ_ROOT}/tmp/${RACINE}_cmds_to_run-*.sh"
|
||||
fi
|
||||
|
||||
# END
|
7
bin/cron-cloud.sh
Executable file
@ -0,0 +1,7 @@
|
||||
#!/bin/bash
|
||||
|
||||
for cloud in $(docker ps | grep -i nextcloudServ |awk '{print $12}')
|
||||
do
|
||||
docker exec -u www-data $cloud php cron.php
|
||||
|
||||
done
|
209
bin/dns.sh
Executable file
@ -0,0 +1,209 @@
|
||||
#!/bin/bash
|
||||
|
||||
# list/ajout/supprime/ un sous-domaine
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
. "${DOCKERS_ENV}"
|
||||
|
||||
cd "${KAZ_ROOT}"
|
||||
export PRG="$0"
|
||||
export IP="127.0.0.1"
|
||||
export ETC_HOSTS="/etc/hosts"
|
||||
|
||||
# no more export in .env
|
||||
export $(set | grep "domain=")
|
||||
|
||||
declare -a forbidenName
|
||||
forbidenName=(${calcHost} calc ${cloudHost} bureau ${dateHost} date ${dokuwikiHost} dokuwiki ${fileHost} file ${ldapHost} ${pahekoHost} ${gitHost} ${gravHost} ${matterHost} ${officeHost} collabora ${padHost} ${sympaHost} listes ${webmailHost} ${wordpressHost} www ${vigiloHost} form)
|
||||
|
||||
export FORCE="NO"
|
||||
export CMD=""
|
||||
export SIMU=""
|
||||
|
||||
usage(){
|
||||
echo "Usage: ${PRG} list [sub-domain...]"
|
||||
echo " ${PRG} [-n] [-f] {add/del} sub-domain..."
|
||||
echo " -h help"
|
||||
echo " -n simulation"
|
||||
echo " -f force protected domain"
|
||||
exit 1
|
||||
}
|
||||
|
||||
for ARG in $@
|
||||
do
|
||||
case "${ARG}" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
'-f' )
|
||||
shift
|
||||
export FORCE="YES"
|
||||
;;
|
||||
'-n' )
|
||||
shift
|
||||
export SIMU="echo"
|
||||
;;
|
||||
'list'|'add'|'del' )
|
||||
shift
|
||||
CMD="${ARG}"
|
||||
break
|
||||
;;
|
||||
* )
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z "${CMD}" ]; then
|
||||
usage
|
||||
fi
|
||||
|
||||
. "${KAZ_KEY_DIR}/env-gandi"
|
||||
|
||||
if [[ -z "${GANDI_KEY}" ]] ; then
|
||||
echo
|
||||
echo "no GANDI_KEY set in ${KAZ_KEY_DIR}/env-gandi"
|
||||
usage
|
||||
fi
|
||||
|
||||
|
||||
waitNet () {
|
||||
if [[ "${domain}" = "kaz.local" ]]; then
|
||||
return
|
||||
fi
|
||||
|
||||
### wait when error code 503
|
||||
if [[ $(curl -H "authorization: Apikey ${GANDI_KEY}" --connect-timeout 2 -s -D - "${GANDI_API}" -o /dev/null 2>/dev/null | head -n1) != *200* ]]; then
|
||||
echo "DNS not available. Please wait..."
|
||||
while [[ $(curl -H "authorization: Apikey ${GANDI_KEY}" --connect-timeout 2 -s -D - "${GANDI_API}" -o /dev/null 2>/dev/null | head -n1) != *200* ]]
|
||||
do
|
||||
sleep 5
|
||||
done
|
||||
exit
|
||||
fi
|
||||
}
|
||||
|
||||
list(){
|
||||
if [[ "${domain}" = "kaz.local" ]]; then
|
||||
grep --perl-regex "^${IP}\s.*${domain}" "${ETC_HOSTS}" 2> /dev/null | sed -e "s|^${IP}\s*\([0-9a-z.-]${domain}\)$|\1|g"
|
||||
return
|
||||
fi
|
||||
waitNet
|
||||
trap 'rm -f "${TMPFILE}"' EXIT
|
||||
TMPFILE="$(mktemp)" || exit 1
|
||||
if [[ -n "${SIMU}" ]] ; then
|
||||
${SIMU} curl -X GET "${GANDI_API}/records" -H "authorization: Apikey ${GANDI_KEY}"
|
||||
else
|
||||
curl -X GET "${GANDI_API}/records" -H "authorization: Apikey ${GANDI_KEY}" 2>/dev/null | \
|
||||
sed "s/,{/\n/g" | \
|
||||
sed 's/.*rrset_name":"\([^"]*\)".*rrset_values":\["\([^"]*\)".*/\1:\2/g'| \
|
||||
grep -v '^[_@]'| \
|
||||
grep -e ":${domain}\.*$" -e ":prod[0-9]*$" > ${TMPFILE}
|
||||
fi
|
||||
if [ $# -lt 1 ]; then
|
||||
cat ${TMPFILE}
|
||||
else
|
||||
for ARG in $@
|
||||
do
|
||||
cat ${TMPFILE} | grep "${ARG}.*:"
|
||||
done
|
||||
fi
|
||||
}
|
||||
|
||||
saveDns () {
|
||||
for ARG in $@ ; do
|
||||
if [[ "${ARG}" =~ .local$ ]] ; then
|
||||
echo "${PRG}: old fasion style (remove .local at the end)"
|
||||
usage;
|
||||
fi
|
||||
if [[ "${ARG}" =~ .bzh$ ]] ; then
|
||||
echo "${PRG}: old fasion style (remove .bzh at the end)"
|
||||
usage;
|
||||
fi
|
||||
if [[ "${ARG}" =~ .dev$ ]] ; then
|
||||
echo "${PRG}: old fasion style (remove .dev at the end)"
|
||||
usage;
|
||||
fi
|
||||
done
|
||||
if [[ "${domain}" = "kaz.local" ]]; then
|
||||
return
|
||||
fi
|
||||
waitNet
|
||||
${SIMU} curl -X POST "${GANDI_API}/snapshots" -H "authorization: Apikey ${GANDI_KEY}" 2>/dev/null
|
||||
}
|
||||
|
||||
badName(){
|
||||
[[ -z "$1" ]] && return 0;
|
||||
for item in "${forbidenName[@]}"; do
|
||||
[[ "${item}" == "$1" ]] && [[ "${FORCE}" == "NO" ]] && return 0
|
||||
done
|
||||
return 1
|
||||
}
|
||||
|
||||
add(){
|
||||
if [ $# -lt 1 ]; then
|
||||
exit
|
||||
fi
|
||||
saveDns $@
|
||||
declare -a ADDED
|
||||
for ARG in $@
|
||||
do
|
||||
if badName "${ARG}" ; then
|
||||
echo "can't manage '${ARG}'. Use -f option"
|
||||
continue
|
||||
fi
|
||||
case "${domain}" in
|
||||
kaz.local )
|
||||
if grep -q --perl-regex "^${IP}.*[ \t]${ARG}.${domain}" "${ETC_HOSTS}" 2> /dev/null ; then
|
||||
break
|
||||
fi
|
||||
if grep -q --perl-regex "^${IP}[ \t]" "${ETC_HOSTS}" 2> /dev/null ; then
|
||||
${SIMU} sudo sed -i -e "0,/^${IP}[ \t]/s/^\(${IP}[ \t]\)/\1${ARG}.${domain} /g" "${ETC_HOSTS}"
|
||||
else
|
||||
${SIMU} sudo sed -i -e "$ a ${IP}\t${ARG}.${domain}" "${ETC_HOSTS}" 2> /dev/null
|
||||
fi
|
||||
;;
|
||||
*)
|
||||
${SIMU} curl -X POST "${GANDI_API}/records" -H "authorization: Apikey ${GANDI_KEY}" -H 'content-type: application/json' -d '{"rrset_type":"CNAME", "rrset_name":"'${ARG}'", "rrset_values":["'${site}'"]}'
|
||||
echo
|
||||
;;
|
||||
esac
|
||||
ADDED+=("${ARG}")
|
||||
done
|
||||
echo "Domains added to ${domain}: ${ADDED[@]}"
|
||||
}
|
||||
|
||||
del(){
|
||||
if [ $# -lt 1 ]; then
|
||||
exit
|
||||
fi
|
||||
saveDns $@
|
||||
declare -a REMOVED
|
||||
for ARG in $@
|
||||
do
|
||||
if badName "${ARG}" ; then
|
||||
echo "can't manage '${ARG}'. Use -f option"
|
||||
continue
|
||||
fi
|
||||
case "${domain}" in
|
||||
kaz.local )
|
||||
if !grep -q --perl-regex "^${IP}.*[ \t]${ARG}.${domain}" "${ETC_HOSTS}" 2> /dev/null ; then
|
||||
break
|
||||
fi
|
||||
${SIMU} sudo sed -i -e "/^${IP}[ \t]*${ARG}.${domain}[ \t]*$/d" \
|
||||
-e "s|^\(${IP}.*\)[ \t]${ARG}.${domain}|\1|g" "${ETC_HOSTS}"
|
||||
;;
|
||||
* )
|
||||
${SIMU} curl -X DELETE "${GANDI_API}/records/${ARG}" -H "authorization: Apikey ${GANDI_KEY}"
|
||||
echo
|
||||
;;
|
||||
esac
|
||||
REMOVED+=("${ARG}")
|
||||
done
|
||||
echo "Domains removed from ${domain}: ${REMOVED[@]}"
|
||||
}
|
||||
|
||||
#echo "CMD: ${CMD} $*"
|
||||
${CMD} $*
|
66
bin/envoiMails.sh
Executable file
@ -0,0 +1,66 @@
|
||||
#!/bin/bash
|
||||
|
||||
#kan: 09/09/2022
|
||||
#koi: envoyer un mail à une liste (sans utiliser sympa sur listes.kaz.bzh)
|
||||
#ki : fab
|
||||
|
||||
#on récupère toutes les variables et mdp
|
||||
# on prend comme source des repertoire le dossier du dessus ( /kaz dans notre cas )
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
SIMULATION=NO
|
||||
|
||||
CMD="/tmp/envoiMail_cmds_to_run.sh"
|
||||
echo "#!/bin/bash" > ${CMD} && chmod +x ${CMD}
|
||||
|
||||
#################################################################################################################
|
||||
MAIL_KAZ="
|
||||
KAZ, association morbihannaise, propose de \"dégoogliser\" l’internet avec des solutions et services numériques libres alternatifs à ceux des GAFAM. Elle invite les habitants et les élus de Vannes et de sa région à une réunion publique d’information et d’échange :
|
||||
Le jeudi 22 septembre 2022 à 18 heures
|
||||
à la Maison des associations, 31 rue Guillaume Le Bartz à Vannes.
|
||||
|
||||
Cette invitation est destinée à toute personne sensible aux enjeux du numérique, aux risques pour la démocratie et les libertés, à sa participation au dérèglement climatique et à l’épuisement des ressources de notre planète.
|
||||
|
||||
Nous dirons qui nous sommes, quelles sont nos valeurs et quels sont concrètement les solutions et services que nous proposons, leurs conditions d’accès et l’accompagnement utile pour leur prise en main.
|
||||
|
||||
Les premières pierres de KAZ ont été posées voilà bientôt deux ans, en pleine pandémie de la COVID, par sept citoyens qui ont voulu répondre à l’appel de Framasoft de dégoogliser l’internet. Ne plus se lamenter sur la puissance des GAFAM, mais proposer des solutions et des services numériques alternatifs à ceux de Google, Amazon, Facebook, Apple et Microsoft en s’appuyant sur les fondamentaux Sobre, Libre, Éthique et Local.
|
||||
|
||||
A ce jour, près de 200 particuliers ont ouvert une adresse @kaz.bz, plus de 80 organisations ont souscrit au bouquet de services de KAZ et près de 800 collaborateurs d’organisations utilisatrices des services de KAZ participent peu ou prou au réseau de KAZ.
|
||||
Beaucoup de services sont gratuits et accessibles sur le site https://kaz.bzh. D’autres demandent l’ouverture d’un compte moyennant une petite participation financière de 10€ par an pour les particuliers et de 30€ par an pour les organisations. Ceci est permis par le bénévolat des membres de la collégiale qui dirigent l’association et administrent les serveurs et les services.
|
||||
|
||||
A ce stade, de nombreuses questions se posent à KAZ. Quelle ambition de couverture de ses services auprès des particuliers et des organisations sur son territoire, le Morbihan ? Quel accompagnement dans la prise en main des outils ? Quels nouveaux services seraient utiles ?
|
||||
|
||||
Nous serions heureux de votre participation à notre réunion publique du 22 septembre d'une durée totale de 2h :
|
||||
* Présentation valeurs / contexte (15mn)
|
||||
* Présentation outils (45mn)
|
||||
* Questions / Réponses (1h)
|
||||
|
||||
En restant à votre disposition,
|
||||
|
||||
La Collégiale de KAZ
|
||||
"
|
||||
#################################################################################################################
|
||||
|
||||
FIC_MAILS="/tmp/fic_mails"
|
||||
|
||||
while read EMAIL_CIBLE
|
||||
do
|
||||
|
||||
echo "docker exec -i mailServ mailx -a 'Content-Type: text/plain; charset=\"UTF-8\"' -r contact@kaz.bzh -s \"Invitation rencontre KAZ jeudi 22 septembre à 18h00 à la Maison des assos à Vannes\" ${EMAIL_CIBLE} << EOF
|
||||
${MAIL_KAZ}
|
||||
EOF" | tee -a ${CMD}
|
||||
|
||||
echo "sleep 2" | tee -a ${CMD}
|
||||
|
||||
done < ${FIC_MAILS}
|
||||
|
||||
# des commandes à lancer ?
|
||||
if [ "${SIMULATION}" == "NO" ];then
|
||||
echo "on exécute"
|
||||
${CMD}
|
||||
else
|
||||
echo "Aucune commande n'a été lancée: Possibilité de le faire à la main. cf ${CMD}"
|
||||
fi
|
||||
|
240
bin/foreign-domain.sh
Executable file
@ -0,0 +1,240 @@
|
||||
#!/bin/bash
|
||||
|
||||
# list/ajout/supprime/ les domaines extérieurs à kaz.bzh
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
export PRG="$0"
|
||||
cd $(dirname $0)
|
||||
|
||||
. "${DOCKERS_ENV}"
|
||||
|
||||
LETS_DIR="/etc/letsencrypt/$([ "${mode}" == "local" ] && echo "local" || echo "live")"
|
||||
|
||||
declare -a availableComposes availableOrga
|
||||
availableComposes=(${pahekoHost} ${cloudHost} ${dokuwikiHost} ${wordpressHost} ${matterHost} ${castopodHost})
|
||||
availableOrga=($(sed -e "s/\(.*\)[ \t]*#.*$/\1/" -e "s/^[ \t]*\(.*\)-orga$/\1/" -e "/^$/d" "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
availableProxyComposes=($(getList "${KAZ_CONF_DIR}/container-proxy.list"))
|
||||
|
||||
# no more export in .env
|
||||
export $(set | grep "domain=")
|
||||
|
||||
export CMD=""
|
||||
export SIMU=""
|
||||
export CHANGE=""
|
||||
|
||||
usage(){
|
||||
echo "Usage: ${PRG} list [friend-domain...]"
|
||||
echo " ${PRG} [-n] add orga [${pahekoHost} ${cloudHost} ${dokuwikiHost} ${wordpressHost} ${matterHost} ${castopodHost}] [friend-domain...] "
|
||||
echo " ${PRG} [-n] del [friend-domain...]"
|
||||
echo " ${PRG} -l"
|
||||
echo " -l short list"
|
||||
echo " -renewAll"
|
||||
echo " -h help"
|
||||
echo " -n simulation"
|
||||
exit 1
|
||||
}
|
||||
|
||||
export CERT_CFG="${KAZ_CONF_PROXY_DIR}/foreign-certificate"
|
||||
|
||||
createCert () {
|
||||
(
|
||||
fileName="${LETS_DIR}/$1-key.pem"
|
||||
#[ -f "${fileName}" ] || return
|
||||
# if [ -f "${fileName}" ]; then
|
||||
# fileTime=$(stat --format='%Y' "${fileName}")
|
||||
# current_time=$(date +%s)
|
||||
# if (( "${fileTime}" > ( "${current_time}" - ( 60 * 60 * 24 * 89 ) ) )); then
|
||||
# exit
|
||||
# fi
|
||||
# fi
|
||||
printKazMsg "create certificat for $1"
|
||||
${SIMU} docker exec -i proxyServ bash -c "/opt/certbot/bin/certbot certonly -n --nginx -d $1"
|
||||
)
|
||||
|
||||
}
|
||||
|
||||
for ARG in $@; do
|
||||
case "${ARG}" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
'-n' )
|
||||
shift
|
||||
export SIMU="echo"
|
||||
;;
|
||||
'-renewAll')
|
||||
for i in $("${KAZ_BIN_DIR}/foreign-domain.sh" -l); do
|
||||
echo "$i"
|
||||
createCert "$i" |grep failed
|
||||
done
|
||||
exit
|
||||
;;
|
||||
'-l')
|
||||
for compose in ${availableComposes[@]} ; do
|
||||
grep "server_name" "${KAZ_CONF_PROXY_DIR}/${compose}_kaz_name" | sed -e "s/[ \t]*\([^#]*\)#.*/\1/g" -e "/^$/d" -e "s/.*server_name[ \t]\([^ ;]*\).*/\1/"
|
||||
done
|
||||
exit
|
||||
;;
|
||||
'list'|'add'|'del' )
|
||||
shift
|
||||
CMD="${ARG}"
|
||||
break
|
||||
;;
|
||||
* )
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z "${CMD}" ]; then
|
||||
echo "Commande missing"
|
||||
usage
|
||||
fi
|
||||
|
||||
########################################
|
||||
badDomaine () {
|
||||
[[ -z "$1" ]] && return 0;
|
||||
[[ ! "$1" =~ ^[-.a-zA-Z0-9]*$ ]] && return 0;
|
||||
return 1
|
||||
}
|
||||
badOrga () {
|
||||
[[ -z "$1" ]] && return 0;
|
||||
[[ ! " ${availableOrga[*]} " =~ " $1 " ]] && return 0
|
||||
return 1
|
||||
}
|
||||
badCompose () {
|
||||
[[ -z "$1" ]] && return 0;
|
||||
[[ ! " ${availableComposes[*]} " =~ " $1 " ]] && return 0
|
||||
return 1
|
||||
}
|
||||
|
||||
########################################
|
||||
listServ () {
|
||||
for compose in ${availableComposes[@]} ; do
|
||||
sed -e "s/[ \t]*\([^#]*\)#.*/\1/g" -e "/^$/d" -e "s/.*server_name[ \t]\([^ ;]*\).*/\1 : ${compose}/" "${KAZ_CONF_PROXY_DIR}/${compose}_kaz_name"
|
||||
done
|
||||
}
|
||||
|
||||
listOrgaServ () {
|
||||
for compose in ${availableComposes[@]} ; do
|
||||
sed -e "s/[ \t]*\([^#]*\)#.*/\1/g" -e "/^$/d" -e "s/\([^ ]*\)[ \t]*\([^ \t;]*\).*/\1 => \2 : ${compose}/" "${KAZ_CONF_PROXY_DIR}/${compose}_kaz_map"
|
||||
done
|
||||
}
|
||||
|
||||
########################################
|
||||
list () {
|
||||
previousOrga=$(listOrgaServ)
|
||||
previousServ=$(listServ)
|
||||
if [ $# -lt 1 ]; then
|
||||
[ -n "${previousOrga}" ] && echo "${previousOrga}"
|
||||
[ -n "${previousServ}" ] && echo "${previousServ}"
|
||||
return
|
||||
fi
|
||||
for ARG in $@
|
||||
do
|
||||
orga=$(echo "${previousOrga}" | grep "${ARG}.* =>")
|
||||
serv=$(echo "${previousServ}" | grep "${ARG}.* =>")
|
||||
[ -n "${orga}" ] && echo "${orga}"
|
||||
[ -n "${serv}" ] && echo "${serv}"
|
||||
done
|
||||
}
|
||||
|
||||
########################################
|
||||
add () {
|
||||
# $1 : orga
|
||||
# $2 : service
|
||||
# $3 : friend-domain
|
||||
[ $# -lt 3 ] && usage
|
||||
badOrga $1 && echo "bad orga: ${RED}$1${NC} not in ${GREEN}${availableOrga[@]}${NC}" && usage
|
||||
badCompose $2 && echo "bad compose: ${RED}$2${NC} not in ${GREEN}${availableComposes[@]}${NC}" && usage
|
||||
ORGA=$1
|
||||
COMPOSE=$2
|
||||
shift; shift
|
||||
CLOUD_SERVNAME="${ORGA}-${nextcloudServName}"
|
||||
CLOUD_CONFIG="${DOCK_VOL}/orga_${ORGA}-cloudConfig/_data/config.php"
|
||||
|
||||
# XXX check compose exist in orga ?
|
||||
# /kaz/bin/kazList.sh service enable ${ORGA}
|
||||
if [ "${COMPOSE}" = "${cloudHost}" ]; then
|
||||
if ! [[ "$(docker ps -f name=${CLOUD_SERVNAME} | grep -w ${CLOUD_SERVNAME})" ]]; then
|
||||
printKazError "${CLOUD_SERVNAME} not running... abort"
|
||||
exit
|
||||
fi
|
||||
fi
|
||||
|
||||
for FRIEND in $@; do
|
||||
badDomaine "${FRIEND}" && echo "bad domaine: ${RED}${FRIEND}${NC}" && usage
|
||||
done
|
||||
|
||||
for FRIEND in $@; do
|
||||
createCert "${FRIEND}"
|
||||
if [ "${COMPOSE}" = "${cloudHost}" ]; then
|
||||
IDX=$(awk 'BEGIN {flag=0; cpt=0} /trusted_domains/ {flag=1} /)/ {if (flag) {print cpt+1; exit 0}} / => / {if (flag && cpt<$1) cpt=$1}' "${CLOUD_CONFIG}")
|
||||
${SIMU} docker exec -ti -u 33 "${CLOUD_SERVNAME}" /var/www/html/occ config:system:set trusted_domains "${IDX}" --value="${FRIEND}"
|
||||
fi
|
||||
|
||||
previousOrga=$(listOrgaServ | grep "${FRIEND}")
|
||||
[[ " ${previousOrga}" =~ " ${FRIEND} => ${ORGA} : ${COMPOSE}" ]] && echo " - already done" && continue
|
||||
[[ " ${previousOrga}" =~ " ${FRIEND} " ]] && echo " - ${YELLOW}${BOLD}$(echo "${previousOrga}" | grep -e "${FRIEND}")${NC} must be deleted before" && return
|
||||
if [[ -n "${SIMU}" ]] ; then
|
||||
echo "${FRIEND} ${ORGA}; => ${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_map"
|
||||
cat <<EOF
|
||||
=> ${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_name
|
||||
server_name ${FRIEND};
|
||||
EOF
|
||||
else
|
||||
echo "${FRIEND} ${ORGA};" >> "${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_map"
|
||||
cat >> "${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_name" <<EOF
|
||||
server_name ${FRIEND};
|
||||
EOF
|
||||
fi
|
||||
echo "${PRG}: ${FRIEND} added"
|
||||
|
||||
CHANGE="add"
|
||||
done
|
||||
#(cd "${KAZ_COMP_DIR}/${ORGA}-orga"; docker-compose restart)
|
||||
}
|
||||
|
||||
########################################
|
||||
del () {
|
||||
[ $# -lt 1 ] && usage
|
||||
|
||||
for FRIEND in $@; do
|
||||
badDomaine "${FRIEND}" && echo "bad domaine: ${RED}${FRIEND}${NC}" && usage
|
||||
previous=$(listOrgaServ | grep -e "${FRIEND}")
|
||||
[[ ! "${previous}" =~ ^${FRIEND} ]] && echo "${FRIEND} not found in ${previous}" && continue
|
||||
# XXX if done OK
|
||||
for COMPOSE in ${availableComposes[@]} ; do
|
||||
if grep -q -e "^[ \t]*${FRIEND}[ \t]" "${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_map" ; then
|
||||
if [ "${COMPOSE}" = "${cloudHost}" ]; then
|
||||
ORGA="$(grep "${FRIEND}" "${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_map" | sed "s/^${FRIEND}\s*\([^;]*\);/\1/")"
|
||||
CLOUD_CONFIG="${DOCK_VOL}/orga_${ORGA}-cloudConfig/_data/config.php"
|
||||
${SIMU} sed -e "/\d*\s*=>\s*'${FRIEND}'/d" -i "${CLOUD_CONFIG}"
|
||||
fi
|
||||
${SIMU} sed -e "/^[ \t]*${FRIEND}[ \t]/d" -i "${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_map"
|
||||
fi
|
||||
if grep -q -e "^[ \t]*server_name ${FRIEND};" "${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_name" ; then
|
||||
${SIMU} sed -i "${KAZ_CONF_PROXY_DIR}/${COMPOSE}_kaz_name" \
|
||||
-e "/^[ \t]*server_name ${FRIEND};/d"
|
||||
fi
|
||||
done
|
||||
echo "${PRG}: ${FRIEND} deleted"
|
||||
CHANGE="del"
|
||||
done
|
||||
}
|
||||
|
||||
########################################
|
||||
${CMD} $@
|
||||
|
||||
if [ -n "${CHANGE}" ] ; then
|
||||
echo "Reload proxy conf"
|
||||
for item in "${availableProxyComposes[@]}"; do
|
||||
${SIMU} ${KAZ_COMP_DIR}/${item}/proxy-gen.sh
|
||||
${SIMU} "${KAZ_COMP_DIR}/proxy/reload.sh"
|
||||
done
|
||||
fi
|
||||
|
||||
########################################
|
648
bin/gestContainers.sh
Executable file
@ -0,0 +1,648 @@
|
||||
#!/bin/bash
|
||||
# Script de manipulation des containers en masse
|
||||
# init /versions / restart ...
|
||||
#
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
PRG=$(basename $0)
|
||||
|
||||
#TODO: ce tab doit être construit à partir de la liste des machines dispos et pas en dur
|
||||
tab_sites_destinations_possibles=("kazoulet" "prod2" "prod1")
|
||||
|
||||
#GLOBAL VARS
|
||||
NAS_VOL="/mnt/disk-nas1/docker/volumes/"
|
||||
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
AVAILABLE_ORGAS=${availableOrga[*]//-orga/}
|
||||
|
||||
availableContainersCommuns=( $(getList "${KAZ_CONF_DIR}/container-withMail.list") $(getList "${KAZ_CONF_DIR}/container-withoutMail.list"))
|
||||
|
||||
OPERATE_ON_MAIN= # par defaut NON on ne traite que des orgas
|
||||
OPERATE_ON_NAS_ORGA="OUI" # par defaut oui, on va aussi sur les orgas du NAS
|
||||
OPERATE_LOCAL_ORGA="OUI" # par defaut oui
|
||||
|
||||
TEMPO_ACTION_STOP=2 # Lors de redémarrage avec tempo, on attend après le stop
|
||||
TEMPO_ACTION_START=60 # Lors de redémarrage avec tempo, avant de reload le proxy
|
||||
|
||||
CONTAINERS_TYPES=
|
||||
|
||||
defaultContainersTypes="cloud agora wp wiki office paheko castopod" # les containers gérés par ce script.
|
||||
|
||||
declare -A DockerServNames # le nom des containers correspondant
|
||||
DockerServNames=( [cloud]="${nextcloudServName}" [agora]="${mattermostServName}" [wiki]="${dokuwikiServName}" [wp]="${wordpressServName}" [office]="${officeServName}" [paheko]="${pahekoServName}" [castopod]="${castopodServName}" )
|
||||
|
||||
declare -A FilterLsVolume # Pour trouver quel volume appartient à quel container
|
||||
FilterLsVolume=( [cloud]="cloudMain" [agora]="matterConfig" [wiki]="wikiConf" [wp]="wordpress" [castopod]="castopodMedia" )
|
||||
|
||||
declare -A composeDirs # Le nom du repertoire compose pour le commun
|
||||
composeDirs=( [cloud]="cloud" [agora]="mattermost" [wiki]="dokuwiki" [office]="collabora" [paheko]="paheko" [castopod]="castopod" )
|
||||
|
||||
declare -A serviceNames # Le nom du du service dans le dockerfile d'orga
|
||||
serviceNames=( [cloud]="cloud" [agora]="agora" [wiki]="dokuwiki" [wp]="wordpress" [office]="collabora" [castopod]="castopod")
|
||||
|
||||
declare -A subScripts
|
||||
subScripts=( [cloud]="manageCloud.sh" [agora]="manageAgora.sh" [wiki]="manageWiki.sh" [wp]="manageWp.sh" [castopod]="manageCastopod.sh" )
|
||||
|
||||
declare -A OrgasOnNAS
|
||||
declare -A OrgasLocales
|
||||
declare -A NbOrgas
|
||||
declare -A RunningOrgas
|
||||
declare -A Posts
|
||||
|
||||
QUIET="1" # redirection des echo
|
||||
|
||||
OCCCOMANDS=()
|
||||
MMCTLCOMANDS=()
|
||||
EXECCOMANDS=()
|
||||
|
||||
# CLOUD
|
||||
APPLIS_PAR_DEFAUT="tasks calendar contacts bookmarks mail richdocuments external drawio snappymail ransomware_protection" #rainloop richdocumentscode
|
||||
|
||||
|
||||
usage() {
|
||||
echo "${PRG} [OPTION] [CONTAINERS_TYPES] [COMMANDES] [ORGAS]
|
||||
Ce script regroupe l'ensemble des opérations que l'on souhaite automatiser sur plusieurs containers
|
||||
Par defaut, sur les orgas, mais on peut aussi ajouter les communs
|
||||
|
||||
OPTIONS
|
||||
-h|--help Cette aide :-)
|
||||
-n|--simu SIMULATION
|
||||
-q|--quiet On ne parle pas (utile avec le -n pour avoir que les commandes)
|
||||
-m|--main Traite aussi le container commun (cloud commun / agora commun / wiki commun)
|
||||
-M Ne traite que le container commun, et pas les orgas
|
||||
--nas Ne traite QUE les orgas sur le NAS
|
||||
--local Ne traite pas les orgas sur le NAS
|
||||
-v|--version Donne la version des containers et signale les MàJ
|
||||
-l|--list Liste des containers (up / down, local ou nas) de cette machine
|
||||
|
||||
CONTAINERS_TYPES
|
||||
-cloud Pour agir sur les clouds
|
||||
-agora Pour agir sur les agoras
|
||||
-wp Les wp
|
||||
-wiki Les wiki
|
||||
-office Les collabora
|
||||
-paheko Le paheko
|
||||
-castopod Les castopod
|
||||
|
||||
COMMANDES (on peut en mettre plusieurs dans l'ordre souhaité)
|
||||
-I|--install L'initialisation du container
|
||||
-t Redémarre avec tempo (docker-compose down puis sleep ${TEMPO_ACTION_STOP} puis up puis sleep ${TEMPO_ACTION_START})
|
||||
-r Redémarre sans tempo (docker restart)
|
||||
-exec \"command\" Envoie une commande docker exec
|
||||
|
||||
--optim Lance la procédure Nextcloud pour optimiser les performances ** **
|
||||
-occ \"command\" Envoie une commande via occ ** **
|
||||
-u Mets à jour les applis ** SPECIFIQUES **
|
||||
-i Install des applis ** CLOUD **
|
||||
-a \"app1 app2 ...\" Choix des appli à installer ou mettre à jour (entre guillemets) ** **
|
||||
-U|--upgrade Upgrade des clouds ** **
|
||||
|
||||
-mmctl \"command\" Envoie une commande via mmctl ** SPECIFIQUES **
|
||||
-p|--post \"team\" \"message\" Poste un message dans une team agora ** AGORA **
|
||||
|
||||
ORGAS
|
||||
[orga1 orga2 ... ] on peut filtrer parmi : ${AVAILABLE_ORGAS}
|
||||
|
||||
|
||||
Exemples :
|
||||
${PRG} -office -m -r restart de tous les collaboras (libére RAM)
|
||||
${PRG} -cloud -u -r -q -n Affiche toutes les commandes (-n -q ) pour mettre à jour toutes les applis des clouds + restart (-u -r)
|
||||
${PRG} -p \"monorga:town-square\" \"Hello\" monorga # envoie Hello sur le centreville de l'orga monorga sur son mattermost dédié
|
||||
|
||||
"
|
||||
}
|
||||
|
||||
|
||||
####################################################
|
||||
################ fonctions clefs ###################
|
||||
####################################################
|
||||
|
||||
_populate_lists(){
|
||||
# récupère les listes d'orga à traiter
|
||||
# on rempli les tableaux OrgasOnNAS / OrgasLocales / NbOrgas ... par type de container
|
||||
|
||||
if [ -z "${CONTAINERS_TYPES}" ]; then
|
||||
# wow, on traite tout le monde d'un coup...
|
||||
CONTAINERS_TYPES="$defaultContainersTypes"
|
||||
fi
|
||||
|
||||
for TYPE in ${CONTAINERS_TYPES}; do
|
||||
if [ -n "${FilterLsVolume[$TYPE]}" ] ; then # on regarde dans les volumes
|
||||
[ -n "$OPERATE_ON_NAS_ORGA" ] && OrgasOnNAS["$TYPE"]=$( _getListOrgas ${NAS_VOL} ${FilterLsVolume[$TYPE]} )
|
||||
[ -n "$OPERATE_LOCAL_ORGA" ] && OrgasLocales["$TYPE"]=$( _getListOrgas ${DOCK_VOL} ${FilterLsVolume[$TYPE]} "SANSLN")
|
||||
else # un docker ps s'il n'y a pas de volumes
|
||||
[ -n "$OPERATE_LOCAL_ORGA" ] && OrgasLocales["$TYPE"]=$(docker ps --format '{{.Names}}' | grep ${DockerServNames[$TYPE]} | sed -e "s/-*${DockerServNames[$TYPE]}//")
|
||||
fi
|
||||
NbOrgas["$TYPE"]=$(($(echo ${OrgasOnNAS["$TYPE"]} | wc -w) + $(echo ${OrgasLocales["$TYPE"]} | wc -w)))
|
||||
RunningOrgas["$TYPE"]=$(docker ps --format '{{.Names}}' | grep ${DockerServNames[$TYPE]} | sed -e "s/-*${DockerServNames[$TYPE]}//")
|
||||
done
|
||||
}
|
||||
|
||||
_getListOrgas(){
|
||||
# retrouve les orgas à partir des volume présents
|
||||
# $1 where to lookup
|
||||
# $2 filter
|
||||
# $3 removeSymbolicLinks
|
||||
[ ! -d $1 ] || [ -z "$2" ] && return 1 # si le repertoire n'existe pas on skip
|
||||
LIST=$(ls "${1}" | grep -i orga | grep -i "$2" | sed -e "s/-${2}$//g" | sed -e 's/^orga_//')
|
||||
[ -n "$3" ] && LIST=$(ls -F "${1}" | grep '/' | grep -i orga | grep -i "$2" | sed -e "s/-${2}\/$//g" | sed -e 's/^orga_//')
|
||||
LIST=$(comm -12 <(printf '%s\n' ${LIST} | sort) <(printf '%s\n' ${AVAILABLE_ORGAS} | sort))
|
||||
echo "$LIST"
|
||||
}
|
||||
|
||||
_executeFunctionForAll(){
|
||||
# Parcours des container et lancement des commandes
|
||||
# Les commandes ont en derniers paramètres le type et l'orga et une string parmi KAZ/ORGANAS/ORGALOCAL pour savoir sur quoi on opère
|
||||
# $1 function
|
||||
# $2 nom de la commande
|
||||
# $3 quel types de containers
|
||||
# $4 params : quels paramètres à passer à la commande (les clefs sont #ORGA# #DOCKERSERVNAME# #SURNAS# #ISMAIN# #TYPE# #COMPOSEDIR# )
|
||||
for TYPE in ${3}; do
|
||||
if [ -n "$OPERATE_ON_MAIN" ]; then
|
||||
if [[ -n "${composeDirs[$TYPE]}" && "${availableContainersCommuns[*]}" =~ "${composeDirs[$TYPE]}" ]]; then # pas de cloud / agora / wp / wiki sur cette instance
|
||||
Dockername=${DockerServNames[$TYPE]}
|
||||
PARAMS=$(echo $4 | sed -e "s/#ORGA#//g;s/#DOCKERSERVNAME#/$Dockername/g;s/#ISMAIN#/OUI/g;s/#SURNAS#/NON/g;s/#TYPE#/$TYPE/g;s%#COMPOSEDIR#%${KAZ_COMP_DIR}/${composeDirs[$TYPE]}%g" )
|
||||
echo "-------- $2 $TYPE COMMUN ----------------------------" >& $QUIET
|
||||
eval "$1" $PARAMS
|
||||
fi
|
||||
fi
|
||||
if [[ ${NbOrgas[$TYPE]} -gt 0 ]]; then
|
||||
echo "-------- $2 des $TYPE des ORGAS ----------------------------" >& $QUIET
|
||||
COMPTEUR=1
|
||||
if [ -n "$OPERATE_LOCAL_ORGA" ]; then
|
||||
for ORGA in ${OrgasLocales[$TYPE]}; do
|
||||
Dockername=${ORGA}-${DockerServNames[$TYPE]}
|
||||
PARAMS=$(echo $4 | sed -e "s/#ORGA#/${ORGA}/g;s/#DOCKERSERVNAME#/$Dockername/g;s/#ISMAIN#/NON/g;s/#SURNAS#/NON/g;s/#TYPE#/$TYPE/g;s%#COMPOSEDIR#%${KAZ_COMP_DIR}/${ORGA}-orga%g" )
|
||||
echo "${RED} ${ORGA}-orga ${NC}($COMPTEUR/${NbOrgas[$TYPE]})" >& $QUIET
|
||||
eval "$1" $PARAMS
|
||||
COMPTEUR=$((COMPTEUR + 1))
|
||||
done
|
||||
fi
|
||||
if [ -n "$OPERATE_ON_NAS_ORGA" ]; then
|
||||
for ORGA in ${OrgasOnNAS[$TYPE]}; do
|
||||
Dockername=${ORGA}-${DockerServNames[$TYPE]}
|
||||
PARAMS=$(echo $4 | sed -e "s/#ORGA#/${ORGA}/g;s/#DOCKERSERVNAME#/$Dockername/g;s/#ISMAIN#/NON/g;s/#SURNAS#/OUI/g;s/#TYPE#/$TYPE/g;s%#COMPOSEDIR#%${KAZ_COMP_DIR}/${ORGA}-orga%g" )
|
||||
echo "${RED} ${ORGA}-orga ${NC}($COMPTEUR/${NbOrgas[$TYPE]})" >& $QUIET
|
||||
eval "$1" $PARAMS
|
||||
COMPTEUR=$((COMPTEUR + 1))
|
||||
done
|
||||
fi
|
||||
fi
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
##############################################
|
||||
################ COMMANDES ###################
|
||||
##############################################
|
||||
Init(){
|
||||
# Initialisation des containers
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Initialisation" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
_executeFunctionForAll "_initContainer" "Initialisation" "${CONTAINERS_TYPES[@]}" "#TYPE# #ISMAIN# #SURNAS# #ORGA# "
|
||||
}
|
||||
|
||||
restart-compose() {
|
||||
# Parcours les containers et redémarre avec tempo
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "DOCKER-COMPOSE DOWN puis sleep ${TEMPO_ACTION_STOP}" >& $QUIET
|
||||
echo "DOCKER-COMPOSE UP puis sleep ${TEMPO_ACTION_START}" >& $QUIET
|
||||
echo "de ${CONTAINERS_TYPES} pour $NB_ORGAS_STR" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
_executeFunctionForAll "_restartContainerAvecTempo" "Restart" "${CONTAINERS_TYPES[@]}" "#TYPE# #ISMAIN# #COMPOSEDIR#"
|
||||
|
||||
${SIMU} sleep ${TEMPO_ACTION_START}
|
||||
_reloadProxy
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
echo "${GREEN}FIN${NC} " >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
}
|
||||
|
||||
restart() {
|
||||
# Parcours les containers et redémarre
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "DOCKER RESTART des ${CONTAINERS_TYPES} pour $NB_ORGAS_STR" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
_executeFunctionForAll "_restartContainer" "Restart" "${CONTAINERS_TYPES[@]}" "#DOCKERSERVNAME#"
|
||||
|
||||
_reloadProxy
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
echo "${GREEN}FIN${NC} " >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
}
|
||||
|
||||
version(){
|
||||
# Parcours les containers et affiche leurs versions
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "VERSIONS" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
_executeFunctionForAll "_versionContainer" "Version" "${CONTAINERS_TYPES[@]}" "#TYPE# #ISMAIN# #ORGA#"
|
||||
}
|
||||
|
||||
|
||||
listContainers(){
|
||||
echo "${NC}--------------------------------------------------------"
|
||||
echo "LISTES"
|
||||
echo "------------------------------------------------------------"
|
||||
for TYPE in ${CONTAINERS_TYPES}; do
|
||||
echo "****************** $TYPE ****************"
|
||||
_listContainer "$TYPE"
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
|
||||
######################## Fonctions génériques #######################
|
||||
|
||||
_initContainer(){
|
||||
# $1 type
|
||||
# $2 COMMUN
|
||||
# $3 ON NAS
|
||||
# $4 orgas
|
||||
if [ -n "${subScripts[$1]}" ] ; then
|
||||
evalStr="${KAZ_BIN_DIR}/${subScripts[$1]} --install"
|
||||
if [ "$3" = "OUI" ]; then evalStr="${evalStr} -nas" ; fi
|
||||
if [ ! "$QUIET" = "1" ]; then evalStr="${evalStr} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then evalStr="${evalStr} -n" ; fi
|
||||
if [ ! "$2" = "OUI" ]; then evalStr="${evalStr} $4" ; fi
|
||||
eval $evalStr
|
||||
fi
|
||||
}
|
||||
|
||||
_restartContainer(){
|
||||
# $1 Dockername
|
||||
echo -n "${NC}Redemarrage ... " >& $QUIET
|
||||
${SIMU}
|
||||
${SIMU} docker restart $1
|
||||
echo "${GREEN}OK${NC}" >& $QUIET
|
||||
}
|
||||
|
||||
_restartContainerAvecTempo(){
|
||||
# $1 type
|
||||
# $2 main container
|
||||
# $2 composeDir
|
||||
dir=$3
|
||||
if [ -z $dir ]; then return 1; fi # le compose n'existe pas ... par exemple wordpress commun
|
||||
cd "$dir"
|
||||
echo -n "${NC}Arrêt ... " >& $QUIET
|
||||
${SIMU}
|
||||
if [ "$2" = "OUI" ]; then ${SIMU} docker-compose stop ;
|
||||
else ${SIMU} docker-compose stop "${serviceNames[$1]}"
|
||||
fi
|
||||
${SIMU} sleep ${TEMPO_ACTION_STOP}
|
||||
echo "${GREEN}OK${NC}" >& $QUIET
|
||||
echo -n "${NC}Démarrage ... " >& $QUIET
|
||||
if [ "$2" = "OUI" ]; then ${SIMU} docker-compose up -d ;
|
||||
else ${SIMU} docker-compose up -d "${serviceNames[$1]}"
|
||||
fi
|
||||
${SIMU} sleep ${TEMPO_ACTION_START}
|
||||
echo "${GREEN}OK${NC}" >& $QUIET
|
||||
}
|
||||
|
||||
|
||||
_reloadProxy() {
|
||||
availableProxyComposes=($(getList "${KAZ_CONF_DIR}/container-proxy.list"))
|
||||
|
||||
for item in "${availableProxyComposes[@]}"; do
|
||||
${SIMU} ${KAZ_COMP_DIR}/${item}/reload.sh
|
||||
done
|
||||
}
|
||||
|
||||
_versionContainer() {
|
||||
# Affiche la version d'un container donné
|
||||
# $1 type
|
||||
# $2 COMMUN
|
||||
# $3 orgas
|
||||
if [ -n "${subScripts[$1]}" ] ; then
|
||||
evalStr="${KAZ_BIN_DIR}/${subScripts[$1]} --version"
|
||||
if [ ! "$2" = "OUI" ]; then evalStr="${evalStr} $3" ; fi
|
||||
eval $evalStr
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
_listContainer(){
|
||||
# pour un type donné (cloud / agora / wiki / wp), fait une synthèse de qui est up et down / nas ou local
|
||||
# $1 type
|
||||
RUNNING_FROM_NAS=$(comm -12 <(printf '%s\n' ${OrgasOnNAS[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort) | sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
RUNNING_LOCAL=$(comm -12 <(printf '%s\n' ${OrgasLocales[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort) | sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
# tu l'a vu la belle commande pour faire une exclusion de liste
|
||||
DOWN_ON_NAS=$(comm -23 <(printf '%s\n' ${OrgasOnNAS[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort) | sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
DOWN_LOCAL=$(comm -23 <(printf '%s\n' ${OrgasLocales[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort)| sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
NB_SUR_NAS=$(echo ${OrgasOnNAS[$1]} | wc -w)
|
||||
NB_LOCAUX=$(echo ${OrgasLocales[$1]} | wc -w)
|
||||
NB_RUNNING_SUR_NAS=$(echo $RUNNING_FROM_NAS | wc -w)
|
||||
NB_RUNNING_LOCALLY=$(echo $RUNNING_LOCAL | wc -w)
|
||||
MAIN_RUNNING="${RED}DOWN${NC}"
|
||||
if docker ps | grep -q " ${DockerServNames[$1]}"
|
||||
then
|
||||
MAIN_RUNNING="${GREEN}UP${NC}"
|
||||
fi
|
||||
|
||||
[ -n "${composeDirs[${1}]}" ] && echo "${NC}Le ${1} commun est $MAIN_RUNNING"
|
||||
if [[ ${NbOrgas[$1]} -gt 0 ]]; then
|
||||
ENLOCALSTR=
|
||||
if [[ ${NB_RUNNING_SUR_NAS[$1]} -gt 0 ]]; then ENLOCALSTR=" en local" ; fi
|
||||
echo "Orgas : $NB_RUNNING_LOCALLY / $NB_LOCAUX running ${1}$ENLOCALSTR"
|
||||
echo "${NC}UP : ${GREEN}${RUNNING_LOCAL}"
|
||||
echo "${NC}DOWN : ${RED}$DOWN_LOCAL${NC}"
|
||||
if [[ ${NB_RUNNING_SUR_NAS[$1]} -gt 0 ]]; then
|
||||
echo "${NC}Orgas : $NB_RUNNING_SUR_NAS / $NB_SUR_NAS running depuis le NAS :"
|
||||
echo "${NC}UP : ${GREEN}${RUNNING_FROM_NAS}"
|
||||
echo "${NC}DOWN : ${RED}$DOWN_ON_NAS${NC}"
|
||||
fi
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
#########################################################
|
||||
############# FONCTIONS SPECIFIQUES #####################
|
||||
#########################################################
|
||||
|
||||
##################################
|
||||
############### CLOUD ############
|
||||
##################################
|
||||
|
||||
UpgradeClouds() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "UPGRADE des cloud" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
RunOCCCommand "upgrade"
|
||||
}
|
||||
|
||||
OptimiseClouds() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Optimisation des cloud" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
RunOCCCommand "db:add-missing-indices" "db:convert-filecache-bigint --no-interaction"
|
||||
}
|
||||
|
||||
InstallApps(){
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "INSTALL DES APPLIS sur les clouds : ${LISTE_APPS}" >& $QUIET
|
||||
echo "-------------------------------------------------------------" >& $QUIET
|
||||
if [ -z "${LISTE_APPS}" ]; then
|
||||
echo "Aucune appli n'est précisée, j'installe les applis par défaut : ${APPLIS_PAR_DEFAUT}" >& $QUIET
|
||||
LISTE_APPS="${APPLIS_PAR_DEFAUT}"
|
||||
fi
|
||||
PARAMS="-a \"$LISTE_APPS\""
|
||||
if [ ! "$QUIET" = "1" ]; then PARAMS="${PARAMS} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then PARAMS="${PARAMS} -n" ; fi
|
||||
_executeFunctionForAll "${KAZ_BIN_DIR}/${subScripts["cloud"]} -i $PARAMS" "Install des applis" "cloud" "#ORGA#"
|
||||
}
|
||||
|
||||
UpdateApplis() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "UPDATE DES APPLIS des cloud : ${LISTE_APPS}" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
PARAMS="-a ${LISTE_APPS}"
|
||||
if [ -z "${LISTE_APPS}" ]; then
|
||||
echo "Aucune appli n'est précisée, je les met toutes à jour! " >& $QUIET
|
||||
PARAMS=
|
||||
fi
|
||||
if [ ! "$QUIET" = "1" ]; then PARAMS="${PARAMS} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then PARAMS="${PARAMS} -n" ; fi
|
||||
_executeFunctionForAll "${KAZ_BIN_DIR}/${subScripts["cloud"]} -u $PARAMS" "Maj des applis" "cloud" "#ORGA#"
|
||||
}
|
||||
|
||||
|
||||
##################################
|
||||
############### AGORA ############
|
||||
##################################
|
||||
|
||||
|
||||
PostMessages(){
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Envoi de messages sur mattermost" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
for TEAM in "${!Posts[@]}"
|
||||
do
|
||||
MSG=${Posts[$TEAM]/\"/\\\"}
|
||||
PARAMS="-p \"$TEAM\" \"$MSG\""
|
||||
if [ ! "$QUIET" = "1" ]; then PARAMS="${PARAMS} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then PARAMS="${PARAMS} -n" ; fi
|
||||
_executeFunctionForAll "${KAZ_BIN_DIR}/${subScripts["agora"]} $PARAMS" "Post vers $TEAM sur l'agora" "agora" "#ORGA#"
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
########## LANCEMENT COMMANDES OCC / MMCTL ############
|
||||
|
||||
RunCommands() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Envoi de commandes en direct" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
# $1 OCC / MMCTL / EXEC
|
||||
# $ suivants : les commandes
|
||||
for command in "${@:2}"
|
||||
do
|
||||
if [ $1 = "OCC" ]; then RunOCCCommand "$command" ; fi
|
||||
if [ $1 = "MMCTL" ]; then RunMMCTLCommand "$command" ; fi
|
||||
if [ $1 = "EXEC" ]; then RunEXECCommand "$command" ; fi
|
||||
done
|
||||
}
|
||||
|
||||
_runSingleOccCommand(){
|
||||
# $1 Command
|
||||
# $2 Dockername
|
||||
${SIMU} docker exec -u 33 $2 /var/www/html/occ $1
|
||||
}
|
||||
|
||||
_runSingleMmctlCommand(){
|
||||
# $1 Command
|
||||
# $2 Dockername
|
||||
${SIMU} docker exec $2 bin/mmctl $1
|
||||
}
|
||||
|
||||
_runSingleExecCommand(){
|
||||
# $1 Command
|
||||
# $2 Dockername
|
||||
${SIMU} docker exec $2 $1
|
||||
}
|
||||
|
||||
RunOCCCommand() {
|
||||
# $1 Command
|
||||
_executeFunctionForAll "_runSingleOccCommand \"${1}\"" "OCC $1" "cloud" "#DOCKERSERVNAME#"
|
||||
}
|
||||
|
||||
RunMMCTLCommand() {
|
||||
# $1 Command
|
||||
_executeFunctionForAll "_runSingleMmctlCommand \"${1}\"" "MMCTL $1" "agora" "#DOCKERSERVNAME#"
|
||||
}
|
||||
|
||||
RunEXECCommand() {
|
||||
# $1 Command
|
||||
_executeFunctionForAll "_runSingleExecCommand \"${1}\"" "docker exec $1" "${CONTAINERS_TYPES[@]}" "#DOCKERSERVNAME#"
|
||||
}
|
||||
|
||||
|
||||
|
||||
########## Main #################
|
||||
for ARG in "$@"; do
|
||||
if [ -n "${GETOCCCOMAND}" ]; then # après un -occ
|
||||
OCCCOMANDS+=("${ARG}")
|
||||
GETOCCCOMAND=
|
||||
elif [ -n "${GETEXECCOMAND}" ]; then # après un -exec
|
||||
EXECCOMANDS+=("${ARG}")
|
||||
GETEXECCOMAND=
|
||||
elif [ -n "${GETAPPS}" ]; then # après un -a
|
||||
LISTE_APPS="${LISTE_APPS} ${ARG}"
|
||||
GETAPPS=""
|
||||
elif [ -n "${GETMMCTLCOMAND}" ]; then # après un -mmctl
|
||||
MMCTLCOMANDS+=("${ARG}")
|
||||
GETMMCTLCOMAND=
|
||||
elif [ -n "${GETTEAM}" ]; then # après un --post
|
||||
GETMESSAGE="now"
|
||||
GETTEAM=""
|
||||
TEAM="${ARG}"
|
||||
elif [ -n "${GETMESSAGE}" ]; then # après un --post "team:channel"
|
||||
if [[ $TEAM == "-*" && ${#TEAM} -le 5 ]]; then echo "J'envoie mon message à \"${TEAM}\" ?? Arf, ça me plait pas j'ai l'impression que tu t'es planté sur la commande."; usage ; exit 1 ; fi
|
||||
if [[ $ARG == "-*" && ${#ARG} -le 5 ]]; then echo "J'envoie le message \"${ARG}\" ?? Arf, ça me plait pas j'ai l'impression que tu t'es planté sur la commande."; usage ; exit 1 ; fi
|
||||
if [[ ! $TEAM =~ .*:.+ ]]; then echo "Il faut mettre un destinataire sous la forme team:channel. Recommence !"; usage ; exit 1 ; fi
|
||||
Posts+=( ["${TEAM}"]="$ARG" )
|
||||
GETMESSAGE=""
|
||||
TEAM=""
|
||||
else
|
||||
case "${ARG}" in
|
||||
'-h' | '--help' )
|
||||
usage && exit ;;
|
||||
'-n' | '--simu')
|
||||
SIMU="echo" ;;
|
||||
'-q' )
|
||||
QUIET="/dev/null" ;;
|
||||
'-m' | '--main' )
|
||||
OPERATE_ON_MAIN="OUI-OUI" ;;
|
||||
'-M' )
|
||||
AVAILABLE_ORGAS= && OPERATE_ON_MAIN="OUI-OUI" ;; #pas d'orgas
|
||||
'--nas' | '-nas' )
|
||||
OPERATE_LOCAL_ORGA= ;; # pas les locales
|
||||
'--local' | '-local' )
|
||||
OPERATE_ON_NAS_ORGA= ;; # pas celles sur NAS
|
||||
'-cloud'|'--cloud')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} cloud" ;;
|
||||
'-agora'|'--agora'|'-mm'|'--mm'|'-matter'*|'--matter'*)
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} agora" ;;
|
||||
'-wiki'|'--wiki')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} wiki" ;;
|
||||
'-wp'|'--wp')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} wp" ;;
|
||||
'-office'|'--office'|'-collab'*|'--collab'*)
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} office" ;;
|
||||
'-paheko'|'--paheko')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} paheko" ;;
|
||||
'-pod'|'--pod'|'-castopod'|'--castopod')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} castopod" ;;
|
||||
'-t' )
|
||||
COMMANDS="${COMMANDS} RESTART-COMPOSE" ;;
|
||||
'-r' )
|
||||
COMMANDS="${COMMANDS} RESTART-DOCKER" ;;
|
||||
'-l' | '--list' )
|
||||
COMMANDS="$(echo "${COMMANDS} LIST" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-v' | '--version')
|
||||
COMMANDS="$(echo "${COMMANDS} VERSION" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-I' | '--install' )
|
||||
COMMANDS="$(echo "${COMMANDS} INIT" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'-U' | '--upgrade')
|
||||
COMMANDS="$(echo "${COMMANDS} UPGRADE" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'--optim' )
|
||||
COMMANDS="$(echo "${COMMANDS} OPTIMISE-CLOUD" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-u' )
|
||||
COMMANDS="$(echo "${COMMANDS} UPDATE-CLOUD-APP" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-i' )
|
||||
COMMANDS="$(echo "${COMMANDS} INSTALL-CLOUD-APP" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-a' )
|
||||
GETAPPS="now" ;;
|
||||
'-occ' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-CLOUD-OCC" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETOCCCOMAND="now" ;;
|
||||
'-mmctl' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-AGORA-MMCTL" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETMMCTLCOMAND="now" ;;
|
||||
'-exec' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-DOCKER-EXEC" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETEXECCOMAND="now" ;;
|
||||
'-p' | '--post' )
|
||||
COMMANDS="$(echo "${COMMANDS} POST-AGORA" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETTEAM="now" ;;
|
||||
'-*' ) # ignore
|
||||
;;
|
||||
*)
|
||||
GIVEN_ORGA="${GIVEN_ORGA} ${ARG%-orga}"
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
done
|
||||
|
||||
|
||||
if [[ "${COMMANDS[*]}" =~ "RESTART-COMPOSE" && "${COMMANDS[*]}" =~ "RESTART-TYPE" ]]; then
|
||||
echo "Je restarte via docker-compose ou via docker mais pas les deux !"
|
||||
usage
|
||||
exit 1
|
||||
fi
|
||||
if [ -z "${COMMANDS}" ]; then
|
||||
usage && exit
|
||||
fi
|
||||
if [ -n "${GIVEN_ORGA}" ]; then
|
||||
# intersection des 2 listes : quelle commande de ouf !!
|
||||
AVAILABLE_ORGAS=$(comm -12 <(printf '%s\n' ${AVAILABLE_ORGAS} | sort) <(printf '%s\n' ${GIVEN_ORGA} | sort))
|
||||
fi
|
||||
|
||||
NB_ORGAS=$(echo "${AVAILABLE_ORGAS}" | wc -w )
|
||||
|
||||
if [[ $NB_ORGAS = 0 && -z "${OPERATE_ON_MAIN}" ]]; then
|
||||
echo "Aucune orga trouvée."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
NB_ORGAS_STR="$NB_ORGAS orgas"
|
||||
[ -n "${OPERATE_ON_MAIN}" ] && NB_ORGAS_STR="$NB_ORGAS_STR + les communs"
|
||||
|
||||
_populate_lists # on récupère les clouds / agora / wiki / wp correspondants aux orga
|
||||
|
||||
if [[ $NB_ORGAS -gt 2 && "${COMMANDS[*]}" =~ 'INIT' ]]; then
|
||||
ETLECLOUDCOMMUN=
|
||||
[ -n "${OPERATE_ON_MAIN}" ] && ETLECLOUDCOMMUN=" ainsi que les containers commun"
|
||||
echo "On s'apprête à initialiser les ${CONTAINERS_TYPES} suivants : ${AVAILABLE_ORGAS}${ETLECLOUDCOMMUN}"
|
||||
checkContinue
|
||||
fi
|
||||
|
||||
for COMMAND in ${COMMANDS}; do
|
||||
case "${COMMAND}" in
|
||||
'LIST' )
|
||||
listContainers && exit ;;
|
||||
'VERSION' )
|
||||
version && exit ;;
|
||||
'OPTIMISE-CLOUD' )
|
||||
OptimiseClouds ;;
|
||||
'RESTART-COMPOSE' )
|
||||
restart-compose ;;
|
||||
'RESTART-DOCKER' )
|
||||
restart ;;
|
||||
'UPDATE-CLOUD-APP' )
|
||||
UpdateApplis ;;
|
||||
'UPGRADE' )
|
||||
UpgradeClouds ;;
|
||||
'INIT' )
|
||||
Init ;;
|
||||
'INSTALL-CLOUD-APP' )
|
||||
InstallApps ;;
|
||||
'RUN-CLOUD-OCC' )
|
||||
RunCommands "OCC" "${OCCCOMANDS[@]}" ;;
|
||||
'RUN-AGORA-MMCTL' )
|
||||
RunCommands "MMCTL" "${MMCTLCOMANDS[@]}" ;;
|
||||
'RUN-DOCKER-EXEC' )
|
||||
RunCommands "EXEC" "${EXECCOMANDS[@]}" ;;
|
||||
'POST-AGORA' )
|
||||
PostMessages ${Posts} ;;
|
||||
esac
|
||||
done
|
659
bin/gestContainers_v2.sh
Executable file
@ -0,0 +1,659 @@
|
||||
#!/bin/bash
|
||||
# Script de manipulation des containers en masse
|
||||
# init /versions / restart ...
|
||||
#
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
PRG=$(basename $0)
|
||||
|
||||
tab_sites_destinations_possibles=($(get_Serveurs_Kaz))
|
||||
# ${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} "${CMD}"
|
||||
# SITE_DST="${tab_sites_destinations_possibles[1]}"
|
||||
# ${tab_sites_destinations_possibles[@]}
|
||||
|
||||
#GLOBAL VARS
|
||||
NAS_VOL="/mnt/disk-nas1/docker/volumes/"
|
||||
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
AVAILABLE_ORGAS=${availableOrga[*]//-orga/}
|
||||
|
||||
availableContainersCommuns=( $(getList "${KAZ_CONF_DIR}/container-withMail.list") $(getList "${KAZ_CONF_DIR}/container-withoutMail.list"))
|
||||
|
||||
OPERATE_ON_MAIN= # par defaut NON on ne traite que des orgas
|
||||
OPERATE_ON_NAS_ORGA="OUI" # par defaut oui, on va aussi sur les orgas du NAS
|
||||
OPERATE_LOCAL_ORGA="OUI" # par defaut oui
|
||||
|
||||
TEMPO_ACTION_STOP=2 # Lors de redémarrage avec tempo, on attend après le stop
|
||||
TEMPO_ACTION_START=120 # Lors de redémarrage avec tempo, avant de reload le proxy
|
||||
|
||||
CONTAINERS_TYPES=
|
||||
|
||||
defaultContainersTypes="cloud agora wp wiki office paheko" # les containers gérés par ce script.
|
||||
|
||||
declare -A DockerServNames # le nom des containers correspondant
|
||||
DockerServNames=( [cloud]="${nextcloudServName}" [agora]="${mattermostServName}" [wiki]="${dokuwikiServName}" [wp]="${wordpressServName}" [office]="${officeServName}" [paheko]="${pahekoServName}" )
|
||||
|
||||
declare -A FilterLsVolume # Pour trouver quel volume appartient à quel container
|
||||
FilterLsVolume=( [cloud]="cloudMain" [agora]="matterConfig" [wiki]="wikiConf" [wp]="wordpress" )
|
||||
|
||||
declare -A composeDirs # Le nom du repertoire compose pour le commun
|
||||
composeDirs=( [cloud]="cloud" [agora]="mattermost" [wiki]="dokuwiki" [office]="collabora" [paheko]="paheko" )
|
||||
|
||||
declare -A serviceNames # Le nom du du service dans le dockerfile d'orga
|
||||
serviceNames=( [cloud]="cloud" [agora]="agora" [wiki]="dokuwiki" [wp]="wordpress" [office]="collabora")
|
||||
|
||||
declare -A subScripts
|
||||
subScripts=( [cloud]="manageCloud.sh" [agora]="manageAgora.sh" [wiki]="manageWiki.sh" [wp]="manageWp.sh" )
|
||||
|
||||
declare -A OrgasOnNAS
|
||||
declare -A OrgasLocales
|
||||
declare -A NbOrgas
|
||||
declare -A RunningOrgas
|
||||
declare -A Posts
|
||||
|
||||
QUIET="1" # redirection des echo
|
||||
|
||||
OCCCOMANDS=()
|
||||
MMCTLCOMANDS=()
|
||||
EXECCOMANDS=()
|
||||
|
||||
# CLOUD
|
||||
APPLIS_PAR_DEFAUT="tasks calendar contacts bookmarks richdocuments external drawio snappymail"
|
||||
|
||||
usage() {
|
||||
echo "${PRG} [OPTION] [CONTAINERS_TYPES] [COMMANDES] [SERVEURS] [ORGAS]
|
||||
Ce script regroupe l'ensemble des opérations que l'on souhaite automatiser sur plusieurs containers, sur un ou plusieurs sites.
|
||||
Par defaut, sur les orgas, mais on peut aussi ajouter les communs
|
||||
|
||||
OPTIONS
|
||||
-h|--help Cette aide :-)
|
||||
-n|--simu SIMULATION
|
||||
-q|--quiet On ne parle pas (utile avec le -n pour avoir que les commandes)
|
||||
-m|--main Traite aussi le container commun (cloud commun / agora commun / wiki commun)
|
||||
-M Ne traite que le container commun, et pas les orgas
|
||||
--nas Ne traite QUE les orgas sur le NAS
|
||||
--local Ne traite pas les orgas sur le NAS
|
||||
-v|--version Donne la version des containers et signale les MàJ
|
||||
-l|--list Liste des containers (up / down, local ou nas) de cette machine
|
||||
|
||||
CONTAINERS_TYPES
|
||||
-cloud Pour agir sur les clouds
|
||||
-agora Pour agir sur les agoras
|
||||
-wp Les wp
|
||||
-wiki Les wiki
|
||||
-office Les collabora
|
||||
|
||||
COMMANDES (on peut en mettre plusieurs dans l'ordre souhaité)
|
||||
-I|--install L'initialisation du container
|
||||
-t Redémarre avec tempo (docker-compose down puis sleep ${TEMPO_ACTION_STOP} puis up puis sleep ${TEMPO_ACTION_START})
|
||||
-r Redémarre sans tempo (docker restart)
|
||||
-exec \"command\" Envoie une commande docker exec
|
||||
|
||||
--optim Lance la procédure Nextcloud pour optimiser les performances ** **
|
||||
-occ \"command\" Envoie une commande via occ ** **
|
||||
-u Mets à jour les applis ** SPECIFIQUES **
|
||||
-i Install des applis ** CLOUD **
|
||||
-a \"app1 app2 ...\" Choix des appli à installer ou mettre à jour (entre guillemets) ** **
|
||||
-U|--upgrade Upgrade des clouds ** **
|
||||
|
||||
-mmctl \"command\" Envoie une commande via mmctl ** SPECIFIQUES **
|
||||
-p|--post \"team\" \"message\" Poste un message dans une team agora ** AGORA **
|
||||
|
||||
SERVEURS
|
||||
--all-srv Lance sur tous les serveurs ${tab_sites_destinations_possibles[@]}, sinon c'est uniquement sur ${site}
|
||||
|
||||
ORGAS sur ${site}
|
||||
[orga1 orga2 ... ] on peut filtrer parmi : ${AVAILABLE_ORGAS}
|
||||
|
||||
|
||||
Exemples :
|
||||
${PRG} -office -m -r # restart de tous les collaboras (libére RAM)
|
||||
${PRG} -cloud -u -r -q -n # affiche toutes les commandes (-n -q ) pour mettre à jour toutes les applis des clouds + restart (-u -r)
|
||||
${PRG} -p \"monorga:town-square\" \"Hello\" monorga # envoie Hello sur le centreville de l'orga monorga sur son mattermost dédié
|
||||
${PRG} -cloud -occ \"config:system:set default_phone_region --value='FR'\" --all-srv # modifie la variable default_phone_region dans le config.php de tous les clouds de tous les serveurs
|
||||
"
|
||||
}
|
||||
|
||||
|
||||
####################################################
|
||||
################ fonctions clefs ###################
|
||||
####################################################
|
||||
|
||||
_populate_lists(){
|
||||
# récupère les listes d'orga à traiter
|
||||
# on rempli les tableaux OrgasOnNAS / OrgasLocales / NbOrgas ... par type de container
|
||||
|
||||
if [ -z "${CONTAINERS_TYPES}" ]; then
|
||||
# wow, on traite tout le monde d'un coup...
|
||||
CONTAINERS_TYPES="$defaultContainersTypes"
|
||||
fi
|
||||
|
||||
for TYPE in ${CONTAINERS_TYPES}; do
|
||||
if [ -n "${FilterLsVolume[$TYPE]}" ] ; then # on regarde dans les volumes
|
||||
[ -n "$OPERATE_ON_NAS_ORGA" ] && OrgasOnNAS["$TYPE"]=$( _getListOrgas ${NAS_VOL} ${FilterLsVolume[$TYPE]} )
|
||||
[ -n "$OPERATE_LOCAL_ORGA" ] && OrgasLocales["$TYPE"]=$( _getListOrgas ${DOCK_VOL} ${FilterLsVolume[$TYPE]} "SANSLN")
|
||||
else # un docker ps s'il n'y a pas de volumes
|
||||
[ -n "$OPERATE_LOCAL_ORGA" ] && OrgasLocales["$TYPE"]=$(docker ps --format '{{.Names}}' | grep ${DockerServNames[$TYPE]} | sed -e "s/-*${DockerServNames[$TYPE]}//")
|
||||
fi
|
||||
NbOrgas["$TYPE"]=$(($(echo ${OrgasOnNAS["$TYPE"]} | wc -w) + $(echo ${OrgasLocales["$TYPE"]} | wc -w)))
|
||||
RunningOrgas["$TYPE"]=$(docker ps --format '{{.Names}}' | grep ${DockerServNames[$TYPE]} | sed -e "s/-*${DockerServNames[$TYPE]}//")
|
||||
done
|
||||
}
|
||||
|
||||
_getListOrgas(){
|
||||
# retrouve les orgas à partir des volume présents
|
||||
# $1 where to lookup
|
||||
# $2 filter
|
||||
# $3 removeSymbolicLinks
|
||||
[ ! -d $1 ] || [ -z "$2" ] && return 1 # si le repertoire n'existe pas on skip
|
||||
LIST=$(ls "${1}" | grep -i orga | grep -i "$2" | sed -e "s/-${2}$//g" | sed -e 's/^orga_//')
|
||||
[ -n "$3" ] && LIST=$(ls -F "${1}" | grep '/' | grep -i orga | grep -i "$2" | sed -e "s/-${2}\/$//g" | sed -e 's/^orga_//')
|
||||
LIST=$(comm -12 <(printf '%s\n' ${LIST} | sort) <(printf '%s\n' ${AVAILABLE_ORGAS} | sort))
|
||||
echo "$LIST"
|
||||
}
|
||||
|
||||
_executeFunctionForAll(){
|
||||
# Parcours des container et lancement des commandes
|
||||
# Les commandes ont en derniers paramètres le type et l'orga et une string parmi KAZ/ORGANAS/ORGALOCAL pour savoir sur quoi on opère
|
||||
# $1 function
|
||||
# $2 nom de la commande
|
||||
# $3 quel types de containers
|
||||
# $4 params : quels paramètres à passer à la commande (les clefs sont #ORGA# #DOCKERSERVNAME# #SURNAS# #ISMAIN# #TYPE# #COMPOSEDIR# )
|
||||
for TYPE in ${3}; do
|
||||
if [ -n "$OPERATE_ON_MAIN" ]; then
|
||||
if [[ -n "${composeDirs[$TYPE]}" && "${availableContainersCommuns[*]}" =~ "${composeDirs[$TYPE]}" ]]; then # pas de cloud / agora / wp / wiki sur cette instance
|
||||
Dockername=${DockerServNames[$TYPE]}
|
||||
PARAMS=$(echo $4 | sed -e "s/#ORGA#//g;s/#DOCKERSERVNAME#/$Dockername/g;s/#ISMAIN#/OUI/g;s/#SURNAS#/NON/g;s/#TYPE#/$TYPE/g;s%#COMPOSEDIR#%${KAZ_COMP_DIR}/${composeDirs[$TYPE]}%g" )
|
||||
echo "-------- $2 $TYPE COMMUN ----------------------------" >& $QUIET
|
||||
eval "$1" $PARAMS
|
||||
fi
|
||||
fi
|
||||
if [[ ${NbOrgas[$TYPE]} -gt 0 ]]; then
|
||||
echo "-------- $2 des $TYPE des ORGAS ----------------------------" >& $QUIET
|
||||
COMPTEUR=1
|
||||
if [ -n "$OPERATE_LOCAL_ORGA" ]; then
|
||||
for ORGA in ${OrgasLocales[$TYPE]}; do
|
||||
Dockername=${ORGA}-${DockerServNames[$TYPE]}
|
||||
PARAMS=$(echo $4 | sed -e "s/#ORGA#/${ORGA}/g;s/#DOCKERSERVNAME#/$Dockername/g;s/#ISMAIN#/NON/g;s/#SURNAS#/NON/g;s/#TYPE#/$TYPE/g;s%#COMPOSEDIR#%${KAZ_COMP_DIR}/${ORGA}-orga%g" )
|
||||
echo "${RED} ${ORGA}-orga ${NC}($COMPTEUR/${NbOrgas[$TYPE]})" >& $QUIET
|
||||
eval "$1" $PARAMS
|
||||
COMPTEUR=$((COMPTEUR + 1))
|
||||
done
|
||||
fi
|
||||
if [ -n "$OPERATE_ON_NAS_ORGA" ]; then
|
||||
for ORGA in ${OrgasOnNAS[$TYPE]}; do
|
||||
Dockername=${ORGA}-${DockerServNames[$TYPE]}
|
||||
PARAMS=$(echo $4 | sed -e "s/#ORGA#/${ORGA}/g;s/#DOCKERSERVNAME#/$Dockername/g;s/#ISMAIN#/NON/g;s/#SURNAS#/OUI/g;s/#TYPE#/$TYPE/g;s%#COMPOSEDIR#%${KAZ_COMP_DIR}/${ORGA}-orga%g" )
|
||||
echo "${RED} ${ORGA}-orga ${NC}($COMPTEUR/${NbOrgas[$TYPE]})" >& $QUIET
|
||||
eval "$1" $PARAMS
|
||||
COMPTEUR=$((COMPTEUR + 1))
|
||||
done
|
||||
fi
|
||||
fi
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
##############################################
|
||||
################ COMMANDES ###################
|
||||
##############################################
|
||||
Init(){
|
||||
# Initialisation des containers
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Initialisation" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
_executeFunctionForAll "_initContainer" "Initialisation" "${CONTAINERS_TYPES[@]}" "#TYPE# #ISMAIN# #SURNAS# #ORGA# "
|
||||
}
|
||||
|
||||
restart-compose() {
|
||||
# Parcours les containers et redémarre avec tempo
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "DOCKER-COMPOSE DOWN puis sleep ${TEMPO_ACTION_STOP}" >& $QUIET
|
||||
echo "DOCKER-COMPOSE UP puis sleep ${TEMPO_ACTION_START}" >& $QUIET
|
||||
echo "de ${CONTAINERS_TYPES} pour $NB_ORGAS_STR" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
_executeFunctionForAll "_restartContainerAvecTempo" "Restart" "${CONTAINERS_TYPES[@]}" "#TYPE# #ISMAIN# #COMPOSEDIR#"
|
||||
|
||||
${SIMU} sleep ${TEMPO_ACTION_START}
|
||||
_reloadProxy
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
echo "${GREEN}FIN${NC} " >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
}
|
||||
|
||||
restart() {
|
||||
# Parcours les containers et redémarre
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "DOCKER RESTART des ${CONTAINERS_TYPES} pour $NB_ORGAS_STR" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
_executeFunctionForAll "_restartContainer" "Restart" "${CONTAINERS_TYPES[@]}" "#DOCKERSERVNAME#"
|
||||
|
||||
_reloadProxy
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
echo "${GREEN}FIN${NC} " >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
}
|
||||
|
||||
version(){
|
||||
# Parcours les containers et affiche leurs versions
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "VERSIONS" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
_executeFunctionForAll "_versionContainer" "Version" "${CONTAINERS_TYPES[@]}" "#TYPE# #ISMAIN# #ORGA#"
|
||||
}
|
||||
|
||||
|
||||
listContainers(){
|
||||
echo "${NC}--------------------------------------------------------"
|
||||
echo "LISTES"
|
||||
echo "------------------------------------------------------------"
|
||||
for TYPE in ${CONTAINERS_TYPES}; do
|
||||
echo "****************** $TYPE ****************"
|
||||
_listContainer "$TYPE"
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
|
||||
######################## Fonctions génériques #######################
|
||||
|
||||
_initContainer(){
|
||||
# $1 type
|
||||
# $2 COMMUN
|
||||
# $3 ON NAS
|
||||
# $4 orgas
|
||||
if [ -n "${subScripts[$1]}" ] ; then
|
||||
evalStr="${KAZ_BIN_DIR}/${subScripts[$1]} --install"
|
||||
if [ "$3" = "OUI" ]; then evalStr="${evalStr} -nas" ; fi
|
||||
if [ ! "$QUIET" = "1" ]; then evalStr="${evalStr} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then evalStr="${evalStr} -n" ; fi
|
||||
if [ ! "$2" = "OUI" ]; then evalStr="${evalStr} $4" ; fi
|
||||
eval $evalStr
|
||||
fi
|
||||
}
|
||||
|
||||
_restartContainer(){
|
||||
# $1 Dockername
|
||||
echo -n "${NC}Redemarrage ... " >& $QUIET
|
||||
${SIMU}
|
||||
${SIMU} docker restart $1
|
||||
echo "${GREEN}OK${NC}" >& $QUIET
|
||||
}
|
||||
|
||||
_restartContainerAvecTempo(){
|
||||
# $1 type
|
||||
# $2 main container
|
||||
# $2 composeDir
|
||||
dir=$3
|
||||
if [ -z $dir ]; then return 1; fi # le compose n'existe pas ... par exemple wordpress commun
|
||||
cd "$dir"
|
||||
echo -n "${NC}Arrêt ... " >& $QUIET
|
||||
${SIMU}
|
||||
if [ "$2" = "OUI" ]; then ${SIMU} docker-compose stop ;
|
||||
else ${SIMU} docker-compose stop "${serviceNames[$1]}"
|
||||
fi
|
||||
${SIMU} sleep ${TEMPO_ACTION_STOP}
|
||||
echo "${GREEN}OK${NC}" >& $QUIET
|
||||
echo -n "${NC}Démarrage ... " >& $QUIET
|
||||
if [ "$2" = "OUI" ]; then ${SIMU} docker-compose up -d ;
|
||||
else ${SIMU} docker-compose up -d "${serviceNames[$1]}"
|
||||
fi
|
||||
${SIMU} sleep ${TEMPO_ACTION_START}
|
||||
echo "${GREEN}OK${NC}" >& $QUIET
|
||||
}
|
||||
|
||||
|
||||
_reloadProxy() {
|
||||
availableProxyComposes=($(getList "${KAZ_CONF_DIR}/container-proxy.list"))
|
||||
|
||||
for item in "${availableProxyComposes[@]}"; do
|
||||
${SIMU} ${KAZ_COMP_DIR}/${item}/reload.sh
|
||||
done
|
||||
}
|
||||
|
||||
_versionContainer() {
|
||||
# Affiche la version d'un container donné
|
||||
# $1 type
|
||||
# $2 COMMUN
|
||||
# $3 orgas
|
||||
if [ -n "${subScripts[$1]}" ] ; then
|
||||
evalStr="${KAZ_BIN_DIR}/${subScripts[$1]} --version"
|
||||
if [ ! "$2" = "OUI" ]; then evalStr="${evalStr} $3" ; fi
|
||||
eval $evalStr
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
_listContainer(){
|
||||
# pour un type donné (cloud / agora / wiki / wp), fait une synthèse de qui est up et down / nas ou local
|
||||
# $1 type
|
||||
RUNNING_FROM_NAS=$(comm -12 <(printf '%s\n' ${OrgasOnNAS[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort) | sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
RUNNING_LOCAL=$(comm -12 <(printf '%s\n' ${OrgasLocales[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort) | sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
# tu l'a vu la belle commande pour faire une exclusion de liste
|
||||
DOWN_ON_NAS=$(comm -23 <(printf '%s\n' ${OrgasOnNAS[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort) | sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
DOWN_LOCAL=$(comm -23 <(printf '%s\n' ${OrgasLocales[$1]} | sort) <(printf '%s\n' ${RunningOrgas[$1]} | sort)| sed -e ':a;N;$!ba;s/\n/ /g')
|
||||
NB_SUR_NAS=$(echo ${OrgasOnNAS[$1]} | wc -w)
|
||||
NB_LOCAUX=$(echo ${OrgasLocales[$1]} | wc -w)
|
||||
NB_RUNNING_SUR_NAS=$(echo $RUNNING_FROM_NAS | wc -w)
|
||||
NB_RUNNING_LOCALLY=$(echo $RUNNING_LOCAL | wc -w)
|
||||
MAIN_RUNNING="${RED}DOWN${NC}"
|
||||
if docker ps | grep -q " ${DockerServNames[$1]}"
|
||||
then
|
||||
MAIN_RUNNING="${GREEN}UP${NC}"
|
||||
fi
|
||||
|
||||
[ -n "${composeDirs[${1}]}" ] && echo "${NC}Le ${1} commun est $MAIN_RUNNING"
|
||||
if [[ ${NbOrgas[$1]} -gt 0 ]]; then
|
||||
ENLOCALSTR=
|
||||
if [[ ${NB_RUNNING_SUR_NAS[$1]} -gt 0 ]]; then ENLOCALSTR=" en local" ; fi
|
||||
echo "Orgas : $NB_RUNNING_LOCALLY / $NB_LOCAUX running ${1}$ENLOCALSTR"
|
||||
echo "${NC}UP : ${GREEN}${RUNNING_LOCAL}"
|
||||
echo "${NC}DOWN : ${RED}$DOWN_LOCAL${NC}"
|
||||
if [[ ${NB_RUNNING_SUR_NAS[$1]} -gt 0 ]]; then
|
||||
echo "${NC}Orgas : $NB_RUNNING_SUR_NAS / $NB_SUR_NAS running depuis le NAS :"
|
||||
echo "${NC}UP : ${GREEN}${RUNNING_FROM_NAS}"
|
||||
echo "${NC}DOWN : ${RED}$DOWN_ON_NAS${NC}"
|
||||
fi
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
#########################################################
|
||||
############# FONCTIONS SPECIFIQUES #####################
|
||||
#########################################################
|
||||
|
||||
##################################
|
||||
############### CLOUD ############
|
||||
##################################
|
||||
|
||||
UpgradeClouds() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "UPGRADE des cloud" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
RunOCCCommand "upgrade"
|
||||
}
|
||||
|
||||
OptimiseClouds() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Optimisation des cloud" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
RunOCCCommands "db:add-missing-indices" "db:convert-filecache-bigint --no-interaction"
|
||||
}
|
||||
|
||||
InstallApps(){
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "INSTALL DES APPLIS sur les clouds : ${LISTE_APPS}" >& $QUIET
|
||||
echo "-------------------------------------------------------------" >& $QUIET
|
||||
if [ -z "${LISTE_APPS}" ]; then
|
||||
echo "Aucune appli n'est précisée, j'installe les applis par défaut : ${APPLIS_PAR_DEFAUT}" >& $QUIET
|
||||
LISTE_APPS="${APPLIS_PAR_DEFAUT}"
|
||||
fi
|
||||
PARAMS="-a \"$LISTE_APPS\""
|
||||
if [ ! "$QUIET" = "1" ]; then PARAMS="${PARAMS} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then PARAMS="${PARAMS} -n" ; fi
|
||||
_executeFunctionForAll "${KAZ_BIN_DIR}/${subScripts["cloud"]} -i $PARAMS" "Install des applis" "cloud" "#ORGA#"
|
||||
}
|
||||
|
||||
UpdateApplis() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "UPDATE DES APPLIS des cloud : ${LISTE_APPS}" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
PARAMS="-a ${LISTE_APPS}"
|
||||
if [ -z "${LISTE_APPS}" ]; then
|
||||
echo "Aucune appli n'est précisée, je les met toutes à jour! " >& $QUIET
|
||||
PARAMS=
|
||||
fi
|
||||
if [ ! "$QUIET" = "1" ]; then PARAMS="${PARAMS} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then PARAMS="${PARAMS} -n" ; fi
|
||||
_executeFunctionForAll "${KAZ_BIN_DIR}/${subScripts["cloud"]} -u $PARAMS" "Maj des applis" "cloud" "#ORGA#"
|
||||
}
|
||||
|
||||
|
||||
##################################
|
||||
############### AGORA ############
|
||||
##################################
|
||||
|
||||
|
||||
PostMessages(){
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Envoi de messages sur mattermost" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
|
||||
for TEAM in "${!Posts[@]}"
|
||||
do
|
||||
MSG=${Posts[$TEAM]/\"/\\\"}
|
||||
PARAMS="-p \"$TEAM\" \"$MSG\""
|
||||
if [ ! "$QUIET" = "1" ]; then PARAMS="${PARAMS} -q" ; fi
|
||||
if [ -n "$SIMU" ]; then PARAMS="${PARAMS} -n" ; fi
|
||||
_executeFunctionForAll "${KAZ_BIN_DIR}/${subScripts["agora"]} $PARAMS" "Post vers $TEAM sur l'agora" "agora" "#ORGA#"
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
########## LANCEMENT COMMANDES OCC / MMCTL ############
|
||||
|
||||
RunCommands() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "Envoi de commandes en direct" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
# $1 OCC / MMCTL / EXEC
|
||||
# $ suivants : les commandes
|
||||
for command in "${@:2}"
|
||||
do
|
||||
if [ $1 = "OCC" ]; then RunOCCCommand "$command" ; fi
|
||||
if [ $1 = "MMCTL" ]; then RunMMCTLCommand "$command" ; fi
|
||||
if [ $1 = "EXEC" ]; then RunEXECCommand "$command" ; fi
|
||||
done
|
||||
}
|
||||
|
||||
_runSingleOccCommand(){
|
||||
# $1 Command
|
||||
# $2 Dockername
|
||||
${SIMU} docker exec -u 33 $2 /var/www/html/occ $1
|
||||
}
|
||||
|
||||
_runSingleMmctlCommand(){
|
||||
# $1 Command
|
||||
# $2 Dockername
|
||||
${SIMU} docker exec $2 bin/mmctl $1
|
||||
}
|
||||
|
||||
_runSingleExecCommand(){
|
||||
# $1 Command
|
||||
# $2 Dockername
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} docker exec $2 $1
|
||||
}
|
||||
|
||||
RunOCCCommand() {
|
||||
# $1 Command
|
||||
_executeFunctionForAll "_runSingleOccCommand \"${1}\"" "OCC $1" "cloud" "#DOCKERSERVNAME#"
|
||||
}
|
||||
|
||||
RunMMCTLCommand() {
|
||||
# $1 Command
|
||||
_executeFunctionForAll "_runSingleMmctlCommand \"${1}\"" "MMCTL $1" "agora" "#DOCKERSERVNAME#"
|
||||
}
|
||||
|
||||
RunEXECCommand() {
|
||||
# $1 Command
|
||||
_executeFunctionForAll "_runSingleExecCommand \"${1}\"" "docker exec $1" "${CONTAINERS_TYPES[@]}" "#DOCKERSERVNAME#"
|
||||
}
|
||||
|
||||
|
||||
|
||||
########## Contrôle #################
|
||||
for ARG in "$@"; do
|
||||
# Seul PROD1 peut attaquer tous les autres serveurs kaz sinon un serveur kaz peut juste s'attaquer lui-même (aie!)
|
||||
if [ "${ARG}" == "--all-srv" -a "${site}" != "prod1" ]; then
|
||||
echo "${RED}--all-srv choisi alors qu'on n'est pas sur prod1 : impossible, on quitte${NC}"
|
||||
# mais pour l'instant on autorise pour les tests
|
||||
# exit
|
||||
fi
|
||||
done
|
||||
|
||||
########## Main #################
|
||||
for ARG in "$@"; do
|
||||
|
||||
#echo "${ARG}"
|
||||
|
||||
if [ -n "${GETOCCCOMAND}" ]; then # après un -occ
|
||||
OCCCOMANDS+=("${ARG}")
|
||||
GETOCCCOMAND=
|
||||
elif [ -n "${GETEXECCOMAND}" ]; then # après un -exec
|
||||
EXECCOMANDS+=("${ARG}")
|
||||
GETEXECCOMAND=
|
||||
elif [ -n "${GETAPPS}" ]; then # après un -a
|
||||
LISTE_APPS="${LISTE_APPS} ${ARG}"
|
||||
GETAPPS=""
|
||||
elif [ -n "${GETMMCTLCOMAND}" ]; then # après un -mmctl
|
||||
MMCTLCOMANDS+=("${ARG}")
|
||||
GETMMCTLCOMAND=
|
||||
elif [ -n "${GETTEAM}" ]; then # après un --post
|
||||
GETMESSAGE="now"
|
||||
GETTEAM=""
|
||||
TEAM="${ARG}"
|
||||
elif [ -n "${GETMESSAGE}" ]; then # après un --post "team:channel"
|
||||
if [[ $TEAM == "-*" && ${#TEAM} -le 5 ]]; then echo "J'envoie mon message à \"${TEAM}\" ?? Arf, ça me plait pas j'ai l'impression que tu t'es planté sur la commande."; usage ; exit 1 ; fi
|
||||
if [[ $ARG == "-*" && ${#ARG} -le 5 ]]; then echo "J'envoie le message \"${ARG}\" ?? Arf, ça me plait pas j'ai l'impression que tu t'es planté sur la commande."; usage ; exit 1 ; fi
|
||||
if [[ ! $TEAM =~ .*:.+ ]]; then echo "Il faut mettre un destinataire sous la forme team:channel. Recommence !"; usage ; exit 1 ; fi
|
||||
Posts+=( ["${TEAM}"]="$ARG" )
|
||||
GETMESSAGE=""
|
||||
TEAM=""
|
||||
else
|
||||
case "${ARG}" in
|
||||
'-h' | '--help' )
|
||||
usage && exit ;;
|
||||
'-n' | '--simu')
|
||||
SIMU="echo" ;;
|
||||
'-q' )
|
||||
QUIET="/dev/null" ;;
|
||||
'-m' | '--main' )
|
||||
OPERATE_ON_MAIN="OUI-OUI" ;;
|
||||
'-M' )
|
||||
AVAILABLE_ORGAS= && OPERATE_ON_MAIN="OUI-OUI" ;; #pas d'orgas
|
||||
'--nas' | '-nas' )
|
||||
OPERATE_LOCAL_ORGA= ;; # pas les locales
|
||||
'--local' | '-local' )
|
||||
OPERATE_ON_NAS_ORGA= ;; # pas celles sur NAS
|
||||
'-cloud'|'--cloud')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} cloud" ;;
|
||||
'-agora'|'--agora')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} agora" ;;
|
||||
'-wiki'|'--wiki')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} wiki" ;;
|
||||
'-wp'|'--wp')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} wp" ;;
|
||||
'-office'|'--office')
|
||||
CONTAINERS_TYPES="${CONTAINERS_TYPES} office" ;;
|
||||
'-t' )
|
||||
COMMANDS="${COMMANDS} RESTART-COMPOSE" ;;
|
||||
'-r' )
|
||||
COMMANDS="${COMMANDS} RESTART-DOCKER" ;;
|
||||
'-l' | '--list' )
|
||||
COMMANDS="$(echo "${COMMANDS} LIST" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-v' | '--version')
|
||||
COMMANDS="$(echo "${COMMANDS} VERSION" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-I' | '--install' )
|
||||
COMMANDS="$(echo "${COMMANDS} INIT" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'-U' | '--upgrade')
|
||||
COMMANDS="$(echo "${COMMANDS} UPGRADE" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'--optim' )
|
||||
COMMANDS="$(echo "${COMMANDS} OPTIMISE-CLOUD" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-u' )
|
||||
COMMANDS="$(echo "${COMMANDS} UPDATE-CLOUD-APP" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-i' )
|
||||
COMMANDS="$(echo "${COMMANDS} INSTALL-CLOUD-APP" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-a' )
|
||||
GETAPPS="now" ;;
|
||||
'-occ' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-CLOUD-OCC" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETOCCCOMAND="now" ;;
|
||||
'-mmctl' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-AGORA-MMCTL" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETMMCTLCOMAND="now" ;;
|
||||
'-exec' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-DOCKER-EXEC" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETEXECCOMAND="now" ;;
|
||||
'-p' | '--post' )
|
||||
COMMANDS="$(echo "${COMMANDS} POST-AGORA" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETTEAM="now" ;;
|
||||
'-*' ) # ignore
|
||||
;;
|
||||
*)
|
||||
GIVEN_ORGA="${GIVEN_ORGA} ${ARG%-orga}"
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
done
|
||||
|
||||
|
||||
if [[ "${COMMANDS[*]}" =~ "RESTART-COMPOSE" && "${COMMANDS[*]}" =~ "RESTART-TYPE" ]]; then
|
||||
echo "Je restarte via docker-compose ou via docker mais pas les deux !"
|
||||
usage
|
||||
exit 1
|
||||
fi
|
||||
if [ -z "${COMMANDS}" ]; then
|
||||
usage && exit
|
||||
fi
|
||||
if [ -n "${GIVEN_ORGA}" ]; then
|
||||
# intersection des 2 listes : quelle commande de ouf !!
|
||||
AVAILABLE_ORGAS=$(comm -12 <(printf '%s\n' ${AVAILABLE_ORGAS} | sort) <(printf '%s\n' ${GIVEN_ORGA} | sort))
|
||||
fi
|
||||
|
||||
NB_ORGAS=$(echo "${AVAILABLE_ORGAS}" | wc -w )
|
||||
|
||||
if [[ $NB_ORGAS = 0 && -z "${OPERATE_ON_MAIN}" ]]; then
|
||||
echo "Aucune orga trouvée."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
NB_ORGAS_STR="$NB_ORGAS orgas"
|
||||
[ -n "${OPERATE_ON_MAIN}" ] && NB_ORGAS_STR="$NB_ORGAS_STR + les communs"
|
||||
|
||||
_populate_lists # on récupère les clouds / agora / wiki / wp correspondants aux orga
|
||||
|
||||
if [[ $NB_ORGAS -gt 2 && "${COMMANDS[*]}" =~ 'INIT' ]]; then
|
||||
ETLECLOUDCOMMUN=
|
||||
[ -n "${OPERATE_ON_MAIN}" ] && ETLECLOUDCOMMUN=" ainsi que les containers commun"
|
||||
echo "On s'apprête à initialiser les ${CONTAINERS_TYPES} suivants : ${AVAILABLE_ORGAS}${ETLECLOUDCOMMUN}"
|
||||
checkContinue
|
||||
fi
|
||||
|
||||
for COMMAND in ${COMMANDS}; do
|
||||
case "${COMMAND}" in
|
||||
'LIST' )
|
||||
listContainers && exit ;;
|
||||
'VERSION' )
|
||||
version && exit ;;
|
||||
'OPTIMISE-CLOUD' )
|
||||
OptimiseClouds ;;
|
||||
'RESTART-COMPOSE' )
|
||||
restart-compose ;;
|
||||
'RESTART-DOCKER' )
|
||||
restart ;;
|
||||
'UPDATE-CLOUD-APP' )
|
||||
UpdateApplis ;;
|
||||
'UPGRADE' )
|
||||
UpgradeClouds ;;
|
||||
'INIT' )
|
||||
Init ;;
|
||||
'INSTALL-CLOUD-APP' )
|
||||
InstallApps ;;
|
||||
'RUN-CLOUD-OCC' )
|
||||
RunCommands "OCC" "${OCCCOMANDS[@]}" ;;
|
||||
'RUN-AGORA-MMCTL' )
|
||||
RunCommands "MMCTL" "${MMCTLCOMANDS[@]}" ;;
|
||||
'RUN-DOCKER-EXEC' )
|
||||
RunCommands "EXEC" "${EXECCOMANDS[@]}" ;;
|
||||
'POST-AGORA' )
|
||||
PostMessages ${Posts} ;;
|
||||
esac
|
||||
done
|
1151
bin/gestUsers.sh
Executable file
77
bin/indicateurs.sh
Executable file
@ -0,0 +1,77 @@
|
||||
#!/bin/bash
|
||||
# 23/04/2021
|
||||
# script de mise a jour du fichier de collecte pour future intégration dans la base de donneyyy
|
||||
# did
|
||||
|
||||
KAZ_ROOT=$(cd $(dirname $0)/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
FIC_COLLECTE=${KAZ_STATE_DIR}/collecte.csv
|
||||
FIC_ACTIVITE_MAILBOX=${KAZ_STATE_DIR}/activites_mailbox.csv
|
||||
|
||||
mkdir -p ${KAZ_STATE_DIR}
|
||||
mkdir -p ${KAZ_STATE_DIR}/metro
|
||||
|
||||
#Jirafeau
|
||||
echo "$(date +%Y-%m-%d-%H-%M-%S);" \
|
||||
"depot-count;" \
|
||||
"$(find ${DOCK_VOL}/jirafeau_fileData/_data/files/ -name \*count| wc -l)" >> "${FIC_COLLECTE}"
|
||||
echo "$(date +%Y-%m-%d-%H-%M-%S);" \
|
||||
"depot-size;" \
|
||||
"$(du -ks ${DOCK_VOL}/jirafeau_fileData/_data/files/ | awk -F " " '{print $1}')" >> "${FIC_COLLECTE}"
|
||||
|
||||
#PLACE DISQUE sur serveur
|
||||
echo "$(date +%Y-%m-%d-%H-%M-%S);" \
|
||||
"disk-system-size-used;" \
|
||||
"$(df | grep sda | awk -F " " '{print $3}')" >> "${FIC_COLLECTE}"
|
||||
echo "$(date +%Y-%m-%d-%H-%M-%S);" \
|
||||
"disk-system-size-used-human;" \
|
||||
"$(df -h | grep sda | awk -F " " '{print $3}')" >> "${FIC_COLLECTE}"
|
||||
|
||||
#nombre de mails kaz:
|
||||
echo "$(date +%Y-%m-%d-%H-%M-%S);" \
|
||||
"mailboxes;" \
|
||||
"$(cat ${KAZ_COMP_DIR}/postfix/config/postfix-accounts.cf | wc -l)" >> "${FIC_COLLECTE}"
|
||||
|
||||
#nombre d'alias kaz:
|
||||
echo "$(date +%Y-%m-%d-%H-%M-%S);" \
|
||||
"mail_alias;" \
|
||||
"$(cat ${KAZ_COMP_DIR}/postfix/config/postfix-virtual.cf | wc -l)" >> "${FIC_COLLECTE}"
|
||||
#Nombre d' orgas
|
||||
echo "$(date +%Y-%m-%d-%H-%M-%S);" \
|
||||
"Orgas;" \
|
||||
"$(ls -l /kaz/dockers/ | grep orga | wc -l)" >> "${FIC_COLLECTE}"
|
||||
|
||||
#stats des 2 postfix (mail+sympa)
|
||||
EXP=$(/usr/bin/hostname -s)
|
||||
|
||||
STATS1=$(cat ${DOCK_VOL}/sympa_sympaLog/_data/mail.log.1 | /usr/sbin/pflogsumm)
|
||||
#docker exec -i mailServ mailx -r $EXP -s "stats Sympa" root <<DEB_MESS
|
||||
#$STATS1
|
||||
#DEB_MESS
|
||||
|
||||
STATS2=$(cat ${DOCK_VOL}/postfix_mailLog/_data/mail.log | /usr/sbin/pflogsumm)
|
||||
#docker exec -i mailServ mailx -r $EXP -s "stats Postfix" root <<DEB_MESS
|
||||
#$STATS2
|
||||
#DEB_MESS
|
||||
|
||||
IFS=''
|
||||
for line in $(ls -lt --time-style=long-iso "${DOCK_VOL}/postfix_mailData/_data/kaz.bzh/"); do
|
||||
echo "${line}" | awk '{print $6";"$7";"$8";"$9}' > "${FIC_ACTIVITE_MAILBOX}"
|
||||
done
|
||||
|
||||
#pour pister les fuites mémoires
|
||||
docker stats --no-stream --format "table {{.Name}}\t{{.Container}}\t{{.MemUsage}}" | sort -k 3 -h > "${KAZ_STATE_DIR}/metro/$(date +"%Y%m%d")_docker_memory_kaz.log"
|
||||
ps aux --sort -rss > "${KAZ_STATE_DIR}/metro/$(date +"%Y%m%d")_ps_kaz.log"
|
||||
free -hlt > "${KAZ_STATE_DIR}/metro/$(date +"%Y%m%d")_mem_kaz.log"
|
||||
systemd-cgls --no-pager > "${KAZ_STATE_DIR}/metro/$(date +"%Y%m%d")_cgls_kaz.log"
|
||||
for i in $(docker container ls --format "{{.ID}}"); do docker inspect -f '{{.State.Pid}} {{.Name}}' $i; done > "${KAZ_STATE_DIR}/metro/$(date +"%Y%m%d")_dockerpid_kaz.log"
|
||||
|
||||
#on piste cette saloperie d'ethercalc
|
||||
#echo $(date +"%Y%m%d") >> "${KAZ_STATE_DIR}/metro/docker_stats_ethercalc.log"
|
||||
#docker stats --no-stream ethercalcServ ethercalcDB >> "${KAZ_STATE_DIR}/metro/docker_stats_ethercalc.log"
|
||||
|
||||
#fab le 04/10/2022
|
||||
#taille des dockers
|
||||
docker system df -v > "${KAZ_STATE_DIR}/metro/$(date +"%Y%m%d")_docker_size_kaz.log"
|
220
bin/init.sh
Executable file
@ -0,0 +1,220 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
cd "${KAZ_ROOT}"
|
||||
|
||||
MY_MAIN_IP=$(ip a | grep "inet " | head -2 | tail -1 | sed "s%.*inet *\([0-9.]*\)/.*%\1%")
|
||||
MY_SECOND_IP=$(ip a | grep "inet " | head -3 | tail -1 | sed "s%.*inet *\([0-9.]*\)/.*%\1%")
|
||||
|
||||
DOMAIN="kaz.local"
|
||||
DOMAIN_SYMPA="kaz.local"
|
||||
HTTP_PROTO="https"
|
||||
MAIN_IP="${MY_MAIN_IP}"
|
||||
SYMPA_IP="MY_SECOND_IP"
|
||||
RESTART_POLICY="no"
|
||||
JIRAFEAU_DIR="/var/jirafeauData/$(apg -n 1 -m 16 -M NCL)/"
|
||||
|
||||
DOCKERS_TMPL_ENV="${KAZ_CONF_DIR}/dockers.tmpl.env"
|
||||
|
||||
RESET_ENV="true"
|
||||
if [ -f "${DOCKERS_ENV}" ]; then
|
||||
DOMAIN=$(getValInFile "${DOCKERS_ENV}" "domain")
|
||||
DOMAIN_SYMPA=$(getValInFile "${DOCKERS_ENV}" "domain_sympa")
|
||||
HTTP_PROTO=$(getValInFile "${DOCKERS_ENV}" "httpProto")
|
||||
MAIN_IP=$(getValInFile "${DOCKERS_ENV}" "MAIN_IP")
|
||||
SYMPA_IP=$(getValInFile "${DOCKERS_ENV}" "SYMPA_IP")
|
||||
RESTART_POLICY=$(getValInFile "${DOCKERS_ENV}" "restartPolicy")
|
||||
JIRAFEAU_DIR=$(getValInFile "${DOCKERS_ENV}" "jirafeauDir")
|
||||
while : ; do
|
||||
read -p "Change '${DOCKERS_ENV}'? " resetEnv
|
||||
case "${resetEnv}" in
|
||||
[yYoO]* )
|
||||
break
|
||||
;;
|
||||
""|[Nn]* )
|
||||
RESET_ENV=""
|
||||
break
|
||||
;;
|
||||
* )
|
||||
echo "Please answer yes no."
|
||||
;;
|
||||
esac
|
||||
done
|
||||
fi
|
||||
|
||||
[ -n "${RESET_ENV}" ] && {
|
||||
echo "Reset '${DOCKERS_ENV}'"
|
||||
read -p " * domain (kaz.bzh / dev.kaz.bzh / kaz.local)? [${YELLOW}${DOMAIN}${NC}] " domain
|
||||
case "${domain}" in
|
||||
"" )
|
||||
DOMAIN="${DOMAIN}"
|
||||
;;
|
||||
* )
|
||||
# XXX ne conserver que .-0-9a-z
|
||||
DOMAIN=$(sed 's/[^a-z0-9.-]//g' <<< "${domain}")
|
||||
;;
|
||||
esac
|
||||
|
||||
read -p " * lists domain (kaz.bzh / kaz2.ovh / kaz.local)? [${YELLOW}${DOMAIN_SYMPA}${NC}] " domain
|
||||
case "${domain}" in
|
||||
"" )
|
||||
DOMAIN_SYMPA="${DOMAIN_SYMPA}"
|
||||
;;
|
||||
* )
|
||||
DOMAIN_SYMPA="${domain}"
|
||||
;;
|
||||
esac
|
||||
|
||||
while : ; do
|
||||
read -p " * protocol (https / http)? [${YELLOW}${HTTP_PROTO}${NC}] " proto
|
||||
case "${proto}" in
|
||||
"" )
|
||||
HTTP_PROTO="${HTTP_PROTO}"
|
||||
break
|
||||
;;
|
||||
"https"|"http" )
|
||||
HTTP_PROTO="${proto}"
|
||||
break
|
||||
;;
|
||||
* ) echo "Please answer joe, emacs, vim or no."
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
while : ; do
|
||||
read -p " * main IP (ip)? [${YELLOW}${MAIN_IP}${NC}] " ip
|
||||
case "${ip}" in
|
||||
"" )
|
||||
MAIN_IP="${MAIN_IP}"
|
||||
break
|
||||
;;
|
||||
* )
|
||||
if testValidIp "${ip}" ; then
|
||||
MAIN_IP="${ip}"
|
||||
break
|
||||
else
|
||||
echo "Please answer x.x.x.x format."
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
while : ; do
|
||||
read -p " * lists IP (ip)? [${YELLOW}${SYMPA_IP}${NC}] " ip
|
||||
case "${ip}" in
|
||||
"" )
|
||||
SYMPA_IP="${SYMPA_IP}"
|
||||
break
|
||||
;;
|
||||
* )
|
||||
if testValidIp "${ip}" ; then
|
||||
SYMPA_IP="${ip}"
|
||||
break
|
||||
else
|
||||
echo "Please answer x.x.x.x format."
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
while : ; do
|
||||
read -p " * restart policy (always / unless-stopped / no)? [${YELLOW}${RESTART_POLICY}${NC}] " policy
|
||||
case "${policy}" in
|
||||
"" )
|
||||
RESTART_POLICY="${RESTART_POLICY}"
|
||||
break
|
||||
;;
|
||||
"always"|"unless-stopped"|"no")
|
||||
RESTART_POLICY="${policy}"
|
||||
break
|
||||
;;
|
||||
* ) echo "Please answer always, unless-stopped or no."
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
while : ; do
|
||||
read -p " * Jirafeau dir? [${YELLOW}${JIRAFEAU_DIR}${NC}] " jirafeauDir
|
||||
case "${jirafeauDir}" in
|
||||
"" )
|
||||
JIRAFEAU_DIR="${JIRAFEAU_DIR}"
|
||||
break
|
||||
;;
|
||||
* )
|
||||
if [[ "${jirafeauDir}" =~ ^/var/jirafeauData/[0-9A-Za-z]{1,16}/$ ]]; then
|
||||
JIRAFEAU_DIR="${jirafeauDir}"
|
||||
break
|
||||
else
|
||||
echo "Please give dir name (/var/jirafeauData/[0-9A-Za-z]{1,3}/)."
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
[ -f "${DOCKERS_ENV}" ] || cp "${DOCKERS_TMPL_ENV}" "${DOCKERS_ENV}"
|
||||
|
||||
sed -i "${DOCKERS_ENV}" \
|
||||
-e "s%^\s*domain\s*=.*$%domain=${DOMAIN}%" \
|
||||
-e "s%^\s*domain_sympa\s*=.*$%domain_sympa=${DOMAIN_SYMPA}%" \
|
||||
-e "s%^\s*httpProto\s*=.*$%httpProto=${HTTP_PROTO}%" \
|
||||
-e "s%^\s*MAIN_IP\s*=.*$%MAIN_IP=${MAIN_IP}%" \
|
||||
-e "s%^\s*SYMPA_IP\s*=.*$%SYMPA_IP=${SYMPA_IP}%" \
|
||||
-e "s%^\s*restartPolicy\s*=.*$%restartPolicy=${RESTART_POLICY}%" \
|
||||
-e "s%^\s*ldapRoot\s*=.*$%ldapRoot=dc=${DOMAIN_SYMPA/\./,dc=}%" \
|
||||
-e "s%^\s*jirafeauDir\s*=.*$%jirafeauDir=${JIRAFEAU_DIR}%"
|
||||
}
|
||||
|
||||
if [ ! -f "${KAZ_CONF_DIR}/container-mail.list" ]; then
|
||||
cat > "${KAZ_CONF_DIR}/container-mail.list" <<EOF
|
||||
# e-mail server composer
|
||||
postfix
|
||||
ldap
|
||||
#sympa
|
||||
EOF
|
||||
fi
|
||||
|
||||
if [ ! -f "${KAZ_CONF_DIR}/container-orga.list" ]; then
|
||||
cat > "${KAZ_CONF_DIR}/container-orga.list" <<EOF
|
||||
# orga composer
|
||||
EOF
|
||||
fi
|
||||
|
||||
if [ ! -f "${KAZ_CONF_DIR}/container-proxy.list" ]; then
|
||||
cat > "${KAZ_CONF_DIR}/container-proxy.list" <<EOF
|
||||
proxy
|
||||
EOF
|
||||
fi
|
||||
|
||||
if [ ! -f "${KAZ_CONF_DIR}/container-withMail.list" ]; then
|
||||
cat > "${KAZ_CONF_DIR}/container-withMail.list" <<EOF
|
||||
web
|
||||
etherpad
|
||||
roundcube
|
||||
framadate
|
||||
paheko
|
||||
dokuwiki
|
||||
gitea
|
||||
mattermost
|
||||
cloud
|
||||
EOF
|
||||
fi
|
||||
|
||||
if [ ! -f "${KAZ_CONF_DIR}/container-withoutMail.list" ]; then
|
||||
cat > "${KAZ_CONF_DIR}/container-withoutMail.list" <<EOF
|
||||
jirafeau
|
||||
ethercalc
|
||||
collabora
|
||||
#vigilo
|
||||
#grav
|
||||
EOF
|
||||
fi
|
||||
|
||||
if [ ! -d "${KAZ_ROOT}/secret" ]; then
|
||||
rsync -a "${KAZ_ROOT}/secret.tmpl/" "${KAZ_ROOT}/secret/"
|
||||
. "${KAZ_ROOT}/secret/SetAllPass.sh"
|
||||
"${KAZ_BIN_DIR}/secretGen.sh"
|
||||
"${KAZ_BIN_DIR}/updateDockerPassword.sh"
|
||||
fi
|
144
bin/install.sh
Executable file
@ -0,0 +1,144 @@
|
||||
#!/bin/bash
|
||||
|
||||
set -e
|
||||
# on pourra inclure le fichier dockers.env pour
|
||||
# gérer l' environnement DEV, PROD ou LOCAL
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
export VAGRANT_SRC_DIR=/vagrant/files
|
||||
cd "${KAZ_ROOT}"
|
||||
|
||||
if [ ! -f "${KAZ_ROOT}/config/dockers.env" ]; then
|
||||
printKazError "dockers.env not found"
|
||||
exit 1
|
||||
fi
|
||||
for type in mail orga proxy withMail withoutMail ; do
|
||||
if [ ! -f "${KAZ_ROOT}/config/container-${type}.list" ]; then
|
||||
printKazError "container-${type}.list not found"
|
||||
exit 1
|
||||
fi
|
||||
done
|
||||
|
||||
mkdir -p "${KAZ_ROOT}/log/"
|
||||
export DebugLog="${KAZ_ROOT}/log/log-install-$(date +%y-%m-%d-%T)-"
|
||||
(
|
||||
declare -a DOCKERS_LIST NEW_SERVICE
|
||||
# dockers à démarrer (manque : sympa, wordpress, orga)
|
||||
DOCKERS_LIST+=($(getList "${KAZ_CONF_DIR}/container-withoutMail.list"))
|
||||
DOCKERS_LIST+=($(getList "${KAZ_CONF_DIR}/container-proxy.list"))
|
||||
DOCKERS_LIST+=($(getList "${KAZ_CONF_DIR}/container-mail.list"))
|
||||
DOCKERS_LIST+=($(getList "${KAZ_CONF_DIR}/container-withMail.list"))
|
||||
# web proxy postfix sympa roundcube jirafeau ldap quotas cachet ethercalc etherpad framadate paheko dokuwiki gitea mattermost cloud collabora
|
||||
# 8080 443 8081 8082 8083 8084 8085 8086 8087 8088 8089 8090 8091 8092 8093 8094
|
||||
# pour ne tester qu'un sous-ensemble de service
|
||||
if [ $# -ne 0 ]; then
|
||||
case $1 in
|
||||
-h*|--h*)
|
||||
echo $(basename "$0") " [-h] [-help] ([1-9]* | {service...})"
|
||||
echo " -h"
|
||||
echo " -help Display this help."
|
||||
echo " service.. service to enable"
|
||||
echo " [1-9]* level of predefined services set selection"
|
||||
exit
|
||||
;;
|
||||
0)
|
||||
echo $(basename "$0"): " level '0' not defined"
|
||||
exit
|
||||
;;
|
||||
[0-9]*)
|
||||
for level in $(seq 1 $1); do
|
||||
case ${level} in
|
||||
1) NEW_SERVICE+=("web" "proxy");;
|
||||
2) NEW_SERVICE+=("postfix");;
|
||||
3) NEW_SERVICE+=("roundcube");;
|
||||
4) NEW_SERVICE+=("sympa");;
|
||||
5) NEW_SERVICE+=("jirafeau");;
|
||||
6) NEW_SERVICE+=("ldap");;
|
||||
7) NEW_SERVICE+=("quotas");;
|
||||
8) NEW_SERVICE+=("cachet");;
|
||||
9) NEW_SERVICE+=("ethercalc");;
|
||||
10) NEW_SERVICE+=("etherpad");;
|
||||
11) NEW_SERVICE+=("framadate");;
|
||||
12) NEW_SERVICE+=("paheko");;
|
||||
13) NEW_SERVICE+=("dokuwiki");;
|
||||
14) NEW_SERVICE+=("gitea");;
|
||||
15) NEW_SERVICE+=("mattermost");;
|
||||
16) NEW_SERVICE+=("collabora");;
|
||||
17) NEW_SERVICE+=("cloud");;
|
||||
*)
|
||||
echo $(basename "$0"): " level '${level}' not defined"
|
||||
exit
|
||||
;;
|
||||
esac
|
||||
done
|
||||
DOCKERS_LIST=(${NEW_SERVICE[@]})
|
||||
printKazMsg "level $1"
|
||||
;;
|
||||
*)
|
||||
# XXX il manque l'extention des noms (jir va fair le start de jirafeau mais pas le download et le first)
|
||||
DOCKERS_LIST=($*)
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
|
||||
DOCKERS_LIST=($(filterAvailableComposes ${DOCKERS_LIST[*]}))
|
||||
|
||||
printKazMsg "dockers: ${DOCKERS_LIST[*]}"
|
||||
|
||||
# on pré-télécharge à l'origine Vagrant (jirafeau...)
|
||||
mkdir -p "${KAZ_ROOT}/git" "${KAZ_ROOT}/download"
|
||||
for DOCKER in ${DOCKERS_LIST[@]}; do
|
||||
if [ -f "${KAZ_ROOT}/dockers/${DOCKER}/download.sh" ]; then
|
||||
cd "${KAZ_ROOT}/dockers/${DOCKER}"
|
||||
./download.sh
|
||||
fi
|
||||
done
|
||||
|
||||
# on pré-télécharge le dépollueur
|
||||
if [[ " ${DOCKERS_LIST[*]} " =~ " "(jirafeau|postfix|sympa)" " ]]; then
|
||||
"${KAZ_BIN_DIR}/installDepollueur.sh"
|
||||
docker volume create filterConfig
|
||||
fi
|
||||
|
||||
# on sauve les pré-téléchargement pour le prochain lancement de Vagrant
|
||||
[ -d "${VAGRANT_SRC_DIR}/kaz/download" ] &&
|
||||
rsync -a "${KAZ_ROOT}/download/" "${VAGRANT_SRC_DIR}/kaz/download/"
|
||||
[ -d "${VAGRANT_SRC_DIR}/kaz/git" ] &&
|
||||
rsync -a "${KAZ_ROOT}/git/" "${VAGRANT_SRC_DIR}/kaz/git/"
|
||||
|
||||
# on construit les dockers qui contiennent un script de création (etherpad, framadate, jirafeau...)
|
||||
for DOCKER in ${DOCKERS_LIST[@]}; do
|
||||
if [ -f "${KAZ_ROOT}/dockers/${DOCKER}/build.sh" ]; then
|
||||
cd "${KAZ_ROOT}/dockers/${DOCKER}"
|
||||
./build.sh
|
||||
fi
|
||||
done
|
||||
|
||||
# on démare les containers de la liste uniquement (en une fois par cohérence de proxy)
|
||||
# "${KAZ_ROOT}/bin/container.sh" stop ${DOCKERS_LIST[*]}
|
||||
"${KAZ_ROOT}/bin/container.sh" start ${DOCKERS_LIST[*]}
|
||||
|
||||
if [[ " ${DOCKERS_LIST[*]} " =~ " etherpad " ]]; then
|
||||
# pb avec la lanteur de démarrage du pad :-(
|
||||
sleep 5
|
||||
"${KAZ_ROOT}/bin/container.sh" start etherpad
|
||||
fi
|
||||
|
||||
|
||||
if [[ " ${DOCKERS_LIST[*]} " =~ " jirafeau " ]]; then
|
||||
# pb avec la lanteur de démarrage du jirafeau :-(
|
||||
(cd "${KAZ_COMP_DIR}/jirafeau" ; docker-compose restart)
|
||||
fi
|
||||
|
||||
# on construit les dockers qui contiennent un script de création (etherpad, framadate, jirafeau...)
|
||||
for DOCKER in ${DOCKERS_LIST[@]}; do
|
||||
if [ -f "${KAZ_ROOT}/dockers/${DOCKER}/first.sh" ]; then
|
||||
cd "${KAZ_ROOT}/dockers/${DOCKER}"
|
||||
./first.sh
|
||||
fi
|
||||
done
|
||||
|
||||
echo "########## ********** End install $(date +%D-%T)"
|
||||
) > >(tee ${DebugLog}stdout.log) 2> >(tee ${DebugLog}stderr.log >&2)
|
27
bin/installDepollueur.sh
Executable file
@ -0,0 +1,27 @@
|
||||
#!/bin/bash
|
||||
|
||||
SRC_DEP=https://git.kaz.bzh/KAZ/depollueur.git
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
if [[ -f "${KAZ_GIT_DIR}/depollueur/test-running" ]]; then
|
||||
exit
|
||||
fi
|
||||
|
||||
printKazMsg "\n *** Installation du dépollueur"
|
||||
|
||||
sudo apt-get install -y --fix-missing build-essential make g++ libboost-program-options-dev libboost-system-dev libboost-filesystem-dev libcurl4-gnutls-dev libssl-dev
|
||||
|
||||
mkdir -p "${KAZ_GIT_DIR}"
|
||||
cd "${KAZ_GIT_DIR}"
|
||||
if [ ! -d "depollueur" ]; then
|
||||
git clone "${SRC_DEP}"
|
||||
fi
|
||||
cd depollueur
|
||||
git reset --hard && git pull
|
||||
make
|
||||
|
||||
. "${DOCKERS_ENV}"
|
||||
echo "${domain}" > "src/bash/domainname"
|
166
bin/interoPaheko.sh
Executable file
@ -0,0 +1,166 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
URL_PAHEKO="$httpProto://${paheko_API_USER}:${paheko_API_PASSWORD}@kaz-paheko.$(echo $domain)"
|
||||
|
||||
PRG=$(basename $0)
|
||||
RACINE=$(echo $PRG | awk '{print $1}')
|
||||
|
||||
TFILE_INT_PAHEKO_ACTION=$(mktemp /tmp/XXXXXXXX_INT_PAHEKO_ACTION.json)
|
||||
TFILE_INT_PAHEKO_IDFILE=$(mktemp /tmp/XXXXXXXX_TFILE_INT_PAHEKO_IDFILE.json)
|
||||
FILE_CREATEUSER="$KAZ_ROOT/tmp/createUser.txt"
|
||||
|
||||
sep=' '
|
||||
|
||||
#trap "rm -f ${TFILE_INT_PAHEKO_IDFILE} ${TFILE_INT_PAHEKO_ACTION} " 0 1 2 3 15
|
||||
|
||||
############################################ Fonctions #######################################################
|
||||
TEXTE="
|
||||
# -- fichier de création des comptes KAZ
|
||||
# --
|
||||
# -- 1 ligne par compte
|
||||
# -- champs séparés par ;. les espaces en début et en fin sont enlevés
|
||||
# -- laisser vide si pas de donnée
|
||||
# -- pas d'espace dans les variables
|
||||
# --
|
||||
# -- ORGA: nom de l'organisation (max 15 car), vide sinon
|
||||
# -- ADMIN_ORGA: O/N indique si le user est admin de l'orga (va le créer comme admin du NC de l'orga et admin de l'équipe agora)
|
||||
# -- NC_ORGA: O/N indique si l'orga a demandé un NC
|
||||
# -- PAHEKO_ORGA: O/N indique si l'orga a demandé un paheko
|
||||
# -- WP_ORGA: O/N indique si l'orga a demandé un wp
|
||||
# -- AGORA_ORGA: O/N indique si l'orga a demandé un mattermost
|
||||
# -- WIKI_ORGA: O/N indique si l'orga a demandé un wiki
|
||||
# -- NC_BASE: O/N indique si le user doit être inscrit dans le NC de base
|
||||
# -- GROUPE_NC_BASE: soit null soit le groupe dans le NC de base
|
||||
# -- EQUIPE_AGORA: soit null soit equipe agora (max 15 car)
|
||||
# -- QUOTA=(1/10/20/...) en GB
|
||||
# --
|
||||
# NOM ; PRENOM ; EMAIL_SOUHAITE ; EMAIL_SECOURS ; ORGA ; ADMIN_ORGA ; NC_ORGA ; PAHEKO_ORGA ; WP_ORGA ; AGORA_ORGA ; WIKI_ORGA ; NC_BASE ; GROUPE_NC_BASE ; EQUIPE_AGORA ; QUOTA
|
||||
#
|
||||
# exemple pour un compte découverte:
|
||||
# dupont ; jean-louis; jean-louis.dupont@kaz.bzh ; gregomondo@kaz.bzh; ; N; N; N; N; N; N; O; ; ;1
|
||||
#
|
||||
# exemple pour un compte asso de l'orga gogol avec le service dédié NC uniquement + une équipe dans l'agora
|
||||
# dupont ; jean-louis; jean-louis.dupont@kaz.bzh ; gregomondo@kaz.bzh; gogol ; O; O; N; N; N; N;N;;gogol_team; 10
|
||||
"
|
||||
|
||||
Int_paheko_Action() {
|
||||
# $1 est une action;
|
||||
ACTION=$1
|
||||
OPTION=$2
|
||||
# on envoie la requête sur le serveur paheko avec la clause à créer
|
||||
# problème de gestion de remontée de données dans la table services_users quand le compte a plus de 2 activités
|
||||
#curl -s ${URL_PAHEKO}/api/sql -d "SELECT * from users cross join services_users on users.id = services_users.id_user where users.action_auto='${ACTION}';" >>${TFILE_INT_PAHEKO_ACTION}
|
||||
curl -s ${URL_PAHEKO}/api/sql -d "SELECT * from users where action_auto='${ACTION}';" >>${TFILE_INT_PAHEKO_ACTION}
|
||||
[ ! -z ${TFILE_INT_PAHEKO_ACTION} ] || { echo "probleme de fichier ${TFILE_INT_PAHEKO_ACTION}" ; exit 1;}
|
||||
REP_ID=$(jq -c '.results[].id ' ${TFILE_INT_PAHEKO_ACTION} 2>/dev/null)
|
||||
if [ ! -z "${REP_ID}" ]
|
||||
then
|
||||
[ "$OPTION" = "silence" ] || echo -e "${RED}Nombre de compte ${ACTION} ${NC}= ${GREEN} $(echo ${REP_ID} | wc -w) ${NC}"
|
||||
if [ -f "$FILE_CREATEUSER" ]
|
||||
then
|
||||
mv $FILE_CREATEUSER $FILE_CREATEUSER.$(date +%d-%m-%Y-%H:%M:%S)
|
||||
fi
|
||||
echo "# -------- Fichier généré le $(date +%d-%m-%Y-%H:%M:%S) ----------">${FILE_CREATEUSER}
|
||||
echo "${TEXTE}" >>${FILE_CREATEUSER}
|
||||
for VAL_ID in ${REP_ID}
|
||||
do
|
||||
jq -c --argjson val "${VAL_ID}" '.results[] | select (.id == $val)' ${TFILE_INT_PAHEKO_ACTION} > ${TFILE_INT_PAHEKO_IDFILE}
|
||||
for VAL_GAR in id_category action_auto nom email email_secours quota_disque admin_orga nom_orga responsable_organisation responsable_email agora cloud wordpress garradin docuwiki id_service
|
||||
do
|
||||
eval $VAL_GAR=$(jq .$VAL_GAR ${TFILE_INT_PAHEKO_IDFILE})
|
||||
done
|
||||
#comme tout va bien on continue
|
||||
#on compte le nom de champs dans la zone nom pour gérer les noms et prénoms composés
|
||||
# si il y a 3 champs, on associe les 2 premieres valeurs avec un - et on laisse le 3ème identique
|
||||
# si il y a 4 champs on associe les 1 et le 2 avec un tiret et le 3 et 4 avec un tiret
|
||||
# on met les champs nom_ok et prenom_ok à blanc
|
||||
nom_ok=""
|
||||
prenom_ok=""
|
||||
# on regarde si le nom de l' orga est renseigné ou si le nom de l' orga est null et l' activité de membre est 7 (membre rattaché)
|
||||
# si c' est le cas alors le nom est le nom de l' orga et le prénom est forcé à la valeur Organisation
|
||||
if [[ "$nom_orga" = null ]] || [[ "$nom_orga" != null && "$id_service" = "7" ]]
|
||||
then
|
||||
[ "$OPTION" = "silence" ] || echo -e "${NC}Abonné ${GREEN}${nom}${NC}"
|
||||
#si lactivité est membre rattaché on affiche a quelle orga il est rattaché
|
||||
if [ "$id_service" = "7" ] && [ "$OPTION" != "silence" ] && [ "$nom_orga" != null ]
|
||||
then
|
||||
echo -e "${NC}Orga Rattachée : ${GREEN}${nom_orga}${NC}"
|
||||
fi
|
||||
COMPTE_NOM=$(echo $nom | awk -F' ' '{for (i=1; i != NF; i++); print i;}')
|
||||
case "${COMPTE_NOM}" in
|
||||
0|1)
|
||||
echo "Il faut corriger le champ nom (il manque un nom ou prénom) de paheko"
|
||||
echo "je quitte et supprime le fichier ${FILE_CREATEUSER}"
|
||||
rm -f $FILE_CREATEUSER
|
||||
exit 2
|
||||
;;
|
||||
2)
|
||||
nom_ok=$(echo $nom | awk -F' ' '{print $1}')
|
||||
prenom_ok=$(echo $nom | awk -F' ' '{print $2}')
|
||||
;;
|
||||
*)
|
||||
nom_ok=
|
||||
prenom_ok=
|
||||
for i in ${nom}; do grep -q '^[A-Z]*$' <<<"${i}" && nom_ok="${nom_ok}${sep}${i}" || prenom_ok="${prenom_ok}${sep}${i}"; done
|
||||
nom_ok="${nom_ok#${sep}}"
|
||||
prenom_ok="${prenom_ok#${sep}}"
|
||||
if [ -z "${nom_ok}" ] || [ -z "${prenom_ok}" ]; then
|
||||
echo "Il faut corriger le champ nom (peut être un nom de famille avec une particule ?) de paheko"
|
||||
echo "je quitte et supprime le fichier ${FILE_CREATEUSER}"
|
||||
rm -f $FILE_CREATEUSER
|
||||
exit
|
||||
fi
|
||||
esac
|
||||
# comme l' orga est à null nom orga est a vide, pas d' admin orga, on met dans l' agora générale
|
||||
# pas d' équipe agora et de groupe nextcloud spécifique
|
||||
|
||||
nom_orga=" "
|
||||
admin_orga="N"
|
||||
nc_base="O"
|
||||
equipe_agora=" "
|
||||
groupe_nc_base=" "
|
||||
else
|
||||
# L' orga est renseigné dans paheko donc les nom et prenoms sont forcé a nom_orga et Organisation
|
||||
# un équipe agora portera le nom de l' orga, le compte ne sera pas créé dans le nextcloud général
|
||||
# et le compte est admin de son orga
|
||||
nom_orga=$(echo $nom_orga | tr [:upper:] [:lower:])
|
||||
[ "$OPTION" = "silence" ] || echo -e "${NC}Orga : ${GREEN}${nom_orga}${NC}"
|
||||
nom_ok=$nom_orga
|
||||
# test des caractères autorisés dans le nom d' orga: lettres, chiffres et/ou le tiret
|
||||
if ! [[ "${nom_ok}" =~ ^[[:alnum:]-]+$ ]]; then
|
||||
echo "Erreur : l' orga doit être avec des lettres et/ou des chiffres. Le séparateur doit être le tiret"
|
||||
rm -f $FILE_CREATEUSER<EFBFBD>
|
||||
exit 2
|
||||
fi
|
||||
prenom_ok=organisation
|
||||
equipe_agora=$nom_orga
|
||||
groupe_nc_base=" "
|
||||
nc_base="N"
|
||||
admin_orga="O"
|
||||
fi
|
||||
# Pour le reste on renomme les null en N ( non ) et les valeurs 1 en O ( Oui)
|
||||
cloud=$(echo $cloud | sed -e 's/0/N/g' | sed -e 's/1/O/g')
|
||||
paheko=$(echo $garradin | sed -e 's/0/N/g' | sed -e 's/1/O/g')
|
||||
wordpress=$(echo $wordpress | sed -e 's/0/N/g' | sed -e 's/1/O/g')
|
||||
agora=$(echo $agora | sed -e 's/0/N/g' | sed -e 's/1/O/g')
|
||||
docuwiki=$(echo $docuwiki | sed -e 's/0/N/g' | sed -e 's/1/O/g')
|
||||
# et enfin on écrit dans le fichier
|
||||
echo "$nom_ok;$prenom_ok;$email;$email_secours;$nom_orga;$admin_orga;$cloud;$paheko;$wordpress;$agora;$docuwiki;$nc_base;$groupe_nc_base;$equipe_agora;$quota_disque">>${FILE_CREATEUSER}
|
||||
done
|
||||
else
|
||||
echo "Rien à créer"
|
||||
exit 2
|
||||
fi
|
||||
}
|
||||
#Int_paheko_Action "A créer" "silence"
|
||||
Int_paheko_Action "A créer"
|
||||
exit 0
|
||||
|
14
bin/iptables.sh
Executable file
@ -0,0 +1,14 @@
|
||||
#!/bin/bash
|
||||
|
||||
#cleaning, may throw errors at first launch
|
||||
#iptables -t nat -D POSTROUTING -o ens18 -j ipbis
|
||||
#iptables -t nat -F ipbis
|
||||
#iptables -t nat -X ipbis
|
||||
|
||||
iptables -t nat -N ipbis
|
||||
iptables -t nat -F ipbis
|
||||
iptables -t nat -I ipbis -o ens18 -p tcp --source `docker inspect -f '{{.NetworkSettings.Networks.sympaNet.IPAddress}}' sympaServ` -j SNAT --to `ifconfig ens18:0 | grep "inet" | awk '{print $2}'`
|
||||
iptables -t nat -I ipbis -o ens18 -p tcp --source `docker inspect -f '{{.NetworkSettings.Networks.jirafeauNet.IPAddress}}' sympaServ` -j SNAT --to `ifconfig ens18:0 | grep "inet" | awk '{print $2}'`
|
||||
iptables -t nat -A ipbis -j RETURN
|
||||
iptables -t nat -D POSTROUTING -o ens18 -j ipbis
|
||||
iptables -t nat -I POSTROUTING -o ens18 -j ipbis
|
110
bin/kazDockerNet.sh
Executable file
@ -0,0 +1,110 @@
|
||||
#!/bin/bash
|
||||
|
||||
# faire un completion avec les composant dispo
|
||||
|
||||
PRG=$(basename $0)
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
usage () {
|
||||
echo "Usage: ${PRG} [-n] [-h] list|add [netName]..."
|
||||
echo " -n : simulation"
|
||||
echo " -h|--help : help"
|
||||
echo
|
||||
echo " create all net : ${PRG} add $(${KAZ_BIN_DIR}/kazList.sh compose validate)"
|
||||
exit 1
|
||||
}
|
||||
|
||||
allNetName=""
|
||||
export CMD=""
|
||||
for ARG in $@; do
|
||||
case "${ARG}" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
'-n' )
|
||||
shift
|
||||
export SIMU="echo"
|
||||
;;
|
||||
-*)
|
||||
usage
|
||||
;;
|
||||
list|add)
|
||||
CMD="${ARG}"
|
||||
shift;
|
||||
;;
|
||||
*)
|
||||
allNetName="${allNetName} ${ARG}"
|
||||
shift
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z "${CMD}" ] ; then
|
||||
usage
|
||||
fi
|
||||
|
||||
# running composes
|
||||
export allBridgeName="$(docker network list | grep bridge | awk '{print $2}')"
|
||||
# running network
|
||||
export allBridgeNet=$(for net in ${allBridgeName} ; do docker inspect ${net} | grep Subnet | sed 's#.*"\([0-9]*\.[0-9]*\.[0-9]*\.[0-9]*/[0-9]*\)".*# \1#'; done)
|
||||
|
||||
minB=0
|
||||
minC=0
|
||||
minD=0
|
||||
|
||||
getNet() {
|
||||
netName="$1Net"
|
||||
|
||||
if [[ "${allBridgeName}" =~ "${netName}" ]]; then
|
||||
echo "${netName} already created"
|
||||
return
|
||||
fi
|
||||
# echo "start 10.${minB}.${minC}.$((${minD}*16))"
|
||||
|
||||
find=""
|
||||
for b in $(eval echo {${minB}..255}); do
|
||||
for c in $(eval echo {${minC}..255}); do
|
||||
for d in $(eval echo {${minD}..15}); do
|
||||
if [ ! -z "${find}" ]; then
|
||||
minB=${b}
|
||||
minC=${c}
|
||||
minD=${d}
|
||||
return
|
||||
fi
|
||||
# to try
|
||||
subnet="10.${b}.${c}.$((d*16))"
|
||||
if [[ "${allBridgeNet}" =~ " ${subnet}/" ]];
|
||||
then
|
||||
# used
|
||||
# XXX check netmask
|
||||
continue
|
||||
fi
|
||||
# the winner is...
|
||||
echo "${netName} => ${subnet}/28"
|
||||
${SIMU} docker network create --subnet "${subnet}/28" "${netName}"
|
||||
find="ok"
|
||||
done
|
||||
minD=0
|
||||
done
|
||||
minC=0
|
||||
done
|
||||
}
|
||||
|
||||
list () {
|
||||
echo "name: " ${allBridgeName}
|
||||
echo "net: " ${allBridgeNet}
|
||||
}
|
||||
|
||||
add () {
|
||||
if [ -z "${allNetName}" ] ; then
|
||||
usage
|
||||
fi
|
||||
for netName in ${allNetName}; do
|
||||
getNet "${netName}"
|
||||
done
|
||||
}
|
||||
|
||||
"${CMD}"
|
181
bin/kazList.sh
Executable file
@ -0,0 +1,181 @@
|
||||
#!/bin/bash
|
||||
|
||||
PRG=$(basename $0)
|
||||
SIMU=""
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
. "${DOCKERS_ENV}"
|
||||
|
||||
cd "$(dirname $0)"
|
||||
|
||||
ALL_STATUS=$(docker ps -a --format "{{.ID}} {{.Names}} {{.Status}}")
|
||||
SERVICES_CHOICE="$(getAvailableServices | tr "\n" "|")"
|
||||
SERVICES_CHOICE="${SERVICES_CHOICE%|}"
|
||||
|
||||
usage () {
|
||||
echo "${RED}${BOLD}" \
|
||||
"Usage: $0 [-h] {compose|orga|service} {available|validate|enable|disable|status} [names]...${NL}" \
|
||||
" -h help${NL}" \
|
||||
" compose {available|validate|enable|disable} : list docker-compose name${NL}" \
|
||||
" compose status : status of docker-compose (default available)${NL}" \
|
||||
" service {available|validate} : list services name${NL}" \
|
||||
" service {enable|disable} : list services in orga${NL}" \
|
||||
" service status : status of services in orga${NL}" \
|
||||
" service {${SERVICES_CHOICE}}${NL}" \
|
||||
" : list of orga enable a service ${NL}" \
|
||||
" [compose] in${NL}" \
|
||||
" ${CYAN}$((getAvailableComposes;getAvailableOrgas) | tr "\n" " ")${NC}${NL}"
|
||||
exit 1
|
||||
}
|
||||
|
||||
# ========================================
|
||||
compose_available () {
|
||||
echo $*
|
||||
}
|
||||
|
||||
getComposeEnableByProxy () {
|
||||
onList=$(
|
||||
for type in ${KAZ_CONF_DIR}/container-*.list ; do
|
||||
getList "${type}"
|
||||
done)
|
||||
local compose
|
||||
for compose in ${onList} ; do
|
||||
composeFlag="proxy_${compose//-/_}"
|
||||
[[ "${!composeFlag}" == "on" ]] && echo ${compose}
|
||||
done
|
||||
}
|
||||
|
||||
compose_validate () {
|
||||
echo $(
|
||||
for type in ${KAZ_CONF_DIR}/container-*.list ; do
|
||||
getList "${type}"
|
||||
done | filterInList $*)
|
||||
}
|
||||
|
||||
compose_enable () {
|
||||
echo $(getComposeEnableByProxy | filterInList $*)
|
||||
}
|
||||
|
||||
compose_disable () {
|
||||
echo $(getAvailableComposes | filterNotInList $(getComposeEnableByProxy) | filterInList $*)
|
||||
}
|
||||
|
||||
compose_status () {
|
||||
for compose in $*; do
|
||||
cd "${KAZ_COMP_DIR}/${compose}"
|
||||
echo "${compose}:"
|
||||
for service in $(docker-compose ps --services 2>/dev/null); do
|
||||
id=$(docker-compose ps -q "${service}" | cut -c 1-12)
|
||||
if [ -z "${id}" ]; then
|
||||
echo " - ${RED}${BOLD}[Down]${NC} ${service}"
|
||||
else
|
||||
status=$(grep "^${id}\b" <<< "${ALL_STATUS}" | sed "s/.*${id}\s\s*\S*\s\s*\(\S*.*\)/\1/")
|
||||
COLOR=$([[ "${status}" =~ Up ]] && echo "${GREEN}" || echo "${RED}")
|
||||
echo " - ${COLOR}${BOLD}[${status}]${NC} ${service}"
|
||||
fi
|
||||
done
|
||||
done
|
||||
}
|
||||
|
||||
# ========================================
|
||||
service_available () {
|
||||
echo $(getAvailableServices)
|
||||
}
|
||||
|
||||
service_validate () {
|
||||
echo $(getAvailableServices)
|
||||
}
|
||||
|
||||
getServiceInOrga () {
|
||||
for orga in $*; do
|
||||
[[ "${orga}" = *"-orga" ]] || continue
|
||||
local ORGA_DIR="${KAZ_COMP_DIR}/${orga}"
|
||||
ORGA_COMPOSE="${ORGA_DIR}/docker-compose.yml"
|
||||
[[ -f "${ORGA_COMPOSE}" ]] || continue
|
||||
for service in $(getAvailableServices); do
|
||||
case "${service}" in
|
||||
paheko)
|
||||
[ -f "${ORGA_DIR}/usePaheko" ] && echo "${service}"
|
||||
;;
|
||||
wiki)
|
||||
grep -q "\s*dokuwiki:" "${ORGA_COMPOSE}" 2>/dev/null && echo "${service}"
|
||||
;;
|
||||
wp)
|
||||
grep -q "\s*wordpress:" "${ORGA_COMPOSE}" 2>/dev/null && echo "${service}"
|
||||
;;
|
||||
*)
|
||||
grep -q "\s*${service}:" "${ORGA_COMPOSE}" 2>/dev/null && echo "${service}"
|
||||
esac
|
||||
done
|
||||
done
|
||||
}
|
||||
|
||||
getOrgaWithService() {
|
||||
service="$1"
|
||||
shift
|
||||
case "${service}" in
|
||||
wiki) keyword="dokuwiki" ;;
|
||||
wp) keyword="wordpress" ;;
|
||||
*) keyword="${service}" ;;
|
||||
esac
|
||||
for orga in $*; do
|
||||
[[ "${orga}" = *"-orga" ]] || continue
|
||||
local ORGA_DIR="${KAZ_COMP_DIR}/${orga}"
|
||||
ORGA_COMPOSE="${ORGA_DIR}/docker-compose.yml"
|
||||
[[ -f "${ORGA_COMPOSE}" ]] || continue
|
||||
if [ "${service}" = "paheko" ]; then
|
||||
[ -f "${ORGA_DIR}/usePaheko" ] && echo "${orga}"
|
||||
else
|
||||
grep -q "\s*${keyword}:" "${ORGA_COMPOSE}" 2>/dev/null && echo "${orga}"
|
||||
fi
|
||||
done
|
||||
}
|
||||
|
||||
service_enable () {
|
||||
echo $(getServiceInOrga $* | sort -u)
|
||||
}
|
||||
|
||||
service_disable () {
|
||||
echo $(getAvailableServices | filterNotInList $(getServiceInOrga $*))
|
||||
}
|
||||
|
||||
service_status () {
|
||||
# ps per enable
|
||||
echo "*** TODO ***"
|
||||
}
|
||||
|
||||
# ========================================
|
||||
|
||||
KAZ_CMD=""
|
||||
case "$1" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
compose|service)
|
||||
KAZ_CMD="$1"
|
||||
shift
|
||||
;;
|
||||
*)
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
|
||||
KAZ_OPT=""
|
||||
case "$1" in
|
||||
available|validate|enable|disable|status)
|
||||
KAZ_OPT="$1"
|
||||
shift
|
||||
;;
|
||||
*)
|
||||
if [ "${KAZ_CMD}" = "service" ] && [[ $1 =~ ^(${SERVICES_CHOICE})$ ]]; then
|
||||
KAZ_OPT="$1"
|
||||
shift
|
||||
getOrgaWithService "${KAZ_OPT}" $(filterAvailableComposes $*)
|
||||
exit
|
||||
fi
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
|
||||
${KAZ_CMD}_${KAZ_OPT} $(filterAvailableComposes $*)
|
12
bin/ldap/ldap_sauve.sh
Executable file
@ -0,0 +1,12 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
|
||||
FILE_LDIF=/home/sauve/ldap.ldif
|
||||
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
docker exec -u 0 -i ${ldapServName} slapcat -F /opt/bitnami/openldap/etc/slapd.d -b ${ldap_root} | gzip >${FILE_LDIF}.gz
|
23
bin/ldap/ldapvi.sh
Executable file
@ -0,0 +1,23 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
LDAP_IP=$(docker inspect -f '{{.NetworkSettings.Networks.ldapNet.IPAddress}}' ldapServ)
|
||||
|
||||
read -p "quel éditeur ? [vi] " EDITOR
|
||||
EDITOR=${EDITOR:-vi}
|
||||
|
||||
# if [ ${EDITOR} = 'emacs' ]; then
|
||||
# echo "ALERTE ALERTE !!! quelqu'un a voulu utiliser emacs :) :) :)"
|
||||
# exit
|
||||
# fi
|
||||
|
||||
EDITOR=${EDITOR:-vi}
|
||||
export EDITOR=${EDITOR}
|
||||
|
||||
ldapvi -h $LDAP_IP -D "cn=${ldap_LDAP_ADMIN_USERNAME},${ldap_root}" -w ${ldap_LDAP_ADMIN_PASSWORD} --discover
|
222
bin/ldap/migrate_to_ldap.sh
Executable file
@ -0,0 +1,222 @@
|
||||
#!/bin/bash
|
||||
|
||||
echo "ATTENTION ! Il ne faut plus utiliser ce script, il est probable qu'il commence à mettre la grouille avec le LDAP qui vit sa vie..."
|
||||
exit 1
|
||||
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
ACCOUNTS=/kaz/dockers/postfix/config/postfix-accounts.cf
|
||||
|
||||
LDAP_IP=$(docker inspect -f '{{.NetworkSettings.Networks.ldapNet.IPAddress}}' ldapServ)
|
||||
URL_GARRADIN="$httpProto://${paheko_API_USER}:${paheko_API_PASSWORD}@kaz-paheko.$(echo $domain)"
|
||||
# docker exec -i nextcloudDB mysql --user=${nextcloud_MYSQL_USER} --password=${nextcloud_MYSQL_PASSWORD} ${nextcloud_MYSQL_DATABASE} <<< "select * from oc_accounts;" > /tmp/oc_accounts
|
||||
|
||||
ERRORS="/tmp/ldap-errors.log"
|
||||
> ${ERRORS}
|
||||
|
||||
mkdir -p /tmp/ldap/
|
||||
|
||||
# curl -s ${URL_GARRADIN}/api/sql -d "SELECT * from membres where emails_rattaches LIKE '%mailrattache%';"
|
||||
|
||||
for line in `cat ${ACCOUNTS}`
|
||||
do
|
||||
mail=$(echo $line | awk -F '|' '{print $1}')
|
||||
user=$(echo $mail | awk -F '@' '{print $1}')
|
||||
domain=$(echo $mail | awk -F '@' '{print $2}')
|
||||
pass=$(echo $line | awk -F '|' '{print $2}' | sed -e "s/SHA512-//")
|
||||
IDENT_KAZ=
|
||||
if [ ${mode} = "prod" ]; then
|
||||
ficheGarradin=$(curl -s ${URL_GARRADIN}/api/sql -d "SELECT * from membres where email='${mail}';")
|
||||
mailDeSecours=$(echo ${ficheGarradin} | jq .results[0].email_secours | sed -e "s/\"//g")
|
||||
quota=$(echo ${ficheGarradin} | jq .results[0].quota_disque | sed -e "s/\"//g")
|
||||
nom=$(echo ${ficheGarradin} | jq .results[0].nom | sed -e "s/\"//g")
|
||||
nom_orga=$(echo ${ficheGarradin} | jq .results[0].nom_orga | sed -e "s/\"//g")
|
||||
else
|
||||
mailDeSecours=${mail}
|
||||
quota=1
|
||||
nom=${mail}
|
||||
nom_orga="null"
|
||||
fi
|
||||
|
||||
if [ "${quota}" = "null" ]; then
|
||||
quota=1
|
||||
fi
|
||||
|
||||
# nextcloudEnabled=MAYBE
|
||||
# IDENT_KAZ=$(grep -i \"${mail}\" /tmp/oc_accounts | cut -f1)
|
||||
#
|
||||
# if [ ! -z "${IDENT_KAZ}" ]; then # ident Kaz trouvé avec le mail Kaz
|
||||
# nextcloudEnabled=TRUE
|
||||
# else # pas trouvé avec le mail Kaz
|
||||
# if [ "${nom_orga}" != "null" ]; then # c'est une orga, pas de NC
|
||||
# IDENT_KAZ="null"
|
||||
# nextcloudEnabled=FALSE
|
||||
# else # pas trouvé avec le mail Kaz, pas une orga, on retente avec le mail de secours
|
||||
# IDENT_KAZ=$(grep -i \"${mailDeSecours}\" /tmp/oc_accounts | cut -f1 | head -n1)
|
||||
# if [ ! -z "${IDENT_KAZ}" ]; then # on a trouvé l'ident kaz chez NC avec le mail de secours
|
||||
# nextcloudEnabled=TRUE
|
||||
# else # pas trouvé avec le mail Kaz, pas une orga, pas trouvé avec le mail de secours
|
||||
# ficheRattache=$(curl -s ${URL_GARRADIN}/api/sql -d "SELECT * from membres where emails_rattaches LIKE '%${mail}%';" | jq ".results | length")
|
||||
# if [ $ficheRattache != "0" ]; then # c'est un mail rattaché, pas de NC c'est normal
|
||||
# IDENT_KAZ="null"
|
||||
# nextcloudEnabled=FALSE
|
||||
# else # pas trouvé, pas une orga, pas mail rattaché donc souci
|
||||
# echo "Pas trouvé l'identifiant Kaz nextcloud pour ${mail} / ${mailDeSecours}, on désactive nextcloud pour ce compte" >> ${ERRORS}
|
||||
# IDENT_KAZ="null"
|
||||
# nextcloudEnabled=FALSE
|
||||
# fi
|
||||
# fi
|
||||
# fi
|
||||
# fi
|
||||
|
||||
|
||||
echo -e "\n\ndn: cn=${mail},ou=users,${ldap_root}\n\
|
||||
changeType: add\n\
|
||||
objectClass: inetOrgPerson\n\
|
||||
sn: ${nom}\n\
|
||||
userPassword: ${pass}\n\
|
||||
\n\n\
|
||||
dn: cn=${mail},ou=users,${ldap_root}\n\
|
||||
changeType: modify\n\
|
||||
replace: objectClass\n\
|
||||
objectClass: inetOrgPerson\n\
|
||||
objectClass: kaznaute\n\
|
||||
objectClass: PostfixBookMailAccount\n\
|
||||
objectClass: nextcloudAccount\n\
|
||||
-\n\
|
||||
replace: sn\n\
|
||||
sn: ${nom}\n\
|
||||
-\n\
|
||||
replace: mail\n\
|
||||
mail: ${mail}\n\
|
||||
-\n\
|
||||
replace: mailEnabled\n\
|
||||
mailEnabled: TRUE\n\
|
||||
-\n\
|
||||
replace: mailGidNumber\n\
|
||||
mailGidNumber: 5000\n\
|
||||
-\n\
|
||||
replace: mailHomeDirectory\n\
|
||||
mailHomeDirectory: /var/mail/${domain}/${user}/\n\
|
||||
-\n\
|
||||
replace: mailQuota\n\
|
||||
mailQuota: ${quota}G\n\
|
||||
-\n\
|
||||
replace: mailStorageDirectory\n\
|
||||
mailStorageDirectory: maildir:/var/mail/${domain}/${user}/\n\
|
||||
-\n\
|
||||
replace: mailUidNumber\n\
|
||||
mailUidNumber: 5000\n\
|
||||
-\n\
|
||||
replace: nextcloudQuota\n\
|
||||
nextcloudQuota: ${quota} GB\n\
|
||||
-\n\
|
||||
replace: mailDeSecours\n\
|
||||
mailDeSecours: ${mailDeSecours}\n\
|
||||
-\n\
|
||||
replace: quota\n\
|
||||
quota: ${quota}\n\
|
||||
-\n\
|
||||
replace: agoraEnabled\n\
|
||||
agoraEnabled: TRUE\n\
|
||||
-\n\
|
||||
replace: mobilizonEnabled\n\
|
||||
mobilizonEnabled: TRUE\n\n" | tee /tmp/ldap/${mail}.ldif | ldapmodify -c -H ldap://${LDAP_IP} -D "cn=${ldap_LDAP_ADMIN_USERNAME},${ldap_root}" -x -w ${ldap_LDAP_ADMIN_PASSWORD}
|
||||
done
|
||||
|
||||
#replace: nextcloudEnabled\n\
|
||||
#nextcloudEnabled: ${nextcloudEnabled}\n\
|
||||
#-\n\
|
||||
#replace: identifiantKaz\n\
|
||||
#identifiantKaz: ${IDENT_KAZ}\n\
|
||||
#-\n\
|
||||
|
||||
OLDIFS=${IFS}
|
||||
IFS=$'\n'
|
||||
# ALIASES est le fichier d'entrée
|
||||
ALIASES="/kaz/dockers/postfix/config/postfix-virtual.cf"
|
||||
# ALIASES_WITHLDAP est le fichier de sortie des forwards qu'on ne met pas dans le ldap
|
||||
ALIASES_WITHLDAP="/kaz/dockers/postfix/config/postfix-virtual-withldap.cf"
|
||||
# On vide le fichier de sortie avant de commencer
|
||||
> ${ALIASES_WITHLDAP}
|
||||
|
||||
|
||||
for line in `cat ${ALIASES}`
|
||||
do
|
||||
echo "Virtual line is $line"
|
||||
if [ `grep -v "," <<< $line` ]
|
||||
then
|
||||
echo "Alias simple"
|
||||
mail=$(echo $line | awk -F '[[:space:]]*' '{print $2}')
|
||||
if [ `grep $mail ${ACCOUNTS}` ]
|
||||
then
|
||||
echo "Alias vers un mail local, go ldap"
|
||||
LIST=""
|
||||
for alias in `grep ${mail} ${ALIASES} | grep -v "," | cut -d' ' -f1`
|
||||
do
|
||||
LIST=${LIST}"mailAlias: $alias\n"
|
||||
done
|
||||
echo -e "dn: cn=${mail},ou=users,${ldap_root}\n\
|
||||
changeType: modify
|
||||
replace: mailAlias\n\
|
||||
$LIST\n\n" | ldapmodify -c -H ldap://${LDAP_IP} -D "cn=${ldap_LDAP_ADMIN_USERNAME},${ldap_root}" -x -w ${ldap_LDAP_ADMIN_PASSWORD}
|
||||
else
|
||||
echo "Alias vers un mail externe, go fichier"
|
||||
echo $line >> ${ALIASES_WITHLDAP}
|
||||
echo " + intégration LDAP"
|
||||
src=$(echo $line | awk -F '[[:space:]]*' '{print $1}')
|
||||
dst=$(echo $line | awk -F '[[:space:]]*' '{print $2}')
|
||||
echo -e "\n\ndn: cn=${src},ou=mailForwardings,${ldap_root}\n\
|
||||
changeType: add\n\
|
||||
objectClass: organizationalRole\n\
|
||||
\n\n\
|
||||
dn: cn=${src},ou=mailForwardings,${ldap_root}\n\
|
||||
changeType: modify\n\
|
||||
replace: objectClass\n\
|
||||
objectClass: organizationalRole\n\
|
||||
objectClass: PostfixBookMailForward\n\
|
||||
-\n\
|
||||
replace: mailAlias\n\
|
||||
mailAlias: ${src}\n\
|
||||
-\n\
|
||||
replace: mail\n\
|
||||
mail: ${dst}\n\n" | ldapmodify -c -H ldap://${LDAP_IP} -D "cn=${ldap_LDAP_ADMIN_USERNAME},${ldap_root}" -x -w ${ldap_LDAP_ADMIN_PASSWORD}
|
||||
fi
|
||||
else
|
||||
echo "Forward vers plusieurs adresses, on met dans le fichier"
|
||||
echo $line >> ${ALIASES_WITHLDAP}
|
||||
echo " + intégration LDAP"
|
||||
src=$(echo $line | awk -F '[[:space:]]*' '{print $1}')
|
||||
dst=$(echo $line | awk -F '[[:space:]]*' '{print $2}')
|
||||
OOLDIFS=${IFS}
|
||||
IFS=","
|
||||
LIST=""
|
||||
for alias in ${dst}
|
||||
do
|
||||
LIST=${LIST}"mail: $alias\n"
|
||||
done
|
||||
IFS=${OOLDIFS}
|
||||
echo -e "\n\ndn: cn=${src},ou=mailForwardings,${ldap_root}\n\
|
||||
changeType: add\n\
|
||||
objectClass: organizationalRole\n\
|
||||
\n\n\
|
||||
dn: cn=${src},ou=mailForwardings,${ldap_root}\n\
|
||||
changeType: modify\n\
|
||||
replace: objectClass\n\
|
||||
objectClass: organizationalRole\n\
|
||||
objectClass: PostfixBookMailForward\n\
|
||||
-\n\
|
||||
replace: mailAlias\n\
|
||||
mailAlias: ${src}\n\
|
||||
-\n\
|
||||
replace: mail\n\
|
||||
${LIST}\n\n" | ldapmodify -c -H ldap://${LDAP_IP} -D "cn=${ldap_LDAP_ADMIN_USERNAME},${ldap_root}" -x -w ${ldap_LDAP_ADMIN_PASSWORD}
|
||||
|
||||
fi
|
||||
done
|
||||
IFS=${OLDIFS}
|
20
bin/ldap/tests/nc_orphans.sh
Executable file
@ -0,0 +1,20 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
LDAP_IP=$(docker inspect -f '{{.NetworkSettings.Networks.ldapNet.IPAddress}}' ldapServ)
|
||||
|
||||
docker exec -i nextcloudDB mysql --user=${nextcloud_MYSQL_USER} --password=${nextcloud_MYSQL_PASSWORD} ${nextcloud_MYSQL_DATABASE} <<< "select uid from oc_users;" > /tmp/nc_users.txt
|
||||
|
||||
OLDIFS=${IFS}
|
||||
IFS=$'\n'
|
||||
for line in `cat /tmp/nc_users.txt`; do
|
||||
result=$(ldapsearch -h $LDAP_IP -D "cn=${ldap_LDAP_ADMIN_USERNAME},${ldap_root}" -w ${ldap_LDAP_ADMIN_PASSWORD} -b $ldap_root -x "(identifiantKaz=${line})" | grep numEntries)
|
||||
echo "${line} ${result}" | grep -v "numEntries: 1" | grep -v "^uid"
|
||||
done
|
||||
IFS=${OLDIFS}
|
BIN
bin/look/feminin/kaz-entier.png
Normal file
After Width: | Height: | Size: 26 KiB |
90
bin/look/feminin/kaz-entier.svg
Normal file
After Width: | Height: | Size: 39 KiB |
BIN
bin/look/feminin/kaz-signature.png
Normal file
After Width: | Height: | Size: 2.4 KiB |
BIN
bin/look/feminin/kaz-tete.png
Normal file
After Width: | Height: | Size: 2.2 KiB |
89
bin/look/feminin/kaz-tete.svg
Normal file
After Width: | Height: | Size: 73 KiB |
BIN
bin/look/feminin/kazdate.png
Normal file
After Width: | Height: | Size: 10 KiB |
BIN
bin/look/feminin/kazmel.png
Normal file
After Width: | Height: | Size: 8.8 KiB |
BIN
bin/look/greve/kaz-entier.png
Normal file
After Width: | Height: | Size: 28 KiB |
100
bin/look/greve/kaz-entier.svg
Normal file
After Width: | Height: | Size: 38 KiB |
BIN
bin/look/greve/kaz-signature.png
Normal file
After Width: | Height: | Size: 2.8 KiB |
BIN
bin/look/greve/kaz-tete.png
Normal file
After Width: | Height: | Size: 3.0 KiB |
98
bin/look/greve/kaz-tete.svg
Normal file
After Width: | Height: | Size: 72 KiB |
BIN
bin/look/greve/kazdate.png
Normal file
After Width: | Height: | Size: 11 KiB |
BIN
bin/look/greve/kazmel.png
Normal file
After Width: | Height: | Size: 9.7 KiB |
BIN
bin/look/kaz/kaz-entier.png
Normal file
After Width: | Height: | Size: 23 KiB |
79
bin/look/kaz/kaz-entier.svg
Normal file
After Width: | Height: | Size: 38 KiB |
BIN
bin/look/kaz/kaz-signature.png
Normal file
After Width: | Height: | Size: 2.3 KiB |
BIN
bin/look/kaz/kaz-tete.png
Normal file
After Width: | Height: | Size: 2.1 KiB |
66
bin/look/kaz/kaz-tete.svg
Normal file
After Width: | Height: | Size: 71 KiB |
BIN
bin/look/kaz/kazdate.png
Normal file
After Width: | Height: | Size: 10 KiB |
BIN
bin/look/kaz/kazmel.png
Normal file
After Width: | Height: | Size: 8.2 KiB |
BIN
bin/look/noel/kaz-entier.png
Normal file
After Width: | Height: | Size: 43 KiB |
87
bin/look/noel/kaz-entier.svg
Normal file
After Width: | Height: | Size: 83 KiB |
BIN
bin/look/noel/kaz-signature.png
Normal file
After Width: | Height: | Size: 2.6 KiB |
BIN
bin/look/noel/kaz-tete.png
Normal file
After Width: | Height: | Size: 11 KiB |
75
bin/look/noel/kaz-tete.svg
Normal file
After Width: | Height: | Size: 116 KiB |
BIN
bin/look/noel/kazdate.png
Normal file
After Width: | Height: | Size: 12 KiB |
BIN
bin/look/noel/kazmel.png
Normal file
After Width: | Height: | Size: 9.6 KiB |
180
bin/manageAgora.sh
Executable file
@ -0,0 +1,180 @@
|
||||
#!/bin/bash
|
||||
# Script de manipulation d'un mattermost'
|
||||
# init /versions / restart ...
|
||||
#
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
#GLOBAL VARS
|
||||
PRG=$(basename $0)
|
||||
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
AVAILABLE_ORGAS=${availableOrga[*]//-orga/}
|
||||
|
||||
QUIET="1"
|
||||
ONNAS=
|
||||
|
||||
AGORACOMMUN="OUI_PAR_DEFAUT"
|
||||
DockerServName=${mattermostServName}
|
||||
|
||||
declare -A Posts
|
||||
|
||||
usage() {
|
||||
echo "${PRG} [OPTION] [COMMANDES] [ORGA]
|
||||
Manipulation d'un mattermost
|
||||
|
||||
OPTIONS
|
||||
-h|--help Cette aide :-)
|
||||
-n|--simu SIMULATION
|
||||
-q|--quiet On ne parle pas (utile avec le -n pour avoir que les commandes)
|
||||
--nas L'orga se trouve sur le NAS !
|
||||
|
||||
COMMANDES (on peut en mettre plusieurs dans l'ordre souhaité)
|
||||
-I|--install L'initialisation du mattermost
|
||||
-v|--version Donne la version du mattermost et signale les MàJ
|
||||
|
||||
|
||||
-mmctl \"command\" Envoie une commande via mmctl ** SPECIFIQUES **
|
||||
-p|--post \"team\" \"message\" Poste un message dans une team agora ** AGORA **
|
||||
|
||||
ORGA parmi : ${AVAILABLE_ORGAS}
|
||||
ou vide si mattermost commun
|
||||
"
|
||||
}
|
||||
|
||||
|
||||
Init(){
|
||||
NOM=$ORGA
|
||||
if [ -n "$AGORACOMMUN" ] ; then NOM="KAZ" ; fi
|
||||
CONF_FILE="${DOCK_VOL}/orga_${ORGA}-matterConfig/_data/config.json"
|
||||
if [ -n "${AGORACOMMUN}" ]; then
|
||||
CONF_FILE="${DOCK_VOL}/mattermost_matterConfig/_data/config.json"
|
||||
elif [ -n "${ONNAS}" ]; then
|
||||
CONF_FILE="${NAS_VOL}/orga_${ORGA}-matterConfig/_data/config.json"
|
||||
fi
|
||||
|
||||
|
||||
${SIMU} sed -i \
|
||||
-e 's|"SiteURL": ".*"|"SiteURL": "'${MATTER_URL}'"|g' \
|
||||
-e 's|"ListenAddress": ".*"|"ListenAddress": ":'${matterPort}'"|g' \
|
||||
-e 's|"WebsocketURL": ".*"|"WebsocketURL": "wss://'${MATTER_URI}'"|g' \
|
||||
-e 's|"AllowCorsFrom": ".*"|"AllowCorsFrom": "'${domain}' '${MATTER_URI}':443 '${MATTER_URI}'"|g' \
|
||||
-e 's|"ConsoleLevel": ".*"|"ConsoleLevel": "ERROR"|g' \
|
||||
-e 's|"SendEmailNotifications": false|"SendEmailNotifications": true|g' \
|
||||
-e 's|"FeedbackEmail": ".*"|"FeedbackEmail": "admin@'${domain}'"|g' \
|
||||
-e 's|"FeedbackOrganization": ".*"|"FeedbackOrganization": "Cochez la KAZ du libre !"|g' \
|
||||
-e 's|"ReplyToAddress": ".*"|"ReplyToAddress": "admin@'${domain}'"|g' \
|
||||
-e 's|"SMTPServer": ".*"|"SMTPServer": "mail.'${domain}'"|g' \
|
||||
-e 's|"SMTPPort": ".*"|"SMTPPort": "25"|g' \
|
||||
-e 's|"DefaultServerLocale": ".*"|"DefaultServerLocale": "fr"|g' \
|
||||
-e 's|"DefaultClientLocale": ".*"|"DefaultClientLocale": "fr"|g' \
|
||||
-e 's|"AvailableLocales": ".*"|"AvailableLocales": "fr"|g' \
|
||||
${CONF_FILE}
|
||||
|
||||
# on redémarre pour prendre en compte (changement de port)
|
||||
${SIMU} docker restart "${DockerServName}"
|
||||
[ $? -ne 0 ] && printKazError "$DockerServName est down : impossible de terminer l'install" && return 1 >& $QUIET
|
||||
|
||||
${SIMU} waitUrl "$MATTER_URL" 300
|
||||
[ $? -ne 0 ] && printKazError "$DockerServName ne parvient pas à démarrer correctement : impossible de terminer l'install" && return 1 >& $QUIET
|
||||
|
||||
# creation compte admin
|
||||
${SIMU} curl -i -d "{\"email\":\"${mattermost_MM_ADMIN_EMAIL}\",\"username\":\"${mattermost_user}\",\"password\":\"${mattermost_pass}\",\"allow_marketing\":true}" "${MATTER_URL}/api/v4/users"
|
||||
|
||||
MM_TOKEN=$(_getMMToken ${MATTER_URL})
|
||||
|
||||
#on crée la team
|
||||
${SIMU} curl -i -H "Authorization: Bearer ${MM_TOKEN}" -d "{\"display_name\":\"${NOM}\",\"name\":\"${NOM,,}\",\"type\":\"O\"}" "${MATTER_URL}/api/v4/teams"
|
||||
}
|
||||
|
||||
Version(){
|
||||
VERSION=$(docker exec "$DockerServName" bin/mmctl version | grep -i version:)
|
||||
echo "Version $DockerServName : ${GREEN}${VERSION}${NC}"
|
||||
}
|
||||
|
||||
_getMMToken(){
|
||||
#$1 MATTER_URL
|
||||
${SIMU} curl -i -s -d "{\"login_id\":\"${mattermost_user}\",\"password\":\"${mattermost_pass}\"}" "${1}/api/v4/users/login" | grep 'token' | sed 's/token:\s*\(.*\)\s*/\1/' | tr -d '\r'
|
||||
}
|
||||
|
||||
PostMessage(){
|
||||
printKazMsg "Envoi à $TEAM : $MESSAGE" >& $QUIET
|
||||
|
||||
${SIMU} docker exec -ti "${DockerServName}" bin/mmctl auth login "${MATTER_URL}" --name local-server --username ${mattermost_user} --password ${mattermost_pass}
|
||||
${SIMU} docker exec -ti "${DockerServName}" bin/mmctl post create "${TEAM}" --message "${MESSAGE}"
|
||||
}
|
||||
|
||||
MmctlCommand(){
|
||||
# $1 command
|
||||
${SIMU} docker exec -u 33 "$DockerServName" bin/mmctl $1
|
||||
}
|
||||
|
||||
|
||||
########## Main #################
|
||||
for ARG in "$@"; do
|
||||
if [ -n "${GETMMCTLCOMAND}" ]; then # après un -mmctl
|
||||
MMCTLCOMAND="${ARG}"
|
||||
GETMMCTLCOMAND=
|
||||
elif [ -n "${GETTEAM}" ]; then # après un --post
|
||||
GETMESSAGE="now"
|
||||
GETTEAM=""
|
||||
TEAM="${ARG}"
|
||||
elif [ -n "${GETMESSAGE}" ]; then # après un --post "team:channel"
|
||||
if [[ $TEAM == "-*" && ${#TEAM} -le 5 ]]; then echo "J'envoie mon message à \"${TEAM}\" ?? Arf, ça me plait pas j'ai l'impression que tu t'es planté sur la commande."; usage ; exit 1 ; fi
|
||||
if [[ $ARG == "-*" && ${#ARG} -le 5 ]]; then echo "J'envoie le message \"${ARG}\" ?? Arf, ça me plait pas j'ai l'impression que tu t'es planté sur la commande."; usage ; exit 1 ; fi
|
||||
if [[ ! $TEAM =~ .*:.+ ]]; then echo "Il faut mettre un destinataire sous la forme team:channel. Recommence !"; usage ; exit 1 ; fi
|
||||
MESSAGE="$ARG"
|
||||
GETMESSAGE=""
|
||||
else
|
||||
case "${ARG}" in
|
||||
'-h' | '--help' )
|
||||
usage && exit ;;
|
||||
'-n' | '--simu')
|
||||
SIMU="echo" ;;
|
||||
'-q' )
|
||||
QUIET="/dev/null" ;;
|
||||
'--nas' | '-nas' )
|
||||
ONNAS="SURNAS" ;;
|
||||
'-v' | '--version')
|
||||
COMMANDS="$(echo "${COMMANDS} VERSION" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-I' | '--install' )
|
||||
COMMANDS="$(echo "${COMMANDS} INIT" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'--mmctl' | '-mmctl' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-AGORA-MMCTL" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETMMCTLCOMAND="now" ;;
|
||||
'-p' | '--post' )
|
||||
COMMANDS="$(echo "${COMMANDS} POST-AGORA" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETTEAM="now" ;;
|
||||
'-*' ) # ignore
|
||||
;;
|
||||
*)
|
||||
ORGA="${ARG%-orga}"
|
||||
DockerServName="${ORGA}-${mattermostServName}"
|
||||
AGORACOMMUN=
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
done
|
||||
|
||||
if [ -z "${COMMANDS}" ]; then usage && exit ; fi
|
||||
|
||||
MATTER_URI="${ORGA}-${matterHost}.${domain}"
|
||||
if [ -n "$AGORACOMMUN" ]; then MATTER_URI="${matterHost}.${domain}" ; fi
|
||||
MATTER_URL="${httpProto}://${MATTER_URI}"
|
||||
|
||||
for COMMAND in ${COMMANDS}; do
|
||||
case "${COMMAND}" in
|
||||
'VERSION' )
|
||||
Version && exit ;;
|
||||
'INIT' )
|
||||
Init ;;
|
||||
'RUN-AGORA-MMCTL' )
|
||||
MmctlCommand "$MMCTLCOMAND" ;;
|
||||
'POST-AGORA' )
|
||||
PostMessage ;;
|
||||
esac
|
||||
done
|
117
bin/manageCastopod.sh
Executable file
@ -0,0 +1,117 @@
|
||||
#!/bin/bash
|
||||
# Script de manipulation d'un wordpress'
|
||||
# init /versions / restart ...
|
||||
#
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
#GLOBAL VARS
|
||||
PRG=$(basename $0)
|
||||
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
AVAILABLE_ORGAS=${availableOrga[*]//-orga/}
|
||||
|
||||
QUIET="1"
|
||||
ONNAS=
|
||||
|
||||
|
||||
CASTOPOD_COMMUN="OUI_PAR_DEFAUT"
|
||||
DockerServName=${castopodServName}
|
||||
|
||||
declare -A Posts
|
||||
|
||||
usage() {
|
||||
echo "${PRG} [OPTION] [COMMANDES] [ORGA]
|
||||
Manipulation d'un castopod
|
||||
|
||||
OPTIONS
|
||||
-h|--help Cette aide :-)
|
||||
-n|--simu SIMULATION
|
||||
-q|--quiet On ne parle pas (utile avec le -n pour avoir que les commandes)
|
||||
--nas L'orga se trouve sur le NAS !
|
||||
|
||||
COMMANDES (on peut en mettre plusieurs dans l'ordre souhaité)
|
||||
-I|--install L'initialisation du castopod
|
||||
-v|--version Donne la version du castopod et signale les MàJ
|
||||
|
||||
ORGA parmi : ${AVAILABLE_ORGAS}
|
||||
ou vide si castopod commun
|
||||
"
|
||||
}
|
||||
|
||||
|
||||
|
||||
Init(){
|
||||
|
||||
POD_URL="${httpProto}://${ORGA}-${castopodHost}.${domain}"
|
||||
[ -n "${CASTOPOD_COMMUN}" ] && POD_URL="${httpProto}://${castopodHost}.${domain}"
|
||||
|
||||
if ! [[ "$(docker ps -f name=${DockerServName} | grep -w ${DockerServName})" ]]; then
|
||||
printKazError "Castopod not running... abort"
|
||||
exit
|
||||
fi
|
||||
|
||||
echo "\n *** Premier lancement de Castopod" >& $QUIET
|
||||
|
||||
${SIMU} waitUrl "${POD_URL}"
|
||||
CI_SESSION=$(echo ${headers} | grep "ci_session" | sed "s/.*ci_session=//")
|
||||
|
||||
cookies=$(curl -c - ${POD_URL})
|
||||
CSRF_TOKEN=$(curl --cookie <(echo "$cookies") ${POD_URL}/cp-install | grep "csrf_test_name" | sed "s/.*value=.//" | sed "s/.>//")
|
||||
|
||||
#echo ${CSRF_TOKEN}
|
||||
${SIMU} curl --cookie <(echo "$cookies") -X POST \
|
||||
-d "username=${castopod_ADMIN_USER}" \
|
||||
-d "password=${castopod_ADMIN_PASSWORD}" \
|
||||
-d "email=${castopod_ADMIN_MAIL}" \
|
||||
-d "csrf_test_name=${CSRF_TOKEN}" \
|
||||
"${POD_URL}/cp-install/create-superadmin"
|
||||
|
||||
}
|
||||
|
||||
Version(){
|
||||
VERSION="TODO"
|
||||
echo "Version $DockerServName : ${GREEN}${VERSION}${NC}"
|
||||
}
|
||||
|
||||
|
||||
|
||||
########## Main #################
|
||||
for ARG in "$@"; do
|
||||
case "${ARG}" in
|
||||
'-h' | '--help' )
|
||||
usage && exit ;;
|
||||
'-n' | '--simu')
|
||||
SIMU="echo" ;;
|
||||
'-q' )
|
||||
QUIET="/dev/null" ;;
|
||||
'--nas' | '-nas' )
|
||||
ONNAS="SURNAS" ;;
|
||||
'-v' | '--version')
|
||||
COMMANDS="$(echo "${COMMANDS} VERSION" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-I' | '--install' )
|
||||
COMMANDS="$(echo "${COMMANDS} INIT" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'-*' ) # ignore
|
||||
;;
|
||||
*)
|
||||
ORGA="${ARG%-orga}"
|
||||
DockerServName="${ORGA}-${castopodServName}"
|
||||
CASTOPOD_COMMUN=
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z "${COMMANDS}" ]; then usage && exit ; fi
|
||||
|
||||
for COMMAND in ${COMMANDS}; do
|
||||
case "${COMMAND}" in
|
||||
'VERSION' )
|
||||
Version && exit ;;
|
||||
'INIT' )
|
||||
Init ;;
|
||||
esac
|
||||
done
|
393
bin/manageCloud.sh
Executable file
@ -0,0 +1,393 @@
|
||||
#!/bin/bash
|
||||
# Script de manipulation d'un cloud'
|
||||
# init /versions / restart ...
|
||||
#
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
#GLOBAL VARS
|
||||
PRG=$(basename $0)
|
||||
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
AVAILABLE_ORGAS=${availableOrga[*]//-orga/}
|
||||
|
||||
# CLOUD
|
||||
APPLIS_PAR_DEFAUT="tasks calendar contacts bookmarks mail richdocuments external drawio snappymail ransomware_protection" #rainloop richdocumentscode
|
||||
QUIET="1"
|
||||
ONNAS=
|
||||
|
||||
CLOUDCOMMUN="OUI_PAR_DEFAUT"
|
||||
DockerServName=${nextcloudServName}
|
||||
|
||||
usage() {
|
||||
echo "${PRG} [OPTION] [COMMANDES] [ORGA]
|
||||
Manipulation d'un cloud
|
||||
|
||||
OPTIONS
|
||||
-h|--help Cette aide :-)
|
||||
-n|--simu SIMULATION
|
||||
-q|--quiet On ne parle pas (utile avec le -n pour avoir que les commandes)
|
||||
--nas L'orga se trouve sur le NAS !
|
||||
|
||||
COMMANDES (on peut en mettre plusieurs dans l'ordre souhaité)
|
||||
-I|--install L'initialisation du cloud
|
||||
-v|--version Donne la version du cloud et signale les MàJ
|
||||
|
||||
--optim Lance la procédure Nextcloud pour optimiser les performances ** **
|
||||
-occ \"command\" Envoie une commande via occ ** **
|
||||
-u Mets à jour les applis ** SPECIFIQUES **
|
||||
-i Install des applis ** CLOUD **
|
||||
-a \"app1 app2 ...\" Choix des appli à installer ou mettre à jour (entre guillemets) ** **
|
||||
-U|--upgrade Upgrade des clouds ** **
|
||||
-O|--officeURL MAJ le office de ce nextcloud ** **
|
||||
|
||||
ORGA parmi : ${AVAILABLE_ORGAS}
|
||||
ou vide si cloud commun
|
||||
"
|
||||
}
|
||||
|
||||
|
||||
##################################
|
||||
############### CLOUD ############
|
||||
##################################
|
||||
|
||||
Init(){
|
||||
NOM=$ORGA
|
||||
[ -n "${CLOUDCOMMUN}" ] && NOM="commun"
|
||||
if [ -z "${LISTE_APPS}" ]; then
|
||||
printKazMsg "Aucune appli n'est précisée, j'installerais les applis par défaut : ${APPLIS_PAR_DEFAUT}" >& $QUIET
|
||||
LISTE_APPS="${APPLIS_PAR_DEFAUT}"
|
||||
fi
|
||||
|
||||
checkDockerRunning "$DockerServName" "$NOM"
|
||||
[ $? -ne 0 ] && echo "${CYAN}\n $DockerServName est down : impossible de terminer l'install${NC}" && return 1 >& $QUIET
|
||||
|
||||
CONF_FILE="${DOCK_VOL}/orga_${ORGA}-cloudConfig/_data/config.php"
|
||||
CLOUD_URL="https://${ORGA}-${cloudHost}.${domain}"
|
||||
if [ -n "$CLOUDCOMMUN" ]; then
|
||||
CONF_FILE="${DOCK_VOL}/cloud-cloudConfig/_data/config.php"
|
||||
CLOUD_URL="https://${cloudHost}.${domain}"
|
||||
elif [ -n "${ONNAS}" ]; then
|
||||
CONF_FILE="${NAS_VOL}/orga_${ORGA}-cloudConfig/_data/config.php"
|
||||
fi
|
||||
|
||||
firstInstall "$CLOUD_URL" "$CONF_FILE" " NextCloud de $NOM"
|
||||
updatePhpConf "$CONF_FILE"
|
||||
InstallApplis
|
||||
echo "${CYAN} *** Paramétrage richdocuments pour $ORGA${NC}" >& $QUIET
|
||||
setOfficeUrl
|
||||
occCommand "config:app:set --value 10.0.0.0/8,172.16.0.0/12,192.168.0.0/16 richdocuments wopi_allowlist"
|
||||
occCommand "config:system:set overwrite.cli.url --value=$CLOUD_URL"
|
||||
occCommand "config:system:set disable_certificate_verification --value=true"
|
||||
|
||||
if [ -n "$CLOUDCOMMUN" ]; then initLdap "$NOM" ; fi
|
||||
}
|
||||
|
||||
Version(){
|
||||
VERSION=$(docker exec -u 33 ${DockerServName} /var/www/html/occ status | grep -i version:)
|
||||
VERSION_UPDATE=$(docker exec -u 33 ${DockerServName} /var/www/html/occ update:check | grep -i "available\." | cut -c 1-17)
|
||||
versionSTR="Version ${DockerServName} : ${GREEN}${VERSION}${NC}"
|
||||
[ -n "${VERSION_UPDATE}" ] && versionSTR="$versionSTR -- Disponible : ${RED} ${VERSION_UPDATE} ${NC}"
|
||||
echo "$versionSTR"
|
||||
}
|
||||
|
||||
firstInstall(){
|
||||
# $1 CLOUD_URL
|
||||
# $2 phpConfFile
|
||||
# $3 orga
|
||||
if ! grep -q "'installed' => true," "$2" 2> /dev/null; then
|
||||
printKazMsg "\n *** Premier lancement de $3" >& $QUIET
|
||||
|
||||
${SIMU} waitUrl "$1"
|
||||
|
||||
${SIMU} curl -X POST \
|
||||
-d "install=true" \
|
||||
-d "adminlogin=${nextcloud_NEXTCLOUD_ADMIN_USER}" \
|
||||
-d "adminpass=${nextcloud_NEXTCLOUD_ADMIN_PASSWORD}" \
|
||||
-d "directory=/var/www/html/data" \
|
||||
-d "dbtype=mysql" \
|
||||
-d "dbuser=${nextcloud_MYSQL_USER}" \
|
||||
-d "dbpass=${nextcloud_MYSQL_PASSWORD}" \
|
||||
-d "dbname=${nextcloud_MYSQL_DATABASE}" \
|
||||
-d "dbhost=${nextcloud_MYSQL_HOST}" \
|
||||
-d "install-recommended-apps=true" \
|
||||
"$1"
|
||||
fi
|
||||
}
|
||||
|
||||
setOfficeUrl(){
|
||||
OFFICE_URL="https://${officeHost}.${domain}"
|
||||
if [ ! "${site}" = "prod1" ]; then
|
||||
OFFICE_URL="https://${site}-${officeHost}.${domain}"
|
||||
fi
|
||||
occCommand "config:app:set --value $OFFICE_URL richdocuments public_wopi_url"
|
||||
occCommand "config:app:set --value $OFFICE_URL richdocuments wopi_url"
|
||||
occCommand "config:app:set --value $OFFICE_URL richdocuments disable_certificate_verification"
|
||||
}
|
||||
|
||||
initLdap(){
|
||||
# $1 Nom du cloud
|
||||
echo "${CYAN} *** Installation LDAP pour $1${NC}" >& $QUIET
|
||||
occCommand "app:enable user_ldap" "${DockerServName}"
|
||||
occCommand "ldap:delete-config s01" "${DockerServName}"
|
||||
occCommand "ldap:create-empty-config" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapAgentName cn=cloud,ou=applications,${ldap_root}" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapAgentPassword ${ldap_LDAP_CLOUD_PASSWORD}" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapBase ${ldap_root}" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapBaseGroups ${ldap_root}" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapBaseUsers ou=users,${ldap_root}" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapExpertUsernameAttr identifiantKaz" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapHost ${ldapServName}" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapPort 389" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapTLS 0" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapLoginFilter \"(&(objectclass=nextcloudAccount)(|(cn=%uid)(identifiantKaz=%uid)))\"" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapQuotaAttribute nextcloudQuota" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapUserFilter \"(&(objectclass=nextcloudAccount)(nextcloudEnabled=TRUE))\"" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapUserFilterObjectclass nextcloudAccount" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapEmailAttribute mail" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapUserDisplayName cn" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapUserFilterMode 1" "${DockerServName}"
|
||||
occCommand "ldap:set-config s01 ldapConfigurationActive 1" "${DockerServName}"
|
||||
|
||||
# Dans le mariadb, pour permettre au ldap de reprendre la main : delete from oc_users where uid<>'admin';
|
||||
# docker exec -i nextcloudDB mysql --user=<user> --password=<password> <db> <<< "delete from oc_users where uid<>'admin';"
|
||||
|
||||
# Doc : https://help.nextcloud.com/t/migration-to-ldap-keeping-users-and-data/13205
|
||||
|
||||
# Exemple de table/clés :
|
||||
# +-------------------------------+----------------------------------------------------------+
|
||||
# | Configuration | s01 |
|
||||
# +-------------------------------+----------------------------------------------------------+
|
||||
# | hasMemberOfFilterSupport | 0 |
|
||||
# | homeFolderNamingRule | |
|
||||
# | lastJpegPhotoLookup | 0 |
|
||||
# | ldapAgentName | cn=cloud,ou=applications,dc=kaz,dc=sns |
|
||||
# | ldapAgentPassword | *** |
|
||||
# | ldapAttributesForGroupSearch | |
|
||||
# | ldapAttributesForUserSearch | |
|
||||
# | ldapBackgroundHost | |
|
||||
# | ldapBackgroundPort | |
|
||||
# | ldapBackupHost | |
|
||||
# | ldapBackupPort | |
|
||||
# | ldapBase | ou=users,dc=kaz,dc=sns |
|
||||
# | ldapBaseGroups | ou=users,dc=kaz,dc=sns |
|
||||
# | ldapBaseUsers | ou=users,dc=kaz,dc=sns |
|
||||
# | ldapCacheTTL | 600 |
|
||||
# | ldapConfigurationActive | 1 |
|
||||
# | ldapConnectionTimeout | 15 |
|
||||
# | ldapDefaultPPolicyDN | |
|
||||
# | ldapDynamicGroupMemberURL | |
|
||||
# | ldapEmailAttribute | mail |
|
||||
# | ldapExperiencedAdmin | 0 |
|
||||
# | ldapExpertUUIDGroupAttr | |
|
||||
# | ldapExpertUUIDUserAttr | |
|
||||
# | ldapExpertUsernameAttr | uid |
|
||||
# | ldapExtStorageHomeAttribute | |
|
||||
# | ldapGidNumber | gidNumber |
|
||||
# | ldapGroupDisplayName | cn |
|
||||
# | ldapGroupFilter | |
|
||||
# | ldapGroupFilterGroups | |
|
||||
# | ldapGroupFilterMode | 0 |
|
||||
# | ldapGroupFilterObjectclass | |
|
||||
# | ldapGroupMemberAssocAttr | |
|
||||
# | ldapHost | ldap |
|
||||
# | ldapIgnoreNamingRules | |
|
||||
# | ldapLoginFilter | (&(|(objectclass=nextcloudAccount))(cn=%uid)) |
|
||||
# | ldapLoginFilterAttributes | |
|
||||
# | ldapLoginFilterEmail | 0 |
|
||||
# | ldapLoginFilterMode | 0 |
|
||||
# | ldapLoginFilterUsername | 1 |
|
||||
# | ldapMatchingRuleInChainState | unknown |
|
||||
# | ldapNestedGroups | 0 |
|
||||
# | ldapOverrideMainServer | |
|
||||
# | ldapPagingSize | 500 |
|
||||
# | ldapPort | 389 |
|
||||
# | ldapQuotaAttribute | nextcloudQuota |
|
||||
# | ldapQuotaDefault | |
|
||||
# | ldapTLS | 0 |
|
||||
# | ldapUserAvatarRule | default |
|
||||
# | ldapUserDisplayName | cn |
|
||||
# | ldapUserDisplayName2 | |
|
||||
# | ldapUserFilter | (&(objectclass=nextcloudAccount)(nextcloudEnabled=TRUE)) |
|
||||
# | ldapUserFilterGroups | |
|
||||
# | ldapUserFilterMode | 1 |
|
||||
# | ldapUserFilterObjectclass | nextcloudAccount |
|
||||
# | ldapUuidGroupAttribute | auto |
|
||||
# | ldapUuidUserAttribute | auto |
|
||||
# | turnOffCertCheck | 0 |
|
||||
# | turnOnPasswordChange | 0 |
|
||||
# | useMemberOfToDetectMembership | 1 |
|
||||
# +-------------------------------+----------------------------------------------------------+
|
||||
}
|
||||
|
||||
updatePhpConf(){
|
||||
# $1 php_conf_file
|
||||
if [ $# -ne 1 ]; then
|
||||
echo "${RED}#Je ne sais pas ou écrire la conf php !${NC}"
|
||||
return 1
|
||||
fi
|
||||
|
||||
echo "${CYAN} *** Maj de la conf $1${NC}" >& $QUIET
|
||||
PHPCONF="$1"
|
||||
_addVarAfterInConf "default_language" " 'default_language' => 'fr'," "CONFIG = array (" "${PHPCONF}"
|
||||
_addVarAfterInConf "theme" " 'theme' => ''," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "default_phone_region" " 'default_phone_region' => 'FR'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "loglevel" " 'loglevel' => 2," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "maintenance" " 'maintenance' => false," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "app_install_overwrite" " 'app_install_overwrite' => \n array (\n 0 => 'documents',\n )," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "overwriteprotocol" " 'overwriteprotocol' => 'https'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "mail_domain" " 'mail_domain' => '${domain}'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "mail_from_address" " 'mail_from_address' => 'admin'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "mail_smtpport" " 'mail_smtpport' => '25'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "mail_sendmailmode" " 'mail_sendmailmode' => 'smtp'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "mail_smtphost" " 'mail_smtphost' => '${smtpHost}.${domain}'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "mail_smtpmode" " 'mail_smtpmode' => 'smtp'," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "enable_previews" " 'enable_previews' => true," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "trashbin_retention_obligation" " 'trashbin_retention_obligation' => '30, auto'," "'installed' => true," "${PHPCONF}"
|
||||
#pour supprimer le message "obtenir un compte gratuit" dans le footer
|
||||
_addVarAfterInConf "simpleSignUpLink.shown" " 'simpleSignUpLink.shown' => false," "'installed' => true," "${PHPCONF}"
|
||||
_addVarAfterInConf "trusted_proxies" " 'trusted_proxies' => array( 0 => '10.0.0.0/8', 1 => '172.16.0.0/12', 2 => '192.168.0.0/16' )," "'installed' => true," "${PHPCONF}"
|
||||
}
|
||||
|
||||
UpgradeClouds() {
|
||||
echo "${NC}--------------------------------------------------------" >& $QUIET
|
||||
echo "UPGRADE des cloud" >& $QUIET
|
||||
echo "--------------------------------------------------------" >& $QUIET
|
||||
occCommand "upgrade"
|
||||
}
|
||||
|
||||
OptimiseClouds() {
|
||||
occCommand "db:add-missing-indices" "db:convert-filecache-bigint --no-interaction"
|
||||
}
|
||||
|
||||
|
||||
UpdateApplis() {
|
||||
printKazMsg "UPDATE DES APPLIS du cloud ${DockerServName} : ${LISTE_APPS}" >& $QUIET
|
||||
if [ -z "${LISTE_APPS}" ]; then
|
||||
occCommand "app:update --all"
|
||||
return
|
||||
fi
|
||||
echo "Mise à jour de ${LISTE_APPS}" >& $QUIET
|
||||
for app in ${LISTE_APPS}
|
||||
do
|
||||
occCommand "app:update ${app}"
|
||||
done
|
||||
}
|
||||
|
||||
InstallApplis(){
|
||||
if [ -z "${LISTE_APPS}" ]; then
|
||||
printKazMsg "Aucune appli n'est précisée, j'installe les applis par défaut : ${APPLIS_PAR_DEFAUT}" >& $QUIET
|
||||
LISTE_APPS="${APPLIS_PAR_DEFAUT}"
|
||||
fi
|
||||
apps=$LISTE_APPS
|
||||
|
||||
if ! [[ "$(docker ps -f name=${DockerServName} | grep -w ${DockerServName})" ]]; then
|
||||
printKazError "${RED}# ${DockerServName} not running... impossible d'installer les applis${NC}" >& $QUIET
|
||||
return 1
|
||||
fi
|
||||
LIST_ALL=$(docker exec -ti -u 33 "${DockerServName}" /var/www/html/occ app:list |
|
||||
awk 'BEGIN {cp=0}
|
||||
/Enabled:/ {cp=1 ; next};
|
||||
/Disabled:/ {cp=0; next};
|
||||
{if (cp) print $0};')
|
||||
|
||||
for app in $apps
|
||||
do
|
||||
grep -wq "${app}" <<<"${LIST_ALL}" 2>/dev/null && echo "${app} dejà installée" >& $QUIET && continue
|
||||
echo " install ${app}" >& $QUIET
|
||||
occCommand "app:install ${app}"
|
||||
done
|
||||
}
|
||||
|
||||
occCommand(){
|
||||
# $1 Command
|
||||
${SIMU} docker exec -u 33 $DockerServName /var/www/html/occ $1
|
||||
}
|
||||
|
||||
_addVarAfterInConf(){
|
||||
# $1 key
|
||||
# $2 val
|
||||
# $3 where
|
||||
# $4 confFile
|
||||
if ! grep -q "$1" "${4}" ; then
|
||||
echo -n " ${CYAN}${BOLD}$1${NC}" >& $QUIET
|
||||
${SIMU} sed -i -e "/$3/a\ $2" "$4"
|
||||
fi
|
||||
}
|
||||
|
||||
########## Main #################
|
||||
for ARG in "$@"; do
|
||||
if [ -n "${GETOCCCOMAND}" ]; then # après un -occ
|
||||
OCCCOMAND="${ARG}"
|
||||
GETOCCCOMAND=
|
||||
elif [ -n "${GETAPPS}" ]; then # après un -a
|
||||
LISTE_APPS="${LISTE_APPS} ${ARG}"
|
||||
GETAPPS=""
|
||||
else
|
||||
case "${ARG}" in
|
||||
'-h' | '--help' )
|
||||
usage && exit ;;
|
||||
'-n' | '--simu')
|
||||
SIMU="echo" ;;
|
||||
'-q' )
|
||||
QUIET="/dev/null" ;;
|
||||
'--nas' | '-nas' )
|
||||
ONNAS="SURNAS" ;;
|
||||
'-v' | '--version')
|
||||
COMMANDS="$(echo "${COMMANDS} VERSION" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-I' | '--install' )
|
||||
COMMANDS="$(echo "${COMMANDS} INIT" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'-U' | '--upgrade')
|
||||
COMMANDS="$(echo "${COMMANDS} UPGRADE" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-O' | '--officeURL')
|
||||
COMMANDS="$(echo "${COMMANDS} OFFICEURL" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'--optim' | '-optim' )
|
||||
COMMANDS="$(echo "${COMMANDS} OPTIMISE-CLOUD" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-u' )
|
||||
COMMANDS="$(echo "${COMMANDS} UPDATE-CLOUD-APP" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-i' )
|
||||
COMMANDS="$(echo "${COMMANDS} INSTALL-CLOUD-APP" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-a' )
|
||||
GETAPPS="now" ;;
|
||||
'--occ' | '-occ' )
|
||||
COMMANDS="$(echo "${COMMANDS} RUN-CLOUD-OCC" | sed "s/\s/\n/g" | sort | uniq)"
|
||||
GETOCCCOMAND="now" ;;
|
||||
'-*' ) # ignore
|
||||
;;
|
||||
*)
|
||||
ORGA="${ARG%-orga}"
|
||||
DockerServName="${ORGA}-${nextcloudServName}"
|
||||
CLOUDCOMMUN=
|
||||
;;
|
||||
esac
|
||||
fi
|
||||
done
|
||||
|
||||
if [ -z "${COMMANDS}" ]; then
|
||||
usage && exit
|
||||
fi
|
||||
|
||||
for COMMAND in ${COMMANDS}; do
|
||||
case "${COMMAND}" in
|
||||
'VERSION' )
|
||||
Version && exit ;;
|
||||
'OPTIMISE-CLOUD' )
|
||||
OptimiseClouds ;;
|
||||
'UPDATE-CLOUD-APP' )
|
||||
UpdateApplis ;;
|
||||
'UPGRADE' )
|
||||
UpgradeClouds ;;
|
||||
'INIT' )
|
||||
Init ;;
|
||||
'INSTALL-CLOUD-APP' )
|
||||
InstallApplis ;;
|
||||
'OFFICEURL' )
|
||||
setOfficeUrl ;;
|
||||
'RUN-CLOUD-OCC' )
|
||||
occCommand "${OCCCOMAND}";;
|
||||
esac
|
||||
done
|
177
bin/manageWiki.sh
Executable file
@ -0,0 +1,177 @@
|
||||
#!/bin/bash
|
||||
# Script de manipulation d'un dokuwiki'
|
||||
# init /versions / restart ...
|
||||
#
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
#GLOBAL VARS
|
||||
PRG=$(basename $0)
|
||||
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
AVAILABLE_ORGAS=${availableOrga[*]//-orga/}
|
||||
DNLD_DIR="${KAZ_DNLD_DIR}/dokuwiki"
|
||||
|
||||
QUIET="1"
|
||||
ONNAS=
|
||||
|
||||
WIKICOMMUN="OUI_PAR_DEFAUT"
|
||||
DockerServName=${dokuwikiServName}
|
||||
|
||||
declare -A Posts
|
||||
|
||||
usage() {
|
||||
echo "${PRG} [OPTION] [COMMANDES] [ORGA]
|
||||
Manipulation d'un dokuwiki
|
||||
|
||||
OPTIONS
|
||||
-h|--help Cette aide :-)
|
||||
-n|--simu SIMULATION
|
||||
-q|--quiet On ne parle pas (utile avec le -n pour avoir que les commandes)
|
||||
--nas L'orga se trouve sur le NAS !
|
||||
|
||||
COMMANDES (on peut en mettre plusieurs dans l'ordre souhaité)
|
||||
-I|--install L'initialisation du dokuwiki
|
||||
-v|--version Donne la version du dokuwiki et signale les MàJ
|
||||
|
||||
--reload kill lighthttpd
|
||||
|
||||
ORGA parmi : ${AVAILABLE_ORGAS}
|
||||
ou vide si dokuwiki commun
|
||||
"
|
||||
}
|
||||
|
||||
|
||||
|
||||
Init(){
|
||||
|
||||
NOM=$ORGA
|
||||
if [ -n "$WIKICOMMUN" ] ; then NOM="KAZ" ; fi
|
||||
TPL_DIR="${VOL_PREFIX}wikiLibtpl/_data"
|
||||
PLG_DIR="${VOL_PREFIX}wikiPlugins/_data"
|
||||
CONF_DIR="${VOL_PREFIX}wikiConf/_data"
|
||||
|
||||
# Gael, j'avais ajouté ça mais j'ai pas test alors je laisse comme avant ...
|
||||
# A charge au prochain qui monte un wiki de faire qque chose
|
||||
#WIKI_ROOT="${dokuwiki_WIKI_ROOT}"
|
||||
#WIKI_EMAIL="${dokuwiki_WIKI_EMAIL}"
|
||||
#WIKI_PASS="${dokuwiki_WIKI_PASSWORD}"
|
||||
|
||||
WIKI_ROOT=Kaz
|
||||
WIKI_EMAIL=wiki@kaz.local
|
||||
WIKI_PASS=azerty
|
||||
|
||||
${SIMU} checkDockerRunning "${DockerServName}" "${NOM}" || exit
|
||||
|
||||
if [ ! -f "${CONF_DIR}/local.php" ] ; then
|
||||
echo "\n *** Premier lancement de Dokuwiki ${NOM}" >& $QUIET
|
||||
|
||||
${SIMU} waitUrl "${WIKI_URL}"
|
||||
|
||||
${SIMU} curl -X POST \
|
||||
-A "Mozilla/5.0 (X11; Linux x86_64)" \
|
||||
-d "l=fr" \
|
||||
-d "d[title]=${NOM}" \
|
||||
-d "d[acl]=true" \
|
||||
-d "d[superuser]=${WIKI_ROOT}" \
|
||||
-d "d[fullname]=Admin"\
|
||||
-d "d[email]=${WIKI_EMAIL}" \
|
||||
-d "d[password]=${WIKI_PASS}" \
|
||||
-d "d[confirm]=${WIKI_PASS}" \
|
||||
-d "d[policy]=1" \
|
||||
-d "d[allowreg]=false" \
|
||||
-d "d[license]=0" \
|
||||
-d "d[pop]=false" \
|
||||
-d "submit=Enregistrer" \
|
||||
"${WIKI_URL}/install.php"
|
||||
|
||||
# XXX initialiser admin:<pass>:admin:<mel>:admin,user
|
||||
#${SIMU} rsync -auHAX local.php users.auth.php acl.auth.php "${CONF_DIR}/"
|
||||
|
||||
${SIMU} sed -i "${CONF_DIR}/local.php" \
|
||||
-e "s|\(.*conf\['title'\].*=.*'\).*';|\1${NOM}';|g" \
|
||||
-e "s|\(.*conf\['title'\].*=.*'\).*';|\1${NOM}';|g" \
|
||||
-e "/conf\['template'\]/d" \
|
||||
-e '$a\'"\$conf['template'] = 'docnavwiki';"''
|
||||
|
||||
${SIMU} sed -i -e "s|\(.*conf\['lang'\].*=.*'\)en';|\1fr';|g" "${CONF_DIR}/dokuwiki.php"
|
||||
|
||||
${SIMU} chown -R www-data: "${CONF_DIR}/"
|
||||
fi
|
||||
|
||||
${SIMU} unzipInDir "${DNLD_DIR}/docnavwiki.zip" "${TPL_DIR}/"
|
||||
${SIMU} chown -R www-data: "${TPL_DIR}/"
|
||||
# ckgedit : bof
|
||||
for plugin in captcha smtp todo wrap wrapadd; do
|
||||
${SIMU} unzipInDir "${DNLD_DIR}/${plugin}.zip" "${PLG_DIR}"
|
||||
done
|
||||
${SIMU} chown -R www-data: "${PLG_DIR}/"
|
||||
}
|
||||
|
||||
Version(){
|
||||
# $1 ContainerName
|
||||
VERSION=$(docker exec $1 cat /dokuwiki/VERSION)
|
||||
echo "Version $1 : ${GREEN}${VERSION}${NC}"
|
||||
}
|
||||
|
||||
|
||||
Reload(){
|
||||
# $1 ContainerName
|
||||
if [ -f "${VOL_PREFIX}wikiData/_data/farms/init.sh" ]; then
|
||||
${SIMU} docker exec -ti "${1}" /dokuwiki/data/farms/init.sh
|
||||
${SIMU} pkill -KILL lighttpd
|
||||
fi
|
||||
}
|
||||
|
||||
########## Main #################
|
||||
for ARG in "$@"; do
|
||||
case "${ARG}" in
|
||||
'-h' | '--help' )
|
||||
usage && exit ;;
|
||||
'-n' | '--simu')
|
||||
SIMU="echo" ;;
|
||||
'-q' )
|
||||
QUIET="/dev/null" ;;
|
||||
'--nas' | '-nas' )
|
||||
ONNAS="SURNAS" ;;
|
||||
'-v' | '--version')
|
||||
COMMANDS="$(echo "${COMMANDS} VERSION" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'--reload' )
|
||||
COMMANDS="$(echo "${COMMANDS} RELOAD" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'-I' | '--install' )
|
||||
COMMANDS="$(echo "${COMMANDS} INIT" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'-*' ) # ignore
|
||||
;;
|
||||
*)
|
||||
ORGA="${ARG%-orga}"
|
||||
DockerServName="${ORGA}-${dokuwikiServName}"
|
||||
WIKICOMMUN=
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z "${COMMANDS}" ]; then usage && exit ; fi
|
||||
|
||||
VOL_PREFIX="${DOCK_VOL}/orga_${ORGA}-"
|
||||
WIKI_URL="${httpProto}://${ORGA}-${dokuwikiHost}.${domain}"
|
||||
if [ -n "${WIKICOMMUN}" ]; then
|
||||
VOL_PREFIX="${DOCK_VOL}/dokuwiki_doku"
|
||||
WIKI_URL="${httpProto}://${dokuwikiHost}.${domain}"
|
||||
elif [ -n "${ONNAS}" ]; then
|
||||
VOL_PREFIX="${NAS_VOL}/orga_${ORGA}-"
|
||||
fi
|
||||
|
||||
for COMMAND in ${COMMANDS}; do
|
||||
case "${COMMAND}" in
|
||||
'VERSION' )
|
||||
Version "${DockerServName}" && exit ;;
|
||||
'INIT' )
|
||||
Init "${DockerServName}" ;;
|
||||
'RELOAD' )
|
||||
Reload "${DockerServName}";;
|
||||
esac
|
||||
done
|
130
bin/manageWp.sh
Executable file
@ -0,0 +1,130 @@
|
||||
#!/bin/bash
|
||||
# Script de manipulation d'un wordpress'
|
||||
# init /versions / restart ...
|
||||
#
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
#GLOBAL VARS
|
||||
PRG=$(basename $0)
|
||||
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
AVAILABLE_ORGAS=${availableOrga[*]//-orga/}
|
||||
|
||||
QUIET="1"
|
||||
ONNAS=
|
||||
|
||||
WPCOMMUN="OUI_PAR_DEFAUT"
|
||||
DockerServName=${wordpressServName}
|
||||
|
||||
declare -A Posts
|
||||
|
||||
usage() {
|
||||
echo "${PRG} [OPTION] [COMMANDES] [ORGA]
|
||||
Manipulation d'un wordpress
|
||||
|
||||
OPTIONS
|
||||
-h|--help Cette aide :-)
|
||||
-n|--simu SIMULATION
|
||||
-q|--quiet On ne parle pas (utile avec le -n pour avoir que les commandes)
|
||||
--nas L'orga se trouve sur le NAS !
|
||||
|
||||
COMMANDES (on peut en mettre plusieurs dans l'ordre souhaité)
|
||||
-I|--install L'initialisation du wordpress
|
||||
-v|--version Donne la version du wordpress et signale les MàJ
|
||||
|
||||
ORGA parmi : ${AVAILABLE_ORGAS}
|
||||
ou vide si wordpress commun
|
||||
"
|
||||
}
|
||||
|
||||
|
||||
|
||||
Init(){
|
||||
PHP_CONF="${DOCK_VOL}/orga_${ORGA}-wordpress/_data/wp-config.php"
|
||||
WP_URL="${httpProto}://${ORGA}-${wordpressHost}.${domain}"
|
||||
|
||||
if [ -n "${ONNAS}" ]; then
|
||||
PHP_CONF="${NAS_VOL}/orga_${ORGA}-wordpress/_data/wp-config.php"
|
||||
fi
|
||||
|
||||
if ! [[ "$(docker ps -f name=${DockerServName} | grep -w ${DockerServName})" ]]; then
|
||||
printKazError "Wordpress not running... abort"
|
||||
exit
|
||||
fi
|
||||
|
||||
# XXX trouver un test du genre if ! grep -q "'installed' => true," "${PHP_CONF}" 2> /dev/null; then
|
||||
echo "\n *** Premier lancement de WP" >& $QUIET
|
||||
|
||||
${SIMU} waitUrl "${WP_URL}"
|
||||
|
||||
${SIMU} curl -X POST \
|
||||
-d "user_name=${wp_WORDPRESS_ADMIN_USER}" \
|
||||
-d "admin_password=${wp_WORDPRESS_ADMIN_PASSWORD}" \
|
||||
-d "admin_password2=${wp_WORDPRESS_ADMIN_PASSWORD}" \
|
||||
-d "pw_weak=true" \
|
||||
-d "admin_email=admin@kaz.bzh" \
|
||||
-d "blog_public=0" \
|
||||
-d "language=fr_FR" \
|
||||
"${WP_URL}/wp-admin/install.php?step=2"
|
||||
|
||||
#/* pour forcer les maj autrement qu'en ftp */
|
||||
_addVarBeforeInConf "FS_METHOD" "define('FS_METHOD', 'direct');" "\/\* That's all, stop editing! Happy publishing. \*\/" "$PHP_CONF"
|
||||
}
|
||||
|
||||
Version(){
|
||||
VERSION=$(docker exec $DockerServName cat /var/www/html/wp-includes/version.php | grep "wp_version " | sed -e "s/.*version\s*=\s*[\"\']//" | sed "s/[\"\'].*//")
|
||||
echo "Version $DockerServName : ${GREEN}${VERSION}${NC}"
|
||||
}
|
||||
|
||||
|
||||
_addVarBeforeInConf(){
|
||||
# $1 key
|
||||
# $2 ligne à ajouter avant la ligne
|
||||
# $3 where
|
||||
# $4 fichier de conf php
|
||||
if ! grep -q "$1" "${4}" ; then
|
||||
echo -n " ${CYAN}${BOLD}$1${NC}" >& $QUIET
|
||||
${SIMU} sed -i -e "s/$3/$2\\n$3/" "${4}"
|
||||
fi
|
||||
}
|
||||
|
||||
########## Main #################
|
||||
for ARG in "$@"; do
|
||||
case "${ARG}" in
|
||||
'-h' | '--help' )
|
||||
usage && exit ;;
|
||||
'-n' | '--simu')
|
||||
SIMU="echo" ;;
|
||||
'-q' )
|
||||
QUIET="/dev/null" ;;
|
||||
'--nas' | '-nas' )
|
||||
ONNAS="SURNAS" ;;
|
||||
'-v' | '--version')
|
||||
COMMANDS="$(echo "${COMMANDS} VERSION" | sed "s/\s/\n/g" | sort | uniq)" ;;
|
||||
'-I' | '--install' )
|
||||
COMMANDS="$(echo "${COMMANDS} INIT" | sed "s/\s/\n/g" | sort | uniq)" ;; # le sed sort uniq, c'est pour pas l'avoir en double
|
||||
'-*' ) # ignore
|
||||
;;
|
||||
*)
|
||||
ORGA="${ARG%-orga}"
|
||||
DockerServName="${ORGA}-${wordpressServName}"
|
||||
WPCOMMUN=
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z "${COMMANDS}" ]; then usage && exit ; fi
|
||||
|
||||
for COMMAND in ${COMMANDS}; do
|
||||
case "${COMMAND}" in
|
||||
'VERSION' )
|
||||
Version && exit ;;
|
||||
'INIT' )
|
||||
Init ;;
|
||||
esac
|
||||
done
|
148
bin/migVersProdX.sh
Executable file
@ -0,0 +1,148 @@
|
||||
#!/bin/bash
|
||||
|
||||
#koi: pouvoir migrer une orga (data+dns) depuis PROD1 vers PRODx
|
||||
#kan: 07/12/2023
|
||||
#ki: françois puis fab (un peu)
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
|
||||
NAS_VOL="/mnt/disk-nas1/docker/volumes/"
|
||||
|
||||
#TODO: ce tab doit être construit à partir de la liste des machines dispos et pas en dur
|
||||
tab_sites_destinations_possibles=("kazoulet" "prod2")
|
||||
|
||||
#par défaut, on prend le premier site
|
||||
SITE_DST="${tab_sites_destinations_possibles[0]}"
|
||||
|
||||
declare -a availableOrga
|
||||
availableOrga=($(getList "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
|
||||
export SIMU=""
|
||||
export COPY=""
|
||||
|
||||
usage () {
|
||||
echo "Usage: $0 [-n] [-d host_distant] [-c] [orga]...[orga]"
|
||||
echo " -h : this help"
|
||||
echo " -d host_distant : ${SITE_DST} par défaut"
|
||||
echo " -n : simulation"
|
||||
echo " -c : only copy data but doesn't stop"
|
||||
echo " [orgas] : in ${availableOrga[@]}"
|
||||
echo " example : migVersProdX.sh -d kazoulet -c splann-orga && migVersProdX.sh -d kazoulet splann-orga"
|
||||
exit 1
|
||||
}
|
||||
|
||||
while getopts "hncd:" option; do
|
||||
case ${option} in
|
||||
h)
|
||||
usage
|
||||
exit 0
|
||||
;;
|
||||
n)
|
||||
SIMU="echo"
|
||||
;;
|
||||
c)
|
||||
COPY="true"
|
||||
;;
|
||||
d)
|
||||
SITE_DST=${OPTARG}
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
# site distant autorisé ?
|
||||
if [[ " ${tab_sites_destinations_possibles[*]} " == *" $SITE_DST "* ]]; then
|
||||
true
|
||||
else
|
||||
echo
|
||||
echo "${RED}${BOLD}Sites distants possibles : ${tab_sites_destinations_possibles[@]}${NC}"
|
||||
echo
|
||||
usage
|
||||
exit 0
|
||||
fi
|
||||
|
||||
# Récupérer les orgas dans un tableau
|
||||
shift $((OPTIND-1))
|
||||
Orgas=("$@")
|
||||
|
||||
#ces orgas existent-elles sur PROD1 ?
|
||||
for orga in "${Orgas[@]}"; do
|
||||
if [[ ! " ${availableOrga[@]} " =~ " ${orga} " ]]; then
|
||||
echo
|
||||
echo "Unknown orga: ${RED}${BOLD}${ARG}${orga}${NC}"
|
||||
echo
|
||||
usage
|
||||
exit 0
|
||||
fi
|
||||
done
|
||||
|
||||
echo
|
||||
echo "Site distant: ${GREEN}${BOLD}${SITE_DST}${NC}"
|
||||
echo
|
||||
|
||||
#for orgaLong in ${Orgas}; do
|
||||
# echo ${Orgas}
|
||||
#done
|
||||
#exit
|
||||
|
||||
for orgaLong in ${Orgas}; do
|
||||
orgaCourt="${orgaLong%-orga}"
|
||||
orgaLong="${orgaCourt}-orga"
|
||||
echo "${BLUE}${BOLD}migration de ${orgaCourt}${NC}"
|
||||
|
||||
# if [ -d "${DOCK_VOL_PAHEKO_ORGA}/${orgaCourt}" ]; then
|
||||
# if ! ssh -p 2201 root@${SITE_DST}.${domain} "test -d ${DOCK_VOL_PAHEKO_ORGA}/${orgaCourt}"; then
|
||||
# echo "${RED}${BOLD} ... can't move paheko to ${SITE_DST}${NC}"
|
||||
# echo " intall paheko in ${SITE_DST}.${domain} before!"
|
||||
# continue
|
||||
# fi
|
||||
# fi
|
||||
|
||||
#on créé le répertoire de l'orga pour paheko sur SITE_DST s'il n'existe pas
|
||||
#pratique quand paheko n'est pas encore installé sur PROD1 mais commandé
|
||||
if [ -f "${KAZ_COMP_DIR}/${orgaLong}/usePaheko" ]; then
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} "mkdir -p ${DOCK_VOL_PAHEKO_ORGA}/${orgaCourt} && chown www-data:www-data ${DOCK_VOL_PAHEKO_ORGA}/${orgaCourt}"
|
||||
|
||||
#ensuite, on peut refaire la liste des routes paheko pour traefik
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} "cd ${KAZ_COMP_DIR}/paheko/ && ./docker-compose-gen.sh"
|
||||
fi
|
||||
|
||||
if [ -z "${COPY}" ]; then
|
||||
cd "${KAZ_COMP_DIR}/${orgaLong}"
|
||||
docker-compose logs --tail 100| grep $(date "+ %Y-%m-%d")
|
||||
checkContinue
|
||||
${SIMU} docker-compose down
|
||||
fi
|
||||
|
||||
if [ $(ls -d ${NAS_VOL}/orga_${orgaCourt}-* 2>/dev/null | wc -l) -gt 0 ]; then
|
||||
echo "${BLUE}${BOLD} ... depuis nas${NC}"
|
||||
${SIMU} rsync -aAhHX --info=progress2 --delete ${NAS_VOL}/orga_${orgaCourt}-* -e "ssh -p 2201" root@${SITE_DST}.${domain}:${DOCK_VOL}
|
||||
else
|
||||
echo "${BLUE}${BOLD} ... depuis disque${NC}"
|
||||
${SIMU} rsync -aAhHX --info=progress2 --delete ${DOCK_VOL}/orga_${orgaCourt}-* -e "ssh -p 2201" root@${SITE_DST}.${domain}:${DOCK_VOL}
|
||||
fi
|
||||
|
||||
if [ -z "${COPY}" ]; then
|
||||
echo "${BLUE}${BOLD} ... config${NC}"
|
||||
if [ -d "${DOCK_VOL_PAHEKO_ORGA}/${orgaCourt}" ]; then
|
||||
${SIMU} rsync -aAhHX --info=progress2 --delete "${DOCK_VOL_PAHEKO_ORGA}/${orgaCourt}" -e "ssh -p 2201" root@${SITE_DST}.${domain}:"${DOCK_VOL_PAHEKO_ORGA}/"
|
||||
fi
|
||||
${SIMU} rsync -aAhHX --info=progress2 --delete ${KAZ_COMP_DIR}/${orgaLong} -e "ssh -p 2201" root@${SITE_DST}.${domain}:${KAZ_COMP_DIR}/
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} "grep -q '^${orgaLong}\$' /kaz/config/container-orga.list || echo ${orgaLong} >> /kaz/config/container-orga.list"
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} ${KAZ_COMP_DIR}/${orgaLong}/init-volume.sh
|
||||
|
||||
cd "${KAZ_COMP_DIR}/${orgaLong}"
|
||||
${SIMU} ./orga-rm.sh
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} "${KAZ_COMP_DIR}/${orgaLong}/orga-gen.sh" --create
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} "${KAZ_BIN_DIR}/container.sh" start "${orgaLong}"
|
||||
|
||||
${SIMU} ssh -p 2201 root@${SITE_DST}.${domain} "${KAZ_BIN_DIR}/manageCloud.sh" --officeURL "${orgaCourt}"
|
||||
fi
|
||||
|
||||
|
||||
done
|
44
bin/migration.sh
Executable file
@ -0,0 +1,44 @@
|
||||
#!/bin/bash
|
||||
|
||||
CV1=/kaz-old/bin/container.sh
|
||||
DV1=/kaz-old/dockers
|
||||
EV1=/kaz-old/config
|
||||
SV1=/kaz-old/secret
|
||||
|
||||
BV2=/kaz/bin
|
||||
DV2=/kaz/dockers
|
||||
EV2=/kaz/config
|
||||
SV2=/kaz/secret
|
||||
OV2=/kaz/config/orgaTmpl/orga-gen.sh
|
||||
|
||||
[ -x "${CV1}" ] || exit
|
||||
[ -d "${BV2}" ] || exit
|
||||
|
||||
SIMU="echo SIMU"
|
||||
|
||||
${SIMU} "${CV1}" stop orga
|
||||
${SIMU} "${CV1}" stop
|
||||
|
||||
${SIMU} rsync "${EV1}/dockers.env" "${EV2}/"
|
||||
${SIMU} rsync "${SV1}/SetAllPass.sh" "${SV2}/"
|
||||
${SIMU} "${BV2}/updateDockerPassword.sh"
|
||||
|
||||
# XXX ? rsync /kaz/secret/allow_admin_ip /kaz-git/secret/allow_admin_ip
|
||||
|
||||
${SIMU} "${BV2}/container.sh" start cloud dokuwiki ethercalc etherpad framadate paheko gitea jirafeau mattermost postfix proxy roundcube web
|
||||
|
||||
${SIMU} rsync -aAHXh --info=progress2 "${DV1}/web/html/" "/var/lib/docker/volumes/web_html/_data/"
|
||||
${SIMU} chown -R www-data: "/var/lib/docker/volumes/web_html/_data/"
|
||||
|
||||
${SIMU} cd "${DV1}"
|
||||
cd "${DV1}"
|
||||
for ORGA_DIR in *-orga; do
|
||||
services=$(echo $([ -x "${ORGA_DIR}/tmpl-gen.sh" ] && "${ORGA_DIR}/tmpl-gen.sh" -l))
|
||||
if [ -n "${services}" ]; then
|
||||
ORGA="${ORGA_DIR%-orga}"
|
||||
|
||||
echo " * ${ORGA}: ${services}"
|
||||
${SIMU} "${OV2}" "${ORGA}" $(for s in ${services}; do echo "+${s}"; done)
|
||||
fi
|
||||
done
|
||||
|
172
bin/mvOrga2Nas.sh
Executable file
@ -0,0 +1,172 @@
|
||||
#!/bin/bash
|
||||
|
||||
# déplace des orga de
|
||||
# /var/lib/docker/volumes/
|
||||
# vers
|
||||
# /mnt/disk-nas1/docker/volumes/
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
DOCK_NAS="/mnt/disk-nas1/docker/volumes"
|
||||
|
||||
DOCK_SRC="${DOCK_VOL}"
|
||||
DOCK_DST="${DOCK_NAS}"
|
||||
|
||||
export PRG="$0"
|
||||
cd $(dirname $0)
|
||||
|
||||
. "${DOCKERS_ENV}"
|
||||
|
||||
declare -a availableOrga
|
||||
availableOrga=($(sed -e "s/\(.*\)[ \t]*#.*$/\1/" -e "s/^[ \t]*\(.*\)-orga$/\1/" -e "/^$/d" "${KAZ_CONF_DIR}/container-orga.list"))
|
||||
|
||||
# no more export in .env
|
||||
export $(set | grep "domain=")
|
||||
|
||||
export SIMU=""
|
||||
export ONLY_SYNC=""
|
||||
export NO_SYNC=""
|
||||
export FORCE=""
|
||||
export REVERSE=""
|
||||
|
||||
usage(){
|
||||
echo "Usage: ${PRG} [orga...]"
|
||||
echo " -h help"
|
||||
echo " -n simulation"
|
||||
echo " -y force"
|
||||
echo " -s phase1 only"
|
||||
echo " -r reverse (${DOCK_NAS} to ${DOCK_VOL})"
|
||||
echo " -ns no pre sync"
|
||||
exit 1
|
||||
}
|
||||
|
||||
for ARG in $@; do
|
||||
case "${ARG}" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
'-n' )
|
||||
shift
|
||||
export SIMU="echo"
|
||||
;;
|
||||
'-y' )
|
||||
shift
|
||||
export FORCE="yes"
|
||||
;;
|
||||
'-s' )
|
||||
shift
|
||||
export ONLY_SYNC="yes"
|
||||
;;
|
||||
'-r' )
|
||||
shift
|
||||
export REVERSE="yes"
|
||||
;;
|
||||
'-ns' )
|
||||
shift
|
||||
export NO_SYNC="yes"
|
||||
;;
|
||||
*)
|
||||
break
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
[[ -z "$1" ]] && usage
|
||||
|
||||
for ARG in $@; do
|
||||
[[ ! " ${availableOrga[*]} " =~ " ${ARG} " ]] && echo "${RED}${ARG}${NC} is not an orga" && usage
|
||||
done
|
||||
|
||||
########################################
|
||||
|
||||
# on copie dans new pour que le changement soit atomique
|
||||
NEW=""
|
||||
if [ -n "${REVERSE}" ]; then
|
||||
DOCK_SRC="${DOCK_NAS}"
|
||||
DOCK_DST="${DOCK_VOL}"
|
||||
NEW="/new"
|
||||
${SIMU} mkdir -p "${DOCK_DST}${NEW}"
|
||||
fi
|
||||
# on garde une copie dans back au cas où
|
||||
BACK="${DOCK_SRC}/old"
|
||||
|
||||
echo " move from ${BLUE}${BOLD}${DOCK_SRC}${NC} to ${BLUE}${BOLD}${DOCK_DST}${NC}"
|
||||
checkContinue
|
||||
|
||||
cd "${DOCK_SRC}"
|
||||
volext=$(ls -d orga* | sed 's%.*-%%' | sort -u)
|
||||
declare -a orgaPhase2
|
||||
|
||||
# Pour que l'interruption de service soit la plus courte possible,
|
||||
# on pré-copie toutes les infos de puis la création de l'orga
|
||||
echo -n "${BLUE}Phase 1: pre sync.${NC} "
|
||||
[[ -z "${FORCE}" ]] && [[ -z "${NO_SYNC}" ]] && checkContinue
|
||||
echo
|
||||
for ARG in $@; do
|
||||
for EXT in ${volext}; do
|
||||
vol="orga_${ARG}-${EXT}"
|
||||
# test le service existe
|
||||
[ -e "${DOCK_SRC}/${vol}" ] || continue
|
||||
# si c'est un lien sur /var/lib c'est déjà fait
|
||||
[ -z "${REVERSE}" ] && [ -L "${DOCK_SRC}/${vol}" ] && echo "${GREEN}${vol}${NC} : done" && continue
|
||||
# si c'est un lien sur le NAS c'est un problème
|
||||
[ -n "${REVERSE}" ] && [ -L "${DOCK_SRC}/${vol}" ] && echo "${GREEN}${vol}${NC} : bug" && continue
|
||||
# si c'est pas un répertoire c'est un problème
|
||||
! [ -d "${DOCK_SRC}/${vol}" ] && echo "${RED}${vol}${NC} : done ?" && continue
|
||||
# si transfert est déjà fait
|
||||
if [ -n "${REVERSE}" ]; then
|
||||
! [ -L "${DOCK_DST}/${vol}" ] && echo "${RED}${vol}${NC} : done" && continue
|
||||
fi
|
||||
echo " - ${YELLOW}${vol}${NC}"
|
||||
[[ -z "${NO_SYNC}" ]] && ${SIMU} rsync -auHAX --info=progress2 "${DOCK_SRC}/${vol}/" "${DOCK_DST}${NEW}/${vol}/"
|
||||
[[ " ${orgaPhase2[@]} " =~ " ${ARG} " ]] || orgaPhase2+=( "${ARG}" )
|
||||
done
|
||||
done
|
||||
|
||||
[ -n "${ONLY_SYNC}" ] && exit 0
|
||||
|
||||
if (( ${#orgaPhase2[@]} == 0 )); then
|
||||
exit 0
|
||||
fi
|
||||
|
||||
echo -n "${BLUE}Phase 2: mv.${NC} "
|
||||
[[ -z "${FORCE}" ]] && checkContinue
|
||||
echo
|
||||
mkdir -p "${BACK}"
|
||||
for ARG in "${orgaPhase2[@]}"; do
|
||||
cd "${KAZ_ROOT}"
|
||||
cd "${KAZ_COMP_DIR}/${ARG}-orga"
|
||||
! [ -e "docker-compose.yml" ] && echo "no docker-compose.yml for ${RED}${ARG}${NC}" && continue
|
||||
${SIMU} docker-compose down
|
||||
# L'arrêt ne durera que le temps de copier les modifications depuis la phase 1.
|
||||
for EXT in ${volext}; do
|
||||
vol="orga_${ARG}-${EXT}"
|
||||
# test le service existe
|
||||
[ -e "${DOCK_SRC}/${vol}" ] || continue
|
||||
# si c'est un lien sur /var/lib c'est déjà fait
|
||||
[ -z "${REVERSE}" ] && [ -L "${DOCK_SRC}/${vol}" ] && echo "${GREEN}${vol}${NC} : done" && continue
|
||||
# si c'est un lien sur le NAS c'est un problème
|
||||
[ -n "${REVERSE}" ] && [ -L "${DOCK_SRC}/${vol}" ] && echo "${GREEN}${vol}${NC} : bug" && continue
|
||||
# si c'est pas un répertoire c'est un problème
|
||||
! [ -d "${DOCK_SRC}/${vol}" ] && echo "${RED}${vol}${NC} : done ?" && continue
|
||||
# si transfert est déjà fait
|
||||
if [ -n "${REVERSE}" ]; then
|
||||
! [ -L "${DOCK_DST}/${vol}" ] && echo "${RED}${vol}${NC} : done" && continue
|
||||
fi
|
||||
echo " - ${YELLOW}${vol}${NC}"
|
||||
${SIMU} rsync -auHAX --info=progress2 --delete "${DOCK_SRC}/${vol}/" "${DOCK_DST}${NEW}/${vol}/" || exit 1
|
||||
${SIMU} mv "${DOCK_SRC}/${vol}" "${BACK}/"
|
||||
if [ -z "${REVERSE}" ]; then
|
||||
# cas de /var/lib vers NAS
|
||||
${SIMU} ln -sf "${DOCK_DST}/${vol}" "${DOCK_SRC}/"
|
||||
else
|
||||
# cas NAS vers /var/lib
|
||||
${SIMU} rm -f "${DOCK_SRC}/${vol}"
|
||||
${SIMU} mv "${DOCK_DST}${NEW}/${vol}" "${DOCK_DST}/"
|
||||
fi
|
||||
done
|
||||
${SIMU} docker-compose up -d
|
||||
[[ -x "reload.sh" ]] && "./reload.sh"
|
||||
echo
|
||||
done
|
83
bin/nettoie
Executable file
@ -0,0 +1,83 @@
|
||||
#!/bin/bash
|
||||
|
||||
POUBELLE="${HOME}/tmp/POUBELLE"
|
||||
mkdir -p "${POUBELLE}"
|
||||
|
||||
usage () {
|
||||
echo `basename "$0"` " [-] [-h] [-help] [-clean] [-wipe] [-n] [directory ...]"
|
||||
echo " remove temporaries files"
|
||||
echo " - Treat the following arguments as filenames \`-\' so that"
|
||||
echo " you can specify filenames starting with a minus."
|
||||
echo " -h"
|
||||
echo " -help Display this help."
|
||||
echo " -n Simulate the remove (juste print files)."
|
||||
echo " directories are the roots where the purge had to be done. If no"
|
||||
echo " roots are given, the root is the home directory."
|
||||
}
|
||||
|
||||
DETRUIT=""
|
||||
ANT_OPT=""
|
||||
ANT_CMD=""
|
||||
case "$1" in
|
||||
'-' )
|
||||
shift;;
|
||||
'-n' )
|
||||
DETRUIT="echo"
|
||||
ANT_OPT="-p"
|
||||
shift;;
|
||||
'-clean' )
|
||||
ANT_CMD="clean"
|
||||
shift;;
|
||||
'-wipe' )
|
||||
ANT_CMD="wipe"
|
||||
shift;;
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
shift
|
||||
exit;;
|
||||
esac
|
||||
|
||||
DIRS=$*
|
||||
if test "$#" -le 1
|
||||
then
|
||||
DIRS="$*"
|
||||
if test -z "$1" -o -d "$1"
|
||||
then
|
||||
cd $1 || exit
|
||||
DIRS=.
|
||||
fi
|
||||
fi
|
||||
|
||||
if test "${ANT_CMD}" != ""
|
||||
then
|
||||
find $DIRS -type f -name build.xml -execdir ant -f {} "${ANT_CMD}" \;
|
||||
find $DIRS -type f -name Makefile\* -execdir make -f {} "${ANT_CMD}" \;
|
||||
exit
|
||||
fi
|
||||
|
||||
find $DIRS -type d -name .xvpics -exec $DETRUIT rm -r {} \; -prune
|
||||
|
||||
find $DIRS '(' \
|
||||
-type d -name POUBELLE -prune \
|
||||
-o \
|
||||
-type f '(' \
|
||||
-name core -o -name '*.BAK' -o -name '*.bak' -o -name '*.CKP' \
|
||||
-o -name '.*.BAK' -o -name '.*.bak' -o -name '.*.CKP' \
|
||||
-o -name '.*.back' -o -name '*.back' \
|
||||
-o -name '*.backup' -o -name '*.backup ' \
|
||||
-o -name '.*.backup' -o -name '.*.backup ' \
|
||||
-o -name .make.state \
|
||||
-o -name 'untitled*' -o -name 'Sansnom' \
|
||||
-o -name '.emacs_*' -o -name '.wi_*' \
|
||||
-o -name 'ws_ftp.log' -o -name 'hs_err*.log' \
|
||||
-o -name '#*' -o -name '*~' -o -name '.*~' -o -name junk \
|
||||
-o -name '.~lock.*#' \
|
||||
-o -name '*%' -o -name '.*%' \
|
||||
')'\
|
||||
-print -exec $DETRUIT mv -f '{}' "${POUBELLE}" \; \
|
||||
')'
|
||||
|
||||
# -o -name '*.ps' -o -name '.*.ps' \
|
||||
# -o -name '*.i' -o -name '*.ixx' \
|
||||
# -o -name '.*.sav' -o -name '*.sav' \
|
||||
|
24
bin/nextcloud_maintenance.sh
Executable file
@ -0,0 +1,24 @@
|
||||
#!/bin/bash
|
||||
#on récupère toutes les variables et mdp
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
URL_AGORA=https://$matterHost.$domain/api/v4
|
||||
EQUIPE=kaz
|
||||
|
||||
PostMattermost() {
|
||||
PostM=$1
|
||||
CHANNEL=$2
|
||||
TEAMID=$(curl -s -H "Authorization: Bearer ${mattermost_token}" "${URL_AGORA}/teams/name/${EQUIPE}" | jq .id | sed -e 's/"//g')
|
||||
CHANNELID=$(curl -s -H "Authorization: Bearer ${mattermost_token}" ${URL_AGORA}/teams/${TEAMID}/channels/name/${CHANNEL} | jq .id | sed -e 's/"//g')
|
||||
curl -s i-X POST -i -H "Authorization: Bearer ${mattermost_token}" -d "{\"channel_id\":\"${CHANNELID}\",\"message\":\"${PostM}\"}" "${URL_AGORA}/posts" >/dev/null 2>&1
|
||||
}
|
||||
|
||||
LISTEORGA=$(ls -F1 /var/lib/docker/volumes/ | grep cloudData | sed -e 's/^orga_//g' -e 's/-cloudData\///g')
|
||||
for CLOUD in ${LISTEORGA}
|
||||
do
|
||||
/kaz/bin/gestContainers.sh -cloud -occ "maintenance:mode" ${CLOUD} | grep -i enable && PostMattermost "ERREUR : Le cloud ${CLOUD} sur ${site} est en mode maintenance" "Sysadmin-alertes"
|
||||
done
|
27
bin/postfix-superviz.sh
Executable file
@ -0,0 +1,27 @@
|
||||
#!/bin/bash
|
||||
# supervision de sympa
|
||||
#KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
URL_AGORA=$(echo $matterHost).$(echo $domain)
|
||||
MAX_QUEUE=50
|
||||
|
||||
OLDIFS=$IFS
|
||||
IFS=" "
|
||||
COUNT_MAILQ=$(docker exec -ti mailServ mailq | tail -n1 | gawk '{print $5}')
|
||||
|
||||
docker exec ${mattermostServName} bin/mmctl --suppress-warnings auth login $httpProto://$URL_AGORA --name local-server --username $mattermost_user --password $mattermost_pass >/dev/null 2>&1
|
||||
|
||||
if [ "${COUNT_MAILQ}" -gt "${MAX_QUEUE}" ]; then
|
||||
echo "---------------------------------------------------------- "
|
||||
echo -e "Mail queue Postfix ALert, Messages: ${RED}${COUNT_MAILQ}${NC}"
|
||||
echo "---------------------------------------------------------- "
|
||||
docker exec mattermostServ bin/mmctl post create kaz:Sysadmin-alertes --message "Alerte mailq Postfix : La file d' attente est de ${COUNT_MAILQ} messages" >/dev/null 2>&1
|
||||
fi
|
||||
|
||||
IFS=${OLDIFS}
|
17
bin/runAlertings.sh
Executable file
@ -0,0 +1,17 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
. "${DOCKERS_ENV}"
|
||||
|
||||
for dockerToTest in $KAZ_ROOT/dockers/*/alerting/; do
|
||||
for aTest in $dockerToTest/*; do
|
||||
res=$($aTest)
|
||||
if [ -n "$res" ]; then
|
||||
echo $res
|
||||
docker exec -ti mattermostServ bin/mmctl post create kaz:Sysadmin-alertes --message "${res:0:1000}"
|
||||
fi
|
||||
done
|
||||
done
|
37
bin/sauve_memory.sh
Executable file
@ -0,0 +1,37 @@
|
||||
#! /bin/sh
|
||||
# date: 30/03/2022
|
||||
# koi: récupérer du swap et de la ram (uniquement sur les services qui laissent filer la mémoire)
|
||||
# ki: fab
|
||||
|
||||
#pour vérifier les process qui prennent du swap : for file in /proc/*/status ; do awk '/Tgid|VmSwap|Name/{printf $2 " " $3}END{ print ""}' $file; done | grep kB | sort -k 3 -n
|
||||
|
||||
# Ces commandes donnent le nom du process, son PID et la taille mémoire en swap. Par exemple :
|
||||
# dans /proc/<PID>/status y'a un VMSwap qui est la taille de swap utilisée par le process.
|
||||
|
||||
#calc
|
||||
docker restart ethercalcDB ethercalcServ
|
||||
|
||||
#sympa
|
||||
#docker restart sympaServ
|
||||
#/kaz/dockers/sympa/reload.sh
|
||||
# --> bof, ça arrête mal le bazar (4 mails d'ano autour de sympa_msg.pl / bounced.pl / task_manager.pl / bulk.pl / archived.pl)
|
||||
|
||||
#sympa
|
||||
# restart de sympa et relance du script de copie des librairies du filtre de messages
|
||||
#docker exec -it sympaServ service sympa stop
|
||||
#sleep 5
|
||||
#docker exec -it sympaServ service sympa start
|
||||
#sleep 5
|
||||
#/kaz/dockers/sympa/reload.sh
|
||||
#sleep 2
|
||||
#docker exec sympaServ chmod 777 /home/filter/filter.sh
|
||||
#docker exec sympaServ sendmail -q
|
||||
|
||||
#pour restart cette s.... de collabora
|
||||
/kaz/bin/gestContainers.sh -office -m -r
|
||||
|
||||
#postfix
|
||||
docker exec -it mailServ supervisorctl restart changedetector
|
||||
|
||||
#proxy
|
||||
docker exec -i proxyServ bash -c "/etc/init.d/nginx reload"
|
18
bin/sauve_serveur.sh
Executable file
@ -0,0 +1,18 @@
|
||||
#! /bin/sh
|
||||
# date: 12/11/2020
|
||||
|
||||
#PATH=/bin:/sbin:/usr/bin:/usr/sbin
|
||||
PATH_SAUVE="/home/sauve/"
|
||||
|
||||
iptables-save > $PATH_SAUVE/iptables.sav
|
||||
|
||||
dpkg --get-selections > $PATH_SAUVE/dpkg_selection
|
||||
tar -clzf $PATH_SAUVE/etc_sauve.tgz /etc 1> /dev/null 2> /dev/null
|
||||
tar -clzf $PATH_SAUVE/var_spool.tgz /var/spool 1> /dev/null 2> /dev/null
|
||||
tar -clzf $PATH_SAUVE/root.tgz /root 1> /dev/null 2> /dev/null
|
||||
tar -clzf $PATH_SAUVE/kaz.tgz /kaz 1> /dev/null 2> /dev/null
|
||||
rsync -a /var/spool/cron/crontabs $PATH_SAUVE
|
||||
|
||||
#sauve les bases
|
||||
/kaz/bin/container.sh save
|
||||
|
385
bin/scriptBorg.sh
Executable file
@ -0,0 +1,385 @@
|
||||
#!/bin/bash
|
||||
# --------------------------------------------------------------------------------------
|
||||
# Didier
|
||||
#
|
||||
# Script de sauvegarde avec BorgBackup
|
||||
# la commande de creation du dépot est : borg init --encryption=repokey /mnt/backup-nas1/BorgRepo
|
||||
# la conf de borg est dans /root/.config/borg
|
||||
# Le repository peut etre distant: BORG_REPO='ssh://user@host:port/path/to/repo'
|
||||
# la clé ssh devra être copiée sur le site distant et l' init se fera sous la forme
|
||||
# borg init --encryption=repokey ssh://user@host:port/path/to/repo
|
||||
# la clé est modifiable avec la commande borg key change-passphrase
|
||||
# toutes les variables sont dans la config générale de KAZ
|
||||
# scripts PRE et Post
|
||||
# les script pre et post doivent s' appelle pre_xxxxx.sh ou post_xxxx.sh
|
||||
# La variable BORGSCRIPTS est le chemin du repertoire des scripts dans la config générale de Kaz
|
||||
#####################################################
|
||||
#KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
KAZ_ROOT=/kaz
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
|
||||
VERSION="V-18-05-2024"
|
||||
PRG=$(basename $0)
|
||||
RACINE=$(echo $PRG | awk '{print $1}')
|
||||
#IFS=' '
|
||||
#####################################################
|
||||
# Variables générales
|
||||
#####################################################
|
||||
# le volume monté ou sera le repo
|
||||
# impérativement laisser vide dans le cas d' un repo distant
|
||||
VOLUME_SAUVEGARDES=${borg_VOLUME_SAUVEGARDES}
|
||||
# le repo Borg
|
||||
export BORG_REPO=${borg_BORG_REPO}
|
||||
# Le mot de passe du repo
|
||||
export BORG_PASSPHRASE=${borg_BORG_PASSPHRASE}
|
||||
# les personnes qui recevront le rapport de sauvegarde
|
||||
MAIL_RAPPORT=${borg_MAIL_RAPPORT}
|
||||
# Recevoir un mail quand la sauvegarde est OK ?
|
||||
MAILOK=${borg_MAILOK}
|
||||
MAILWARNING=${borg_MAILWARNING}
|
||||
MAILDETAIL=${borg_MAILDETAIL}
|
||||
# repertoire de montage des sauvegardes pour les restaurations
|
||||
BORGMOUNT=${borg_BORGMOUNT}
|
||||
# - la liste des repertoires à sauver séparés par un espace
|
||||
LISTREPSAUV=${borg_LISTREPSAUV}
|
||||
# - Les sauvegardes à garder jour, semaines, mois
|
||||
NB_BACKUPS_JOUR=${borg_NB_BACKUPS_JOUR}
|
||||
NB_BACKUPS_SEM=${borg_NB_BACKUPS_SEM}
|
||||
NB_BACKUPS_MOIS=${borg_NB_BACKUPS_MOIS}
|
||||
# Le Repertoire ou sont les pré traitement
|
||||
BORGSCRIPTS=${borg_BORGSCRIPTS}
|
||||
BORGLOG="${borg_BORGLOG}/BorgBackup-$(date +%d-%m-%Y-%H-%M-%S).log"
|
||||
DEBUG=false
|
||||
|
||||
#####################################################
|
||||
#
|
||||
FICLOG="/var/log/${PRG}.log"
|
||||
|
||||
#####################################################
|
||||
|
||||
trap 'LogFic "script stoppé sur un SIGTERM ou SIGINT" >&2; exit 2' INT TERM
|
||||
|
||||
LogFic() {
|
||||
[ ! -w ${FICLOG} ] && { echo "Probleme d' ecriture dans $FICLOG" ; exit 1 ;}
|
||||
echo "$(date +%d-%m-%Y-%H-%M-%S) : $1" >> ${FICLOG}
|
||||
}
|
||||
#
|
||||
ExpMail() {
|
||||
MAIL_SOURCE=$1
|
||||
MAIL_SUJET=$2
|
||||
MAIL_DEST=$3
|
||||
MAIL_TEXTE=$4
|
||||
# a mettre ailleurs
|
||||
mailexp=${borg_MAILEXP}
|
||||
mailpassword=${borg_MAILPASSWORD}
|
||||
mailserveur=${borg_MAILSERVEUR}
|
||||
#
|
||||
#sendemail -t ${MAIL_DEST} -u ${MAIL_SUJET} -m ${MAIL_TEXTE} -f $mailexp -s $mailserveur:587 -xu $mailexp -xp $mailpassword -o tls=yes >/dev/null 2>&1
|
||||
printf "Subject:${MAIL_SUJET}\n${MAIL_TEXTE}" | msmtp ${MAIL_DEST}
|
||||
#docker exec -i mailServ mailx -a 'Content-Type: text/plain; charset="UTF-8"' -r ${MAIL_SOURCE} -s "${MAIL_SUJET}" ${MAIL_DEST} << EOF
|
||||
#${MAIL_TEXTE}
|
||||
#EOF
|
||||
}
|
||||
|
||||
Pre_Sauvegarde() {
|
||||
if [ -d ${BORGSCRIPTS} ]
|
||||
then
|
||||
cd ${BORGSCRIPTS}
|
||||
for FicPre in $(ls )
|
||||
do
|
||||
if [ -x ${FicPre} ] && [ $(echo ${FicPre} | grep -i ^pre_) ]
|
||||
then
|
||||
LogFic " - Pré traitement de la sauvegarde : ${FicPre}"
|
||||
[ "$DEBUG" = true ] && echo " - Pré traitement de la sauvegarde : ${FicPre}"
|
||||
./${FicPre}
|
||||
fi
|
||||
done
|
||||
fi
|
||||
}
|
||||
|
||||
Post_Sauvegarde() {
|
||||
if [ -d ${BORGSCRIPTS} ]
|
||||
then
|
||||
cd ${BORGSCRIPTS}
|
||||
for FicPre in $(ls )
|
||||
do
|
||||
if [ -x ${FicPre} ] && [ $(echo ${FicPre} | grep -i ^post_) ]
|
||||
then
|
||||
LogFic " - Post traitement de la sauvegarde : ${FicPre}"
|
||||
[ "$DEBUG" = true ] && echo " - Post traitement de la sauvegarde : ${FicPre}"
|
||||
./${FicPre}
|
||||
fi
|
||||
done
|
||||
fi
|
||||
|
||||
}
|
||||
|
||||
Sauvegarde() {
|
||||
Pre_Sauvegarde
|
||||
BACKUP_PRE=$?
|
||||
borg create \
|
||||
--filter AME \
|
||||
--exclude-caches \
|
||||
--stats \
|
||||
--show-rc \
|
||||
--exclude 'home/*/.cache/*' \
|
||||
--exclude 'var/tmp/*' \
|
||||
::$(date +%Y-%m-%d-%H-%M-%S-%A)-{hostname} \
|
||||
${LISTREPSAUV} >>${BORGLOG} 2>>${BORGLOG}
|
||||
BACKUP_EXIT=$?
|
||||
Post_Sauvegarde
|
||||
BACKUP_POST=$?
|
||||
}
|
||||
|
||||
Purge() {
|
||||
borg prune \
|
||||
--prefix '{hostname}-' \
|
||||
--keep-daily ${NB_BACKUPS_JOUR} \
|
||||
--keep-weekly ${NB_BACKUPS_SEM} \
|
||||
--keep-monthly ${NB_BACKUPS_MOIS} \
|
||||
|
||||
PRUNE_EXIT=$?
|
||||
}
|
||||
|
||||
Compact() {
|
||||
|
||||
borg compact --progress ${BORG_REPO}
|
||||
|
||||
}
|
||||
|
||||
usage() {
|
||||
|
||||
echo "-h : Usage"
|
||||
echo "-c : Permet de compacter ${BORG_REPO}"
|
||||
echo "-d : Permet de verifier les variables de sauvegarde"
|
||||
echo "-i : Mode interractif"
|
||||
echo "-l : Liste les sauvegardes sans monter ${BORG_REPO}"
|
||||
echo "-m : Monte le REPO (${BORG_REPO} sur ${BORGMOUNT})"
|
||||
echo "-p : Permet de lancer la phase de purge des backup en fonctions des variables: jour=${NB_BACKUPS_JOUR},semaine=${NB_BACKUPS_SEM},mois=${NB_BACKUPS_MOIS}"
|
||||
echo "-s : Lance la sauvegarde"
|
||||
echo "-u : Demonte le REPO (${BORG_REPO} de ${BORGMOUNT})"
|
||||
echo "-v : Version"
|
||||
exit
|
||||
}
|
||||
|
||||
Borgvariables() {
|
||||
echo "-----------------------------------------------------------"
|
||||
echo " Variables applicatives pour le site ${site}"
|
||||
echo "-----------------------------------------------------------"
|
||||
for borgvar in $(set | grep borg_ | sed -e 's/borg_//' -e 's/=.*$//' | grep ^[A-Z])
|
||||
do
|
||||
echo "$borgvar=${!borgvar}"
|
||||
done
|
||||
if grep borgfs /etc/mtab >/dev/null 2>&1
|
||||
then
|
||||
echo -e "${RED}WARNING${NC}: ${BORG_REPO} est monté sur ${BORGMOUNT}"
|
||||
fi
|
||||
}
|
||||
|
||||
Borgmount() {
|
||||
LogFic "Montage du repo ${BORG_REPO} sur ${BORGMOUNT} .. "
|
||||
echo -en "Montage du repo ${BORG_REPO} sur ${BORGMOUNT} .. "
|
||||
borg mount ${BORG_REPO} ${BORGMOUNT} >/dev/null 2>&1
|
||||
if [ $? = 0 ]
|
||||
then
|
||||
LogFic "Ok"
|
||||
echo -e "${GREEN}Ok${NC}"
|
||||
else
|
||||
LogFic "Error"
|
||||
echo -e "${RED}Error $?${NC}"
|
||||
fi
|
||||
exit
|
||||
}
|
||||
|
||||
Borgumount() {
|
||||
LogFic "Demontage du repo ${BORG_REPO} sur ${BORGMOUNT} .. "
|
||||
echo -en "Demontage du repo ${BORG_REPO} sur ${BORGMOUNT} .. "
|
||||
borg umount ${BORGMOUNT} >/dev/null 2>&1
|
||||
if [ $? = 0 ]
|
||||
then
|
||||
LogFic "Ok"
|
||||
echo -e "${GREEN}Ok${NC}"
|
||||
else
|
||||
LogFic "Error"
|
||||
echo -e "${RED}Error $?${NC}"
|
||||
fi
|
||||
exit
|
||||
}
|
||||
|
||||
Borglist() {
|
||||
LogFic "Borg list demandé"
|
||||
borg list --short ${BORG_REPO}
|
||||
exit
|
||||
}
|
||||
|
||||
main() {
|
||||
|
||||
# ****************************************************** Main *******************************************************************
|
||||
# Création du fichier de log
|
||||
touch ${FICLOG}
|
||||
type -P sendemail || { echo "sendemail non trouvé";exit 1;}
|
||||
#
|
||||
LogFic "#########################################################################################################################"
|
||||
LogFic " *************** ${PRG} Version ${VERSION} ***************"
|
||||
LogFic "#########################################################################################################################"
|
||||
# test si les variables importantes sont renseignées et sortie si tel n' est pas le cas
|
||||
if [ -z "${VOLUME_SAUVEGARDES}" ] && [ -z "${BORG_REPO}" ] || [ -z "${BORG_REPO}" ] || [ -z "${BORG_PASSPHRASE}" ] || [ -z "${MAIL_RAPPORT}" ]
|
||||
then
|
||||
echo "Les variables VOLUME_SAUVEGARDES, BORG_REPO, BORG_PASSPHRASE, MAIL_RAPPORT sont à verifier"
|
||||
LogFic "Les variables VOLUME_SAUVEGARDES, BORG_REPO, BORG_PASSPHRASE, MAIL_RAPPORT sont à verifier"
|
||||
LogFic "Sortie du script"
|
||||
exit 1
|
||||
fi
|
||||
# test si le volume de sauvegarde est ok
|
||||
if [ ! -z ${VOLUME_SAUVEGARDES} ]
|
||||
then
|
||||
[ !$(grep "${VOLUME_SAUVEGARDES}" /etc/mtab >/dev/null 2>&1) ] || { echo "le volume de sauvegarde ${VOLUME_SAUVEGARDES} n' est pas monté"; LogFic "Erreur de montage du volume ${VOLUME_SAUVEGARDES} de sauvegarde" ; exit 1;}
|
||||
else
|
||||
[ ! $(echo ${BORG_REPO} | grep -i ssh 2>/dev/null) ] && { echo "Problème avec le repo distant ";exit 1;}
|
||||
fi
|
||||
|
||||
# Test si le REPO est monté : on sort
|
||||
if grep borgfs /etc/mtab >/dev/null 2>&1
|
||||
then
|
||||
echo "le REPO : ${BORG_REPO} est monté , je sors"
|
||||
LogFic "le REPO : ${BORG_REPO} est monté , je sors"
|
||||
ExpMail borg@${domain} "${site} : Sauvegarde en Erreur !!!!" ${MAIL_RAPPORT} "le REPO : ${BORG_REPO} est monté, sauvegarde impossible"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Tout se passe bien on continue
|
||||
LogFic " - Repertoire a sauver : ${LISTREPSAUV}"
|
||||
LogFic " - Volume Nfs monté : ${VOLUME_SAUVEGARDES}"
|
||||
LogFic " - Repertoire des sauvegardes : ${BORG_REPO}"
|
||||
[ ! -d ${BORGSCRIPTS} ] && LogFic "Pas de repertoire de PRE et POST" || LogFic " - Repertoire des scripts Post/Pré : ${BORGSCRIPTS}"
|
||||
[ "${DEBUG}" = true ] && [ -d ${BORGSCRIPTS} ] && echo "Rep des scripts PRE/POST :${BORGSCRIPTS}"
|
||||
LogFic " - Rapport par Mail : ${MAIL_RAPPORT}"
|
||||
LogFic " - Backups jour : ${NB_BACKUPS_JOUR} , Backups semaines : ${NB_BACKUPS_SEM} , Backups mois : ${NB_BACKUPS_MOIS}"
|
||||
[ "${DEBUG}" = true ] && echo "${LISTREPSAUV} sauvé dans ${BORG_REPO}, Rapport vers : ${MAIL_RAPPORT}"
|
||||
LogFic "#########################################################################################################################"
|
||||
LogFic " - Démarrage de la sauvegarde"
|
||||
[ "$DEBUG" = true ] && echo "Demarrage de la sauvegarde : "
|
||||
LogFic " - Log dans ${BORGLOG}"
|
||||
Sauvegarde
|
||||
[ "$DEBUG" = true ] && echo "code de retour de backup : ${BACKUP_EXIT}"
|
||||
LogFic " - Code de retour de la commande sauvegarde : ${BACKUP_EXIT}"
|
||||
LogFic " - Démarrage du nettoyage des sauvegardes"
|
||||
[ "$DEBUG" = true ] && echo "Nettoyage des sauvegardes: "
|
||||
Purge
|
||||
LogFic " - Code retour du Nettoyage des sauvegardes (0=OK; 1=WARNING, 2=ERROR) : ${PRUNE_EXIT}"
|
||||
[ "$DEBUG" = true ] && echo "code de retour prune : ${PRUNE_EXIT}"
|
||||
#
|
||||
########################################################################################
|
||||
# si la variable MAILDETAIL est true alors on affecte le contenu du log sinon LOGDATA est VIDE
|
||||
LOGDATA=""
|
||||
[ "$MAILDETAIL" = true ] && LOGDATA=$(cat ${BORGLOG})
|
||||
[ "$DEBUG" = true ] && [ "$MAILDETAIL" = true ] && echo "Envoi du mail à ${MAIL_RAPPORT}"
|
||||
# On teste le code retour de la sauvegarde, on log et on envoie des mails
|
||||
case "${BACKUP_EXIT}" in
|
||||
'0' )
|
||||
IFS=''
|
||||
MESS_SAUVE_OK="
|
||||
Salut
|
||||
|
||||
La sauvegarde est ok, ce message peut être enlevé avec la variable MAILOK=false
|
||||
Que la force soit avec toi
|
||||
|
||||
BorgBackup
|
||||
|
||||
"
|
||||
LogFic " - la sauvegarde est OK"
|
||||
[ "$MAILOK" = true ] && ExpMail borg@${domain} "${site} : Sauvegarde Ok" ${MAIL_RAPPORT} ${MESS_SAUVE_OK}${LOGDATA}
|
||||
IFS=' '
|
||||
;;
|
||||
'1' )
|
||||
IFS=''
|
||||
MESS_SAUVE_ERR="
|
||||
Salut
|
||||
|
||||
La sauvegarde est en warning
|
||||
Code de retour de la commande sauvegarde : ${BACKUP_EXIT}
|
||||
Le log contenant les infos est ${BORGLOG}
|
||||
|
||||
BorgBackup
|
||||
|
||||
"
|
||||
LogFic " - Sauvegarde en Warning: ${BACKUP_EXIT}"
|
||||
[ "$MAILWARNING" = true ] && ExpMail borg@${domain} "${site} : Sauvegarde en Warning: ${BACKUP_EXIT}" ${MAIL_RAPPORT} ${MESS_SAUVE_ERR}${LOGDATA}
|
||||
IFS=' '
|
||||
;;
|
||||
* )
|
||||
IFS=''
|
||||
MESS_SAUVE_ERR="
|
||||
Salut
|
||||
|
||||
La sauvegarde est en Erreur
|
||||
Code de retour de la commande sauvegarde : ${BACKUP_EXIT}
|
||||
Le log à consulter est ${BORGLOG}
|
||||
|
||||
BorgBackup
|
||||
|
||||
"
|
||||
LogFic " - !!!!! Sauvegarde en Erreur !!!!! : ${BACKUP_EXIT}"
|
||||
ExpMail borg@${domain} "${site} : Sauvegarde en Erreur !!!! : ${BACKUP_EXIT}" ${MAIL_RAPPORT} ${MESS_SAUVE_ERR}${LOGDATA}
|
||||
IFS=' '
|
||||
;;
|
||||
esac
|
||||
LogFic " - Fin de la sauvegarde"
|
||||
exit
|
||||
}
|
||||
|
||||
[ ! "$#" -eq "0" ] || usage
|
||||
# On teste les arguments pour le script
|
||||
for ARG in $@; do
|
||||
case "${ARG}" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
'-m' )
|
||||
shift
|
||||
Borgmount
|
||||
;;
|
||||
'-u' )
|
||||
shift
|
||||
Borgumount
|
||||
;;
|
||||
'-l' )
|
||||
shift
|
||||
Borglist
|
||||
;;
|
||||
'-i' )
|
||||
shift
|
||||
DEBUG=true
|
||||
;;
|
||||
'-v' )
|
||||
shift
|
||||
echo "Version : ${VERSION}"
|
||||
exit
|
||||
;;
|
||||
'-d' )
|
||||
shift
|
||||
Borgvariables
|
||||
exit
|
||||
;;
|
||||
'-c' )
|
||||
shift
|
||||
Compact
|
||||
exit
|
||||
;;
|
||||
'-s' )
|
||||
main
|
||||
;;
|
||||
'-p' )
|
||||
shift
|
||||
read -p "Ok pour lancer la purge en fonction de ces valeurs : jour=${NB_BACKUPS_JOUR},semaine=${NB_BACKUPS_SEM},mois=${NB_BACKUPS_MOIS} ? O/N : " READPURGE
|
||||
[[ ${READPURGE} =~ ^[oO]$ ]] && Purge || echo "pas de purge"
|
||||
exit
|
||||
;;
|
||||
* | ' ' )
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
161
bin/scriptSauve.sh
Executable file
@ -0,0 +1,161 @@
|
||||
#!/bin/bash
|
||||
# Didier le 14 Septembre 2022
|
||||
#
|
||||
# TODO : Inclure un script post et pre.
|
||||
#
|
||||
#####################################################
|
||||
#KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
KAZ_ROOT="/kaz"
|
||||
. $KAZ_ROOT/bin/.commonFunctions.sh
|
||||
setKazVars
|
||||
. $DOCKERS_ENV
|
||||
. $KAZ_ROOT/secret/SetAllPass.sh
|
||||
VERSION="1.0"
|
||||
PRG=$(basename $0)
|
||||
RACINE=$(echo $PRG | awk '{print $1}')
|
||||
IFS=' '
|
||||
#
|
||||
#####################################################
|
||||
MAILSOURCE="sauve@kaz.bzh"
|
||||
VOLUME_SAUVEGARDES="/mnt/backup-nas1"
|
||||
#SAUVE_REPO=${VOLUME_SAUVEGARDES}/SauveRepo
|
||||
SAUVE_REPO=admin@nas-kaz1:/share/Backup/SauveRepo
|
||||
MAIL_RAPPORT="didier@kaz.bzh;francois@kaz.bzh;fab@kaz.bzh;fanch@kaz.bzh"
|
||||
#####################################################
|
||||
SCRIPTLOG="/mnt/disk-nas1/log/${PRG}-$(date +%d-%m-%Y-%H-%M-%S).log"
|
||||
FICLOG="/var/log/${PRG}.log"
|
||||
#####################################################
|
||||
# - la liste des repertoires à sauver séparés par un espace
|
||||
LISTREPSAUV="/var/lib/docker/volumes /kaz"
|
||||
#####################################################
|
||||
# - Les sauvegardes à garder jour, semaines, mois
|
||||
NB_BACKUPS_JOUR=15
|
||||
NB_BACKUPS_SEM=8
|
||||
NB_BACKUPS_MOIS=12
|
||||
#####################################################
|
||||
# Recevoir un mail quand la sauvegarde est OK ?
|
||||
MAILOK=true
|
||||
#####################################################
|
||||
|
||||
trap 'LogFic "script stoppé sur un SIGTERM ou SIGINT" >&2; exit 2' INT TERM
|
||||
|
||||
LogFic() {
|
||||
[ ! -w ${FICLOG} ] && { echo "Probleme d' ecriture dans $FICLOG" ; exit 1 ;}
|
||||
echo "$(date +%d-%m-%Y-%H-%M-%S) : $1" >> ${FICLOG}
|
||||
}
|
||||
#
|
||||
ExpMail() {
|
||||
MAIL_SOURCE=$1
|
||||
MAIL_SUJET=$2
|
||||
MAIL_DEST=$3
|
||||
MAIL_TEXTE=$4
|
||||
docker exec -i mailServ mailx -a 'Content-Type: text/plain; charset="UTF-8"' -r ${MAIL_SOURCE} -s "${MAIL_SUJET}" ${MAIL_DEST} << EOF
|
||||
${MAIL_TEXTE}
|
||||
EOF
|
||||
}
|
||||
Sauvegarde() {
|
||||
#$1 est le repertoire à sauver, on créé le sous repertoire dans le repo
|
||||
CODE_TMP=""
|
||||
if [ -r $1 ]
|
||||
then
|
||||
echo "Sauvegarde $1" >>${SCRIPTLOG}
|
||||
#mkdir -p ${SAUVE_REPO}/$1 >/dev/null 2>&1
|
||||
#rdiff-backup --verbosity 3 $1 ${SAUVE_REPO}/$1 >>${SCRIPTLOG} 2>>${SCRIPTLOG}
|
||||
rsync -aAHXh --del --stats --exclude 'files_trashbin' $1 ${SAUVE_REPO} >>${SCRIPTLOG} 2>>${SCRIPTLOG}
|
||||
CODE_TMP=$?
|
||||
else
|
||||
LogFic "$1 n' existe pas ou n' est pas accessible en lecture"
|
||||
CODE_TMP=1
|
||||
fi
|
||||
LogFic "Code Retour de la sauvegarde de $1 : ${CODE_TMP}"
|
||||
BACKUP_EXIT=$(expr ${BACKUP_EXIT} + ${CODE_TMP} )
|
||||
}
|
||||
#
|
||||
Purge() {
|
||||
echo "Commande prune de rdiff-backup"
|
||||
PRUNE_EXIT=$?
|
||||
}
|
||||
# ****************************************************** Main *******************************************************************
|
||||
# Création du fichier de log
|
||||
touch ${FICLOG}
|
||||
#
|
||||
LogFic "#########################################################################################################################"
|
||||
LogFic " *************** ${PRG} Version ${VERSION} ***************"
|
||||
LogFic "#########################################################################################################################"
|
||||
# test si les variables importantes sont renseignées et sortie si tel n' est pas le cas
|
||||
if [ -z "${VOLUME_SAUVEGARDES}" ] || [ -z "${SAUVE_REPO}" ]
|
||||
then
|
||||
echo "VOLUME_SAUVEGARDES et SAUVE_REPO à verifier"
|
||||
LogFic "VOLUME_SAUVEGARDES et SAUVE_REPO à verifier"
|
||||
LogFic "Sortie du script"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
#####################################################################################################################################################
|
||||
################### Mise en commentaire de cette section puisque le repo est en rsync ( voir plus tard comment gérer ça )
|
||||
# test si le volume de sauvegarde est ok
|
||||
#grep "${VOLUME_SAUVEGARDES}" /etc/mtab >/dev/null 2>&1
|
||||
#if [ "$?" -ne "0" ]
|
||||
#then
|
||||
# echo "le volume de sauvegarde ${VOLUME_SAUVEGARDES} n' est pas monté"
|
||||
# LogFic "Erreur de montage du volume ${VOLUME_SAUVEGARDES} de sauvegarde"
|
||||
# exit 1
|
||||
#fi
|
||||
# Test si j' ai le droit d' écrire dans le Repo
|
||||
# [ ! -w ${SAUVE_REPO} ] && { echo "Verifier le droit d' écriture dans ${SAUVE_REPO}" ; LogFic "Verifier le droit d' écriture dans ${SAUVE_REPO}"; exit 1;}
|
||||
#####################################################################################################################################################
|
||||
|
||||
# Tout se passe bien on continue
|
||||
LogFic " - Repertoire a sauver : ${LISTREPSAUV}"
|
||||
#LogFic " - Volume Nfs monté : ${VOLUME_SAUVEGARDES}"
|
||||
LogFic " - Destination des sauvegardes : ${SAUVE_REPO}"
|
||||
LogFic " - Rapport par Mail : ${MAIL_RAPPORT}"
|
||||
#LogFic " - Backups jour : ${NB_BACKUPS_JOUR} , Backups semaines : ${NB_BACKUPS_SEM} , Backups mois : ${NB_BACKUPS_MOIS}"
|
||||
LogFic "#########################################################################################################################"
|
||||
LogFic " - Démarrage de la sauvegarde"
|
||||
LogFic " - Log dans ${SCRIPTLOG}"
|
||||
BACKUP_EXIT=0
|
||||
PRUNE_EXIT=0
|
||||
for REPS in ${LISTREPSAUV}
|
||||
do
|
||||
LogFic "Sauvegarde de ${REPS}"
|
||||
Sauvegarde ${REPS}
|
||||
done
|
||||
LogFic "Code retour compilé de toutes les sauvegardes : ${BACKUP_EXIT}"
|
||||
################################## a gérer plus tard
|
||||
#LogFic " - Démarrage du nettoyage des sauvegardes"
|
||||
#Purge
|
||||
#LogFic " - Code retour du Nettoyage des sauvegardes (0=OK; 1=WARNING, 2=ERROR) : ${PRUNE_EXIT}"
|
||||
#
|
||||
########################################################################################
|
||||
# On teste le code retour de la sauvegarde, on log et on envoie des mails
|
||||
case "${BACKUP_EXIT}" in
|
||||
'0' )
|
||||
IFS=''
|
||||
MESS_SAUVE_OK="
|
||||
Salut
|
||||
|
||||
La sauvegarde est ok, ce message peut être enlevé avec la variable MAILOK=false
|
||||
Que la force soit avec toi
|
||||
|
||||
Ton esclave des sauvegardes"
|
||||
LogFic " - la sauvegarde est OK"
|
||||
[ "$MAILOK" = true ] && ExpMail ${MAILSOURCE} "Sauvegarde Ok" ${MAIL_RAPPORT} ${MESS_SAUVE_OK}
|
||||
IFS=' '
|
||||
;;
|
||||
* )
|
||||
IFS=''
|
||||
MESS_SAUVE_ERR="
|
||||
Salut
|
||||
|
||||
La sauvegarde est en Erreur
|
||||
Le log à consulter est ${SCRIPTLOG}
|
||||
Code retour de la Sauvegarde ( code Rsync ): ${BACKUP_EXIT}
|
||||
|
||||
Ton esclave des sauvegardes"
|
||||
LogFic " - !!!!! Sauvegarde en Erreur !!!!!"
|
||||
ExpMail ${MAILSOURCE} "!!!! Sauvegarde en Erreur !!!!" ${MAIL_RAPPORT} ${MESS_SAUVE_ERR}
|
||||
IFS=' '
|
||||
;;
|
||||
esac
|
||||
LogFic " - Fin de la sauvegarde"
|
72
bin/secretGen.sh
Executable file
@ -0,0 +1,72 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
cd "${KAZ_ROOT}"
|
||||
|
||||
NEW_DIR="secret"
|
||||
TMPL_DIR="secret.tmpl"
|
||||
|
||||
if [ ! -d "${NEW_DIR}/" ]; then
|
||||
rsync -a "${TMPL_DIR}/" "${NEW_DIR}/"
|
||||
fi
|
||||
|
||||
NEW_FILE="${NEW_DIR}/SetAllPass-new.sh"
|
||||
TMPL_FILE="${NEW_DIR}/SetAllPass.sh"
|
||||
|
||||
while read line ; do
|
||||
if [[ "${line}" =~ ^# ]] || [ -z "${line}" ] ; then
|
||||
echo "${line}"
|
||||
continue
|
||||
fi
|
||||
if [[ "${line}" =~ "--clean_val--" ]] ; then
|
||||
case "${line}" in
|
||||
*jirafeau_DATA_DIR*)
|
||||
JIRAFEAU_DIR=$(getValInFile "${DOCKERS_ENV}" "jirafeauDir")
|
||||
[ -z "${JIRAFEAU_DIR}" ] &&
|
||||
echo "${line}" ||
|
||||
sed "s%\(.*\)--clean_val--\(.*\)%\1${JIRAFEAU_DIR}\2%" <<< ${line}
|
||||
continue
|
||||
;;
|
||||
*DATABASE*)
|
||||
dbName="$(sed "s/\([^_]*\)_.*/\1/" <<< ${line})_$(apg -n 1 -m 2 -M NCL | cut -c 1-2)"
|
||||
sed "s/\(.*\)--clean_val--\(.*\)/\1${dbName}\2/" <<< ${line}
|
||||
continue
|
||||
;;
|
||||
*ROOT_PASSWORD*|*PASSWORD*)
|
||||
pass="$(apg -n 1 -m 16 -M NCL)"
|
||||
sed "s/\(.*\)--clean_val--\(.*\)/\1${pass}\2/" <<< ${line}
|
||||
continue
|
||||
;;
|
||||
*USER*)
|
||||
user="$(sed "s/\([^_]*\)_.*/\1/" <<< ${line})_$(apg -n 1 -m 2 -M NCL | cut -c 1-2)"
|
||||
sed "s/\(.*\)--clean_val--\(.*\)/\1${user}\2/" <<< ${line}
|
||||
continue
|
||||
;;
|
||||
*RAIN_LOOP*|*office_password*|*mattermost_*|*sympa_*|*gitea_*)
|
||||
pass="$(apg -n 1 -m 16 -M NCL)"
|
||||
sed "s/\(.*\)--clean_val--\(.*\)/\1${pass}\2/" <<< ${line}
|
||||
continue
|
||||
;;
|
||||
*vaultwarden_ADMIN_TOKEN*)
|
||||
pass="$(apg -n 1 -m 32 -M NCL)"
|
||||
sed "s/\(.*\)--clean_val--\(.*\)/\1${pass}\2/" <<< ${line}
|
||||
continue
|
||||
;;
|
||||
esac
|
||||
else
|
||||
echo "${line}"
|
||||
continue
|
||||
fi
|
||||
printKazError "${line}" >&2
|
||||
done < "${TMPL_FILE}" > "${NEW_FILE}"
|
||||
|
||||
mv "${NEW_FILE}" "${TMPL_FILE}"
|
||||
|
||||
chmod a+x "${TMPL_FILE}"
|
||||
. "${TMPL_FILE}"
|
||||
"${KAZ_BIN_DIR}/updateDockerPassword.sh"
|
||||
|
||||
exit 0
|
58
bin/setOwner.sh
Executable file
@ -0,0 +1,58 @@
|
||||
#!/bin/bash
|
||||
|
||||
cd $(dirname $0)/..
|
||||
KAZ=$(pwd)
|
||||
owner=root
|
||||
|
||||
usage(){
|
||||
echo "Usage: $0 [root|user]"
|
||||
exit 1
|
||||
}
|
||||
|
||||
case $# in
|
||||
0)
|
||||
;;
|
||||
1)
|
||||
owner=$1
|
||||
;;
|
||||
*)
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
|
||||
####################
|
||||
# config
|
||||
cd ${KAZ}
|
||||
DIRS="config secret bin"
|
||||
|
||||
chown -hR ${owner}: ${DIRS}
|
||||
find ${DIRS} -type f -exec chmod a-x {} \;
|
||||
find ${DIRS} -type f -name \*.sh -exec chmod a+x {} \;
|
||||
chmod -R a+X ${DIRS}
|
||||
chmod -R go= ${DIRS}
|
||||
|
||||
chmod a+x bin/*.sh
|
||||
chown -hR www-data: config/orgaTmpl/wiki-conf/
|
||||
|
||||
####################
|
||||
# dockers
|
||||
cd ${KAZ}/dockers
|
||||
|
||||
chown -h ${owner}: . * */.env */* */config/*
|
||||
chmod a-x,a+r * */*
|
||||
chmod a+X . * */*
|
||||
chmod a+x */*.sh
|
||||
|
||||
chown -hR ${owner}: \
|
||||
etherpad/etherpad-lite/ \
|
||||
paheko/extensions paheko/paheko-* \
|
||||
jirafeau/Jirafeau \
|
||||
mattermost/app
|
||||
|
||||
chown -hR www-data: \
|
||||
vigilo \
|
||||
web/html
|
||||
|
||||
chmod -R a+rX web/html
|
||||
|
||||
|
13
bin/updateAllOrga.sh
Executable file
@ -0,0 +1,13 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
cd "${KAZ_COMP_DIR}"
|
||||
for orga in *-orga
|
||||
do
|
||||
${orga}/orga-gen.sh
|
||||
"${KAZ_ROOT}/bin/container.sh" stop "${orga}"
|
||||
"${KAZ_ROOT}/bin/container.sh" start "${orga}"
|
||||
done
|
121
bin/updateDockerPassword.sh
Executable file
@ -0,0 +1,121 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd $(dirname $0)/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
# pour mise au point
|
||||
# SIMU=echo
|
||||
|
||||
# Améliorations à prévoir
|
||||
# - donner en paramètre les services concernés (pour limité les modifications)
|
||||
# - pour les DB si on déclare un nouveau login, alors les privilèges sont créé mais les anciens pas révoqués
|
||||
|
||||
. "${DOCKERS_ENV}"
|
||||
. "${KAZ_KEY_DIR}/SetAllPass.sh"
|
||||
|
||||
updateEnvDB(){
|
||||
# $1 = prefix
|
||||
# $2 = envName
|
||||
# $3 = containerName of DB
|
||||
rootPass="$1_MYSQL_ROOT_PASSWORD"
|
||||
dbName="$1_MYSQL_DATABASE"
|
||||
userName="$1_MYSQL_USER"
|
||||
userPass="$1_MYSQL_PASSWORD"
|
||||
|
||||
${SIMU} sed -i \
|
||||
-e "s/MYSQL_ROOT_PASSWORD=.*/MYSQL_ROOT_PASSWORD=${!rootPass}/g" \
|
||||
-e "s/MYSQL_DATABASE=.*/MYSQL_DATABASE=${!dbName}/g" \
|
||||
-e "s/MYSQL_USER=.*/MYSQL_USER=${!userName}/g" \
|
||||
-e "s/MYSQL_PASSWORD=.*/MYSQL_PASSWORD=${!userPass}/g" \
|
||||
"$2"
|
||||
|
||||
# seulement si pas de mdp pour root
|
||||
# pb oeuf et poule (il faudrait les anciennes valeurs) :
|
||||
# * si rootPass change, faire à la main
|
||||
# * si dbName change, faire à la main
|
||||
checkDockerRunning "$3" "$3" || return
|
||||
echo "change DB pass on docker $3"
|
||||
echo "grant all privileges on ${!dbName}.* to '${!userName}' identified by '${!userPass}';" | \
|
||||
docker exec -i $3 bash -c "mysql --user=root --password=${!rootPass}"
|
||||
}
|
||||
|
||||
updateEnv(){
|
||||
# $1 = prefix
|
||||
# $2 = envName
|
||||
|
||||
for varName in $(grep "^[a-zA-Z_]*=" $2 | sed "s/^\([^=]*\)=.*/\1/g")
|
||||
do
|
||||
srcName="$1_${varName}"
|
||||
srcVal=$(echo "${!srcName}" | sed -e "s/[&]/\\\&/g")
|
||||
${SIMU} sed -i \
|
||||
-e "s%^[ ]*${varName}=.*\$%${varName}=${srcVal}%" \
|
||||
"$2"
|
||||
done
|
||||
}
|
||||
|
||||
framadateUpdate(){
|
||||
[[ "${COMP_ENABLE}" =~ " framadate " ]] || return
|
||||
if [ ! -f "${DOCK_LIB}/volumes/framadate_dateConfig/_data/config.php" ]; then
|
||||
return 0
|
||||
fi
|
||||
checkDockerRunning "${framadateServName}" "Framadate" &&
|
||||
${SIMU} docker exec -ti "${framadateServName}" bash -c -i "htpasswd -bc /var/framadate/admin/.htpasswd ${framadate_HTTPD_USER} ${framadate_HTTPD_PASSWORD}"
|
||||
${SIMU} sed -i \
|
||||
-e "s/^#*const DB_USER[ ]*=.*$/const DB_USER= '${framadate_MYSQL_USER}';/g" \
|
||||
-e "s/^#*const DB_PASSWORD[ ]*=.*$/const DB_PASSWORD= '${framadate_MYSQL_PASSWORD}';/g" \
|
||||
"${DOCK_LIB}/volumes/framadate_dateConfig/_data/config.php"
|
||||
}
|
||||
|
||||
jirafeauUpdate(){
|
||||
[[ "${COMP_ENABLE}" =~ " jirafeau " ]] || return
|
||||
if [ ! -f "${DOCK_LIB}/volumes/jirafeau_fileConfig/_data/config.local.php" ]; then
|
||||
return 0
|
||||
fi
|
||||
SHA=$(echo -n "${jirafeau_HTTPD_PASSWORD}" | sha256sum | cut -d \ -f 1)
|
||||
${SIMU} sed -i \
|
||||
-e "s/'admin_password'[ ]*=>[ ]*'[^']*'/'admin_password' => '${SHA}'/g" \
|
||||
"${DOCK_LIB}/volumes/jirafeau_fileConfig/_data/config.local.php"
|
||||
}
|
||||
|
||||
####################
|
||||
# main
|
||||
|
||||
updateEnvDB "etherpad" "${KAZ_KEY_DIR}/env-${etherpadDBName}" "${etherpadDBName}"
|
||||
updateEnvDB "framadate" "${KAZ_KEY_DIR}/env-${framadateDBName}" "${framadateDBName}"
|
||||
updateEnvDB "gitea" "${KAZ_KEY_DIR}/env-${gitDBName}" "${gitDBName}"
|
||||
updateEnvDB "mattermost" "${KAZ_KEY_DIR}/env-${mattermostDBName}" "${mattermostDBName}"
|
||||
updateEnvDB "nextcloud" "${KAZ_KEY_DIR}/env-${nextcloudDBName}" "${nextcloudDBName}"
|
||||
updateEnvDB "roundcube" "${KAZ_KEY_DIR}/env-${roundcubeDBName}" "${roundcubeDBName}"
|
||||
updateEnvDB "sympa" "${KAZ_KEY_DIR}/env-${sympaDBName}" "${sympaDBName}"
|
||||
updateEnvDB "vigilo" "${KAZ_KEY_DIR}/env-${vigiloDBName}" "${vigiloDBName}"
|
||||
updateEnvDB "wp" "${KAZ_KEY_DIR}/env-${wordpressDBName}" "${wordpressDBName}"
|
||||
updateEnvDB "vaultwarden" "${KAZ_KEY_DIR}/env-${vaultwardenDBName}" "${vaultwardenDBName}"
|
||||
updateEnvDB "castopod" "${KAZ_KEY_DIR}/env-${castopodDBName}" "${castopodDBName}"
|
||||
|
||||
updateEnv "apikaz" "${KAZ_KEY_DIR}/env-${apikazServName}"
|
||||
updateEnv "ethercalc" "${KAZ_KEY_DIR}/env-${ethercalcServName}"
|
||||
updateEnv "etherpad" "${KAZ_KEY_DIR}/env-${etherpadServName}"
|
||||
updateEnv "framadate" "${KAZ_KEY_DIR}/env-${framadateServName}"
|
||||
updateEnv "gandi" "${KAZ_KEY_DIR}/env-gandi"
|
||||
updateEnv "gitea" "${KAZ_KEY_DIR}/env-${gitServName}"
|
||||
updateEnv "jirafeau" "${KAZ_KEY_DIR}/env-${jirafeauServName}"
|
||||
updateEnv "mattermost" "${KAZ_KEY_DIR}/env-${mattermostServName}"
|
||||
updateEnv "nextcloud" "${KAZ_KEY_DIR}/env-${nextcloudServName}"
|
||||
updateEnv "office" "${KAZ_KEY_DIR}/env-${officeServName}"
|
||||
updateEnv "roundcube" "${KAZ_KEY_DIR}/env-${roundcubeServName}"
|
||||
updateEnv "vigilo" "${KAZ_KEY_DIR}/env-${vigiloServName}"
|
||||
updateEnv "wp" "${KAZ_KEY_DIR}/env-${wordpressServName}"
|
||||
updateEnv "ldap" "${KAZ_KEY_DIR}/env-${ldapServName}"
|
||||
updateEnv "sympa" "${KAZ_KEY_DIR}/env-${sympaServName}"
|
||||
updateEnv "mail" "${KAZ_KEY_DIR}/env-${smtpServName}"
|
||||
updateEnv "mobilizon" "${KAZ_KEY_DIR}/env-${mobilizonServName}"
|
||||
updateEnv "mobilizon" "${KAZ_KEY_DIR}/env-${mobilizonDBName}"
|
||||
updateEnv "vaultwarden" "${KAZ_KEY_DIR}/env-${vaultwardenServName}"
|
||||
updateEnv "castopod" "${KAZ_KEY_DIR}/env-${castopodServName}"
|
||||
updateEnv "ldap" "${KAZ_KEY_DIR}/env-${ldapUIName}"
|
||||
|
||||
|
||||
framadateUpdate
|
||||
jirafeauUpdate
|
||||
exit 0
|
447
bin/updateGit.sh
Executable file
@ -0,0 +1,447 @@
|
||||
#!/bin/bash
|
||||
|
||||
# l'idée et de faire un rsync dans un répertoir provisoire et de téléverser les différences.
|
||||
|
||||
# initialilisation :
|
||||
# cd /MonRepDeTest
|
||||
# mkdir -p kazdev kazprod
|
||||
# rsync -rlptDEHAX --delete --info=progress2 root@kazdev:/kaz/ ./kazdev/
|
||||
# rsync -rlptDEHAX --delete --info=progress2 root@kazprod:/kaz/ ./kazprod/
|
||||
|
||||
# exemple :
|
||||
# cd /MonRepDeTest/kazdev/
|
||||
# ./dockers/rdiff.sh /MonRepDeTest/kazprod/ root@kazprod
|
||||
# cd /MonRepDeTest/kazprod/
|
||||
# ./dockers/rdiff.sh /MonRepDeTest/kazdev/ root@kazdev
|
||||
|
||||
export KAZ_ROOT=$(cd "$(dirname $0)/.."; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
export REF_DIRS="bin config dockers secret.tmpl"
|
||||
export SIMU=""
|
||||
|
||||
usage () {
|
||||
echo "Usage: $0 [-n] [-h]"
|
||||
echo " -h help"
|
||||
echo " -n simulation"
|
||||
exit 1
|
||||
}
|
||||
|
||||
for ARG in $@
|
||||
do
|
||||
case "${ARG}" in
|
||||
'-h' | '-help' )
|
||||
usage
|
||||
;;
|
||||
'-n' )
|
||||
shift
|
||||
export SIMU="echo"
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [[ $# -ne 0 ]]; then
|
||||
echo "Illegal number of parameters"
|
||||
usage
|
||||
fi
|
||||
|
||||
########################################
|
||||
# check system
|
||||
|
||||
for prg in git ssh rsync kompare; do
|
||||
if ! type "${prg}" > /dev/null; then
|
||||
printKazError "$0 need ${prg}"
|
||||
echo "please run \"apt-get install ${prg}\""
|
||||
exit
|
||||
fi
|
||||
done
|
||||
|
||||
|
||||
########################################
|
||||
# config
|
||||
|
||||
declare -a SKIP_FILE
|
||||
export SKIP_FILE=$(grep -v -e ^# -e ^$ "${KAZ_CONF_DIR}/skip-file.txt")
|
||||
KAZ_CFG_UPDATED=""
|
||||
KAZ_UPDATE_ENV="${KAZ_CONF_DIR}/updateGit.conf"
|
||||
if [ -f "${KAZ_UPDATE_ENV}" ]; then
|
||||
. "${KAZ_UPDATE_ENV}"
|
||||
else
|
||||
KAZ_SRC_TYPE="VAGRANT"
|
||||
KAZ_VAGRANT_ROOT="~/kaz-vagrant"
|
||||
KAZ_DEV_REMOTE="root@kazdev"
|
||||
KAZ_DEV_ROOT="/kaz"
|
||||
KAZ_PROD_REMOTE="root@kazprod"
|
||||
KAZ_PROD_ROOT="/kaz"
|
||||
KAZ_OTHER_REMOTE="192.168.1.1"
|
||||
KAZ_OTHER_ROOT="~/git/kaz"
|
||||
fi
|
||||
|
||||
while : ; do
|
||||
read -p " Form which tested server ([Vagrant|DEV|PROD|OTHER]) you want updaye git KAZ? [${KAZ_SRC_TYPE}]: " rep
|
||||
case "${rep}" in
|
||||
"")
|
||||
break
|
||||
;;
|
||||
[vV]*)
|
||||
KAZ_SRC_TYPE="VAGRANT"
|
||||
KAZ_CFG_UPDATED="true"
|
||||
break
|
||||
;;
|
||||
[dD]*)
|
||||
KAZ_SRC_TYPE="DEV"
|
||||
KAZ_CFG_UPDATED="true"
|
||||
break
|
||||
;;
|
||||
[pP]*)
|
||||
KAZ_SRC_TYPE="PROD"
|
||||
KAZ_CFG_UPDATED="true"
|
||||
break
|
||||
;;
|
||||
[oO]*)
|
||||
KAZ_SRC_TYPE="OTHER"
|
||||
KAZ_CFG_UPDATED="true"
|
||||
break
|
||||
;;
|
||||
* )
|
||||
printKazError "\"${rep}\" not match with [Vagrant|DEV|PROD|OTHER]."
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
case "${KAZ_SRC_TYPE}" in
|
||||
VAGRANT)
|
||||
while : ; do
|
||||
read -p " Give kaz-vagrant root? [${KAZ_VAGRANT_ROOT}]: " vagrantPath
|
||||
if [ -z "${vagrantPath}" ]; then
|
||||
vagrantPath="${KAZ_VAGRANT_ROOT}"
|
||||
else
|
||||
KAZ_CFG_UPDATED="true"
|
||||
fi
|
||||
if [ ! -d "${vagrantPath/#\~/${HOME}}" ]; then
|
||||
printKazError "${vagrantPath} doesn't exist"
|
||||
continue
|
||||
fi
|
||||
KAZ_VAGRANT_ROOT="${vagrantPath}"
|
||||
KAZ_VAGRANT_PAHT="$(cd "${vagrantPath/#\~/${HOME}}" 2>/dev/null; pwd)"
|
||||
(for sign in .git .vagrant; do
|
||||
if [ ! -d "${KAZ_VAGRANT_PAHT}/${sign}" ]; then
|
||||
printKazError "${KAZ_VAGRANT_PAHT} not contains ${sign}"
|
||||
exit 1
|
||||
fi
|
||||
done
|
||||
exit 0
|
||||
) && break;
|
||||
done
|
||||
;;
|
||||
DEV|PROD|OTHER)
|
||||
case "${KAZ_SRC_TYPE}" in
|
||||
DEV)
|
||||
remoteUser="${KAZ_DEV_REMOTE}"
|
||||
remotePath="${KAZ_DEV_ROOT}"
|
||||
;;
|
||||
PROD)
|
||||
remoteUser="${KAZ_PROD_REMOTE}"
|
||||
remotePath="${KAZ_PROD_ROOT}"
|
||||
;;
|
||||
OTHER)
|
||||
remoteUser="${KAZ_OTHER_REMOTE}"
|
||||
remotePath="${KAZ_OTHER_ROOT}"
|
||||
;;
|
||||
esac
|
||||
while : ; do
|
||||
read -p "Give remote access? [${remoteUser}]: " rep
|
||||
case "${rep}" in
|
||||
"" )
|
||||
break
|
||||
;;
|
||||
* )
|
||||
if [[ "${rep}" =~ ^([a-zA-Z0-9._%+-]+@)?[a-zA-Z0-9.-]+$ ]]; then
|
||||
remoteUser="${rep}"
|
||||
break
|
||||
else
|
||||
printKazError "${rep} not match with [user@server]"
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
done
|
||||
while : ; do
|
||||
read -p "Give remote path? [${remotePath}]: " rep
|
||||
case "${rep}" in
|
||||
"" )
|
||||
break
|
||||
;;
|
||||
* )
|
||||
if [[ "${rep}" =~ ^~?[a-zA-Z0-9/._-]*/$ ]]; then
|
||||
remotePath="${rep}"
|
||||
break
|
||||
else
|
||||
printKazError "${rep} not match with [path]"
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
done
|
||||
case "${KAZ_SRC_TYPE}" in
|
||||
DEV)
|
||||
if [ "${remoteUser}" != "${KAZ_DEV_REMOTE}" ]; then
|
||||
KAZ_DEV_REMOTE="${remoteUser}"; KAZ_CFG_UPDATED="true"
|
||||
fi
|
||||
if [ "${remotePath}" != "${KAZ_DEV_ROOT}" ]; then
|
||||
KAZ_DEV_ROOT="${remotePath}"; KAZ_CFG_UPDATED="true"
|
||||
fi
|
||||
;;
|
||||
PROD)
|
||||
if [ "${remoteUser}" != "${KAZ_PROD_REMOTE}" ]; then
|
||||
KAZ_PROD_REMOTE="${remoteUser}"; KAZ_CFG_UPDATED="true"
|
||||
fi
|
||||
if [ "${remotePath}" != "${KAZ_PROD_ROOT}" ]; then
|
||||
KAZ_PROD_ROOT="${remotePath}"; KAZ_CFG_UPDATED="true"
|
||||
fi
|
||||
;;
|
||||
OTHER)
|
||||
if [ "${remoteUser}" != "${KAZ_OTHER_REMOTE}" ]; then
|
||||
KAZ_OTHER_REMOTE="${remoteUser}"; KAZ_CFG_UPDATED="true"
|
||||
fi
|
||||
if [ "${remotePath}" != "${KAZ_OTHER_ROOT}" ]; then
|
||||
KAZ_OTHER_ROOT="${remotePath}"; KAZ_CFG_UPDATED="true"
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
;;
|
||||
esac
|
||||
|
||||
if [ -n "${KAZ_CFG_UPDATED}" ]; then
|
||||
printKazMsg "Update ${KAZ_UPDATE_ENV}"
|
||||
cat > "${KAZ_UPDATE_ENV}" <<EOF
|
||||
KAZ_SRC_TYPE="${KAZ_SRC_TYPE}"
|
||||
KAZ_VAGRANT_ROOT="${KAZ_VAGRANT_ROOT}"
|
||||
KAZ_DEV_REMOTE="${KAZ_DEV_REMOTE}"
|
||||
KAZ_DEV_ROOT="${KAZ_DEV_ROOT}"
|
||||
KAZ_PROD_REMOTE="${KAZ_PROD_REMOTE}"
|
||||
KAZ_PROD_ROOT="${KAZ_PROD_ROOT}"
|
||||
KAZ_OTHER_REMOTE="${KAZ_OTHER_REMOTE}"
|
||||
KAZ_OTHER_ROOT="${KAZ_OTHER_ROOT}"
|
||||
EOF
|
||||
fi
|
||||
|
||||
########################################
|
||||
# check git/kaz
|
||||
|
||||
cd "${KAZ_ROOT}"
|
||||
CURRENT_BRANCH="$(git branch | grep "*")"
|
||||
if [ "${CURRENT_BRANCH}" == "* develop" ]; then
|
||||
printKazMsg "You are on ${CURRENT_BRANCH}."
|
||||
else
|
||||
printKazError "You supposed to be on * develop"
|
||||
checkContinue
|
||||
fi
|
||||
|
||||
if [ "$(git status | grep "git restore")" ]; then
|
||||
echo "You didn't commit previous change."
|
||||
checkContinue
|
||||
|
||||
fi
|
||||
|
||||
########################################
|
||||
# download valide source from Vagrant, DEV, PROD or OTHER
|
||||
|
||||
export TESTED_DIR="${KAZ_ROOT}/tmp/kaz"
|
||||
mkdir -p "${TESTED_DIR}"
|
||||
|
||||
printKazMsg "Download from ${KAZ_SRC_TYPE} to ${TESTED_DIR}"
|
||||
checkContinue
|
||||
case "${KAZ_SRC_TYPE}" in
|
||||
VAGRANT)
|
||||
(
|
||||
echo "check vagrant status (must be launch with vagrant)"
|
||||
cd "${KAZ_VAGRANT_ROOT/#\~/${HOME}}"
|
||||
while :; do
|
||||
if [ -n "$(vagrant status | grep running)" ]; then
|
||||
exit
|
||||
fi
|
||||
printKazMsg "Please start vagrant"
|
||||
checkContinue
|
||||
done
|
||||
)
|
||||
|
||||
printKazMsg check key
|
||||
while grep -q "@@@@@@@@@@" <<<"$(ssh -p 2222 -i ${KAZ_VAGRANT_ROOT/#\~/${HOME}}/.vagrant/machines/default/virtualbox/private_key vagrant@127.0.0.1 date 2>&1 >/dev/null)"; do
|
||||
printKazError "ssh key has changed"
|
||||
echo "you must call :"
|
||||
echo "${YELLOW} ssh-keygen -f ~/.ssh/known_hosts -R \"[127.0.0.1]:2222\"${NC}"
|
||||
checkContinue
|
||||
done
|
||||
# XXX remote root
|
||||
${SIMU} rsync -rlptDEHAX --no-o --delete --info=progress2 \
|
||||
-e "ssh -p 2222 -i ${KAZ_VAGRANT_ROOT/#\~/${HOME}}/.vagrant/machines/default/virtualbox/private_key" \
|
||||
$(for i in ${REF_DIRS} git download ; do echo "vagrant@127.0.0.1:/kaz/$i" ; done) \
|
||||
"${TESTED_DIR}"
|
||||
;;
|
||||
DEV|PROD|OTHER)
|
||||
# remoteUser is already set
|
||||
case "${KAZ_SRC_TYPE}" in
|
||||
DEV)
|
||||
remoteUser="${KAZ_DEV_REMOTE}"; remotePath="${KAZ_DEV_ROOT}"
|
||||
;;
|
||||
PROD)
|
||||
remoteUser="${KAZ_PROD_REMOTE}"; remotePath="${KAZ_PROD_ROOT}"
|
||||
;;
|
||||
OTHER)
|
||||
remoteUser="${KAZ_OTHER_REMOTE}"; remotePath="${KAZ_OTHER_ROOT}"
|
||||
;;
|
||||
esac
|
||||
|
||||
${SIMU} rsync -rlptDEHAX --no-o --delete --info=progress2 \
|
||||
$(for i in ${REF_DIRS} ; do echo "${remoteUser}:${remotePath}$i" ; done) \
|
||||
"${TESTED_DIR}"
|
||||
;;
|
||||
esac
|
||||
|
||||
|
||||
cd "${TESTED_DIR}"
|
||||
|
||||
badName(){
|
||||
[[ -z "$1" ]] && return 0
|
||||
for item in ${SKIP_FILE[@]}; do
|
||||
[[ "$1/" =~ "${item}" ]] && return 0
|
||||
done
|
||||
return 1
|
||||
}
|
||||
|
||||
|
||||
declare -a CHANGED_DIRS
|
||||
CHANGED_DIRS=$(find ${REF_DIRS} -type d ! -exec /bin/test -d "${KAZ_ROOT}/{}" \; -print -prune)
|
||||
for file in ${CHANGED_DIRS[@]}; do
|
||||
if badName "${file}" ; then
|
||||
echo SKIP ${file}
|
||||
continue
|
||||
fi
|
||||
printKazMsg "New dir ${file}"
|
||||
while true; do
|
||||
read -p "Synchronize ${GREEN}${file}/${NC} to ${GREEN}${KAZ_ROOT}/${file}/${NC}? [y/n/i/help]: " yn
|
||||
case $yn in
|
||||
[Yy]*)
|
||||
${SIMU} rsync -rlptDEHAX --info=progress2 "${file}/" "${KAZ_ROOT}/${file}/"
|
||||
(cd "${KAZ_ROOT}" ; git add "${file}/" )
|
||||
break
|
||||
;;
|
||||
""|[Nn]*)
|
||||
break
|
||||
;;
|
||||
[Ii]*)
|
||||
# add to skip
|
||||
echo "${file}" >> "${KAZ_CONF_DIR}/skip-file.txt"
|
||||
break
|
||||
;;
|
||||
*)
|
||||
echo -e \
|
||||
" yes: add all subdir ${file} in git\n" \
|
||||
" no: don't update git\n" \
|
||||
" ignore: never ask this question\n" \
|
||||
" help: print this help"
|
||||
;;
|
||||
esac
|
||||
done
|
||||
done
|
||||
|
||||
declare -a NEW_FILES
|
||||
NEW_FILES=$(find ${REF_DIRS} '(' -type d ! -exec /bin/test -d "${KAZ_ROOT}/{}" \; -prune ')' -o '(' -type f ! -exec /bin/test -f "${KAZ_ROOT}/{}" \; -print ')')
|
||||
for file in ${NEW_FILES[@]}; do
|
||||
if badName "${file}" ; then
|
||||
echo SKIP ${file}
|
||||
continue
|
||||
fi
|
||||
echo "New file ${file}"
|
||||
while true; do
|
||||
read -p "Synchronize ${GREEN}${file}${NC} to ${GREEN}${KAZ_ROOT}/${file}${NC}? [y/n/i/help]: " yn
|
||||
case $yn in
|
||||
[Yy]*)
|
||||
${SIMU} rsync -rlptDEHAX --info=progress2 "${file}" "${KAZ_ROOT}/${file}"
|
||||
(cd "${KAZ_ROOT}" ; git add "${file}" )
|
||||
break
|
||||
;;
|
||||
[Nn]*)
|
||||
break
|
||||
;;
|
||||
[Ii]*)
|
||||
# add to skip
|
||||
echo "${file}" >> "${KAZ_CONF_DIR}/skip-file.txt"
|
||||
break
|
||||
;;
|
||||
*)
|
||||
echo -e \
|
||||
" yes: add all subdir ${file} in git\n" \
|
||||
" no: don't update git\n" \
|
||||
" ignore: never ask this question\n" \
|
||||
" help: print this help"
|
||||
;;
|
||||
esac
|
||||
done
|
||||
done
|
||||
|
||||
trap 'rm -f "${TMPFILE}"' EXIT
|
||||
export TMPFILE="$(mktemp)" || exit 1
|
||||
|
||||
CHANGED_FILES=$(find ${REF_DIRS} '(' -type d ! -exec /bin/test -d "${KAZ_ROOT}/{}" \; -prune ')' -o '(' -type f -exec /bin/test -f "${KAZ_ROOT}/{}" \; ! -exec cmp -s "{}" "${KAZ_ROOT}/{}" \; -print ')')
|
||||
for file in ${CHANGED_FILES[@]} ; do
|
||||
if badName "${file}" ; then
|
||||
echo SKIP ${file}
|
||||
continue
|
||||
fi
|
||||
echo "TEST ${file}"
|
||||
kompare "${file}" "${KAZ_ROOT}/${file}"
|
||||
if [ "${KAZ_ROOT}/${file}" -ot "${TMPFILE}" ]; then
|
||||
echo "No change of ${KAZ_ROOT}/${file}"
|
||||
continue
|
||||
fi
|
||||
chmod --reference="${file}" "${KAZ_ROOT}/${file}"
|
||||
done
|
||||
|
||||
echo
|
||||
while : ; do
|
||||
read -p "Do you want to keep ${TESTED_DIR} to speed up next download? [yes]" rep
|
||||
case "${rep}" in
|
||||
""|[yYoO]* )
|
||||
break
|
||||
;;
|
||||
[Nn]* )
|
||||
rm -r "${TESTED_DIR}"
|
||||
;;
|
||||
* )
|
||||
echo "Please answer yes no."
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
cd "${KAZ_ROOT}"
|
||||
echo -e "\nThe git will now commit in ${CURRENT_BRANCH}"
|
||||
checkContinue
|
||||
git commit -a
|
||||
|
||||
echo -e "\nThe git will now pull in ${CURRENT_BRANCH}"
|
||||
checkContinue
|
||||
git pull
|
||||
|
||||
printKazError "\nCheck if any confict"
|
||||
#XXX check conflict
|
||||
|
||||
echo -e "\nThe git will now push in ${CURRENT_BRANCH}"
|
||||
checkContinue
|
||||
git push
|
||||
|
||||
printKazMsg "\nYou have to logged on ${KAZ_SRC_TYPE}, and launch:\n"
|
||||
echo -e \
|
||||
" ssh root@host\n" \
|
||||
" cd /kaz\n" \
|
||||
" git reset --hard\n" \
|
||||
" git pull"
|
||||
|
||||
printKazMsg "\nIf you want to promote in master branch:\n" \
|
||||
echo -e \
|
||||
" git checkout master\n" \
|
||||
" git pull\n" \
|
||||
" git merge develop\n" \
|
||||
"${RED}check if confict${NC}\n" \
|
||||
" git commit -a \n" \
|
||||
" git push\n"
|
36
bin/updateLook.sh
Executable file
@ -0,0 +1,36 @@
|
||||
#!/bin/bash
|
||||
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
cd "${KAZ_BIN_DIR}/look"
|
||||
|
||||
THEMES=$(ls -F -- '.' | grep '/$' | sed 's%/%%' | tr '\n' '|' | sed 's%|$%%')
|
||||
|
||||
usage () {
|
||||
echo "usage $0 {${THEMES}}"
|
||||
exit
|
||||
}
|
||||
|
||||
[ -z "$1" ] && usage
|
||||
[ -d "$1" ] || usage
|
||||
|
||||
cd $1
|
||||
|
||||
docker cp kaz-tete.png jirafeauServ:/var/jirafeau/media/kaz/kaz.png
|
||||
docker cp kazdate.png framadateServ:/var/framadate/images/logo-framadate.png
|
||||
docker cp kazmel.png roundcubeServ:/var/www/html/skins/elastic/images/kazmel.png
|
||||
docker cp kaz-tete.png sympaServ:/usr/share/sympa/static_content/icons/logo_sympa.png
|
||||
docker cp kaz-tete.png dokuwikiServ:/dokuwiki/lib/tpl/docnavwiki/images/logo.png
|
||||
docker cp kaz-tete.png ldapUI:/var/www/html/images/ltb-logo.png
|
||||
docker cp kaz-entier.svg webServ:/usr/share/nginx/html/images/logo.svg
|
||||
docker cp kaz-signature.png webServ:/usr/share/nginx/html/m/logo.png
|
||||
|
||||
for cloud in nextcloudServ kaz-nextcloudServ; do
|
||||
docker cp kaz-entier.svg "${cloud}":/var/www/html/themes/kaz-entier.svg
|
||||
docker cp kaz-tete.svg "${cloud}":/var/www/html/themes/kaz-tete.svg
|
||||
docker exec -ti -u 33 "${cloud}" /var/www/html/occ theming:config logo /var/www/html/themes/kaz-tete.svg # tete
|
||||
docker exec -ti -u 33 "${cloud}" /var/www/html/occ theming:config logoheader /var/www/html/themes/kaz-entier.svg # entier
|
||||
# non #exec -ti -u 33 "${cloud}" /var/www/html/occ theming:config favicon /var/www/html/themes/kaz-patte.svg # patte
|
||||
done
|
13
bin/upgradeDockerCompose.sh
Executable file
@ -0,0 +1,13 @@
|
||||
#!/bin/bash
|
||||
|
||||
OLDVERSION=$(docker-compose -v | sed -En 's/.*version ([a-z0-9\.]*).*/\1/p')
|
||||
DOCKERCOMPOSE_VERSION="v2.17.3"
|
||||
|
||||
if [ "$OLDVERSION" = "$DOCKERCOMPOSE_VERSION" ]
|
||||
then
|
||||
echo -e "Docker Compose déjà en version $DOCKERCOMPOSE_VERSION"
|
||||
exit
|
||||
fi
|
||||
|
||||
curl -SL https://github.com/docker/compose/releases/download/$DOCKERCOMPOSE_VERSION/docker-compose-linux-x86_64 -o /usr/local/bin/docker-compose
|
||||
chmod +x /usr/local/bin/docker-compose
|
72
bin/verifExistenceMails.sh
Executable file
@ -0,0 +1,72 @@
|
||||
#!/bin/bash
|
||||
|
||||
#Koi: on vérifie que chaque email possède son répertoire et vice et versa (on supprime sinon)
|
||||
#Kan: 20/06/2022
|
||||
#Ki: fab
|
||||
|
||||
#on récupère toutes les variables et mdp
|
||||
# on prend comme source des repertoire le dossier du dessus ( /kaz dans notre cas )
|
||||
KAZ_ROOT=$(cd "$(dirname $0)"/..; pwd)
|
||||
. "${KAZ_ROOT}/bin/.commonFunctions.sh"
|
||||
setKazVars
|
||||
|
||||
cd $(dirname $0)/..
|
||||
. "${DOCKERS_ENV}"
|
||||
. "${KAZ_KEY_DIR}/SetAllPass.sh"
|
||||
|
||||
DOCK_DIR=$KAZ_COMP_DIR
|
||||
|
||||
SETUP_MAIL="docker exec -ti mailServ setup"
|
||||
|
||||
#on détermine le script appelant, le fichier log et le fichier source, tous issus de la même racine
|
||||
PRG=$(basename $0)
|
||||
RACINE=${PRG%.sh}
|
||||
|
||||
# emails et les alias KAZ déjà créés
|
||||
TFILE_EMAIL=$(mktemp /tmp/test_email.XXXXXXXXX.TFILE_EMAIL)
|
||||
|
||||
#on stocke les emails et alias déjà créés
|
||||
(
|
||||
${SETUP_MAIL} email list | cut -d ' ' -f 2 | grep @
|
||||
${SETUP_MAIL} alias list | cut -d ' ' -f 2 | grep @
|
||||
) > ${TFILE_EMAIL}
|
||||
|
||||
|
||||
#did on supprime le ^M en fin de fichier pour pas faire planter les grep
|
||||
sed -i -e 's/\r//g' ${TFILE_EMAIL}
|
||||
|
||||
rep_email="/var/lib/docker/volumes/postfix_mailData/_data"
|
||||
|
||||
#étape n°1: pour chaque répertoire, on vérifie que l'email existe
|
||||
echo "Début Etape n°1: on liste les répertoires des emails et on vérifie que les emails correspondant existent"
|
||||
|
||||
ls -Ifilter -Itmp ${rep_email} | while read fin_email; do
|
||||
ls ${rep_email}/${fin_email} | while read debut_email; do
|
||||
email=`echo ${debut_email}@${fin_email}`
|
||||
#est-ce que l'email existe ?
|
||||
nb_ligne=$(grep "^${email}$" ${TFILE_EMAIL} | wc -l)
|
||||
if [ ${nb_ligne} -gt 0 ];then
|
||||
false
|
||||
else
|
||||
#suppression du répertoire
|
||||
echo rm ${rep_email}/${fin_email}/${debut_email} -rf
|
||||
fi
|
||||
done
|
||||
#si le répertoire domaine est vide, on le supprime
|
||||
find ${rep_email}/${fin_email} -maxdepth 0 -type d -empty -delete
|
||||
done
|
||||
echo "aucune commande n'a été lancée, possible de le faire à la main"
|
||||
echo "Fin Etape n°1"
|
||||
|
||||
#Etape n°2: pour chaque email, on vérifie que le répertoire existe
|
||||
echo "Début Etape n°2 n°2: on liste les emails et on vérifie que les répertoires correspondant existent"
|
||||
cat ${TFILE_EMAIL} | while read email; do
|
||||
debut_email=$(echo ${email} | awk -F '@' '{print $1}')
|
||||
fin_email=$(echo ${email} | awk -F '@' '{print $2}')
|
||||
if [ -d ${rep_email}/${fin_email}/${debut_email} ];then
|
||||
true
|
||||
else
|
||||
echo "Attention, le répertoire ${fin_email}/${debut_email} n'existe pas alors que l'email existe!"
|
||||
fi
|
||||
done
|
||||
echo "Fin Etape n°2"
|
11
bin/vide_poubelle
Executable file
@ -0,0 +1,11 @@
|
||||
#!/bin/sh
|
||||
|
||||
cd "${HOME}/tmp/POUBELLE" >/dev/null 2>&1
|
||||
|
||||
if test "$?" -eq 0
|
||||
then
|
||||
rm -f * .* 2>/dev/null
|
||||
else
|
||||
echo "$0 pas de poubelle a vider !"
|
||||
fi
|
||||
|