forked from akash-network/cosmos-omnibus
-
Notifications
You must be signed in to change notification settings - Fork 4
/
snapshot.sh
executable file
·156 lines (144 loc) · 7.16 KB
/
snapshot.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
#!/bin/bash
set -e
[ "$DEBUG" == "2" ] && set -x
SNAPSHOT_TIME="${SNAPSHOT_TIME:-00:00:00}"
SNAPSHOT_DAY="${SNAPSHOT_DAY:-*}"
SNAPSHOT_DIR="${SNAPSHOT_DIR:-$PROJECT_ROOT/data}"
SNAPSHOT_CMD="${SNAPSHOT_CMD:-$@}"
SNAPSHOT_PATH="${SNAPSHOT_PATH%/}"
SNAPSHOT_PREFIX="${SNAPSHOT_PREFIX:-$CHAIN_ID}"
SNAPSHOT_RETAIN="${SNAPSHOT_RETAIN:-2 days}"
SNAPSHOT_METADATA="${SNAPSHOT_METADATA:-1}"
SNAPSHOT_SAVE_FORMAT="${SNAPSHOT_SAVE_FORMAT:-$SNAPSHOT_FORMAT}"
valid_snapshot_formats=(tar tar.gz tar.zst)
# If not one of valid format values, set it to default value
if ! echo "${valid_snapshot_formats[@]}" | grep -qiw -- "$SNAPSHOT_SAVE_FORMAT"; then
SNAPSHOT_SAVE_FORMAT=tar.gz
fi
# Actual valid values not documented
# 27 is default value
# 31 is max value mentioned in project issues
# Since value > 27 requires special handling on decompression
# Only 27 is allowed at the moment when enabled
# See https://github.com/facebook/zstd/blob/v1.5.2/programs/zstd.1.md for more info
valid_zstd_long_values=(27)
# If non empty string and invalid value detected
# Set to default value assuming long should be enabled
if [ -n "$ZSTD_LONG" ] && ! echo "${valid_zstd_long_values[@]}" | grep -qiw -- "$ZSTD_LONG"; then
ZSTD_LONG=27
fi
zstd_extra_args=""
if [ -n "$ZSTD_LONG" ]; then
zstd_extra_arg="--long=$ZSTD_LONG"
fi
TIME=$(date -u +%T)
DOW=$(date +%u)
echo "$TIME: Starting server"
echo "$TIME: Snapshot will run at $SNAPSHOT_TIME on day $SNAPSHOT_DAY"
exec $SNAPSHOT_CMD &
PID=$!
while true; do
TIME=$(date -u +%T)
DOW=$(date +%u)
if [[ ($SNAPSHOT_DAY == "*") || ($SNAPSHOT_DAY == $DOW) ]] && [[ $SNAPSHOT_TIME == $TIME ]]; then
echo "$TIME: Stopping server"
kill -15 $PID
wait
echo "$TIME: Running snapshot"
aws_args="--endpoint-url ${S3_HOST}"
storj_args="${STORJ_UPLINK_ARGS:--p 4 --progress=false}"
s3_uri_base="s3://${SNAPSHOT_PATH}"
storj_uri_base="sj://${SNAPSHOT_PATH}"
timestamp=$(date +"%Y-%m-%dT%H:%M:%S")
s3_uri="${s3_uri_base}/${SNAPSHOT_PREFIX}_${timestamp}.${SNAPSHOT_SAVE_FORMAT}"
storj_uri="${storj_uri_base}/${SNAPSHOT_PREFIX}_${timestamp}.${SNAPSHOT_SAVE_FORMAT}"
SNAPSHOT_SIZE=$(du -sb $SNAPSHOT_DIR | cut -f1)
if [ -n "$STORJ_ACCESS_GRANT" ]; then
case "${SNAPSHOT_SAVE_FORMAT,,}" in
tar.gz) (tar c -C $SNAPSHOT_DIR . | gzip -1 | pv -petrafb -i 5 -s $SNAPSHOT_SIZE | uplink cp $storj_args - "$storj_uri") 2>&1 | stdbuf -o0 tr '\r' '\n';;
# Compress level can be set via `ZSTD_CLEVEL`, default `3`
# No. of threads can be set via `ZSTD_NBTHREADS`, default `1`, `0` = detected no. of cpu cores
tar.zst) (tar c -C $SNAPSHOT_DIR . | zstd -c $zstd_extra_arg | pv -petrafb -i 5 -s $SNAPSHOT_SIZE | uplink cp $storj_args - "$storj_uri") 2>&1 | stdbuf -o0 tr '\r' '\n';;
# Catchall, assume to be tar
*) (tar c -C $SNAPSHOT_DIR . | pv -petrafb -i 5 -s $SNAPSHOT_SIZE | uplink cp $storj_args - "$storj_uri") 2>&1 | stdbuf -o0 tr '\r' '\n';;
esac
else
case "${SNAPSHOT_SAVE_FORMAT,,}" in
tar.gz) (tar c -C $SNAPSHOT_DIR . | gzip -1 | pv -petrafb -i 5 -s $SNAPSHOT_SIZE | aws $aws_args s3 cp - "$s3_uri" --expected-size $SNAPSHOT_SIZE) 2>&1 | stdbuf -o0 tr '\r' '\n';;
# Compress level can be set via `ZSTD_CLEVEL`, default `3`
# No. of threads can be set via `ZSTD_NBTHREADS`, default `1`, `0` = detected no. of cpu cores
tar.zst) (tar c -C $SNAPSHOT_DIR . | zstd -c $zstd_extra_arg | pv -petrafb -i 5 -s $SNAPSHOT_SIZE | aws $aws_args s3 cp - "$s3_uri" --expected-size $SNAPSHOT_SIZE) 2>&1 | stdbuf -o0 tr '\r' '\n';;
# Catchall, assume to be tar
*) (tar c -C $SNAPSHOT_DIR . | pv -petrafb -i 5 -s $SNAPSHOT_SIZE | aws $aws_args s3 cp - "$s3_uri" --expected-size $SNAPSHOT_SIZE) 2>&1 | stdbuf -o0 tr '\r' '\n';;
esac
fi
if [[ $SNAPSHOT_RETAIN != "0" || $SNAPSHOT_METADATA != "0" ]]; then
if [ -n "$STORJ_ACCESS_GRANT" ]; then
SNAPSHOT_METADATA_URL=$(uplink share --url --not-after=none ${storj_uri_base}/ | grep ^URL | awk '{print $NF}')
readarray -t s3Files < <(uplink ls ${storj_uri_base}/ | grep "${SNAPSHOT_PREFIX}_" | awk '{print $2,$3,$4,$5}' | sort -d -k4,4)
else
readarray -t s3Files < <(aws $aws_args s3 ls "${s3_uri_base}/${SNAPSHOT_PREFIX}_")
fi
snapshots=()
for line in "${s3Files[@]}"; do
createDate=`echo $line|awk {'print $1" "$2'}`
createDate=`date -d"$createDate" +%s`
fileName=`echo $line|awk '{$1=$2=$3=""; print $0}' | sed 's/^[ \t]*//'`
if [[ -n $SNAPSHOT_METADATA_URL && $SNAPSHOT_METADATA_URL != */ ]]; then
fileUrl="${SNAPSHOT_METADATA_URL}/${fileName}"
else
fileUrl="${SNAPSHOT_METADATA_URL}${fileName}"
fi
if [ -n "$STORJ_ACCESS_GRANT" ]; then
fileUrl="${fileUrl}?download=1"
fi
if [ "$SNAPSHOT_RETAIN" != "0" ]; then
olderThan=`date -d"-$SNAPSHOT_RETAIN" +%s`
if [[ $createDate -lt $olderThan ]]; then
if [[ $fileName != "" ]]; then
echo "$TIME: Deleting snapshot $fileName"
if [ -n "$STORJ_ACCESS_GRANT" ]; then
uplink rm "${storj_uri_base}/$fileName"
else
aws $aws_args s3 rm "${s3_uri_base}/$fileName"
fi
fi
else
snapshots+=("$fileUrl")
fi
else
snapshots+=("$fileUrl")
fi
done;
if [ "$SNAPSHOT_METADATA" != "0" ]; then
echo "$TIME: Uploading metadata"
snapshotJson="[]"
for url in ${snapshots[@]}; do
snapshotJson="$(echo $snapshotJson | jq ".+[\"$url\"]")"
done
if [ -n "$STORJ_ACCESS_GRANT" ]; then
echo $snapshotJson | jq '{chain_id: $c, snapshots: ., latest: $l}' \
--arg c "$CHAIN_ID" --arg l "${snapshots[-1]}" | \
uplink cp - "${storj_uri_base}/snapshot.json"
echo "=== Use the following as SNAPSHOT_JSON to restore the DCS Storj backup ==="
##uplink share --url --not-after=none "${storj_uri_base}/snapshot.json" | grep ^URL | awk '{print $NF"?download=1"}'
echo "${SNAPSHOT_METADATA_URL%/}/snapshot.json?download=1"
echo "=== === ==="
else
echo $snapshotJson | jq '{chain_id: $c, snapshots: ., latest: $l}' \
--arg c "$CHAIN_ID" --arg l "${snapshots[-1]}" | \
aws $aws_args s3 cp - "${s3_uri_base}/snapshot.json"
fi
fi
fi
echo "$TIME: Restarting server"
exec $SNAPSHOT_CMD &
PID=$!
sleep 1s
else
if ! kill -0 $PID; then
echo "$TIME: Process has died. Exiting"
break;
fi
fi
done