forked from allada/bsc-archive-snapshot
-
Notifications
You must be signed in to change notification settings - Fork 0
/
build_archive_node.sh
executable file
·539 lines (465 loc) · 20.7 KB
/
build_archive_node.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
#!/bin/bash
# Copyright 2021 Nathan (Blaise) Bruer
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.W
set -euxo pipefail
if [[ $EUID -ne 0 ]]; then
echo "This script must be run as root"
exit 1
fi
# If set to "1", will create a crontab entry to upload a snapshot daily.
# This requires write permission to `s3://public-blockchain-snapshots`.
# You may also set this through an environmental variable at startup.
# SHOULD_AUTO_UPLOAD_SNAPSHOT="0"
# Below is just a comment. It is done this way to make copy and paste much easier.
# These are the commands used to create an image of a snapshot node.
# When it's done the instance will reboot.
# Make sure you launch the instance with the EC2 tag attached to the instance of:
# `no-launch` set to something and the `Allow tags in metadata` flag checked.
cat <<EOF > /dev/null
sudo sh -c 'curl https://raw.githubusercontent.com/allada/bsc-archive-snapshot/master/build_archive_node.sh > /home/ubuntu/build_archive_node.sh'
sudo sh -c "echo \"@reboot root sh -c 'curl --fail http://169.254.169.254/latest/meta-data/tags/instance/no-launch || SHOULD_AUTO_UPLOAD_SNAPSHOT=1 /home/ubuntu/build_archive_node.sh || shutdown +5 now'\" >> /etc/crontab"
sudo chmod +x /home/ubuntu/build_archive_node.sh
sudo CREATE_SNAPSHOT_MODE=1 /home/ubuntu/build_archive_node.sh
EOF
function safe_wait() {
BACKGROUND_PIDS=( $(jobs -p) )
for PID in "${BACKGROUND_PIDS[@]}"; do
wait -f $PID
done
}
# This is pretty much the same as: `aws s3 sync s3://foo/bar /foo/bar`, but with better
# parallelization.
function parallel_sync_download() {
set -euo pipefail
full_s3_path=$1
shift
local_path=$1
shift
s3_bucket=$(echo "$full_s3_path" | cut -d'/' -f3)
s3_path="${full_s3_path#s3://$s3_bucket/}"
num_cores=$(nproc)
# This is an inverse log10(). The lower the number of cores you have the more processes you'll
# spawn. The logic here is that on less powerful machines you'll almost certainly want more
# than 1 download going on at a time. The same in reverse, on 128 core machines, you will be
# limited by network instead of cpu ability. 128 cores = 89 jobs, 64 cores = 56 jobs,
# 32 cores = 36 jobs, 16 cores = 25 jobs, 4 cores = 15 jobs, exc...
parallel_count=$(echo "x = $num_cores / (l(($num_cores + 8) / 8) / l(10)); scale=0; x / 1" | bc -l)
set +x # Reduces the noise of commands being generated.
# Download all the individual files from aws, decompress them, then place them into the snapshots
# folder. This is similar to running:
# aws s3 sync --request-payer=requester s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/snapshots/ /erigon/data/bsc/snapshots/
# The major difference is that it will, while downloading, decompress each file.
commands=()
for aws_path in $(aws s3 ls --request-payer=requester --recursive "$full_s3_path" | tr -s ' ' ' ' | cut -d' ' -f4); do
relative_path=$(dirname "${aws_path#$s3_path}")
file=$(basename $aws_path)
read_remote_file="aws s3 cp --request-payer=requester s3://$s3_bucket/$aws_path -"
mkdir -p "$local_path/$relative_path"
decompress="pzstd -d -q --stdout"
save_to_file="cat > $local_path/$relative_path/${file%.zstd}"
commands+=("sh -c '$read_remote_file | $decompress | $save_to_file'")
done
( IFS=$'\n'; echo "${commands[*]}" ) | \
pjoin --parallel-count $parallel_count
}
function install_prereq() {
set -euxo pipefail
# Basic installs.
apt update
DEBIAN_FRONTEND=noninteractive apt install -y zfsutils-linux unzip pv clang-12 make jq python3-boto3 super
# Use clang as our compiler by default if needed.
ln -s $(which clang-12) /usr/bin/cc || true
snap install --classic go
if ! cargo --version 2>&1 >/dev/null ; then
# Install cargo.
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | bash /dev/stdin -y
source "$HOME/.cargo/env"
fi
}
function setup_drives() {
set -euxo pipefail
if zfs list tank ; then
return # Our drives are probably already setup.
fi
# Creates a new pool with the default device.
DEVICES=( $(lsblk --fs --json | jq -r '.blockdevices[] | select(.children == null and .fstype == null) | .name') )
DEVICES_FULLNAME=()
for DEVICE in "${DEVICES[@]}"; do
DEVICES_FULLNAME+=("/dev/$DEVICE")
done
zpool create -o ashift=12 tank "${DEVICES_FULLNAME[@]}"
# The root tank dataset does not get mounted.
zfs set mountpoint=none tank
# Configures ZFS to be slightly more optimal for our use case.
zfs set compression=lz4 tank
# Note: You might be able to get better erigon performance by changing this to 16k.
zfs set recordsize=128k tank
zfs set sync=disabled tank
zfs set redundant_metadata=most tank
zfs set atime=off tank
zfs set logbias=throughput tank
# By creating a swap it won't hurt much unless it's running on a small instance.
# Under rare cases erigon might want to use an insane amount of ram (like if parlia database is
# missing). This will allow us to at least get beyond that point. Measuring shows it only uses
# about 48gb of ram when this happens. The vast majority of the time the swap will not be used.
zfs create -s -V 48G -b $(getconf PAGESIZE) \
-o compression=zle \
-o sync=always \
-o primarycache=metadata \
-o secondarycache=none \
tank/swap
sleep 3 # It takes a moment for our zvol to be created.
mkswap -f /dev/zvol/tank/swap
swapon /dev/zvol/tank/swap
# Set zfs's arc to 4GB. Erigon uses mmap() to map files into memory which is a cache system itself.
echo 4147483648 > /sys/module/zfs/parameters/zfs_arc_max
}
function install_zstd() {
set -euxo pipefail
if pzstd --help ; then
return # pzstd is already installed.
fi
# Download, setup and install zstd v1.5.2.
# We use an upgraded version rather than what ubuntu uses because
# 1.5.0+ greatly improved performance (3-5x faster for compression/decompression).
mkdir -p /zstd
cd /zstd
wget -q -O- https://github.com/facebook/zstd/releases/download/v1.5.2/zstd-1.5.2.tar.gz | tar xzf -
cd /zstd/zstd-1.5.2
CC=clang-12 CXX=clang++-12 CFLAGS="-O3" make zstd -j$(nproc)
ln -s /zstd/zstd-1.5.2/zstd /usr/bin/zstd || true
cd /zstd/zstd-1.5.2/contrib/pzstd
CC=clang-12 CXX=clang++-12 CFLAGS="-O3" make pzstd -j$(nproc)
rm -rf /usr/bin/pzstd || true
ln -s /zstd/zstd-1.5.2/contrib/pzstd/pzstd /usr/bin/pzstd
}
function install_aws_cli() {
set -euxo pipefail
if aws --version ; then
return # Aws cli already installed.
fi
temp_dir=$(mktemp -d)
cd $temp_dir
curl "https://awscli.amazonaws.com/awscli-exe-linux-$(uname -m).zip" -o "awscliv2.zip"
unzip awscliv2.zip
./aws/install
cd /
rm -rf $temp_dir
ln -s /usr/local/bin/aws /usr/bin/aws
}
function install_s3pcp() {
set -euxo pipefail
if s3pcp --help ; then
return # putils already installed.
fi
temp_dir=$(mktemp -d)
trap 'rm -rf $temp_dir' EXIT
cd $temp_dir
git clone https://github.com/allada/s3pcp.git
cd $temp_dir/s3pcp
make s3pcp
}
function install_putils() {
set -euxo pipefail
if psplit --help && pjoin --help ; then
return # `putils` already installed.
fi
temp_dir=$(mktemp -d)
trap 'rm -rf $temp_dir' EXIT
cd $temp_dir
git clone https://github.com/allada/putils.git
cd $temp_dir/putils/psplit
cargo build --release &
cd $temp_dir/putils/pjoin
cargo build --release &
safe_wait
mv $temp_dir/putils/psplit/target/release/psplit /usr/bin/psplit
mv $temp_dir/putils/pjoin/target/release/pjoin /usr/bin/pjoin
}
function install_erigon() {
set -euxo pipefail
if erigon --help ; then
return; # Erigon already installed.
fi
# Download, setup and install erigon.
mkdir -p /erigon
cd /erigon
git clone https://github.com/ledgerwatch/erigon.git
cd /erigon/erigon
git checkout v2.30.0
CC=clang-12 CXX=clang++-12 CFLAGS="-O3" make erigon
ln -s /erigon/erigon/build/bin/erigon /usr/bin/erigon
# Stop the service if it exists.
systemctl stop erigon-bsc || true
}
function prepare_zfs_datasets() {
set -euxo pipefail
# Create datasets if needed.
zfs create -o mountpoint=/erigon/data tank/erigon_data || true
zfs create -o mountpoint=/erigon/data/bsc tank/erigon_data/bsc || true
}
function download_snapshots() {
set -euxo pipefail
if ! zfs list tank/erigon_data/bsc/snapshots ; then
# Setup zfs dataset and download the latest erigon snapshots into it if needed.
zfs create -o mountpoint=/erigon/data/bsc/snapshots tank/erigon_data/bsc/snapshots
fi
mkdir -p /erigon/data/bsc/snapshots/
parallel_sync_download s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/snapshots/ /erigon/data/bsc/snapshots/
# We then need to touch each .idx file. This is because erigon needs each .idx file to have an
# mtime greater than the .seq file.
find /erigon/data/bsc/snapshots/ -type f -name "*.idx" -exec touch {} \;
}
# This is not strictly required, but it will make it much faster for a node to join the pool.
function download_nodes() {
set -euxo pipefail
if ! zfs list tank/erigon_data/bsc/nodes ; then
zfs create -o mountpoint=/erigon/data/bsc/nodes tank/erigon_data/bsc/nodes
fi
# This command is allowed to fail.
parallel_sync_download s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/nodes/ /erigon/data/bsc/nodes/nodes/ || true
}
# This is not strictly required, but it will make it much faster to start the Execution phase
# because it will be able to find a snapshot in the parlia database. If this fails it just means
# it'll have to traverse many blocks backwards which takes an insane amount of ram and time (usually
# a few hours).
function download_parlia() {
set -euxo pipefail
if ! zfs list tank/erigon_data/bsc/parlia ; then
zfs create -o mountpoint=/erigon/data/bsc/parlia tank/erigon_data/bsc/parlia
fi
parallel_sync_download s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/parlia/ /erigon/data/bsc/parlia/ || true
}
# This complicated bit of code accomplishes 2 goals.
# 1. In the event that the current file being downloaded gets updated while a user is
# downloading the file, this configuration will be pinned to a specific version,
# so it won't get interrupted in the middle of the download unless it takes over
# ~24 hours.
# 2. Downloads many parts at a time and runs a parallelized decompressor. This is
# about 3-4x faster than using normal `aws s3 cp` + `zstd -d`.
function download_database_file() {
set -euxo pipefail
if zfs list tank/erigon_data/bsc/chaindata ; then
return # Already have chaindata.
fi
zfs create -o mountpoint=/erigon/data/bsc/chaindata tank/erigon_data/bsc/chaindata
s3pcp --requester-pays s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/chaindata/mdbx.dat.zstd \
| pv \
| pzstd -p $(nproc) -q -d -f -o /erigon/data/bsc/chaindata/mdbx.dat
}
function prepare_erigon() {
set -euxo pipefail
# Create erigon user if needed.
useradd erigon || true
chown -R erigon:erigon /erigon/data/
# Stop the service if it exists.
systemctl stop erigon-bsc || true
echo '[Unit]
Description=Erigon BSC daemon
After=network.target
[Service]
Type=simple
Restart=always
RestartSec=1
User=erigon
ExecStart=/erigon/start_erigon_service.sh
[Install]
WantedBy=multi-user.target
' > /etc/systemd/system/erigon-bsc.service
echo '#!/bin/bash' > /erigon/start_erigon_service.sh
if [[ "${SHOULD_AUTO_UPLOAD_SNAPSHOT:-}" == "1" ]]; then
# Run erigon in a subshell but append the subshell's process id to first stdout.
# Also give special environ that tells erigon to stop processing new blocks after Finish stage.
echo -n "sh -c '>&2 echo \$\$; STOP_AFTER_STAGE=Finish " >> /erigon/start_erigon_service.sh
fi
echo -n "exec erigon --chain bsc --snapshots=true --db.pagesize=16k --datadir=/erigon/data/bsc --txpool.disable" >> /erigon/start_erigon_service.sh
if [[ "${SHOULD_AUTO_UPLOAD_SNAPSHOT:-}" == "1" ]]; then
# Create a subshell that will get the process id of the erigon and forward stdout as it comes in
# but if it has the magic string 'STOP_AFTER_STAGE env flag forced to stop app' it will shutdown
# erigon and start a snapshot.
# Sadly erigon does not properly shutdown everything when the flag is set, it only stops
# processing new blocks.
echo "' 2> >(PARENT_PID=\$(head -1); while read -r line; do echo >&2 \"\$line\"; if [[ \"\$line\" == *'STOP_AFTER_STAGE env flag forced to stop app'* ]]; then kill \$PARENT_PID || true; fi done)" >> /erigon/start_erigon_service.sh
echo 'super create-bsc-snapshot-and-shutdown' >> /erigon/start_erigon_service.sh
fi
chmod +x /erigon/start_erigon_service.sh
systemctl daemon-reload
systemctl enable erigon-bsc
}
function run_erigon() {
set -euxo pipefail
systemctl start erigon-bsc
}
function add_create_snapshot_script() {
set -euxo pipefail
# Create script that can be used to upload a snapshot quickly.
cat <<'EOT' > /erigon/create-bsc-snapshot-and-shutdown.sh
#!/bin/bash
set -ex
export PATH="$PATH:/usr/sbin"
# Once the snapshot is created shutdown our node.
# A cronjob should start up the node again.
trap 'shutdown now' EXIT
function upload_mdbx_file() {
upload_id=$(aws s3api create-multipart-upload \
--bucket public-blockchain-snapshots \
--key bsc/erigon/archive/latest/v1/chaindata/mdbx.dat.zstd \
--request-payer requester \
| jq -r ".UploadId")
bytes_per_chunk=$(( 1024 * 1024 * 512 )) # 500mib.
avail_mem_kb=$(grep MemAvailable /proc/meminfo | awk '{print $2}')
# Reduce the theoretical size by about 60% because there are 2 copies in memory at all times.
parallel_downloads=$(( avail_mem_kb * 1000 / bytes_per_chunk * 10 / 25 ))
num_cores=$(nproc)
# We want more than the number of cores but not by a lot.
max_parallel_downloads=$(echo "x = ($num_cores + 5) * 1.5; scale=0; x / 1" | bc -l)
if [ $parallel_downloads -gt $max_parallel_downloads ]; then
parallel_downloads=$max_parallel_downloads
fi
mkdir -p /erigon_upload_tmp
mount -t tmpfs -o rw,size=$(( parallel_downloads * bytes_per_chunk + 1024 * 1024 )) tmpfs /erigon_upload_tmp
trap "umount /erigon_upload_tmp" EXIT
mkdir -p /erigon_upload_tmp/working_stdout
mkdir -p /erigon_upload_tmp/upload_part_results
# Sadly s3api/boto3 does not support streaming file descriptors. This means we need to write
# our entire chunk to a file then upload that file. This probably isn't a big deal since the
# data is in memory anyway
pzstd \
-p $parallel_downloads \
-3 \
-v \
--stdout \
/erigon/data/bsc/chaindata/mdbx.dat \
| psplit \
-b $bytes_per_chunk \
"bash -euo pipefail -c ' \
SEQ=\$(( \$SEQ + 1 )) && \
md5_value=\$(tee /erigon_upload_tmp/working_stdout/\$(printf %05d \$SEQ) | md5sum | cut -c -32) && \
trap \"rm -rf /erigon_upload_tmp/working_stdout/\$(printf %05d \$SEQ)\" EXIT && \
etag_result=\$(aws s3api upload-part \
--body /erigon_upload_tmp/working_stdout/\$(printf %05d \$SEQ) \
--request-payer requester \
--bucket public-blockchain-snapshots \
--key bsc/erigon/archive/latest/v1/chaindata/mdbx.dat.zstd \
--upload-id $upload_id \
--part-number \$SEQ \
| jq -r .ETag | tr -d \\\" | tee > /erigon_upload_tmp/upload_part_results/\$(printf %05d \$SEQ)') && \
if [ \$md5_value -ne \$etag_result ]; then echo \"md5 did not match \$md5_value -ne \$etag_result\" >&2 ; exit 1; fi"
# Sadly `aws s3api complete-multipart-upload` requires the `multipart-upload` field be sent as an
# argument which is too large to send over an argument, so we use a short python script to finish.
python3 -c "
import boto3, os
part_nums=os.listdir('/erigon_upload_tmp/upload_part_results/')
part_nums.sort()
boto3.client('s3').complete_multipart_upload(
Bucket='public-blockchain-snapshots',
Key='bsc/erigon/archive/latest/v1/chaindata/mdbx.dat.zstd',
UploadId='$upload_id',
RequestPayer='requester',
MultipartUpload={
'Parts': [{'PartNumber': int(name), 'ETag': open('/erigon_upload_tmp/upload_part_results/' + name).readline().strip()} for name in part_nums]
}
)"
}
# Note: This will also delete remote files that are not local.
# This is pretty much the same as `aws s3 sync /local/folder/ s3://foo/bar/`, but better
# parallization.
function parallel_sync_upload() {
set +x
local_path=$1
shift
full_s3_path=$1
shift
s3_bucket=$(echo "$full_s3_path" | cut -d'/' -f3)
s3_path="${full_s3_path#s3://$s3_bucket/}"
s3_path="${s3_path%/}"
local_files=$(find $local_path -type f | cut -c$((${#local_path}+1))- | sed -e 's/$/.zstd/' | sort)
remote_files=$(aws s3 ls --recursive "$full_s3_path" | tr -s ' ' ' ' | cut -d' ' -f4 | cut -c$((${#s3_path}+2))- | sort)
files_to_remove=$(comm -13 <(echo "$local_files") <(echo "$remote_files"))
files_to_upload=$(comm -23 <(echo "$local_files") <(echo "$remote_files"))
set -x
for s3_delete_file in $files_to_remove ; do
aws s3 rm "s3://$s3_bucket/$s3_path/$s3_delete_file"
done
num_cores=$(nproc)
# This is an inverse log10(). The lower the number of cores you have the more processes you'll
# spawn. The logic here is that on less powerful machines you'll almost certainly want more
# than 1 download going on at a time. The same in reverse, on 128 core machines, you will be
# limited by network instead of cpu ability. 128 cores = 89 jobs, 64 cores = 56 jobs,
# 32 cores = 36 jobs, 16 cores = 25 jobs, 4 cores = 15 jobs, exc...
parallel_count=$(echo "x = $num_cores / (l(($num_cores + 8) / 8) / l(10)); scale=0; x / 1" | bc -l)
# Download all the individual files from aws, decompress them, then place them into the snapshots
# folder. This is similar to running:
# aws s3 sync --request-payer=requester $full_s3_path $local_path
# The major difference is that it will, while downloading, decompress each file.
commands=()
for relative_path_with_zstd in $files_to_upload ; do
relative_path="${relative_path_with_zstd%.zstd}"
compress_file="pzstd -6 -q --stdout $local_path/$relative_path"
upload_file="aws s3 cp - ${full_s3_path%}${relative_path}.zstd"
commands+=("sh -c '$compress_file | $upload_file'")
done
if [ "${#commands[@]}" -gt 0 ] ; then
( IFS=$'\n'; echo "${commands[*]}" ) | pjoin --parallel-count $parallel_count
fi
}
zfs set readonly=on tank/erigon_data/bsc/snapshots
parallel_sync_upload /erigon/data/bsc/snapshots/ s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/snapshots/ &
zfs set readonly=on tank/erigon_data/bsc/nodes
parallel_sync_upload /erigon/data/bsc/nodes/ s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/nodes/ &
zfs set readonly=on tank/erigon_data/bsc/parlia
parallel_sync_upload /erigon/data/bsc/parlia/ s3://public-blockchain-snapshots/bsc/erigon/archive/latest/v1/parlia/ &
zfs set readonly=on tank/erigon_data/bsc/chaindata
upload_mdbx_file &
# If one of the background tasks has a bad exit code it's ok.
wait # Wait for all background tasks to finish.
EOT
chmod 0744 /erigon/create-bsc-snapshot-and-shutdown.sh
chown root:root /erigon/create-bsc-snapshot-and-shutdown.sh
echo "create-bsc-snapshot-and-shutdown /erigon/create-bsc-snapshot-and-shutdown.sh uid=root erigon" >> /etc/super.tab
}
install_prereq
# These installations can happen in parallel.
install_zstd &
install_aws_cli &
install_s3pcp &
install_putils &
install_erigon &
safe_wait # Wait for our parallel jobs finish.
# This should only be set if we are only configuring the instance for an EBS snapshot.
# Only set this global if you want to create your own snapshots and create an image of
# this instance as a template for faster startup.
if [[ "${CREATE_SNAPSHOT_MODE:-}" == "1" ]]; then
apt update
# This fixes an error when upgrading default ubuntu 22.04 instance.
DEBIAN_FRONTEND=noninteractive apt install -y grub-efi-arm64
DEBIAN_FRONTEND=noninteractive apt upgrade -y
shutdown -r +1
exit
fi
setup_drives
prepare_zfs_datasets
download_snapshots & # Download just the snapshots folder.
download_nodes & # Downloads the last known list of nodes.
download_parlia & # Downloads the last known parlia snapshots.
download_database_file & # Download the database file. This is the bulk of the downloads.
safe_wait # Wait for download_snapshot to finish.
prepare_erigon
run_erigon
add_create_snapshot_script