+ -w "https://planet.openstreetmap.org/${web_path}" \
+ -w "https://osm-planet-eu-central-1.s3.dualstack.eu-central-1.amazonaws.com/${s3_web_path}" \
+ -w "https://osm-planet-us-west-2.s3.dualstack.us-west-2.amazonaws.com/${s3_web_path}" \
+ -w "https://ftp5.gwdg.de/pub/misc/openstreetmap/planet.openstreetmap.org/${web_path}" \
+ -w "https://ftpmirror.your.org/pub/openstreetmap/${web_path}" \
+ -w "https://mirror.init7.net/openstreetmap/${web_path}" \
+ -w "https://ftp.fau.de/osm-planet/${web_path}" \
+ -w "https://ftp.spline.de/pub/openstreetmap/${web_path}" \
+ -w "https://downloads.opencagedata.com/planet/${name}" \
+ -w "https://planet.osm-hr.org/${web_path}" \
+ -w "https://planet.maps.mail.ru/${web_path}" \
+ -c "OpenStreetMap ${type} data export, licensed under https://opendatacommons.org/licenses/odbl/ by OpenStreetMap contributors" \
+ -o "${torrent_file}" > /dev/null
+
+ # create .xml global RSS headers if missing
+ torrent_time_rfc="$(date -R -r ${torrent_file})"
+ test -f "${rss_file}" || echo "<x/>" | xmlstarlet select --xml-decl --indent \
+ -N "atom=http://www.w3.org/2005/Atom" \
+ -N "dcterms=http://purl.org/dc/terms/" \
+ -N "content=http://purl.org/rss/1.0/modules/content/" \
+ --encode "UTF-8" \
+ --template \
+ --match / \
+ --elem "rss" \
+ --attr "version" --output "2.0" --break \
+ --attr "atom:DUMMY" --break \
+ --elem "channel" \
+ --elem "title" --output "OpenStreetMap ${type} ${format} torrent RSS" --break \
+ --elem "link" --output "${rss_web_dir}" --break \
+ --elem "atom:link" \
+ --attr "href" --output "${rss_web_dir}/${rss_file}" --break \
+ --attr "rel" --output "self" --break \
+ --attr "type" --output "application/rss+xml" --break \
+ --break \
+ --elem "description" --output "${type}.osm.${format}.torrent RSS feed" --break \
+ --elem "copyright" --output "Source: OpenStreetMap contributors, under ODbL 1.0 licence" --break \
+ --elem "generator" --output "OpenStreetMap xmlstarlet powered shell script v1.0" --break \
+ --elem "language" --output "en" --break \
+ --elem "lastBuildDate" --output "${torrent_time_rfc}" \
+ > "${rss_file}"
+
+ # add newly created .torrent file as new entry to .xml RSS feed, removing excess entries
+ torrent_size="$(stat --format="%s" ${torrent_file})"
+ xmlstarlet edit --inplace \
+ -a "//lastBuildDate" -t elem -n item -v "" \
+ -s "//item[1]" -t elem -n "title" -v "${torrent_file}" \
+ -s "//item[1]" -t elem -n "guid" -v "${torrent_url}" \
+ -s "//item[1]" -t elem -n "link" -v "${torrent_url}" \
+ -s "//item[1]" -t elem -n "pubDate" -v "${torrent_time_rfc}" \
+ -s "//item[1]" -t elem -n "category" -v "OpenStreetMap data" \
+ -s "//item[1]" -t elem -n "enclosure" \
+ -s "//item[1]"/enclosure -t attr -n "type" -v "application/x-bittorrent" \
+ -s "//item[1]"/enclosure -t attr -n "length" -v "${torrent_size}" \
+ -s "//item[1]"/enclosure -t attr -n "url" -v "${torrent_url}" \
+ -s "//item[1]" -t elem -n "description" -v "OpenStreetMap torrent ${torrent_file}" \
+ -u /rss/channel/lastBuildDate -v "${torrent_time_rfc}" \
+ -d /rss/@atom:DUMMY \
+ -d "//item[position()>5]" \
+ "${rss_file}"
+}
+
+function replication_status_wait {
+ local s3_url="$1"
+ for i in {1..3600}; do
+ local replication_status=$(curl -sI --location "${s3_url}" | grep -F 'x-amz-replication-status' | awk '{print $2}' | tr -d '\r' )
+
+ if [[ "${replication_status}" == "COMPLETED" ]]; then
+ return 0 # success
+ fi
+
+ sleep 1
+ done
+ echo "Timeout waiting for ${s3_url} to complete replication status: ${replication_status}"