X-Git-Url: https://git.openstreetmap.org./chef.git/blobdiff_plain/20b2c44f11ad74f2e34bc9f8aee0be22ebe58231..d18cd92a98384aebf371376383a4408485f12374:/cookbooks/planet/templates/default/planetdump.erb diff --git a/cookbooks/planet/templates/default/planetdump.erb b/cookbooks/planet/templates/default/planetdump.erb index ea474ed93..5c0d38045 100644 --- a/cookbooks/planet/templates/default/planetdump.erb +++ b/cookbooks/planet/templates/default/planetdump.erb @@ -26,13 +26,6 @@ if [ -f /tmp/planetdump.lock ]; then fi fi -# Redirect this shell's output to a file. This is so that it -# can be emailed later, since this script is run from incron -# and incron doesn't yet support MAILTO like cron does. The -# command below appears to work in bash as well as dash. -logfile="/tmp/planetdump.log.$$" -exec > "${logfile}" 2>&1 - # Create lock file echo $$ > /tmp/planetdump.lock @@ -40,16 +33,6 @@ echo $$ > /tmp/planetdump.lock function cleanup { # Remove the lock file rm /tmp/planetdump.lock - - # Send an email with the output, since incron doesn't yet - # support doing this in the incrontab - if [[ -s "$logfile" ]] - then - mailx -s "Planet dump output: ${file}" zerebubuth@gmail.com < "${logfile}" - fi - - # Remove the log file - rm -f "${logfile}" } # Remove lock on exit @@ -76,17 +59,19 @@ time nice -n 19 /opt/planet-dump-ng/planet-dump-ng \ # Function to create bittorrent files function mk_torrent { - type="$1" - format="$2" - dir="$3" - s_year="$4" - web_dir="${dir}${s_year}" - name="${type}-${date}.osm.${format}" - web_path="${web_dir}/${name}" - rss_web_dir="https://planet.openstreetmap.org/${dir}" - rss_file="${type}-${format}-rss.xml" - torrent_file="${name}.torrent" - torrent_url="${rss_web_dir}${s_year}/${torrent_file}" + local type="$1" + local format="$2" + local dir="$3" + local s3path="$4" + local s_year="$5" + local web_dir="${dir}${s_year}" + local name="${type}-${date}.osm.${format}" + local web_path="${web_dir}/${name}" + local s3_web_path="${s3path}/${name}" + local rss_web_dir="https://planet.openstreetmap.org/${dir}" + local rss_file="${type}-${format}-rss.xml" + local torrent_file="${name}.torrent" + local torrent_url="${rss_web_dir}${s_year}/${torrent_file}" # create .torrent file mktorrent -l 22 "${name}" \ @@ -96,15 +81,16 @@ function mk_torrent { -a udp://tracker-udp.gbitt.info:80/announce,http://tracker.gbitt.info/announce,https://tracker.gbitt.info/announce \ -a http://retracker.local/announce \ -w "https://planet.openstreetmap.org/${web_path}" \ + -w "https://osm-planet-eu-central-1.s3.dualstack.eu-central-1.amazonaws.com/${s3_web_path}" \ + -w "https://osm-planet-us-west-2.s3.dualstack.us-west-2.amazonaws.com/${s3_web_path}" \ -w "https://ftp5.gwdg.de/pub/misc/openstreetmap/planet.openstreetmap.org/${web_path}" \ -w "https://ftpmirror.your.org/pub/openstreetmap/${web_path}" \ -w "https://mirror.init7.net/openstreetmap/${web_path}" \ - -w "https://free.nchc.org.tw/osm.planet/${web_path}" \ -w "https://ftp.fau.de/osm-planet/${web_path}" \ -w "https://ftp.spline.de/pub/openstreetmap/${web_path}" \ - -w "https://osm.openarchive.site/${name}" \ -w "https://downloads.opencagedata.com/planet/${name}" \ -w "https://planet.osm-hr.org/${web_path}" \ + -w "https://planet.maps.mail.ru/${web_path}" \ -c "OpenStreetMap ${type} data export, licensed under https://opendatacommons.org/licenses/odbl/ by OpenStreetMap contributors" \ -o "${torrent_file}" > /dev/null @@ -129,14 +115,16 @@ function mk_torrent { --attr "type" --output "application/rss+xml" --break \ --break \ --elem "description" --output "${type}.osm.${format}.torrent RSS feed" --break \ + --elem "copyright" --output "Source: OpenStreetMap contributors, under ODbL 1.0 licence" --break \ --elem "generator" --output "OpenStreetMap xmlstarlet powered shell script v1.0" --break \ --elem "language" --output "en" --break \ --elem "lastBuildDate" --output "${torrent_time_rfc}" \ > "${rss_file}" # add newly created .torrent file as new entry to .xml RSS feed, removing excess entries + torrent_size="$(stat --format="%s" ${torrent_file})" xmlstarlet edit --inplace \ - -a "//channel" -t elem -n item -v "" \ + -a "//lastBuildDate" -t elem -n item -v "" \ -s "//item[1]" -t elem -n "title" -v "${torrent_file}" \ -s "//item[1]" -t elem -n "guid" -v "${torrent_url}" \ -s "//item[1]" -t elem -n "link" -v "${torrent_url}" \ @@ -144,26 +132,52 @@ function mk_torrent { -s "//item[1]" -t elem -n "category" -v "OpenStreetMap data" \ -s "//item[1]" -t elem -n "enclosure" \ -s "//item[1]"/enclosure -t attr -n "type" -v "application/x-bittorrent" \ + -s "//item[1]"/enclosure -t attr -n "length" -v "${torrent_size}" \ -s "//item[1]"/enclosure -t attr -n "url" -v "${torrent_url}" \ -s "//item[1]" -t elem -n "description" -v "OpenStreetMap torrent ${torrent_file}" \ - -s "//item[1]" -t elem -n "comments" -v "Source: OpenStreetMap contributors, under ODbL 1.0 licence" \ -u /rss/channel/lastBuildDate -v "${torrent_time_rfc}" \ -d /rss/@atom:DUMMY \ -d "//item[position()>5]" \ "${rss_file}" } +function replication_status_wait { + local s3_url="$1" + for i in {1..3600}; do + local replication_status=$(curl -sI --location "${s3_url}" | grep -F 'x-amz-replication-status' | awk '{print $2}' | tr -d '\r' ) + + if [[ "${replication_status}" == "COMPLETED" ]]; then + return 0 # success + fi + + sleep 1 + done + echo "Timeout waiting for ${s3_url} to complete replication status: ${replication_status}" +} + # Function to install a dump in place function install_dump { - type="$1" - format="$2" - dir="$3" - year="$4" - name="${type}-${date}.osm.${format}" - latest="${type}-latest.osm.${format}" - rss_file="${type}-${format}-rss.xml" + local type="$1" + local format="$2" + local dir="$3" + local s3dir="$4" + local year="$5" + local name="${type}-${date}.osm.${format}" + local latest="${type}-latest.osm.${format}" + local rss_file="${type}-${format}-rss.xml" md5sum "${name}" > "${name}.md5" + + # Upload all files to S3 + /opt/awscli/v2/current/bin/aws --profile=osm-pds-upload s3 cp --storage-class=INTELLIGENT_TIERING --no-progress "${name}.md5" "s3://osm-planet-eu-central-1/${s3dir}/${name}.md5" + /opt/awscli/v2/current/bin/aws --profile=osm-pds-upload s3 cp --storage-class=INTELLIGENT_TIERING --no-progress "${name}.torrent" "s3://osm-planet-eu-central-1/${s3dir}/${name}.torrent" + /opt/awscli/v2/current/bin/aws --profile=osm-pds-upload s3 cp --storage-class=INTELLIGENT_TIERING --no-progress "${name}" "s3://osm-planet-eu-central-1/${s3dir}/${name}" + + # Waiting for S3 replication to complete + replication_status_wait "https://osm-planet-eu-central-1.s3.dualstack.eu-central-1.amazonaws.com/${s3dir}/${name}.md5" + replication_status_wait "https://osm-planet-eu-central-1.s3.dualstack.eu-central-1.amazonaws.com/${s3dir}/${name}.torrent" + replication_status_wait "https://osm-planet-eu-central-1.s3.dualstack.eu-central-1.amazonaws.com/${s3dir}/${name}" + mkdir -p "${dir}/${year}" mv "${name}" "${name}.md5" "${dir}/${year}" ln -sf "${year:-.}/${name}" "${dir}/${latest}" @@ -174,28 +188,17 @@ function install_dump { } # Create *.torrent files -mk_torrent "changesets" "bz2" "planet" "/${year}" -mk_torrent "discussions" "bz2" "planet" "/${year}" -mk_torrent "planet" "bz2" "planet" "/${year}" -mk_torrent "history" "bz2" "planet/full-history" "/${year}" -mk_torrent "planet" "pbf" "pbf" -mk_torrent "history" "pbf" "pbf/full-history" +mk_torrent "changesets" "bz2" "planet" "changesets/osm/${year}" "/${year}" +mk_torrent "discussions" "bz2" "planet" "discussions/osm/${year}" "/${year}" +mk_torrent "planet" "bz2" "planet" "planet/osm/${year}" "/${year}" +mk_torrent "history" "bz2" "planet/full-history" "planet-full-history/osm/${year}" "/${year}" +mk_torrent "planet" "pbf" "pbf" "planet/pbf/${year}" +mk_torrent "history" "pbf" "pbf/full-history" "planet-full-history/pbf/${year}" # Move dumps into place -install_dump "changesets" "bz2" "<%= node[:planet][:dump][:xml_directory] %>" "${year}" -install_dump "discussions" "bz2" "<%= node[:planet][:dump][:xml_directory] %>" "${year}" -install_dump "planet" "bz2" "<%= node[:planet][:dump][:xml_directory] %>" "${year}" -install_dump "history" "bz2" "<%= node[:planet][:dump][:xml_history_directory] %>" "${year}" -install_dump "planet" "pbf" "<%= node[:planet][:dump][:pbf_directory] %>" -install_dump "history" "pbf" "<%= node[:planet][:dump][:pbf_history_directory] %>" - -# Remove pbf dumps older than 90 days -find "<%= node[:planet][:dump][:pbf_directory] %>" "<%= node[:planet][:dump][:pbf_history_directory] %>" \ - -maxdepth 1 -mindepth 1 -type f -mtime +90 \ - \( \ - -iname 'planet-*.pbf' \ - -o -iname 'history-*.pbf' \ - -o -iname 'planet-*.pbf.md5' \ - -o -iname 'history-*.pbf.md5' \ - \) \ - -delete +install_dump "changesets" "bz2" "<%= node[:planet][:dump][:xml_directory] %>" "changesets/osm/${year}" "${year}" +install_dump "discussions" "bz2" "<%= node[:planet][:dump][:xml_directory] %>" "discussions/osm/${year}" "${year}" +install_dump "planet" "bz2" "<%= node[:planet][:dump][:xml_directory] %>" "planet/osm/${year}" "${year}" +install_dump "history" "bz2" "<%= node[:planet][:dump][:xml_history_directory] %>" "planet-full-history/osm/${year}" "${year}" +install_dump "planet" "pbf" "<%= node[:planet][:dump][:pbf_directory] %>" "planet/pbf/${year}" +install_dump "history" "pbf" "<%= node[:planet][:dump][:pbf_history_directory] %>" "planet-full-history/pbf/${year}"