Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

update data acquisition notebook #11

Merged
merged 3 commits into from
Apr 3, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 4 additions & 2 deletions HISTORY.rst
Original file line number Diff line number Diff line change
@@ -1,9 +1,11 @@
History
=======

unreleased
----------
v1.2.1
------
* Update GitHub actions versions
* Update data acquisition notebook, using helper functions to download input data from
zenodo

v1.2.0
------
Expand Down
93 changes: 9 additions & 84 deletions notebooks/data-acquisition.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -39,18 +39,17 @@
},
"outputs": [],
"source": [
"import tempfile\n",
"from io import BytesIO\n",
"from os import environ\n",
"from pathlib import Path\n",
"from zipfile import ZipFile\n",
"\n",
"import numpy as np\n",
"import pandas as pd\n",
"import requests\n",
"from cartopy.io import shapereader\n",
"from fsspec import FSTimeoutError\n",
"from fsspec.implementations.zip import ZipFileSystem\n",
"from pyCIAM.io import (\n",
" download_and_extract_from_zenodo,\n",
" download_and_extract_partial_zip,\n",
" get_zenodo_file_list,\n",
")\n",
"from pyCIAM.utils import copy\n",
"from shared import (\n",
" DIR_SHP,\n",
" DIR_SLR_AR5_IFILES_RAW,\n",
Expand All @@ -71,9 +70,7 @@
" PATH_SLR_HIST_TREND_MAP,\n",
" PATHS_SURGE_LOOKUP,\n",
" save,\n",
")\n",
"\n",
"from pyCIAM.utils import copy"
")"
]
},
{
Expand Down Expand Up @@ -123,76 +120,6 @@
"Z_URL_SLIIDERS_PC = Z_URL_RECORDS"
]
},
{
"cell_type": "code",
"execution_count": 49,
"id": "8d519f83-eb91-4cb0-b2e7-5918b91d5143",
"metadata": {
"tags": []
},
"outputs": [],
"source": [
"def get_download_link(files, prefix):\n",
" links = [\n",
" i[\"links\"]\n",
" for i in files\n",
" if i.get(\"filename\", \"\").startswith(prefix)\n",
" or i.get(\"key\", \"\").startswith(prefix)\n",
" ]\n",
" assert len(links) == 1\n",
" links = links[0]\n",
" return links.get(\"download\", links[\"self\"])\n",
"\n",
"\n",
"def download_and_extract_full_zip(lpath, url):\n",
" if lpath.exists():\n",
" return None\n",
" lpath.parent.mkdir(exist_ok=True, parents=True)\n",
"\n",
" content = BytesIO(requests.get(url, params=PARAMS).content)\n",
" if isinstance(lpath, Path):\n",
" with ZipFile(content, \"r\") as zip_ref:\n",
" zip_ref.extractall(lpath)\n",
" else:\n",
" with tempfile.TemporaryDirectory() as tmpdir:\n",
" with ZipFile(content, \"r\") as zip_ref:\n",
" zip_ref.extractall(tmpdir)\n",
" copy(Path(tmpdir), lpath)\n",
"\n",
"\n",
"def download_and_extract_partial_zip(lpath, url, zip_glob, n_retries=5):\n",
" lpath.mkdir(exist_ok=True, parents=True)\n",
" z = ZipFileSystem(url)\n",
" if isinstance(zip_glob, (list, set, tuple, np.ndarray)):\n",
" files_remote = zip_glob\n",
" else:\n",
" files_remote = [p for p in z.glob(zip_glob) if not p.endswith(\"/\")]\n",
" files_local = [lpath / Path(f).name for f in files_remote]\n",
" for fr, fl in list(zip(files_remote, files_local)):\n",
" if not fl.is_file():\n",
" retries = 0\n",
" while retries < n_retries:\n",
" print(f\"...Downloading {fl.name} (attempt {retries+1}/{n_retries})\")\n",
" try:\n",
" data = z.cat_file(fr)\n",
" break\n",
" except FSTimeoutError:\n",
" if retries < (n_retries - 1):\n",
" retries += 1\n",
" else:\n",
" raise\n",
" print(f\"...Writing {fl.name}\")\n",
" fl.write_bytes(data)\n",
"\n",
"\n",
"def download_and_extract_from_zenodo(lpath, files, prefix, zip_glob=None):\n",
" dl = get_download_link(files, prefix)\n",
" if zip_glob is None:\n",
" return download_and_extract_full_zip(lpath, dl)\n",
" else:\n",
" return download_and_extract_partial_zip(lpath, dl, zip_glob)"
]
},
{
"cell_type": "code",
"execution_count": 5,
Expand All @@ -202,9 +129,7 @@
},
"outputs": [],
"source": [
"pyciam_files = requests.get(\n",
" Z_URL_SLIIDERS_PC.format(doi=Z_PYCIAM_DOI), params=PARAMS\n",
").json()[\"files\"]"
"pyciam_files = get_zenodo_file_list(Z_PYCIAM_DOI)"
]
},
{
Expand Down Expand Up @@ -628,7 +553,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.10.8"
"version": "3.12.2"
},
"widgets": {
"application/vnd.jupyter.widget-state+json": {
Expand Down
77 changes: 77 additions & 0 deletions pyCIAM/io.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,12 +8,22 @@
* load_diaz_inputs
"""

import tempfile
from collections.abc import Iterable
from io import BytesIO
from pathlib import Path
from zipfile import ZipFile

import dask.array as da
import numpy as np
import pandas as pd
import pint_xarray # noqa: F401
import requests
import xarray as xr
from fsspec import FSTimeoutError
from fsspec.implementations.zip import ZipFileSystem

from pyCIAM.utils import copy
from pyCIAM.utils import spherical_nearest_neighbor as snn

from .utils import _s2d
Expand Down Expand Up @@ -783,3 +793,70 @@ def load_diaz_inputs(

inputs = inputs.drop_dims("rcp_pt")
return inputs, slr


def get_zenodo_file_list(doi, params={}):
return requests.get(f"https://zenodo.org/api/records/{doi}", params=params).json()[
"files"
]


def get_download_link(files, prefix):
links = [
i["links"]
for i in files
if i.get("filename", "").startswith(prefix)
or i.get("key", "").startswith(prefix)
]
assert len(links) == 1
links = links[0]
return links.get("download", links["self"])


def _download_and_extract_full_zip(lpath, url, params={}):
if lpath.exists():
return None
lpath.parent.mkdir(exist_ok=True, parents=True)

content = BytesIO(requests.get(url, params=params).content)
if isinstance(lpath, Path):
with ZipFile(content, "r") as zip_ref:
zip_ref.extractall(lpath)
else:
with tempfile.TemporaryDirectory() as tmpdir:
with ZipFile(content, "r") as zip_ref:
zip_ref.extractall(tmpdir)
copy(Path(tmpdir), lpath)


def download_and_extract_partial_zip(lpath, url, zip_glob, n_retries=5):
lpath.mkdir(exist_ok=True, parents=True)
z = ZipFileSystem(url)
if isinstance(zip_glob, (list, set, tuple, np.ndarray)):
files_remote = zip_glob
else:
files_remote = [p for p in z.glob(zip_glob) if not p.endswith("/")]
files_local = [lpath / Path(f).name for f in files_remote]
for fr, fl in list(zip(files_remote, files_local)):
if not fl.is_file():
retries = 0
while retries < n_retries:
print(f"...Downloading {fl.name} (attempt {retries+1}/{n_retries})")
try:
data = z.cat_file(fr)
break
except FSTimeoutError:
if retries < (n_retries - 1):
retries += 1
else:
raise
print(f"...Writing {fl.name}")
fl.write_bytes(data)


def download_and_extract_from_zenodo(lpath, files, prefix, zip_glob=None):
dl = get_download_link(files, prefix)
if zip_glob is None:
return _download_and_extract_full_zip(lpath, dl)
else:
return download_and_extract_partial_zip(lpath, dl, zip_glob)