From 73bec7e8e5b260f38f7a907ad1b5e17457596dbf Mon Sep 17 00:00:00 2001 From: Scott Worley Date: Wed, 8 Apr 2020 17:12:03 -0700 Subject: [PATCH] Fetch with nix-prefetch-url for the caching --- pinch.py | 85 +++++++++++++++++++++++++++++++++++++------------------- 1 file changed, 57 insertions(+), 28 deletions(-) diff --git a/pinch.py b/pinch.py index cdc0ba8..dfdedac 100644 --- a/pinch.py +++ b/pinch.py @@ -5,6 +5,7 @@ import operator import os import os.path import shutil +import subprocess import tempfile import types import urllib.parse @@ -15,14 +16,17 @@ from typing import ( Dict, Iterable, List, + NewType, Sequence, Tuple, ) +Digest16 = NewType('Digest16', str) +Digest32 = NewType('Digest32', str) + class InfoTableEntry(types.SimpleNamespace): - content: bytes - digest: str + digest: Digest16 file: str size: int url: str @@ -115,7 +119,7 @@ def fetch(v: Verification, channel_url: str) -> Info: return info -def parse(v: Verification, info: Info) -> None: +def parse_table(v: Verification, info: Info) -> None: v.status('Parsing channel description as XML') d = xml.dom.minidom.parseString(info.channel_html) v.ok() @@ -133,38 +137,63 @@ def parse(v: Verification, info: Info) -> None: name = row.childNodes[0].firstChild.firstChild.nodeValue url = row.childNodes[0].firstChild.getAttribute('href') size = int(row.childNodes[1].firstChild.nodeValue) - digest = row.childNodes[2].firstChild.firstChild.nodeValue + digest = Digest16(row.childNodes[2].firstChild.firstChild.nodeValue) info.table[name] = InfoTableEntry(url=url, digest=digest, size=size) v.ok() -def fetch_resources(v: Verification, info: Info) -> None: +def digest_file(filename: str) -> Digest16: + hasher = hashlib.sha256() + with open(filename, 'rb') as f: + # pylint: disable=cell-var-from-loop + for block in iter(lambda: f.read(4096), b''): + hasher.update(block) + return Digest16(hasher.hexdigest()) + + +def to_Digest16(v: Verification, digest32: Digest32) -> Digest16: + v.status('Converting digest to base16') + process = subprocess.run( + ['nix', 'to-base16', '--type', 'sha256', digest32], capture_output=True) + v.result(process.returncode == 0) + return Digest16(process.stdout.decode().strip()) + + +def to_Digest32(v: Verification, digest16: Digest16) -> Digest32: + v.status('Converting digest to base32') + process = subprocess.run( + ['nix', 'to-base32', '--type', 'sha256', digest16], capture_output=True) + v.result(process.returncode == 0) + return Digest32(process.stdout.decode().strip()) + + +def fetch_with_nix_prefetch_url( + v: Verification, + url: str, + digest: Digest16) -> str: + v.status('Fetching %s' % url) + process = subprocess.run( + ['nix-prefetch-url', '--print-path', url, digest], capture_output=True) + v.result(process.returncode == 0) + prefetch_digest, path, empty = process.stdout.decode().split('\n') + assert empty == '' + v.check("Verifying nix-prefetch-url's digest", + to_Digest16(v, Digest32(prefetch_digest)) == digest) + v.status("Verifying file digest") + file_digest = digest_file(path) + v.result(file_digest == digest) + return path + +def fetch_resources(v: Verification, info: Info) -> None: for resource in ['git-revision', 'nixexprs.tar.xz']: fields = info.table[resource] - v.status('Fetching resource "%s"' % resource) url = urllib.parse.urljoin(info.forwarded_url, fields.url) - request = urllib.request.urlopen(url, timeout=10) - if fields.size < 4096: - fields.content = request.read() - else: - with tempfile.NamedTemporaryFile(suffix='.nixexprs.tar.xz', delete=False) as tmp_file: - shutil.copyfileobj(request, tmp_file) - fields.file = tmp_file.name - v.result(request.status == 200) - v.status('Verifying digest for "%s"' % resource) - if fields.size < 4096: - actual_hash = hashlib.sha256(fields.content).hexdigest() - else: - hasher = hashlib.sha256() - with open(fields.file, 'rb') as f: - # pylint: disable=cell-var-from-loop - for block in iter(lambda: f.read(4096), b''): - hasher.update(block) - actual_hash = hasher.hexdigest() - v.result(actual_hash == fields.digest) - v.check('Verifying git commit on main page matches git commit in table', - info.table['git-revision'].content.decode() == info.git_commit) + fields.file = fetch_with_nix_prefetch_url(v, url, fields.digest) + v.status('Verifying git commit on main page matches git commit in table') + v.result( + open( + info.table['git-revision'].file).read(999) == info.git_commit) def extract_channel(v: Verification, info: Info) -> None: @@ -179,7 +208,7 @@ def extract_channel(v: Verification, info: Info) -> None: def main() -> None: v = Verification() info = fetch(v, 'https://channels.nixos.org/nixos-20.03') - parse(v, info) + parse_table(v, info) fetch_resources(v, info) extract_channel(v, info) print(info) -- 2.44.1