mirror of
https://github.com/Abdess/retroarch_system.git
synced 2026-04-13 20:32:32 -05:00
Moved shared functions to common.py (single source of truth): - check_inside_zip (was in verify.py, imported by generate_pack) - build_zip_contents_index (was duplicated in verify + generate_pack) - load_emulator_profiles (was in verify, cross_reference, generate_site) - group_identical_platforms (was in verify + generate_pack) Added tests/ with 83 unit tests covering: - resolve_local_file: SHA1, MD5, name, alias, truncated, zip_contents - verify: existence, md5, zipped_file, multi-hash, severity mapping - aliases: field parsing, by_name indexing, beetle_psx field rename - pack: dedup, file_status, zipped_file inner check, EmuDeck entries - severity: all 12 combinations, platform-native behavior 0 regressions: pipeline.py --all produces identical results.
239 lines
8.2 KiB
Python
239 lines
8.2 KiB
Python
"""Tests for pack generation logic in generate_pack.py."""
|
|
|
|
from __future__ import annotations
|
|
|
|
import hashlib
|
|
import os
|
|
import sys
|
|
import tempfile
|
|
import unittest
|
|
import zipfile
|
|
|
|
sys.path.insert(0, os.path.join(os.path.dirname(__file__), "..", "scripts"))
|
|
from common import compute_hashes
|
|
from generate_pack import build_zip_contents_index
|
|
|
|
|
|
class TestBuildZipContentsIndex(unittest.TestCase):
|
|
"""Test build_zip_contents_index: maps inner ROM MD5 to container SHA1."""
|
|
|
|
def setUp(self):
|
|
self.tmpdir = tempfile.mkdtemp()
|
|
self.inner_content = b"inner rom data for index test"
|
|
self.inner_md5 = hashlib.md5(self.inner_content).hexdigest()
|
|
|
|
self.zip_path = os.path.join(self.tmpdir, "container.zip")
|
|
with zipfile.ZipFile(self.zip_path, "w") as zf:
|
|
zf.writestr("rom.bin", self.inner_content)
|
|
|
|
hashes = compute_hashes(self.zip_path)
|
|
self.zip_sha1 = hashes["sha1"]
|
|
self.zip_md5 = hashes["md5"]
|
|
|
|
self.db = {
|
|
"files": {
|
|
self.zip_sha1: {
|
|
"path": self.zip_path,
|
|
"name": "container.zip",
|
|
"md5": self.zip_md5,
|
|
"size": os.path.getsize(self.zip_path),
|
|
},
|
|
},
|
|
"indexes": {
|
|
"by_md5": {self.zip_md5: self.zip_sha1},
|
|
"by_name": {"container.zip": [self.zip_sha1]},
|
|
"by_crc32": {},
|
|
},
|
|
}
|
|
|
|
def tearDown(self):
|
|
import shutil
|
|
shutil.rmtree(self.tmpdir, ignore_errors=True)
|
|
|
|
def test_inner_md5_maps_to_container_sha1(self):
|
|
index = build_zip_contents_index(self.db)
|
|
self.assertIn(self.inner_md5, index)
|
|
self.assertEqual(index[self.inner_md5], self.zip_sha1)
|
|
|
|
def test_non_zip_files_skipped(self):
|
|
"""Non-ZIP files in db don't appear in index."""
|
|
plain_path = os.path.join(self.tmpdir, "plain.bin")
|
|
with open(plain_path, "wb") as f:
|
|
f.write(b"not a zip")
|
|
hashes = compute_hashes(plain_path)
|
|
self.db["files"][hashes["sha1"]] = {
|
|
"path": plain_path,
|
|
"name": "plain.bin",
|
|
"md5": hashes["md5"],
|
|
"size": 9,
|
|
}
|
|
index = build_zip_contents_index(self.db)
|
|
# Only the inner_md5 from the ZIP should be present
|
|
self.assertEqual(len(index), 1)
|
|
|
|
def test_missing_file_skipped(self):
|
|
"""ZIP path that doesn't exist on disk is skipped."""
|
|
self.db["files"]["fake_sha1"] = {
|
|
"path": "/nonexistent/file.zip",
|
|
"name": "file.zip",
|
|
"md5": "a" * 32,
|
|
"size": 0,
|
|
}
|
|
index = build_zip_contents_index(self.db)
|
|
self.assertEqual(len(index), 1)
|
|
|
|
def test_bad_zip_skipped(self):
|
|
"""Corrupt ZIP file is skipped without error."""
|
|
bad_path = os.path.join(self.tmpdir, "bad.zip")
|
|
with open(bad_path, "wb") as f:
|
|
f.write(b"corrupt data")
|
|
hashes = compute_hashes(bad_path)
|
|
self.db["files"][hashes["sha1"]] = {
|
|
"path": bad_path,
|
|
"name": "bad.zip",
|
|
"md5": hashes["md5"],
|
|
"size": 12,
|
|
}
|
|
index = build_zip_contents_index(self.db)
|
|
self.assertEqual(len(index), 1)
|
|
|
|
|
|
class TestFileStatusAggregation(unittest.TestCase):
|
|
"""Test worst-status-wins logic for pack file aggregation."""
|
|
|
|
def test_worst_status_wins(self):
|
|
"""Simulate the worst-status-wins dict pattern from generate_pack."""
|
|
sev_order = {"ok": 0, "untested": 1, "missing": 2}
|
|
file_status = {}
|
|
|
|
def update_status(dest, status):
|
|
prev = file_status.get(dest)
|
|
if prev is None or sev_order.get(status, 0) > sev_order.get(prev, 0):
|
|
file_status[dest] = status
|
|
|
|
update_status("system/bios.bin", "ok")
|
|
update_status("system/bios.bin", "missing")
|
|
self.assertEqual(file_status["system/bios.bin"], "missing")
|
|
|
|
update_status("system/other.bin", "untested")
|
|
update_status("system/other.bin", "ok")
|
|
self.assertEqual(file_status["system/other.bin"], "untested")
|
|
|
|
def test_dedup_same_destination_packed_once(self):
|
|
"""Same destination from multiple systems: only first is packed."""
|
|
seen = set()
|
|
packed = []
|
|
entries = [
|
|
{"dest": "shared/bios.bin", "source": "sys1"},
|
|
{"dest": "shared/bios.bin", "source": "sys2"},
|
|
{"dest": "unique/other.bin", "source": "sys3"},
|
|
]
|
|
for e in entries:
|
|
if e["dest"] in seen:
|
|
continue
|
|
seen.add(e["dest"])
|
|
packed.append(e["dest"])
|
|
self.assertEqual(len(packed), 2)
|
|
self.assertIn("shared/bios.bin", packed)
|
|
self.assertIn("unique/other.bin", packed)
|
|
|
|
|
|
class TestEmuDeckNoDestination(unittest.TestCase):
|
|
"""EmuDeck entries with no destination are counted as checks."""
|
|
|
|
def test_no_destination_counted_as_check(self):
|
|
"""EmuDeck-style entries (md5 whitelist, no filename) are tracked."""
|
|
file_status = {}
|
|
# Simulate generate_pack logic for empty dest
|
|
sys_id = "psx"
|
|
name = ""
|
|
md5 = "abc123"
|
|
by_md5 = {"abc123": "sha1_match"}
|
|
|
|
dest = "" # empty destination
|
|
if not dest:
|
|
fkey = f"{sys_id}/{name}"
|
|
if md5 and md5 in by_md5:
|
|
file_status.setdefault(fkey, "ok")
|
|
else:
|
|
file_status[fkey] = "missing"
|
|
|
|
self.assertIn("psx/", file_status)
|
|
self.assertEqual(file_status["psx/"], "ok")
|
|
|
|
def test_no_destination_missing(self):
|
|
file_status = {}
|
|
sys_id = "psx"
|
|
name = ""
|
|
md5 = "abc123"
|
|
by_md5 = {}
|
|
|
|
dest = ""
|
|
if not dest:
|
|
fkey = f"{sys_id}/{name}"
|
|
if md5 and md5 in by_md5:
|
|
file_status.setdefault(fkey, "ok")
|
|
else:
|
|
file_status[fkey] = "missing"
|
|
|
|
self.assertEqual(file_status["psx/"], "missing")
|
|
|
|
|
|
class TestUserProvidedEntries(unittest.TestCase):
|
|
"""Test user_provided storage handling."""
|
|
|
|
def test_user_provided_creates_instruction_file(self):
|
|
"""Simulate user_provided entry packing logic."""
|
|
tmpdir = tempfile.mkdtemp()
|
|
try:
|
|
zip_path = os.path.join(tmpdir, "test_pack.zip")
|
|
with zipfile.ZipFile(zip_path, "w") as zf:
|
|
entry = {
|
|
"name": "PS3UPDAT.PUP",
|
|
"storage": "user_provided",
|
|
"instructions": "Download from sony.com",
|
|
}
|
|
instr_name = f"INSTRUCTIONS_{entry['name']}.txt"
|
|
zf.writestr(instr_name, f"File needed: {entry['name']}\n\n{entry['instructions']}\n")
|
|
|
|
with zipfile.ZipFile(zip_path, "r") as zf:
|
|
names = zf.namelist()
|
|
self.assertIn("INSTRUCTIONS_PS3UPDAT.PUP.txt", names)
|
|
content = zf.read("INSTRUCTIONS_PS3UPDAT.PUP.txt").decode()
|
|
self.assertIn("PS3UPDAT.PUP", content)
|
|
self.assertIn("sony.com", content)
|
|
finally:
|
|
import shutil
|
|
shutil.rmtree(tmpdir, ignore_errors=True)
|
|
|
|
|
|
class TestZippedFileHashMismatch(unittest.TestCase):
|
|
"""Test zipped_file with hash_mismatch triggers check_inside_zip."""
|
|
|
|
def setUp(self):
|
|
self.tmpdir = tempfile.mkdtemp()
|
|
self.inner_content = b"correct inner rom"
|
|
self.inner_md5 = hashlib.md5(self.inner_content).hexdigest()
|
|
self.zip_path = os.path.join(self.tmpdir, "game.zip")
|
|
with zipfile.ZipFile(self.zip_path, "w") as zf:
|
|
zf.writestr("rom.bin", self.inner_content)
|
|
|
|
def tearDown(self):
|
|
import shutil
|
|
shutil.rmtree(self.tmpdir, ignore_errors=True)
|
|
|
|
def test_hash_mismatch_zip_inner_ok(self):
|
|
"""hash_mismatch on container, but inner ROM MD5 matches."""
|
|
from verify import check_inside_zip, Status
|
|
result = check_inside_zip(self.zip_path, "rom.bin", self.inner_md5)
|
|
self.assertEqual(result, Status.OK)
|
|
|
|
def test_hash_mismatch_zip_inner_not_found(self):
|
|
from verify import check_inside_zip
|
|
result = check_inside_zip(self.zip_path, "missing.bin", self.inner_md5)
|
|
self.assertEqual(result, "not_in_zip")
|
|
|
|
|
|
if __name__ == "__main__":
|
|
unittest.main()
|