This commit is contained in:
Nathan
2025-12-11 12:12:10 -07:00
commit 01599a8f48
7 changed files with 2205 additions and 0 deletions

3
.cursorindexingignore Normal file
View File

@@ -0,0 +1,3 @@
# Don't index SpecStory auto-save files, but allow explicit context inclusion via @ references
.specstory/**

3
.gitignore vendored Normal file
View File

@@ -0,0 +1,3 @@
input/
output/

4
.specstory/.gitignore vendored Normal file
View File

@@ -0,0 +1,4 @@
# SpecStory project identity file
/.project.json
# SpecStory explanation file
/.what-is-this.md

101
check_sequences.py Normal file
View File

@@ -0,0 +1,101 @@
#!/usr/bin/env python3
"""
Diagnostic script to check which sequences are in input vs output.
"""
from pathlib import Path
from collections import defaultdict
def get_sequences(directory):
"""Get all sequence folders (directories containing PNG files)."""
sequences = defaultdict(list)
root = Path(directory)
if not root.exists():
return sequences
for png_file in root.rglob('*.png'):
# Get the sequence folder (parent directory)
seq_folder = png_file.parent
relative_seq = seq_folder.relative_to(root)
sequences[str(relative_seq)].append(png_file)
return sequences
def main():
input_dir = Path('input')
output_dir = Path('output')
print("=" * 80)
print("SEQUENCE DIAGNOSTIC REPORT")
print("=" * 80)
input_sequences = get_sequences(input_dir)
output_sequences = get_sequences(output_dir)
print(f"\nInput sequences: {len(input_sequences)}")
print(f"Output sequences: {len(output_sequences)}")
# Find missing sequences
missing = set(input_sequences.keys()) - set(output_sequences.keys())
if missing:
print(f"\n{'=' * 80}")
print(f"MISSING SEQUENCES ({len(missing)}):")
print(f"{'=' * 80}")
for seq in sorted(missing):
png_count = len(input_sequences[seq])
print(f" - {seq} ({png_count} PNG files)")
# Check if folder exists in output
output_seq_path = output_dir / seq
if output_seq_path.exists():
files_in_output = list(output_seq_path.iterdir())
print(f" Output folder exists but has {len(files_in_output)} files")
if files_in_output:
print(f" Files: {[f.name for f in files_in_output[:5]]}")
else:
print(f" Output folder does not exist")
else:
print("\nAll input sequences have corresponding output sequences.")
# Find sequences with different file counts
print(f"\n{'=' * 80}")
print("SEQUENCE FILE COUNT COMPARISON:")
print(f"{'=' * 80}")
for seq in sorted(input_sequences.keys()):
input_count = len(input_sequences[seq])
output_count = len(output_sequences.get(seq, []))
status = "OK" if input_count == output_count else "DIFF"
print(f"{status:4s} {seq}: {input_count} input -> {output_count} output")
if input_count != output_count and seq in output_sequences:
print(f" Difference: {input_count - output_count} files missing")
# Check for folders without PNG files
print(f"\n{'=' * 80}")
print("ALL FOLDERS IN INPUT DIRECTORY:")
print(f"{'=' * 80}")
input_root = Path('input')
if input_root.exists():
# Get all top-level directories
top_level_dirs = [d for d in input_root.iterdir() if d.is_dir()]
print(f"Total top-level folders: {len(top_level_dirs)}")
print()
folders_with_pngs = set()
for seq_name, png_files in input_sequences.items():
if png_files:
# Get the folder path from the first PNG file
folder_path = png_files[0].parent.relative_to(input_root)
folders_with_pngs.add(str(folder_path))
for folder in sorted(top_level_dirs):
rel_folder = folder.relative_to(input_root)
has_pngs = str(rel_folder) in folders_with_pngs
png_count = len(list(folder.rglob('*.png')))
status = "HAS PNGs" if has_pngs else "NO PNGs"
print(f"{status:10s} {rel_folder} ({png_count} PNG files)")
if __name__ == '__main__':
main()

4
compress_pngs.bat Normal file
View File

@@ -0,0 +1,4 @@
@echo off
python compress_pngs.py
pause

194
compress_pngs.py Normal file
View File

@@ -0,0 +1,194 @@
#!/usr/bin/env python3
"""
Multithreaded PNG compression script.
Compresses all PNG files in subdirectories with maximum parallelism.
"""
import os
import sys
from pathlib import Path
from concurrent.futures import ProcessPoolExecutor, as_completed
from PIL import Image
import multiprocessing
import time
def compress_png(input_path, output_path):
"""Compress a single PNG file."""
try:
input_path = Path(input_path)
output_path = Path(output_path)
# Skip if output file already exists
if output_path.exists():
original_size = os.path.getsize(input_path)
new_size = os.path.getsize(output_path)
savings = original_size - new_size
savings_pct = (savings / original_size * 100) if original_size > 0 else 0
return (str(input_path), True, None, original_size, new_size, savings_pct, True)
# Ensure output directory exists
output_path.parent.mkdir(parents=True, exist_ok=True)
original_size = os.path.getsize(input_path)
img = Image.open(input_path)
# Convert to RGB if RGBA and no transparency needed
if img.mode == 'RGBA':
# Keep RGBA for transparency support
img = img.convert('RGBA')
elif img.mode != 'RGB':
img = img.convert('RGB')
# Save with maximum compression to output path
img.save(str(output_path), 'PNG', optimize=True, compress_level=9)
new_size = os.path.getsize(output_path)
savings = original_size - new_size
savings_pct = (savings / original_size * 100) if original_size > 0 else 0
return (str(input_path), True, None, original_size, new_size, savings_pct, False)
except Exception as e:
return (str(input_path), False, str(e), 0, 0, 0, False)
def find_image_files(input_dir):
"""Find all PNG and JPG files in subdirectories."""
png_files = []
jpg_files = []
root = Path(input_dir)
for img_file in root.rglob('*'):
if img_file.suffix.lower() == '.png':
png_files.append(img_file)
elif img_file.suffix.lower() in ['.jpg', '.jpeg']:
jpg_files.append(img_file)
return png_files, jpg_files
def get_output_path(input_path, input_dir, output_dir):
"""Convert input path to output path preserving directory structure."""
input_path = Path(input_path)
input_dir = Path(input_dir)
output_dir = Path(output_dir)
# Get relative path from input directory
try:
relative_path = input_path.relative_to(input_dir)
except ValueError:
# If input_path is not relative to input_dir, use the full path
relative_path = input_path
# Create output path
return output_dir / relative_path
def format_size(size_bytes):
"""Format file size in human readable format."""
for unit in ['B', 'KB', 'MB', 'GB']:
if size_bytes < 1024.0:
return f"{size_bytes:.2f} {unit}"
size_bytes /= 1024.0
return f"{size_bytes:.2f} TB"
def main():
input_dir = Path('input')
output_dir = Path('output')
# Check if input directory exists
if not input_dir.exists():
print(f"Error: Input directory '{input_dir}' does not exist.")
print("Please create an 'input' folder and place your PNG files there.")
return
print(f"Input directory: {input_dir}")
print(f"Output directory: {output_dir}")
print("Scanning for image files...")
png_files, jpg_files = find_image_files(input_dir)
if jpg_files:
print(f"Found {len(jpg_files)} JPG/JPEG files - ignoring (skipping)")
if not png_files:
print("No PNG files found in input directory.")
return
print(f"Found {len(png_files)} PNG files to process.")
# Create output directory
output_dir.mkdir(exist_ok=True)
# Use all available CPU cores
max_workers = multiprocessing.cpu_count()
print(f"Using {max_workers} worker processes for compression...")
print("-" * 80)
compressed = 0
skipped = 0
failed = 0
total_original_size = 0
total_new_size = 0
start_time = time.time()
last_update_time = start_time
with ProcessPoolExecutor(max_workers=max_workers) as executor:
# Submit all tasks
future_to_file = {
executor.submit(compress_png, str(f), str(get_output_path(f, input_dir, output_dir))): f
for f in png_files
}
# Process results as they complete
for future in as_completed(future_to_file):
result = future.result()
file_path, success, error, orig_size, new_size, savings_pct, was_skipped = result
if success:
if was_skipped:
skipped += 1
else:
compressed += 1
total_original_size += orig_size
total_new_size += new_size
current_time = time.time()
elapsed = current_time - start_time
time_since_update = current_time - last_update_time
processed = compressed + skipped
# Update every file or every 0.5 seconds, whichever comes first
if processed == 1 or time_since_update >= 0.5:
rate = processed / elapsed if elapsed > 0 else 0
remaining = len(png_files) - processed
eta = remaining / rate if rate > 0 else 0
total_savings = total_original_size - total_new_size
total_savings_pct = (total_savings / total_original_size * 100) if total_original_size > 0 else 0
print(f"[{processed:5d}/{len(png_files)}] "
f"Compressed: {compressed} | Skipped: {skipped} | "
f"Speed: {rate:.1f} files/sec | "
f"ETA: {eta:.1f}s | "
f"Saved: {format_size(total_savings)} ({total_savings_pct:.1f}%) | "
f"Elapsed: {elapsed:.1f}s", end='\r')
last_update_time = current_time
else:
failed += 1
print(f"\n[ERROR] Failed to compress {file_path}: {error}")
total_time = time.time() - start_time
total_savings = total_original_size - total_new_size
total_savings_pct = (total_savings / total_original_size * 100) if total_original_size > 0 else 0
processed = compressed + skipped
avg_rate = processed / total_time if total_time > 0 else 0
print("\n" + "=" * 80)
print(f"Compression complete!")
print(f"Successfully compressed: {compressed} files")
print(f"Skipped (already exist): {skipped} files")
if jpg_files:
print(f"Ignored (JPG/JPEG): {len(jpg_files)} files")
print(f"Failed: {failed} files")
print(f"Total time: {total_time:.2f} seconds")
print(f"Average speed: {avg_rate:.2f} files/second")
print(f"Original size: {format_size(total_original_size)}")
print(f"Compressed size: {format_size(total_new_size)}")
print(f"Total savings: {format_size(total_savings)} ({total_savings_pct:.1f}%)")
if __name__ == '__main__':
main()