| import torch |
| import numpy as np |
| from pathlib import Path |
| from transnetv2_pytorch import TransNetV2 |
| import json |
| import re |
|
|
| |
| K = 3 |
| MIN_DURATION_FRAMES = 2 |
| MIN_CONFIDENCE = 0.02 |
|
|
| data_dir = Path("data/animations") |
| files = sorted(data_dir.glob("sample-*.webp")) |
| print(f"Found {len(files)} files to process.") |
| |
| def get_best_device(): |
| if torch.cuda.is_available(): |
| return torch.device("cuda") |
| elif torch.backends.mps.is_available(): |
| return torch.device("mps") |
| |
| else: |
| return torch.device("cpu") |
|
|
| def load_original_frames(filepath): |
| """Load original frames from an animated webp file as PIL Images.""" |
| from PIL import Image |
| im = Image.open(filepath) |
| frames = [] |
| try: |
| while True: |
| frames.append(im.convert("RGB")) |
| im.seek(im.tell() + 1) |
| except EOFError: |
| pass |
| return frames |
|
|
| def save_prediction_plot(single_frame_pred, original_frames, filename, interval=5, title=None): |
| """ |
| Save a plot of single frame predictions with thumbnails annotated at regular intervals. |
| """ |
| import matplotlib.pyplot as plt |
| from matplotlib.offsetbox import OffsetImage, AnnotationBbox |
| import numpy as np |
|
|
| plt.figure(figsize=(12, 4)) |
| if title: |
| plt.title(title) |
| plt.plot(single_frame_pred) |
| ax = plt.gca() |
| |
| for idx in range(0, len(original_frames), interval): |
| thumb = original_frames[idx].resize((64, 64)) |
| imagebox = OffsetImage(np.array(thumb), zoom=0.5) |
| ab = AnnotationBbox(imagebox, (idx, single_frame_pred[idx]), frameon=False, box_alignment=(0.5, -0.1)) |
| ax.add_artist(ab) |
| plt.xlabel("Frame") |
| plt.ylabel("Prediction") |
| plt.tight_layout() |
| plt.savefig(filename) |
| plt.close() |
|
|
| def save_timeline_jpg(frames, scene_change_indices, filename, interval=5, roi_radius=2, title=None, single_frame_pred=None): |
| """ |
| Save a timeline JPG with thumbnails every `interval` frames and every frame near scene changes. |
| Scene change regions are highlighted. Each thumbnail is annotated with its frame index. |
| """ |
| import matplotlib.pyplot as plt |
| from matplotlib.offsetbox import OffsetImage, AnnotationBbox |
| import matplotlib.patches as mpatches |
| import numpy as np |
|
|
| |
| frames_to_render = set(range(0, len(frames), interval)) |
| for idx in scene_change_indices: |
| for offset in range(-roi_radius, roi_radius+1): |
| fidx = idx + offset |
| if 0 <= fidx < len(frames): |
| frames_to_render.add(fidx) |
| frames_to_render = sorted(frames_to_render) |
|
|
| |
| n = len(frames_to_render) |
| x_positions = list(range(n)) |
|
|
| fig, ax = plt.subplots(figsize=(max(8, n*0.5), 3)) |
| ax.set_xlim(-1, n) |
| ax.set_ylim(0, 1) |
| ax.axis('off') |
|
|
| |
| for idx in scene_change_indices: |
| region = [i for i, fidx in enumerate(frames_to_render) if abs(fidx-idx) <= roi_radius] |
| if region: |
| start, end = region[0], region[-1] |
| rect = mpatches.Rectangle((start-0.5, 0.05), end-start+1, 0.9, color='yellow', alpha=0.2) |
| ax.add_patch(rect) |
|
|
| |
| last_frames = set(scene_change_indices) |
| first_frames = set(idx + 1 for idx in scene_change_indices if idx + 1 < len(frames)) |
|
|
| |
| for i, fidx in enumerate(frames_to_render): |
| thumb = frames[fidx].resize((32, 32)) |
| imagebox = OffsetImage(np.array(thumb), zoom=0.7) |
| |
| if fidx in last_frames: |
| bboxprops = dict(edgecolor='red', linewidth=2) |
| elif fidx in first_frames: |
| bboxprops = dict(edgecolor='green', linewidth=2) |
| else: |
| bboxprops = None |
| ab = AnnotationBbox( |
| imagebox, |
| (x_positions[i], 0.6), |
| frameon=True, |
| box_alignment=(0.5, 0.5), |
| bboxprops=bboxprops |
| ) |
| ax.add_artist(ab) |
| |
| ax.text(x_positions[i], 0.32, str(fidx), ha='center', va='center', fontsize=9, color='black', bbox=dict(facecolor='white', edgecolor='none', alpha=0.8, boxstyle='round,pad=0.2')) |
| |
| if single_frame_pred is not None: |
| pred_val = single_frame_pred[fidx] |
| |
| if isinstance(pred_val, np.ndarray): |
| pred_val = float(pred_val.squeeze()) |
| ax.text(x_positions[i], 0.18, f"{pred_val:.2f}", ha='center', va='center', fontsize=8, color='blue', bbox=dict(facecolor='white', edgecolor='none', alpha=0.7, boxstyle='round,pad=0.2')) |
|
|
| if title: |
| ax.text(0, 0.95, title, fontsize=12, ha='left', va='top', color='navy') |
| plt.tight_layout() |
| plt.savefig(filename, dpi=150) |
| plt.close(fig) |
|
|
| def frames_to_video_tensor(frames): |
| """Convert a list of PIL frames to a torch tensor of shape (num_frames, 27, 48, 3) and dtype uint8.""" |
| import numpy as np |
| from PIL import Image |
| processed = [] |
| for frame in frames: |
| arr = np.array(frame.resize((48, 27), resample=Image.Resampling.BILINEAR), dtype=np.uint8) |
| processed.append(torch.from_numpy(arr)) |
| return torch.stack(processed) |
|
|
| def detect_scene_changes(frames): |
| video_tensor = frames_to_video_tensor(frames) |
| video_tensor = video_tensor.unsqueeze(0).to(device) |
| with torch.no_grad(): |
| single_frame_logits, all_frame_logits = model(video_tensor) |
| |
| single_frame_logits_np = single_frame_logits.cpu().numpy().squeeze() |
| all_frame_logits_np = all_frame_logits["many_hot"].cpu().numpy().squeeze() |
| single_frame_pred = torch.sigmoid(single_frame_logits).cpu().numpy().squeeze() |
| all_frame_pred_np = torch.sigmoid(all_frame_logits["many_hot"]).cpu().numpy().squeeze() |
| return { |
| "single_frame_pred": single_frame_pred, |
| "all_frame_pred": all_frame_pred_np, |
| "single_frame_logits": single_frame_logits_np, |
| "all_frame_logits": all_frame_logits_np, |
| } |
|
|
| def cached_detect_scene_changes(file, original_frames): |
| """Detect scene changes with caching to avoid redundant computation.""" |
| match = re.search(r"sample-(\d+)", file.name) |
| sample_num = match.group(1) if match else "unknown" |
| transnetv2_json = file.parent / f"sample-{sample_num}.transnetv2.json" |
| if transnetv2_json.exists(): |
| with open(transnetv2_json, "r") as f: |
| result = json.load(f) |
| result["single_frame_pred"] = np.array(result["single_frame_pred"]) |
| result["all_frame_pred"] = np.array(result["all_frame_pred"]) |
| result["single_frame_logits"] = np.array(result["single_frame_logits"]) |
| result["all_frame_logits"] = np.array(result["all_frame_logits"]) |
| else: |
| result = detect_scene_changes(original_frames) |
| |
| with open(transnetv2_json, "w") as f: |
| json.dump({ |
| "single_frame_pred": result["single_frame_pred"].tolist(), |
| "all_frame_pred": result["all_frame_pred"].tolist(), |
| "single_frame_logits": result["single_frame_logits"].tolist(), |
| "all_frame_logits": result["all_frame_logits"].tolist() |
| }, f, indent=2) |
| return result |
|
|
| if __name__ == "__main__": |
| device = get_best_device() |
| print(f"Using device: {device}") |
| model = TransNetV2() |
| state_dict = torch.load("transnetv2-pytorch-weights.pth") |
| model.load_state_dict(state_dict) |
| model.eval().to(device) |
|
|
| for file in files: |
| match = re.search(r"sample-(\d+)", file.name) |
| sample_num = match.group(1) if match else "unknown" |
| original_frames = load_original_frames(file) |
| result = cached_detect_scene_changes(file, original_frames) |
| |
| |
| single_frame_pred = result["single_frame_pred"] |
| |
| |
| valid_indices = np.arange(1, len(single_frame_pred) - 1) |
| valid_preds = single_frame_pred[1:-1] |
| |
| sorted_indices = valid_indices[np.argsort(valid_preds)[::-1]] |
| scene_change_indices = [] |
| scene_cut_confidences = [] |
| for idx in sorted_indices: |
| if all(abs(idx - prev) >= MIN_DURATION_FRAMES for prev in scene_change_indices): |
| scene_change_indices.append(int(idx)) |
| scene_cut_confidences.append(float(single_frame_pred[idx])) |
| if len(scene_change_indices) >= (K - 1): |
| break |
|
|
| |
| failed = any(conf < MIN_CONFIDENCE for conf in scene_cut_confidences) |
|
|
| print(f"File: {file.name}, Frames: {len(original_frames)}, Scene Changes: {len(scene_change_indices)}, Success: {not failed}") |
| |
| json_filename = file.parent / f"sample-{sample_num}.json" |
| with open(json_filename, "w") as f: |
| json.dump({ |
| "num_frames": len(original_frames), |
| "scene_change_indices": scene_change_indices, |
| "scene_cut_confidences": scene_cut_confidences, |
| |
| "params": { |
| "k": K, |
| "min_duration_frames": MIN_DURATION_FRAMES, |
| "min_confidence": MIN_CONFIDENCE, |
| }, |
| "success": not failed |
| }, f, indent=2) |
| |
| timeline_filename = file.parent / f"sample-{sample_num}.timeline.jpg" |
| plot_filename = file.parent / f"sample-{sample_num}.plot.jpg" |
| save_timeline_jpg( |
| frames=original_frames, |
| scene_change_indices=scene_change_indices, |
| filename=timeline_filename, |
| interval=10, |
| roi_radius=2, |
| title=f"Timeline: {file.name}", |
| single_frame_pred=result["single_frame_pred"] |
| ) |
| |
| save_prediction_plot( |
| single_frame_pred=result["single_frame_pred"], |
| original_frames=original_frames, |
| filename=plot_filename, |
| interval=5, |
| title=f"Single Frame Predictions: {file.name}" |
| ) |
|
|
|
|