| |
| r""" |
| Builds a WebDataset from the Cityscapes Video dataset. |
| |
| Adapted from the `WebDataset documentation<https://github.com/webdataset/webdataset/>`_. |
| """ |
|
|
| import itertools |
| import collections |
| import typing as T |
| from pprint import pformat |
| import argparse |
| import multiprocessing as mp |
| import tarfile |
| import pandas as pd |
| from io import BytesIO |
| import json |
|
|
| from pathlib import Path |
| from tqdm import tqdm |
|
|
|
|
| def parse_args(): |
| ap = argparse.ArgumentParser( |
| description="Build a WebDataset from the Cityscapes Video dataset." |
| ) |
|
|
| |
| ap.add_argument( |
| "--shard-size", |
| "-s", |
| type=int, |
| default=10, |
| help=("Number of sequences per shard."), |
| ) |
| ap.add_argument( |
| "--name", |
| "-n", |
| type=str, |
| default="csvps", |
| help=( |
| "Name of the dataset. This will be used as the prefix for the tar files." |
| ), |
| ) |
| ap.add_argument( |
| "--variant", |
| type=str, |
| default="", |
| help=( |
| "When passing different manifest variants, this will be used to postfix " |
| "each split such that the resulting dataset name is unique." |
| ), |
| ) |
| ap.add_argument( |
| "--force", "-f", action="store_true", help="Overwrite existing data." |
| ) |
| ap.add_argument( |
| "--splits", nargs="+", default=["train", "val", "test"], help="Splits to build." |
| ) |
| ap.add_argument("--compression", "-c", default="", help="Compression to use") |
|
|
| |
| ap.add_argument("manifest", type=Path, help="Path to the manifest CSV file.") |
| ap.add_argument("data", type=Path, help="Path to the Cityscapes Video dataset.") |
| ap.add_argument("output", type=Path, help="Path to the output directory.") |
|
|
| rt = ap.parse_args() |
|
|
| |
| if rt.shard_size < 1: |
| ap.error("Shard size must be a positive integer.") |
| if rt.name == "": |
| ap.error("Name must be a non-empty string.") |
| if not rt.name.isalnum() and not rt.name.islower(): |
| ap.error("Name must be a lowercase alpha-numeric string.") |
| if rt.variant != "" and not rt.variant.isalnum() and not rt.variant.islower(): |
| ap.error("Variant must be a lowercase alpha-numeric string.") |
| if not rt.manifest.exists(): |
| ap.error(f"Manifest file not found: {rt.manifest}") |
| if not rt.data.exists(): |
| ap.error(f"Data directory not found: {rt.data}") |
| if not rt.output.exists(): |
| rt.output.mkdir(parents=True) |
| print(f"Created output directory: {rt.output}") |
|
|
| return rt |
|
|
|
|
| PAD_TO: T.Final[int] = 6 |
|
|
|
|
| def pad_number(n: int) -> str: |
| r""" |
| For sorting, numbers are padded with zeros to a fixed width. |
| """ |
| if not isinstance(n, int): |
| msg = f"Expected an integer, got {n} of type {type(n)}" |
| raise TypeError(msg) |
| return f"{n:0{PAD_TO}d}" |
|
|
|
|
| def read_timestamp(path: Path) -> int: |
| with path.open("r") as f: |
| ts = f.read().strip() |
| if not ts.isdigit(): |
| msg = f"Expected a timestamp, got {ts} from {path}" |
| raise ValueError(msg) |
| return int(ts) |
|
|
|
|
| def write_bytes(tar: tarfile.TarFile, bt: bytes, arc: str): |
| r""" " |
| Simple utility to write the bytes (e.g. metadata json) directly from memory to |
| the tarfile, since these do not exist as a file. |
| """ |
| with BytesIO() as buf: |
| buf.write(bt) |
|
|
| |
| |
| tar_info = tarfile.TarInfo(arc) |
| tar_info.size = buf.tell() |
|
|
| |
| buf.seek(0) |
|
|
| tar.addfile(tar_info, buf) |
|
|
|
|
| def find_sequence_files( |
| seq: int, |
| group: pd.DataFrame, |
| *, |
| data_dir: Path, |
| dataset_name: str, |
| compression: str, |
| missing_ok: bool = False, |
| frame_inputs: T.Sequence[str] = ("image.png", "vehicle.json"), |
| frame_annotations: T.Sequence[str] = ("panoptic.png", "depth.tiff"), |
| sequence_data: T.Sequence[str] = ("camera.json",), |
| separator: str = "/", |
| ) -> T.Iterator[tuple[Path | bytes, str]]: |
| seq_pad = pad_number(seq) |
| seq_dir = data_dir / seq_pad |
|
|
| group = group.sort_values("frame") |
|
|
| |
| primary_keys = group.index.tolist() |
| frame_numbers = list(map(pad_number, group["frame"].tolist())) |
|
|
| for i, meta in enumerate( |
| group.drop(columns=["sequence", "frame", "split"]).to_dict( |
| orient="records", index=True |
| ) |
| ): |
| frame_06 = frame_numbers[i] |
| is_ann = meta["is_annotated"] |
|
|
| |
| meta["primary_key"] = primary_keys[i] |
|
|
| |
| for var in frame_inputs + frame_annotations: |
| path_file = seq_dir / f"{frame_06}.{var}" |
| if not path_file.exists(): |
| if missing_ok or (var in frame_annotations and not is_ann): |
| continue |
| msg = f"File not found: {path_file}" |
| raise FileNotFoundError(msg) |
|
|
| yield ( |
| path_file, |
| separator.join( |
| ( |
| dataset_name, |
| |
| path_file.relative_to(data_dir).as_posix().replace("/", "."), |
| ) |
| ), |
| ) |
|
|
| |
| path_ts = seq_dir / f"{frame_06}.timestamp.txt" |
| if not path_ts.exists(): |
| if not missing_ok: |
| msg = f"Timestamp file not found: {path_ts}" |
| raise FileNotFoundError(msg) |
| meta["timestamp"] = None |
| else: |
| meta["timestamp"] = read_timestamp(path_ts) |
|
|
| |
| yield ( |
| json.dumps(meta).encode("utf-8"), |
| f"{dataset_name}/{seq_pad}.{frame_06}.metadata.json", |
| ) |
|
|
| |
| for var in sequence_data: |
| path_file = seq_dir.with_suffix("." + var) |
| if not path_file.exists(): |
| if missing_ok: |
| continue |
| msg = f"File not found: {path_file}" |
| raise FileNotFoundError(msg) |
|
|
| yield ( |
| path_file, |
| separator.join( |
| ( |
| dataset_name, |
| |
| path_file.relative_to(data_dir).as_posix(), |
| ) |
| ), |
| ) |
|
|
| |
| yield ( |
| json.dumps(frame_numbers).encode("utf-8"), |
| f"{dataset_name}/{seq_pad}.frames.json", |
| ) |
|
|
|
|
| def run_collector( |
| seq: int, group: pd.DataFrame, kwargs: dict |
| ) -> tuple[int, list[tuple[Path | bytes, str]]]: |
| r""" |
| Worker that collects the files for a single sequence. |
| """ |
| return (seq, list(find_sequence_files(seq, group, **kwargs))) |
|
|
|
|
| def run_writer( |
| tar_path: Path, items: list[list[tuple[Path | bytes, str]]], compression: str = "gz" |
| ) -> None: |
| r""" |
| Worker that writes the files to a tar archive. |
| """ |
| if compression != "": |
| tar_path = tar_path.with_suffix(f".tar.{compression}") |
| write_mode = f"w:{compression}" |
| else: |
| tar_path = tar_path.with_suffix(".tar") |
| write_mode = "w" |
|
|
| with tarfile.open(tar_path, write_mode) as tar: |
| for item in itertools.chain.from_iterable(items): |
| try: |
| path, arc = item |
| except ValueError: |
| msg = f"Expected a tuple of length 2, got {item}" |
| raise ValueError(msg) |
|
|
| if isinstance(path, Path): |
| tar.add(path, arcname=arc) |
| else: |
| write_bytes(tar, path, arc) |
|
|
|
|
| def build_shard( |
| mfst: pd.DataFrame, |
| *, |
| tar_dir: Path, |
| shard_size: int, |
| **kwargs, |
| ): |
| |
| tar_dir.mkdir(exist_ok=True, parents=True) |
|
|
| write_log = collections.defaultdict(list) |
|
|
| |
| |
| |
| n_groups = len(mfst["sequence"].unique()) |
| n_shards = n_groups // shard_size |
|
|
| targets = [None] * n_groups |
|
|
| |
| n_proc = min(mp.cpu_count(), 16) |
| with mp.Pool(n_proc) as pool: |
| write_jobs: list[mp.AsyncResult] = [] |
|
|
| |
| with tqdm(total=n_groups, desc="Collecting data") as pbar_group: |
| for seq, files in pool.starmap( |
| run_collector, |
| [(seq, group, kwargs) for seq, group in mfst.groupby("sequence")], |
| chunksize=min(8, shard_size), |
| ): |
| assert targets[seq] is None, f"Duplicate sequence: {seq}" |
|
|
| pbar_group.update() |
|
|
| |
| targets[seq] = files |
|
|
| |
| shard_index = seq // shard_size |
| shard_offset = shard_index * shard_size |
| shard_specs = targets[shard_offset : shard_offset + shard_size] |
|
|
| |
| shard_06 = pad_number(shard_index) |
|
|
| write_log[shard_06].append(pad_number(seq)) |
|
|
| |
| if all(s is not None for s in shard_specs): |
| tar_path = tar_dir / shard_06 |
|
|
| write_jobs.append( |
| pool.apply_async( |
| run_writer, |
| (tar_path, shard_specs, ""), |
| ) |
| ) |
|
|
| |
| with tqdm(total=n_shards, desc="Writing shards") as pbar_shard: |
| for j in write_jobs: |
| j.get() |
| pbar_shard.update() |
|
|
| pool.close() |
| pool.join() |
|
|
| print("Created shard files:\n" + pformat(dict(write_log))) |
|
|
|
|
| def main(): |
| args = parse_args() |
| manifest = pd.read_csv(args.manifest, index_col="primary_key") |
|
|
| |
| for split in args.splits: |
| split_out = "-".join([s for s in (split, args.variant) if len(s) > 0]) |
| tar_dir = args.output / split_out |
|
|
| if tar_dir.exists(): |
| if args.force: |
| print(f"Removing existing dataset: {tar_dir}") |
| for f in tar_dir.glob("*.tar"): |
| f.unlink() |
| else: |
| msg = f"Dataset already exists: {tar_dir}" |
| raise FileExistsError(msg) |
|
|
| print(f"Generating {split_out} split...") |
|
|
| build_shard( |
| manifest[manifest["split"] == split], |
| tar_dir=tar_dir, |
| data_dir=args.data / split, |
| shard_size=args.shard_size, |
| dataset_name=f"{args.name}-{split_out}", |
| missing_ok=True, |
| compression=args.compression |
| ) |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|