liorda commited on
Commit
011fbb6
·
verified ·
1 Parent(s): 22bb8c0

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -2,17 +2,160 @@
2
  license: fair-noncommercial-research-license
3
  pretty_name: Ego-1K
4
  size_categories:
5
- - 1M<n<10M
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  ---
7
 
8
- # Ego-1K A Large-Scale Multiview Video Dataset for Egocentric Vision
9
 
10
- We present Ego-1K, a large-scale, time-synchronized collection of egocentric multiview videos designed to advance neural 3D video synthesis,
11
- dynamic scene understanding, and embodied perception. The dataset contains 956 short (8-10s) egocentric videos taken with a custom rig with
12
- 12 synchronous cameras surrounding a VR headset worn by the user, for a total of 512k frames and 6.17M images. Scene content focuses on hand
13
- motions and hand-object interactions in different settings. Our dataset enables new ways to benchmark egocentric scene reconstruction methods,
14
- and presents unique challenges for existing 3D and 4D novel view synthesis methods due to high disparities and image motion caused by close
15
  dynamic objects and rig egomotion.
16
 
17
 
18
- ![Ego-1k-figure](https://cdn-uploads.huggingface.co/production/uploads/6978f9db14938f877a3b8b02/NglN-sPc8ZwJ8y11wBZdn.png)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: fair-noncommercial-research-license
3
  pretty_name: Ego-1K
4
  size_categories:
5
+ - 100K<n<1M
6
+ task_categories:
7
+ - depth-estimation
8
+ - image-to-3d
9
+ tags:
10
+ - egocentric
11
+ - multiview
12
+ - 3d-reconstruction
13
+ - novel-view-synthesis
14
+ - multi-camera
15
+ configs:
16
+ - config_name: default
17
+ data_files:
18
+ - split: train
19
+ path: data/train-*.parquet
20
+ - split: test
21
+ path: data/test-*.parquet
22
  ---
23
 
24
+ # Ego-1K A Large-Scale Multiview Video Dataset for Egocentric Vision
25
 
26
+ We present Ego-1K, a large-scale, time-synchronized collection of egocentric multiview videos designed to advance neural 3D video synthesis,
27
+ dynamic scene understanding, and embodied perception. The dataset contains 956 short (6.7-9.7s) egocentric videos taken with a custom rig with
28
+ 12 synchronous cameras surrounding a VR headset worn by the user, for a total of 491K frames and 5.9M images. Scene content focuses on hand
29
+ motions and hand-object interactions in different settings. Our dataset enables new ways to benchmark egocentric scene reconstruction methods,
30
+ and presents unique challenges for existing 3D and 4D novel view synthesis methods due to high disparities and image motion caused by close
31
  dynamic objects and rig egomotion.
32
 
33
 
34
+ ![Ego-1k-figure](figures/rig.png)
35
+
36
+ ## Getting Started
37
+
38
+ See [`quickstart.ipynb`](quickstart.ipynb) for a runnable walkthrough that loads metadata, streams images, and visualizes a 12-camera multiview frame.
39
+
40
+ ## Key Statistics
41
+
42
+ | Property | Value |
43
+ |----------|-------|
44
+ | Total recordings | 956 |
45
+ | Train split | 860 recordings |
46
+ | Test split | 96 recordings |
47
+ | Duration per recording | 6.7-9.7 seconds at 60 Hz |
48
+ | Frames per recording | 404-583 (mean 514, median 530) |
49
+ | Cameras | 12 (6 rectified stereo pairs) |
50
+ | Image resolution | 1280 x 1280 pixels (rectified pinhole, 120 deg HFOV) |
51
+ | Total frames | 490,966 |
52
+ | Total images | 5,891,592 |
53
+ | Total duration | 2.3 hours |
54
+ | Total shards | 10,216 |
55
+ | Total dataset size | ~11 TB (WebDataset tar shards) |
56
+ | Typical shard size | ~1.5 GB |
57
+
58
+ ![distributions](figures/distributions.png)
59
+
60
+ ## Dataset Structure
61
+
62
+ ```text
63
+ ego-1k/
64
+ ├── data/
65
+ │ ├── train-<scene_id>.parquet # Per-scene metadata index
66
+ │ └── test-<scene_id>.parquet
67
+ └── shards/
68
+ ├── train/
69
+ │ └── <scene_id>/
70
+ │ ├── <scene_id>-0000.tar # WebDataset tar shards (~1.5 GB each)
71
+ │ ├── <scene_id>-0001.tar
72
+ │ └── ...
73
+ └── test/
74
+ └── <scene_id>/
75
+ └── <scene_id>-0000.tar
76
+ ```
77
+
78
+ ### Tar Shard Contents
79
+
80
+ Each tar sample represents one frame across all 12 cameras:
81
+
82
+ ```text
83
+ <scene_id>/<frame_id:06d>.200-1.png # Raw PNG bytes (1280x1280)
84
+ <scene_id>/<frame_id:06d>.200-2.png
85
+ ...
86
+ <scene_id>/<frame_id:06d>.200-12.png
87
+ <scene_id>/<frame_id:06d>.metadata.json # Pose, rig calibration, scene info
88
+ ```
89
+
90
+ The `metadata.json` per sample contains:
91
+
92
+ | Field | Type | Description |
93
+ |-------|------|-------------|
94
+ | `scene_id` | string | Recording identifier |
95
+ | `frame_id` | int | Frame index (0-indexed) |
96
+ | `timestamp_ns` | int | Frame timestamp in nanoseconds |
97
+ | `pose` | list | 4x4 device-to-world transform (key absent if unavailable) |
98
+ | `rig_calibration` | object | Per-camera intrinsics (`K`) and extrinsics (`E`) |
99
+ | `source` | string | Capture campaign: `OVD_M1` (lab, 513 recordings), `OVD_M2` (apartment, 414), `DD4` (29) |
100
+ | `lux_bins` | string | Lighting level: `51-75`, `76-100`, `101-200`, `201-400`, `401-1000`, `1001+` |
101
+ | `tags` | list | Scene diversity tags |
102
+
103
+ ## Parquet Schema
104
+
105
+ Each row represents a single frame (one timestamp across all 12 cameras):
106
+
107
+ | Column | Type | Description |
108
+ |--------|------|-------------|
109
+ | `scene_id` | string | Recording identifier |
110
+ | `frame_id` | int32 | Frame index within the recording (0-indexed; number of frames varies per scene, range 404-583) |
111
+ | `timestamp_ns` | int64 | Frame timestamp in nanoseconds |
112
+ | `source` | string | Capture campaign: `OVD_M1` (lab), `OVD_M2` (apartment), `DD4` |
113
+ | `lux_bins` | string | Lighting level: `51-75`, `76-100`, `101-200`, `201-400`, `401-1000`, `1001+` |
114
+ | `tags` | string | JSON list of scene diversity tags (85 unique tags covering garments, furnishings, lighting, pose, objects) |
115
+ | `shard_name` | string | Relative path to the tar shard containing this frame's images (e.g., `shards/train/<scene_id>/<scene_id>-0002.tar`) |
116
+ | `pose` | string | JSON: 4x4 device-to-world transform matrix for this frame (null if pose unavailable) |
117
+ | `rig_calibration` | string | JSON: per-camera intrinsics (`K`: 3x3) and extrinsics (`E`: 4x4), static per scene (repeated for each frame for convenience) |
118
+
119
+ ### Calibration Details
120
+
121
+ The `rig_calibration` column contains a JSON object keyed by camera name (`200-1` through `200-12`), each with:
122
+ - **`K`**: 3x3 intrinsic matrix (rectified pinhole projection, 120 deg horizontal FOV)
123
+ - **`E`**: 4x4 extrinsic matrix (camera-to-device transform)
124
+
125
+ The `pose` column contains the 4x4 device-to-world transform, which changes per frame as the headset moves.
126
+
127
+ ## Usage
128
+
129
+ `load_dataset` returns frame-level **metadata only** (poses, calibration, scene info). Images are stored in WebDataset tar shards — use the `webdataset` library to stream them. See [`quickstart.ipynb`](quickstart.ipynb) for a full working example.
130
+
131
+ ### WebDataset (Recommended for Training)
132
+
133
+ Stream tar shards for high-throughput sequential access — no per-file API calls. See the notebook for the full `decode_sample` implementation. To wrap it in a PyTorch DataLoader:
134
+
135
+ ```python
136
+ dataset = wds.WebDataset(shard_urls, nodesplitter=wds.split_by_node, shardshuffle=True).map(decode_sample)
137
+ loader = torch.utils.data.DataLoader(dataset, batch_size=4, num_workers=4)
138
+
139
+ for batch in loader:
140
+ images = batch["images"] # (B, N_cams, 3, 1280, 1280)
141
+ break
142
+ ```
143
+
144
+ ### Parquet Metadata (Random Access)
145
+
146
+ The Parquet files contain frame-level metadata only (poses, calibration, scene info) — images are stored in the tar shards. Use the `shard_name` column to locate which tar file contains a given frame's images.
147
+
148
+ ```python
149
+ shard_url = f"https://huggingface.co/datasets/facebook/ego-1k/resolve/main/{example['shard_name']}"
150
+ ```
151
+
152
+ ## Citation
153
+
154
+ ```bibtex
155
+ @inproceedings{ego1k2026,
156
+ title={{Ego-1K}: A Large-Scale Multiview Video Dataset for Egocentric Vision},
157
+ author={Jae Yong Lee and Daniel Scharstein and Akash Bapat and Hao Hu and Andrew Fu and Haoru Zhao and Paul Sammut and Xiang Li and Stephen Jeapes and Anik Gupta and Lior David and Saketh Madhuvarasu and Jay Girish Joshi and Jason Wither},
158
+ booktitle={CVPR},
159
+ year={2026}
160
+ }
161
+ ```
data/train-1731264974453484.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13763e5c941beb67c4c9d6fb1497992971e793519a6e751fb8f5137ee03bfc01
3
+ size 72238
figures/distributions.png ADDED

Git LFS Details

  • SHA256: 039e20654445b0040434747e7f03a76a2c682c8a8edd77d3893fc626743b0ba3
  • Pointer size: 131 Bytes
  • Size of remote file: 176 kB
figures/rig.png ADDED

Git LFS Details

  • SHA256: 58d7d02a454bcea9f4df20524a354361f0e95f8a056c47891623c5b32f0ad822
  • Pointer size: 132 Bytes
  • Size of remote file: 4.14 MB
quickstart.ipynb ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "cell_type": "markdown",
5
+ "metadata": {},
6
+ "source": "# Ego-1K Quickstart\n\nThis notebook demonstrates how to load and explore the [Ego-1K](https://huggingface.co/datasets/facebook/ego-1k) dataset — a large-scale, time-synchronized collection of egocentric multiview videos with 12 cameras, 956 recordings, and ~491K frames.\n\n**What you'll learn:**\n1. Load frame-level metadata (poses, calibration, scene info) from Parquet\n2. Parse rig calibration and device pose\n3. Stream images from WebDataset tar shards\n4. Visualize a multiview frame (12 cameras)\n\n**Prerequisites:**\n```bash\npip install datasets webdataset torch numpy Pillow huggingface_hub matplotlib\n```"
7
+ },
8
+ {
9
+ "cell_type": "code",
10
+ "execution_count": null,
11
+ "metadata": {},
12
+ "outputs": [],
13
+ "source": "import io\nimport json\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport torch\nimport webdataset as wds\nfrom datasets import load_dataset\nfrom huggingface_hub import HfApi\nfrom PIL import Image\n\nREPO_ID = \"facebook/ego-1k\"\nCAMERAS = [f\"200-{i}\" for i in range(1, 13)] # 12 cameras"
14
+ },
15
+ {
16
+ "cell_type": "code",
17
+ "execution_count": null,
18
+ "metadata": {},
19
+ "outputs": [],
20
+ "source": "# Load frame-level metadata from Parquet (no images downloaded)\nds = load_dataset(REPO_ID, split=\"test\")\nprint(f\"{len(ds)} frames, columns: {ds.column_names}\")\n\nexample = ds[0]\nprint(f\"Example: scene={example['scene_id']}, frame={example['frame_id']}, \"\n f\"source={example['source']}, lux={example['lux_bins']}\")"
21
+ },
22
+ {
23
+ "cell_type": "code",
24
+ "execution_count": null,
25
+ "metadata": {},
26
+ "outputs": [],
27
+ "source": "# Parse rig calibration and device pose from one example\nrig = json.loads(example[\"rig_calibration\"])\npose = json.loads(example[\"pose\"]) if example[\"pose\"] else None\n\nK = np.array(rig[\"200-1\"][\"K\"])\nE = np.array(rig[\"200-1\"][\"E\"])\nprint(f\"Cameras: {sorted(rig.keys())}\")\nprint(f\"Intrinsics K: {K.shape}, Extrinsics E: {E.shape}\")\nif pose is not None:\n print(f\"Device pose: {np.array(pose).shape}\")"
28
+ },
29
+ {
30
+ "cell_type": "code",
31
+ "execution_count": null,
32
+ "metadata": {},
33
+ "outputs": [],
34
+ "source": "def decode_wds_sample(sample, cameras):\n \"\"\"Decode a WebDataset sample into images, intrinsics, extrinsics, and pose.\"\"\"\n meta = json.loads(sample[\"metadata.json\"])\n rig = meta[\"rig_calibration\"]\n\n images, intrinsics, extrinsics = [], [], []\n for cam in cameras:\n img = Image.open(io.BytesIO(sample[f\"{cam}.png\"])).convert(\"RGB\")\n img_tensor = torch.from_numpy(np.array(img)).permute(2, 0, 1).float() / 255.0\n images.append(img_tensor)\n intrinsics.append(torch.tensor(rig[cam][\"K\"], dtype=torch.float32))\n extrinsics.append(torch.tensor(rig[cam][\"E\"], dtype=torch.float32))\n\n pose_raw = meta.get(\"pose\")\n pose = (\n torch.tensor(pose_raw, dtype=torch.float32)\n if pose_raw is not None\n else torch.zeros(4, 4)\n )\n\n return {\n \"images\": torch.stack(images),\n \"intrinsics\": torch.stack(intrinsics),\n \"extrinsics\": torch.stack(extrinsics),\n \"pose\": pose,\n \"scene_id\": meta[\"scene_id\"],\n \"frame_id\": meta[\"frame_id\"],\n }\n\n\n# Discover tar shards for the test split\napi = HfApi()\nall_files = api.list_repo_files(repo_id=REPO_ID, repo_type=\"dataset\")\nshard_files = sorted(\n f for f in all_files if f.startswith(\"shards/test/\") and f.endswith(\".tar\")\n)\n\n# Stream one frame from the first shard (each shard is ~1.5 GB)\nshard = shard_files[0]\nbase_url = f\"https://huggingface.co/datasets/{REPO_ID}/resolve/main\"\nshard_url = f\"pipe:curl -sfL '{base_url}/{shard}'\"\n\npipeline = wds.WebDataset(\n [shard_url], nodesplitter=wds.split_by_node, shardshuffle=False\n).map(lambda s: decode_wds_sample(s, CAMERAS))\n\nsample = next(iter(pipeline))\nprint(f\"scene={sample['scene_id']}, frame={sample['frame_id']}, \"\n f\"images={sample['images'].shape}\")"
35
+ },
36
+ {
37
+ "cell_type": "code",
38
+ "execution_count": null,
39
+ "metadata": {},
40
+ "outputs": [],
41
+ "source": [
42
+ "# Visualize all 12 camera views from the streamed frame\n",
43
+ "fig, axes = plt.subplots(2, 6, figsize=(24, 8))\n",
44
+ "fig.suptitle(\n",
45
+ " f\"Ego-1K — Scene {sample['scene_id']}, Frame {sample['frame_id']}\",\n",
46
+ " fontsize=14,\n",
47
+ ")\n",
48
+ "\n",
49
+ "for idx, ax in enumerate(axes.flat):\n",
50
+ " # Convert from (C, H, W) float tensor to (H, W, C) numpy for display\n",
51
+ " img = sample[\"images\"][idx].permute(1, 2, 0).numpy()\n",
52
+ " ax.imshow(img)\n",
53
+ " ax.set_title(CAMERAS[idx], fontsize=10)\n",
54
+ " ax.axis(\"off\")\n",
55
+ "\n",
56
+ "plt.tight_layout()\n",
57
+ "plt.show()"
58
+ ]
59
+ }
60
+ ],
61
+ "metadata": {
62
+ "kernelspec": {
63
+ "display_name": "Python 3",
64
+ "language": "python",
65
+ "name": "python3"
66
+ },
67
+ "language_info": {
68
+ "name": "python",
69
+ "version": "3.10.0"
70
+ }
71
+ },
72
+ "nbformat": 4,
73
+ "nbformat_minor": 4
74
+ }
shards/train/1731264974453484/1731264974453484-0000.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06c2176845a98b13c2d83e7a170c57ebb760c45c5b7f084d7c57237e03eb2028
3
+ size 1781473280
shards/train/1731264974453484/1731264974453484-0001.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d67ddf3e8afb5d493bf5b29c1788ba3eb3826ffe34ff0838ac02330fc99e3feb
3
+ size 1785630720
shards/train/1731264974453484/1731264974453484-0002.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ada8961b330e85651316e41a0a6f0bbed7abf21cdb7d231b5e921d01f3abb4a
3
+ size 1785190400
shards/train/1731264974453484/1731264974453484-0003.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3bfa001017a3825898655562a8ec290ed3a2bb7fe823d68dac75d369fdbd77e
3
+ size 1783889920
shards/train/1731264974453484/1731264974453484-0004.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13711ed46e893e01d9839a41f90e1a6d1be071f69033ff8fcdf3307b54db144c
3
+ size 1785774080
shards/train/1731264974453484/1731264974453484-0005.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d020fa14fd30948e13f890f8dec90b17e29b8b78cbe3849a77b6508ce3e6bdd
3
+ size 1779752960
shards/train/1731264974453484/1731264974453484-0006.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d021102a56a682c1c3361bb453ee8883fc3c231b7f8643499d1960615038223
3
+ size 1778759680
shards/train/1731264974453484/1731264974453484-0007.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2544553e695dfab9a51641070fe84cbd200d739a6c2686ef8e67a16433d331c
3
+ size 1777100800
shards/train/1731264974453484/1731264974453484-0008.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1db18796dc90d73335341ca18ce85b3754903f3e660dd13d2f1c6f1885a701e7
3
+ size 1778841600
shards/train/1731264974453484/1731264974453484-0009.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9fe6f55b24c36bf1438bd3a53b6219d79fe25986c06a9dc980c96764db41de7
3
+ size 1792399360
shards/train/1731264974453484/1731264974453484-0010.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ef4423d7b4b97c04a403d489f75323a71da017a5a804a8778253aa9aaa405b8
3
+ size 1589811200