|
|
--- |
|
|
dataset_info: |
|
|
features: |
|
|
- name: id |
|
|
dtype: int64 |
|
|
- name: input |
|
|
list: |
|
|
- name: type |
|
|
dtype: string |
|
|
- name: content |
|
|
dtype: string |
|
|
- name: output |
|
|
struct: |
|
|
- name: veo3 |
|
|
list: string |
|
|
- name: framepack |
|
|
list: string |
|
|
- name: framepack_seleted_video |
|
|
dtype: string |
|
|
- name: hunyuan |
|
|
list: string |
|
|
- name: hunyuan_seleted_video |
|
|
dtype: string |
|
|
- name: wan2.2-14b |
|
|
list: string |
|
|
- name: wan2.2-14b_seleted_video |
|
|
dtype: string |
|
|
- name: wan2.2-5b |
|
|
list: string |
|
|
- name: wan2.2-5b_seleted_video |
|
|
dtype: string |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 98746 |
|
|
num_examples: 99 |
|
|
download_size: 36034 |
|
|
dataset_size: 98746 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/train-* |
|
|
--- |
|
|
|
|
|
# Visual-Intelligence |
|
|
|
|
|
## π Links |
|
|
|
|
|
- [πΎ Github Repo](https://github.com/Entroplay/Visual-Intelligence) |
|
|
- [π€ HF Dataset](https://huggingface.co/datasets/Entroplay/Visual-Intelligence) |
|
|
- [π Blog](https://entroplay.ai/research/video-intelligence) |
|
|
|
|
|
## π Dataset Introduction |
|
|
|
|
|
### Dataset Schema |
|
|
|
|
|
- **id**: Unique sample identifier. |
|
|
- **input**: Ordered list describing the input context. |
|
|
- **type**: Either "image" or "text". |
|
|
- **content**: For "image", a relative path to the first-frame image. For "text", the prompt text. |
|
|
- **output**: Generated candidates and final selections by model. |
|
|
- **veo3**: Relative paths to videos generated by the VEO3 pipeline. |
|
|
- **framepack**: Relative paths to videos generated by FramePack across multiple runs. |
|
|
- **hunyuan**: Relative paths to videos generated by Hunyuan across multiple runs. |
|
|
- **wan2.2-5b**: Relative paths to videos generated by Wan-2.2-5B across multiple runs. |
|
|
- **wan2.2-14b**: Relative paths to videos generated by Wan-2.2-14B across multiple runs. |
|
|
- **framepack_seleted_video**: Selected best video among FramePack candidates. |
|
|
- **hunyuan_seleted_video**: Selected best video among Hunyuan candidates. |
|
|
- **wan2.2-5b_seleted_video**: Selected best video among Wan 2.2 5B candidates. |
|
|
- **wan2.2-14b_seleted_video**: Selected best video among Wan 2.2 14B candidates. |
|
|
|
|
|
### Data Format: |
|
|
|
|
|
```json |
|
|
{ |
|
|
"id": 1, |
|
|
"input": [ |
|
|
{ "type": "image", "content": "thumbnails/mp4/keypoint_localization.jpg" }, |
|
|
{ "type": "text", "content": "Add a bright blue dot at the tip of the branch on which the macaw is sitting. ..." } |
|
|
], |
|
|
"output": { |
|
|
"veo3": ["videos/mp4/keypoint_localization.mp4"], |
|
|
"framepack": [ |
|
|
"videos/1_framepack_1.mp4", |
|
|
"videos/1_framepack_2.mp4" |
|
|
], |
|
|
"hunyuan": [ |
|
|
"videos/1_hunyuan_1.mp4", |
|
|
"videos/1_hunyuan_2.mp4" |
|
|
], |
|
|
"wan2.2-5b": [ |
|
|
"videos/1_wan2.2-5b_1.mp4", |
|
|
"videos/1_wan2.2-5b_2.mp4" |
|
|
], |
|
|
"wan2.2-14b": [ |
|
|
"videos/1_wan2.2-14b_1.mp4", |
|
|
"videos/1_wan2.2-14b_2.mp4" |
|
|
], |
|
|
"framepack_seleted_video": "videos/1_framepack_1.mp4", |
|
|
"hunyuan_seleted_video": "videos/1_hunyuan_1.mp4", |
|
|
"wan2.2-5b_seleted_video": "videos/1_wan2.2-5b_1.mp4", |
|
|
"wan2.2-14b_seleted_video": "videos/1_wan2.2-14b_1.mp4" |
|
|
} |
|
|
} |
|
|
``` |
|
|
|
|
|
## π About project |
|
|
|
|
|
Google' Veo 3 shows extreme promise in visual intelligence, demonstrating strong visual commonsense and reasoning in visual generation. We aim to construct a fully open-source evaluation suite to measure current progress in video generative intelligence across various dimensions among several state-of-the-art proprietary and open-source models. |