| | # Working with the Metadata |
| |
|
| | ## Downloading all the metadata files at once |
| |
|
| | Install the huggingface-cli utility (via pip). You may then use the following command: |
| |
|
| | huggingface-cli download Spawning/PD12M --repo-type dataset --local-dir metadata --include "metadata/*" |
| | |
| | ## metadata format |
| |
|
| | The metadata files are in parquet format, and contain the following attributes: |
| | - `id`: A unique identifier for the image. |
| | - `url`: The URL of the image. |
| | - `s3_key`: The S3 file key of the image. |
| | - `caption`: A caption for the image. |
| | - `hash`: The MD5 hash of the image file. |
| | - `width`: The width of the image in pixels. |
| | - `height`: The height of the image in pixels. |
| | - `mime_type`: The MIME type of the image file. |
| | - `license`: The URL of the license. |
| | - `source`: The source organization of the image. |
| |
|
| | #### Open a metadata file |
| | The files are in parquet format, and can be opened with a tool like `pandas` in Python. |
| | ```python |
| | import pandas as pd |
| | df = pd.read_parquet('pd12m.000.parquet') |
| | ``` |
| |
|
| | #### Get URLs from metadata |
| | Once you have opened a maetadata file with pandas, you can get the URLs of the images with the following command: |
| | ```python |
| | urls = df['url'] |
| | ``` |
| |
|
| | ### Download all files mentioned in metadata |
| |
|
| | If you want to just grab all files referenced by a metadata collection, you may try this (adjust to taste): |
| |
|
| | img2dataset --url_list $file --input_format "parquet" \ |
| | --url_col "url" --caption_col "caption" --output_format files \ |
| | --output_folder $dir --processes_count 16 --thread_count 64 \ |
| | --skip_reencode true --min_image_sizel 654 --max_aspect_ratio=1.77 |
| | |
| |
|
| |
|
| |
|