Skip to content

Niji v6 520k

Details

Required storage

This dataset contains all image data, and as such, it will be difficult to extract without adequate disk space. Ensure you have at least 1.5TB of disk space available to extract it.

T5-XXL text embeds for this model will consume ~520GB even with --compress_disk_cache enabled. The VAE embeds will consume just under 80 to 100GB of space, depending on the model being trained and the resolution of the embeds.

Download

huggingface-cli download --repo-type=dataset terminusresearch/nijijourney-v6-520k-raw --local-dir=nijijourney-v6-520k-raw

This will simultaneously download the chunked tar segments from Hugging Face Hub.

Extract

cd nijijourney-v6-520k-raw
cat *.tar | tar x

This will create a folder containing all of the samples inside the current directory.

Dataloader configuration example

{
    "id": "nijijourney-v6-520k-raw",
    "type": "local",
    "cache_dir_vae": "cache/vae-nj-520k/",
    "crop": true,
    "crop_aspect": "square",
    "resolution": 1.0,
    "maximum_image_size": 1.0,
    "minimum_image_size": 0.75,
    "target_downsample_size": 1.00,
    "resolution_type": "area",
    "caption_strategy": "parquet",
    "metadata_backend": "parquet",
    "parquet": {
        "path": "/path/to/nijijourney-v6-520k-raw/train.parquet",
        "caption_column": "gpt_caption",
        "filename_column": "id",
        "width_column": "width",
        "height_column": "height",
        "identifier_includes_extension": false
    }
}