Skip to content

Handling large buckets using less disk space for the database #4

@chrowe

Description

@chrowe

I am working with a bucket that has ~60 million objects.
I am able to get this to import into the inventory table in the db but when it tries to create the inventory_flatten materialized view, I run out of disk space. My rdb container has ~300GB of disk space, but this doesn't seem to be enough because I keep getting ERROR - Failed to ingest inventory files: could not write to file "base/pgsql_tmp/pgsql_tmp1074.81": No space left on device. After my last try I have ~360 millions rows and it is using 194GB of disk space.

I also tried just generating the PREFIXES_MATERIALIZED_VIEW materialized view for the partially completed inventory_flatten table but that also ran out of disk space.

I am able to create the as just basic views, but for this obviously will not perform well on such a large inventory table.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions