Skip to content

Fails to index huge tars >120 million files  #125

@blacha

Description

@blacha

I have a ~100GB 262Million file SQLite database (MBTiles), that I am attempting to convert into a cotar.

The tar creation process fails after 120M files as the hash map containing the previous hashes overflows (NodeJS limit)

Some options:

  1. Split the output tar into multiple smaller chunks (20-30GB ea) so that they contain roughly 100M files each
  2. Switch the hash table to be a tree to support much larger table sizes (Would have creation performance imapacts)
  3. Rewrite the creation into a language that supports BigMaps by default.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions