Importing Large Data Set from Mongo DB

bdx
bdx ⚪️

Hi, I am having a lot of errors importing data from my mongoDB (using the MongoDB V2 connector). The data contains around 10 million Json documents. The connector is working as I have successfully imported previously and I am able to run the importer when the row count is limited to 100,000.


What is the best path to take to get the whole data set imported?


Thanks

Answers

  • GrantSmith
    GrantSmith Indiana 🔴

    Hi @bdx

    You'll want to try and import your data in batches if possible.

  • bdx
    bdx ⚪️

    Thanks Grant.

    What would you say is the best method?

    I already set the batch size to 5000 with the limit row to 9000000.

    Is it a good option to apply a JSON query filter and run multiple data imports or is there a function in the importer that will check for documents not yet imported?

  • jaeW_at_Onyx
    jaeW_at_Onyx Budapest / Portland, OR 🟤

    @bdx it's interesting thaat Domo is choking on 100k rows. I wonder if that's a MongoDB limitation. Domo can ingest multi M row datasets at a time so it's not likely to be a domo infrastructure problem.

    that said, yes, use filters to try to limit just the new documents you're bringing in. You may need to implement a recursive dataflow in Domo to handle the rest of the data.

  • bdx
    bdx ⚪️

    Thanks JaeW