How to efficiently store JSON file objects into MongoDB using NodeJs stream?

View other answers to this thread
Start a personal dev blog on your domain for free and grow your readership.

3.4K+ developers have started their personal blogs on Hashnode in the last one month.

Write in Markdown · Publish articles on custom domain · Gain readership on day zero · Automatic GitHub backup and more

Ashok Dey's photo

So I came up with a very simpe code snippet which is as follow:

const fs = require('fs');
const mongoose = require('mongoose');
const JSONStream = require('JSONStream');
const User = require('./src/models/User');
const config = require('./config.json');

mongoose.connect(config.MONGODB_URI, { poolSize: config.DB_POOL_SIZE });
mongoose.Promise = global.Promise;

const db = mongoose.connection;
let arrayOfUsers = [];

db.on('open', () => {
  console.log('Connected to mongo server.\n');
  const dataStreamFromFile = fs.createReadStream(`${__dirname}/users_large.json`);
  dataStreamFromFile.pipe(JSONStream.parse('*')).on('data', async (userData) => {
    if (arrayOfUsers.length === config.BATCH_INSERT_VALUE) {
      await User.insertMany(arrayOfUsers);
      arrayOfUsers = [];

  dataStreamFromFile.on('end', async () => {
    await User.insertMany(arrayOfUsers); // left over data
    console.log('\nImport complete, closing connection...');

db.on('error', (err) => {
  console.error('MongoDB connection error: ', err);

Note: Inserting 50k records is also a bit slow. Will keep on trying !

Show +1 replies
Ashok Dey's photo

Thanks! I changed the value of BATCH_INSERT_VALUE from 50000 to 1000 and it was faster. Just read the limit of insertMany() is 100

raj rawat's photo

Hi could you please share git repo for the running code Ashok Dey