Sign Up

Sign Up to our social questions and Answers Engine to ask questions, answer people’s questions, and connect with other people.

Have an account? Sign In

Have an account? Sign In Now

Sign In

Login to our social questions & Answers Engine to ask questions answer people’s questions & connect with other people.

Sign Up Here

Forgot Password?

Don't have account, Sign Up Here

Forgot Password

Lost your password? Please enter your email address. You will receive a link and will create a new password via email.

Have an account? Sign In Now

You must login to ask a question.

Forgot Password?

Need An Account, Sign Up Here

Please briefly explain why you feel this question should be reported.

Please briefly explain why you feel this answer should be reported.

Please briefly explain why you feel this user should be reported.

Sign InSign Up

The Archive Base

The Archive Base Logo The Archive Base Logo

The Archive Base Navigation

  • Home
  • SEARCH
  • About Us
  • Blog
  • Contact Us
Search
Ask A Question

Mobile menu

Close
Ask a Question
  • Home
  • Add group
  • Groups page
  • Feed
  • User Profile
  • Communities
  • Questions
    • New Questions
    • Trending Questions
    • Must read Questions
    • Hot Questions
  • Polls
  • Tags
  • Badges
  • Buy Points
  • Users
  • Help
  • Buy Theme
  • SEARCH
Home/ Questions/Q 94601
In Process

The Archive Base Latest Questions

Editorial Team
  • 0
Editorial Team
Asked: May 10, 20262026-05-10T23:31:38+00:00 2026-05-10T23:31:38+00:00

I’ve got a C++ program that’s likely to generate a HUGE amount of data

  • 0

I’ve got a C++ program that’s likely to generate a HUGE amount of data — billions of binary records of varying sizes, most probably less than 256 bytes but a few stretching to several K. Most of the records will seldom be looked at by the program after they’re created, but some will be accessed and modified regularly. There’s no way to tell which are which when they’re created.

Considering the volume of data, there’s no way I can store it all in memory. But as the data only needs to be indexed and accessed by its number (a 64-bit integer), I don’t want the overhead of a full-fledged database program. Ideally I’d like to treat it as an std::map with its data stored on disk until requested.

Is there an already-written library that will do what I’m looking for, or do I need to write it myself?

EDIT: After some thought, I realized that Rob Walker’s answer had a valid point: I’d be hard-pressed to get anywhere near the same kind of data integrity out of a home-brew class that I’d get from a real database.

Although BerkeleyDB (as suggested by RHM) looks like it would do exactly what we’re looking for, the dual-licensing is a headache that we don’t want to deal with. When we’re done with the code and can prove that it would benefit noticeably from BerkeleyDB (which it probably would), we’ll reexamine the issue.

I did look at Ferruccio’s suggestion of stxxl, but I wasn’t able to tell how it would handle the program being interrupted and restarted (maybe with changes). With that much data, I’d hate to just scrap what it had already completed and start over every time, if some of the data could be saved.

So we’ve decided to use an SQLite database, at least for the initial development. Thanks to everyone who answered or voted.

  • 1 1 Answer
  • 0 Views
  • 0 Followers
  • 0
Share
  • Facebook
  • Report

Leave an answer
Cancel reply

You must login to add an answer.

Forgot Password?

Need An Account, Sign Up Here

1 Answer

  • Voted
  • Oldest
  • Recent
  • Random
  1. 2026-05-10T23:31:38+00:00Added an answer on May 10, 2026 at 11:31 pm

    I doubt you will find a library that meets your requirements exactly, so you’ll have to decide on what ‘features’ are really important to you and then decide if an existing DB solution comes close enough.

    Billions of records is a large dataset by any stretch. What rate are records generated at? How long do they persist? Does the access pattern change over time?

    Are updates always with the same amount of data as the original?

    I would suggest proving definitively that a DB solution isn’t going to work before starting to roll your own, particularly if integrity of the data is paramount (and it usually is…) Maintaining that volume of data on disk reliably can definitely be a challenge. Do you need any kind of transaction semantics when changing the data? Is the client multithreaded?

    • 0
    • Reply
    • Share
      Share
      • Share on Facebook
      • Share on Twitter
      • Share on LinkedIn
      • Share on WhatsApp
      • Report

Sidebar

Ask A Question

Stats

  • Questions 58k
  • Answers 58k
  • Best Answers 0
  • User 1
  • Popular
  • Answers
  • Editorial Team

    How to approach applying for a job at a company ...

    • 7 Answers
  • Editorial Team

    How to handle personal stress caused by utterly incompetent and ...

    • 5 Answers
  • Editorial Team

    What is a programmer’s life like?

    • 5 Answers
  • added an answer That code is for Macintosh Common Lisp (MCL). It will… May 11, 2026 at 8:40 am
  • added an answer Your two partial classes are defined in different namespaces, so… May 11, 2026 at 8:40 am
  • added an answer You can also split a string by a substring, using… May 11, 2026 at 8:40 am

Top Members

Trending Tags

analytics british company computer developers django employee employer english facebook french google interview javascript language life php programmer programs salary

Explore

  • Home
  • Add group
  • Groups page
  • Communities
  • Questions
    • New Questions
    • Trending Questions
    • Must read Questions
    • Hot Questions
  • Polls
  • Tags
  • Badges
  • Users
  • Help
  • SEARCH

Footer

© 2021 The Archive Base. All Rights Reserved
With Love by The Archive Base

Insert/edit link

Enter the destination URL

Or link to existing content

    No search term specified. Showing recent items. Search or use up and down arrow keys to select an item.