Fastest way to read a set of files (kernel v5.14) #1160
Unanswered
MartyMcFlyInTheSky
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Our company uses a custom database that we maintain that stores up to 4 terabytes of data in memory. When the application is deployed, the data needs to be dumped on disk and reloaded and as you can imagine, loading 4TB from disk will take some time. My goal is now to make this as fast as possible.
The data is already categorised internally, so my thought was dumping every categories worth of data into seperate files, that could then be loaded (even seperately if wished).
Looking into this documentation already gave me some ideas how to go about it:
I need further clarification on how this is supposed to be meant?
I guess I just don't have a good enough grasp on what this array actually is and how it is looked at by the kernel conceptually. It would be helpful if someone could elaborate on this a bit.
For testing purposes some of our devs will run the application on their devices, some of which are MAC's. I suppose they wouldn't be able to compile the binary since the system calls are not supported, is that true? If so, my colleague made a reference implementation that uses the POSIX read and write commands, how could #ifdef conditionally compile for the availability of these features and fall back to the standard interface if they're not available?
Our HPC cluster currently runs with kernel version 5.14. Which of the above features can I use? Are there restrictions that apply or extensions that are not documented in the above script?
General feedback on my approach would be greatly appreciated! It's really an exciting new interface that I'd like to see in production in our HPC cluster :)
Beta Was this translation helpful? Give feedback.
All reactions