openmolecules.org

 
Home » DataWarrior » Functionality » Can DW handle big data? (multithreading, low cost of memory)
Re: Can DW handle big data? [message #688 is a reply to message #682] Thu, 24 October 2019 20:59 Go to previous messageGo to previous message
thomas is currently offline  thomas
Messages: 661
Registered: June 2014
Senior Member
Dear DaRong,

3 million rows is already a lot. I recommend for very large files to use DataWarrior on Linux, because you can easily increase the memory maximum that there is at least no memory problem. DataWarrior uses multithreading for most functions, which benefit from it. However, reading a file is a serial process and cannot easily be parallelized. Possibly I could gain some performance, when distributing the data analysis after file loading on multiple cores. I will put it on the agenda, but not before the next release, which I anticipate before the end of the year.

Thanks and best wishes,

Thomas
 
Read Message
Read Message
Read Message
Previous Topic: Synchronise Colours
Next Topic: Exclude function in the structure filter does not work with 2 exclude groups
Goto Forum:
  


Current Time: Fri May 17 11:01:20 CEST 2024

Total time taken to generate the page: 0.00715 seconds