linux-c-programming.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* Handling large files
@ 2005-04-22 17:03 Anindya Mozumdar
  2005-04-22 19:20 ` Glynn Clements
  0 siblings, 1 reply; 2+ messages in thread
From: Anindya Mozumdar @ 2005-04-22 17:03 UTC (permalink / raw)
  To: linux-c-programming

Hi,

   Recently I was dealing with large csv ( comma separated value )
   files, of size around 500M.

   I was using perl to parse such files, and it took around 40 minutes
   for perl to read the file, and duplicate it using the csv module.
   Python's module took 1 hr. I am sure even if I had written c code,
   opened the file and parsed it, it would have taken a lot of time.

   However, I used MySQL to create a database from the file, and the
   entire creation took around 2 minutes. I would like to know how is
   this possible - is it a case of threading, memory mapping or some
   good algorithm ?

   I would be thankful to anyone who can give me a good answer to the
   question, as I cant think of a way myself to solve the problem.

Anindya.

^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2005-04-22 19:20 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2005-04-22 17:03 Handling large files Anindya Mozumdar
2005-04-22 19:20 ` Glynn Clements

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).