FW: [jdom-interest] parse large xml file using jdom

New, Cecil (GEAE) cecil.new at ae.ge.com
Tue Jun 10 04:44:35 PDT 2003

If your process can tolerate detaching, perhaps you should consider using SAX.
Takes more work, but you'll have the fastest and least memory approach possible.

-----Original Message-----
From: Eric Chi-hsuan Lai [mailto:lai at physics.utexas.edu]
Sent: Monday, June 09, 2003 5:43 PM
To: jdom-interest at jdom.org
Subject: [jdom-interest] parse large xml file using jdom

	I am in a project that has to process large size xml files, i.e. 
50M-100M in size. Last time I tried to use jdom to parse it ends up with 
out of memory. I am wondering if the new beta 9 has new facility/features 
that can handle such big xml file. I saw on dom4j FAQ has a feature that 
can "detach" a node/element after you are done with it. I hope jdom has 
something similar features.


To control your jdom-interest membership:

More information about the jdom-interest mailing list