[XSL-LIST Mailing List Archive Home] [By Thread] [By Date]

Re: [xsl] FOP : consumption memory

Subject: Re: [xsl] FOP : consumption memory
From: "Imsieke, Gerrit, le-tex gerrit.imsieke@xxxxxxxxx" <xsl-list-service@xxxxxxxxxxxxxxxxxxxxxx>
Date: Sat, 16 Aug 2014 11:38:20 -0000

Ibm not a FOP expert, but canbt you separate the pass(es) that generate the index, resolve cross-refs, etc., from the pass that actually layouts the FO representation? By separate I mean make it separate Java invocations, storing intermediate results as files.

On August 16, 2014 1:04:46 PM CEST, "Eliot Kimber ekimber@xxxxxxxxxxxx" <xsl-list-service@xxxxxxxxxxxxxxxxxxxxxx> wrote:
>I see: the two-column layout means there are no natural breakpoints in
>content before the index. The index has break points but by then it
>be too late. The back-of-the-book index could also be
>quite long and FOP may need to keep the entire area tree in memory in
>order to then resolve the index references.
>As Peter says, I would suspect a naive implementation on FOP's part (I
>haven't looked at the code). Would be useful to try both RenderX XEP
>Antenna House XSL Formatter--I'm sure they would both do better. If
>project can bear the cost, either product would be a good investment.
>Eliot Kimber, Owner
>Contrext, LLC
>On 8/16/14, 2:26 AM, "Jean-Pierre Lamon jpl@xxxxxxxxxx"
><xsl-list-service@xxxxxxxxxxxxxxxxxxxxxx> wrote:
>>Thank you for your response Eliot.
>>I don't know if I can share these files, I must ask the principal
>>it's a mandate.
>>The result is under :
>>I'll let you know
>>-----Message d'origine-----
>>De : Eliot Kimber ekimber@xxxxxxxxxxxx
>>EnvoyC) : vendredi 15 aoC;t 2014 23:44
>>C : xsl-list
>>Objet : Re: [xsl] FOP : consumption memory
>>If your content has natural page breaks (meaning elements that always
>>start a new page) you can always start a new page sequence at that
>>If your content does not have such nature page breaks then of course
>>can't. In that case, one solution would be to generate the
>>area tree (a feature of FOP and all the other FO engines) and then use
>>to find elements that happen to start on new pages and regenerate the
>>with page sequences started at those points. But that seems like
>rather a
>>lot of effort.
>>It might be easier to just give the Java VM running FOP more memory.
>>If this is XML that can be shared publicly I'd be interested in
>>diagnose this issue in exchange for the ability to use the XML for
>>Eliot Kimber, Owner
>>Contrext, LLC
>>On 8/15/14, 1:32 PM, "Jean-Pierre Lamon jpl@xxxxxxxxxx"
>><xsl-list-service@xxxxxxxxxxxxxxxxxxxxxx> wrote:
>>>Thx Geert but I can't spread and break pages. It's a bibliography
>>>national library bibliography).
>>>If someone wants the XML and XSL to test, no problem :-) I'm not very
>>>professional with XSL, I maybe have done some horrors in my
>>>my question is only : why FOP hangs and the little tool works
>>>With absolute respect for people working for free tools like FOP.
>>>-----Message d'origine-----
>>>De : Geert Bormans geert@xxxxxxxxxxxxxxxxxxx
>>>EnvoyC) : vendredi 15 aoC;t 2014 17:12
>>>C : xsl-list@xxxxxxxxxxxxxxxxxxxxxx
>>>Objet : Re: [xsl] FOP : consumption memory
>>>In my experience FOP does a poor thing with long page sequences.
>>>It seems to keep them in memory (for repagination maybe?) completely
>>>Memory footprint for FOP goes down dramatically
>>>if you have a logic that cuts the pages
>>>Rather than using mechanisms such as break before
>>>or similar, create new page sequences when you can
>>>(eg. per chapter, ...)
>>>That has helped me in the past
>>>At 16:33 15/08/2014, you wrote:
>>>>Hi All,
>>>>I know, difficult to say without having the
>>>>source, but could someone explain me why FOP
>>>>crashes, hangs (memory ?) for relative big
>>>>documents and a free small tool like XML2PDF
>>>>render the PDF perfectly and this, dramatically quicker compare to
>>>>IBve tried to play with JAVA memory etcB> no way.
>>>>Thanks and regards
>>>><http://www.mulberrytech.com/xsl/xsl-list>XSL-List info and archive
>>>>(<>by email)

Sent from my Android device with K-9 Mail. Please excuse my brevity.

Current Thread