Hi experts, Have you ever experienced Out of memory problem while using HTML::TableExtract. I'm having little large html files, still i didn't expect this to happen Would you be able to suggest some workarounds for this. I'm using this subroutine in another for loop.
sub zParseHTMLFiles ($$) { my ( $lrefFileList, $lrefColNames ) = @_; my @ldata; foreach my $lFile (@$lrefFileList) { my $lTableExtract = HTML::TableExtract->new( headers => [@$lre +fColNames] ); chomp($lFile); $lTableExtract->parse_file($lFile); foreach my $ls ( $lTableExtract->tables ) { foreach my $lrow ( $lTableExtract->rows ) { chomp( @$lrow[$#$lrow] ); push( @ldata, $lrow ); } } } return \@ldata; }
In reply to Out of memory while using HTML::TableExtract by jthomas
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |