Hi,
you'd need as well to convert these document to pure text in order to
process them; you can try to spawn OOo for conversion in a batch mode
but the easier option is to use unzip in a script, and take content.xml
only from the files. Then process the files using awk (define the field
separator just like you would define the word boundary) and filter out
all words that match <[a-z]+>. This should kill all xml from the files.
Regards
Marcin
ge pisze:
> Hello,
>
> I did word collection several times using different sources( web sources)
>
> I use linux, but these tools are also available
> for windows as gnu tools.
>
> I used awk, like:
> for (i = 1; i <= $NF; i++)
> print $i;
>
> This prints each word in a single line.
>
> Then I sorted the file using sort <infile > outfile
> and then used further awk scripts to get rid of word endings,
> this is probably much easier for Danish, than for Hungarian.
>
> Good luck! Eleonora
>
>
> [lingu-dev] Help needed - bulk extraction of words
>
> Hi all,
> The Danish project has been so fortunate to receive a bunch of articles
> from a news magazine. These are odt files and we would like to extract
> the words from these documents. We have programs for this purpose, but
> we usually get donations one document at the time. This time we have
> several thousand documents and I believe it would take about a year to
> load these documents one by one.
>
> Do any of you have a program that can extract words from several documents ?
>
> The words will be loaded into our workflow for linguistic processing and
> at the end be a part of the Danish spelling directory.
>
> Thanks in advance.
>
---------------------------------------------------------------------
To unsubscribe, e-mail:
[hidden email]
For additional commands, e-mail:
[hidden email]