[Moin-user] wget a wiki page and it's subpage
Ralf-Lists at ralfgross.de
Tue Jan 31 05:44:01 EST 2006
Thomas Waldmann schrieb:
> >I'm trying to get a wiki page and its subpage with many images through
> Some user agents (including wget) receive special treatment as people
> use them often to DOS wikis. So make sure you change the user agent it
> uses (and if it is not your own wiki: USE CAREFULLY).
Changing the agent string didn't help.
> >I need to do this, because we want to give the pages as documentation
> >to a customer. After retrieving the pages in html I'm going to try to
> >convert them to pdf.
> Maybe look at moin-dump, too.
moin-dump doesn't care about the attachments. But I found
http://moinmoin.wikiwikiweb.de/MoinDump, which dumps the whole wiki
including attachments and fixing of the attachment paths. It'd be
nice if I could limit the output to the pages I really need, but it's
ok for now.
More information about the Moin-user