Plugin: aggregate
Author: Joey
Included in ikiwiki: yes
Enabled by default: no
Included in goodstuff: no
Currently enabled: no

This plugin allows content from other feeds to be aggregated into the wiki. To specify feeds to aggregate, use the aggregate directive.

requirements

The meta and tag plugins are also recommended to be used with this one. Either the htmltidy or htmlbalance plugin is suggested, since feeds can easily contain html problems, some of which these plugins can fix.

Installing the LWPx::ParanoidAgent Perl module is strongly recommended. The LWP module can also be used, but is susceptible to server-side request forgery.

triggering aggregation

You will need to run ikiwiki periodically from a cron job, passing it the --aggregate parameter, to make it check for new posts. Here's an example crontab entry:

*/15 * * * * ikiwiki --setup my.wiki --aggregate --refresh

The plugin updates a file .ikiwiki/aggregatetime with the unix time stamp when the next aggregation run could occur. (The file may be empty, if no aggregation is required.) This can be integrated into more complex cron jobs or systems to trigger aggregation only when needed.

Alternatively, you can allow ikiwiki.cgi to trigger the aggregation. You should only need this if for some reason you cannot use cron, and instead want to use a service such as WebCron. To enable this, turn on aggregate_webtrigger in your setup file. The url to visit is http://whatever/ikiwiki.cgi?do=aggregate_webtrigger. Anyone can visit the url to trigger an aggregation run, but it will only check each feed if its updateinterval has passed.

aggregated pages

This plugin creates a page for each aggregated item.

If the aggregateinternal option is enabled in the setup file (which is the default), aggregated pages are stored in the source directory with a "._aggregated" extension. These pages cannot be edited by web users, and do not generate first-class wiki pages. They can still be inlined into a blog, but you have to use internal in PageSpecs, like internal(blog/*).

If aggregateinternal is disabled, you will need to enable the html plugin as well as aggregate itself, since feed entries will be stored as HTML, and as first-class wiki pages -- each one generates a separate HTML page in the output, and they can even be edited. This option is provided only for backwards compatability.

cookies

The cookiejar option can be used to configure how LWP::UserAgent handles cookies. The default is to read them from a file ~/.ikiwiki/cookies, which can be populated using standard perl cookie tools like HTTP::Cookies.