aboutsummaryrefslogtreecommitdiff
path: root/doc/plugins/aggregate.mdwn
blob: b1db828d107643f72c6dc2d736110f5782ed0bb1 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
[[!template id=plugin name=aggregate author="[[Joey]]"]]
[[!tag type/special-purpose]]

This plugin allows content from other feeds to be aggregated into the
wiki. To specify feeds to aggregate, use the
[[ikiwiki/directive/aggregate]] [[ikiwiki/directive]].

## requirements

The [[meta]] and [[tag]] plugins are also recommended to be used with this
one. Either the [[htmltidy]] or [[htmlbalance]] plugin is suggested, since
feeds can easily contain html problems, some of which these plugins can fix.

Installing the [[!cpan LWPx::ParanoidAgent]] Perl module is strongly
recommended. The [[!cpan LWP]] module can also be used, but is susceptible
to server-side request forgery.

## triggering aggregation

You will need to run ikiwiki periodically from a cron job, passing it the
--aggregate parameter, to make it check for new posts. Here's an example
crontab entry:

	*/15 * * * * ikiwiki --setup my.wiki --aggregate --refresh

The plugin updates a file `.ikiwiki/aggregatetime` with the unix time stamp
when the next aggregation run could occur. (The file may be empty, if no
aggregation is required.) This can be integrated into more complex cron
jobs or systems to trigger aggregation only when needed.

Alternatively, you can allow `ikiwiki.cgi` to trigger the aggregation. You
should only need this if for some reason you cannot use cron, and instead
want to use a service such as [WebCron](http://webcron.org). To enable
this, turn on `aggregate_webtrigger` in your setup file. The url to
visit is `http://whatever/ikiwiki.cgi?do=aggregate_webtrigger`. Anyone
can visit the url to trigger an aggregation run, but it will only check
each feed if its `updateinterval` has passed.

## aggregated pages

This plugin creates a page for each aggregated item. 

If the `aggregateinternal` option is enabled in the setup file (which is
the default), aggregated pages are stored in the source directory with a
"._aggregated" extension. These pages cannot be edited by web users, and
do not generate first-class wiki pages. They can still be inlined into a
blog, but you have to use `internal` in [[PageSpecs|IkiWiki/PageSpec]],
like `internal(blog/*)`.

If `aggregateinternal` is disabled, you will need to enable the [[html]]
plugin as well as aggregate itself, since feed entries will be stored as
HTML, and as first-class wiki pages -- each one generates
a separate HTML page in the output, and they can even be edited. This
option is provided only for backwards compatability.

## cookies

The `cookiejar` option can be used to configure how [[!cpan LWP::UserAgent]]
handles cookies. The default is to read them from a file
`~/.ikiwiki/cookies`, which can be populated using standard perl cookie
tools like [[!cpan HTTP::Cookies]].