Parses URLs from html page. Useful on sites which have direct download links of any type (mp3, jpg, torrent, ...).
Many anime-fansubbers do not provide an RSS-feed, this usually works for those cases.
Note: This returns ALL links from the page so you need to configure some filters to match only to desired content.
Example with regexp
html: <url> regexp: accept: - podcast.*\.mp3 rest: filter
This would download all links that have mp3 link with word podcast in them.
By default html plugin will try to guess where link titles should be captured from but in some cases this goes wrong and you end up with useless titles like 'DL', '1', '2'. In case automatic selection does not work you will need to specify where titles should be captured from. Note that it may not always be possible to get good titles if the the title is nowhere near the link.
html: url: <url> title_from: url
Other possible values are: auto (default), url, title and link.
If captured URL is "http://some.domain.com/download/Podcast.2010-01-02.mp3" the extracted title would be "Podcast.2010-01-02.mp3"
If html contains links in form of <a href="http://some.domain.com/download?id=1245932" title="Podcast.2010-01-02.mp3">Download</a> the extracted title would be "Podcast.2010-01-02.mp3"
If html contains links in form of <a href="http://some.domain.com/download?id=1245932">Podcast.2010-01-02.mp3</a> the extracted title would be "Podcast.2010-01-02.mp3"
Titles are gibberish
In some cases the titles are completely useless, however you may still get good filenames when entry is downloaded. This is because servers often send the correct filename in http headers. Of course when titles are useless filtering becomes much more troublesome.
Get only certain links
In case page contains too many links, you can use regexps to select only certain links.
html: url: <url> links_re: - domain\.com
This will create only entries from links which match any of given regexps. Do NOT use this for filtering content, just limit selecting to correct type of links and use real filtering plugins like regexp to do actual filtering.
This plugin supports baisic authentication, it can be specified in the url:
Or as parameters:
html: url: <url> username: <username> password: <password>
You can dump the received HTML into a file by using parameter dump. Useful for debugging.
html: url: <url> dump: file.html