// CrawlStarterFromSraper.java
// ---------------------------
// Copyright 2016 by luccioman; https://github.com/luccioman
//
// This is a part of YaCy, a peer-to-peer based web search engine
//
// LICENSE
//
// This program is free software; you can redistribute it and/or modify
// it under the terms of the GNU General Public License as published by
// the Free Software Foundation; either version 2 of the License, or
// (at your option) any later version.
//
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU General Public License for more details.
//
// You should have received a copy of the GNU General Public License
// along with this program; if not, write to the Free Software
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
package net.yacy.crawler;
import java.net.MalformedURLException;
import java.util.ArrayList;
import java.util.List;
import java.util.Properties;
import net.yacy.cora.document.id.AnchorURL;
import net.yacy.cora.util.ConcurrentLog;
import net.yacy.crawler.data.CrawlProfile;
import net.yacy.document.parser.html.ContentScraperListener;
/**
* Enqueue an entry to the crawlStacker each time an anchor is discovered by the ContentScraper
* @author luccioman
*
*/
public class CrawlStarterFromSraper implements ContentScraperListener {
private final static ConcurrentLog log = new ConcurrentLog(CrawlStarterFromSraper.class.getSimpleName());
/** CrawlStacker instance : will receive anchor links used as crawl starting points */
private CrawlStacker crawlStacker;
/** Hash of the peer initiating the crawl */
private final byte[] initiatorHash;
/** Active crawl profile */
private CrawlProfile profile;
/** Specify whether old indexed entries should be replaced */
private final boolean replace;
/**
* Constructor
* @param crawlStacker CrawlStacker instance : will receive anchor links used as crawl starting points
* @param initiatorHash Hash of the peer initiating the crawl (must not be null)
* @param profile active crawl profile (must not be null)
* @param replace Specify whether old indexed entries should be replaced
* @throws IllegalArgumentException when a required parameter is null
*/
public CrawlStarterFromSraper(final CrawlStacker crawlStacker, final byte[] initiatorHash,
final CrawlProfile profile,
final boolean replace) {
if(crawlStacker == null) {
throw new IllegalArgumentException("crawlStacker parameter must not be null");
}
this.crawlStacker = crawlStacker;
if(initiatorHash == null) {
throw new IllegalArgumentException("initiatorHash parameter must not be null");
}
this.initiatorHash = initiatorHash;
this.replace = replace;
if(profile == null) {
throw new IllegalArgumentException("profile parameter must not be null");
}
this.profile = profile;
}
@Override
public void scrapeTag0(String tagname, Properties tagopts) {
// Nothing to do on this event
}
@Override
public void scrapeTag1(String tagname, Properties tagopts, char[] text) {
// Nothing to do on this event
}
@Override
public void anchorAdded(String anchorURL) {
List<AnchorURL> urls = new ArrayList<>();
try {
urls.add(new AnchorURL(anchorURL));
this.crawlStacker.enqueueEntries(this.initiatorHash, this.profile.handle(), urls, this.replace, this.profile.timezoneOffset());
} catch (MalformedURLException e) {
log.warn("Malformed URL : " + anchorURL);
}
}
}