},{
"blocker": "union.place",
"csv_url": "mastodon/union.place.csv",
+ },{
+ "blocker": "oliphant.social",
+ "csv_url": "mastodon/birdsite.csv",
}
)
elif args.domain in domains:
logger.debug("args.domain='%s' already handled - SKIPPED!", args.domain)
continue
- elif instances.is_recent(block["blocker"]):
- logger.debug("block[blocker]='%s' has been recently crawled - SKIPPED!", block["blocker"])
- continue
# Fetch this URL
logger.info("Fetching csv_url='%s' for blocker='%s' ...", block["csv_url"], block["blocker"])
processed = utils.process_domain(domain, block["blocker"], inspect.currentframe().f_code.co_name)
logger.debug("processed='%s'", processed)
- if utils.process_block(block["blocker"], domain, None, "reject") and config.get("bot_enabled"):
+ if utils.process_block(block["blocker"], domain, None, severity) and config.get("bot_enabled"):
logger.debug("Appending blocked='%s',reason='%s' for blocker='%s' ...", domain, block["block_level"], block["blocker"])
blockdict.append({
"blocked": domain,