forked from fictive-kin/openrecipes
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Added spiders for LoveAndOliveOil.com - issue fictive-kin#100
- Loading branch information
1 parent
dd71587
commit f4b028d
Showing
2 changed files
with
89 additions
and
0 deletions.
There are no files selected for viewing
24 changes: 24 additions & 0 deletions
24
scrapy_proj/openrecipes/spiders/loveandoliveoil_feedspider.py
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,24 @@ | ||
from scrapy.spider import BaseSpider | ||
from scrapy.http import Request | ||
from scrapy.selector import XmlXPathSelector | ||
from openrecipes.spiders.loveandoliveoil_spider import Loveandoliveoil_Mixin | ||
|
||
|
||
class LoveandoliveoilfeedSpider(BaseSpider, Loveandoliveoil_Mixin): | ||
|
||
name = "loveandoliveoil.feed" | ||
|
||
allowed_domains = [ | ||
"loveandoliveoil.com", | ||
"feeds.feedburner.com" | ||
] | ||
start_urls = [ | ||
"http://feeds.feedburner.com/loveandoliveoil", | ||
] | ||
|
||
def parse(self, response): | ||
|
||
xxs = XmlXPathSelector(response) | ||
links = xxs.select("//link/text()").extract() | ||
|
||
return [Request(x, callback=self.parse_item) for x in links] |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,65 @@ | ||
from scrapy.contrib.spiders import CrawlSpider, Rule | ||
from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor | ||
from scrapy.selector import HtmlXPathSelector | ||
from openrecipes.items import RecipeItem, RecipeItemLoader | ||
|
||
|
||
class Loveandoliveoil_Mixin(object): | ||
source = 'loveandoliveoil_spider' | ||
|
||
def parse_item(self, response): | ||
|
||
hxs = HtmlXPathSelector(response) | ||
|
||
base_path = '//blockquote[@class="recipe hrecipe"]' | ||
|
||
recipes_scopes = hxs.select(base_path) | ||
|
||
name_path = '//*[@class="fn"]/text()' | ||
description_path = '//*[@class="summary"]/p/text()' | ||
image_path = '//img[@class="photo"]/@src' | ||
prepTime_path = '//*[@class="preptime"]/text()' | ||
cookTime_path = '//*[@class="cooktime"]/text()' | ||
recipeYield_path = '//*[@class="yield"]/text()' | ||
ingredients_path = '//*[@class="ingredient"]/p/text()' | ||
|
||
recipes = [] | ||
|
||
for r_scope in recipes_scopes: | ||
il = RecipeItemLoader(item=RecipeItem()) | ||
|
||
il.add_value('source', self.source) | ||
|
||
il.add_value('name', r_scope.select(name_path).extract()) | ||
il.add_value('image', r_scope.select(image_path).extract()) | ||
il.add_value('url', response.url) | ||
il.add_value('description', | ||
r_scope.select(description_path).extract()) | ||
|
||
il.add_value('prepTime', r_scope.select(prepTime_path).extract()) | ||
il.add_value('cookTime', r_scope.select(cookTime_path).extract()) | ||
il.add_value('recipeYield', | ||
r_scope.select(recipeYield_path).extract()) | ||
|
||
il.add_value('ingredients', | ||
r_scope.select(ingredients_path).extract()) | ||
|
||
recipes.append(il.load_item()) | ||
|
||
return recipes | ||
|
||
|
||
class Loveandoliveoil_crawlSpider(CrawlSpider, Loveandoliveoil_Mixin): | ||
|
||
name = "www.loveandoliveoil.com" | ||
|
||
allowed_domains = ["www.loveandoliveoil.com"] | ||
|
||
start_urls = ["http://www.loveandoliveoil.com/recipe-index", ] | ||
|
||
rules = ( | ||
Rule(SgmlLinkExtractor(allow=('/tag/.+'))), | ||
|
||
Rule(SgmlLinkExtractor(allow=('/\d{4}/\d{2}/.+\.html')), | ||
callback='parse_item'), | ||
) |