Skip to content

Commit

Permalink
Upgrade scrapy 2.3.0 (#67)
Browse files Browse the repository at this point in the history
* Remove custom context factory

Update scrapy to latest version

* Fix Changelog after update

* Remove nb_hits update in config file

* Remove commented out import

* Remove update_nb_hits related logic since is no longer used

* Fix pylint complaint

* Update scrapy to version 2.3.0

Co-authored-by: renehernandez <rene.hernandez@fullscript.com>
  • Loading branch information
renehernandez and renehernandez authored Sep 21, 2020
1 parent 72033da commit d973df7
Show file tree
Hide file tree
Showing 3 changed files with 27 additions and 9 deletions.
2 changes: 1 addition & 1 deletion Pipfile
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ verify_ssl = true
name = "pypi"

[packages]
Scrapy = "==2.2.1"
Scrapy = "==2.3.0"
selenium = "==3.141.0"
pytest = "==6.0.0"
meilisearch = "==0.12.3"
Expand Down
30 changes: 23 additions & 7 deletions Pipfile.lock

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

4 changes: 3 additions & 1 deletion scraper/src/documentation_spider.py
Original file line number Diff line number Diff line change
Expand Up @@ -145,6 +145,9 @@ def start_requests(self):
},
errback=self.errback_alternative_link)

def parse(self, response, **kwargs):
return super()._parse(response, **kwargs)

def add_records(self, response, from_sitemap):
records = self.strategy.get_records_from_response(response)
self.meilisearch_helper.add_records(records, response.url, from_sitemap)
Expand Down Expand Up @@ -176,7 +179,6 @@ def parse_from_start_url(self, response):

if self.is_rules_compliant(response):
self.add_records(response, from_sitemap=False)

else:
print("\033[94m> Ignored: from start url\033[0m " + response.url)

Expand Down

0 comments on commit d973df7

Please sign in to comment.