-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathscraper.py
260 lines (218 loc) · 9.66 KB
/
scraper.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
from curses import meta
import requests
from bs4 import BeautifulSoup
from urllib.parse import urljoin, urldefrag
import log_setup
import trafilatura
import mdformat
import json
from database_manager import DatabaseManager
from tqdm import tqdm
import coloredlogs
import time
logger = log_setup.get_logger()
logger.name = "Scraper"
class Scraper:
def __init__(self, base_url, exclude_patterns, db_manager: DatabaseManager, rate_limit=0, delay=0):
"""
Initialize the Scraper object with base URL, exclude patterns, and database manager.
Log the initialization process.
Args:
base_url (str): The base URL to start scraping from.
exclude_patterns (list): List of patterns to exclude from scraping.
db_manager (DatabaseManager): The database manager object for storing scraped data.
rate_limit (int): Maximum number of requests per minute.
delay (float): Delay between requests in seconds.
"""
logger.debug(f"Initializing Scraper with base URL: {base_url}")
self.base_url = base_url
self.exclude_patterns = exclude_patterns or []
self.db_manager = db_manager
self.rate_limit = rate_limit
self.delay = delay
def is_valid_link(self, link):
"""
Check if the given link is valid for scraping.
Log the result of the validation.
Args:
link (str): The link to be checked.
Returns:
bool: True if the link is valid, False otherwise.
"""
valid = True
if self.base_url and not link.startswith(self.base_url):
valid = False
for pattern in self.exclude_patterns:
if pattern in link:
valid = False
logger.debug(f"Link validation for {link}: {valid}")
return valid
def fetch_links(self, url, html=None):
"""
Fetch all valid links from the given URL.
Log the fetching process and outcome.
Args:
url (str): The URL to fetch links from.
html (str, optional): The HTML content of the page.
Returns:
set: Set of valid links found on the page.
"""
logger.debug(f"Fetching links from {url}")
try:
if not html:
# Send a GET request to the URL
response = requests.get(url)
if response.status_code != 200:
logger.warning(
f"Failed to fetch {url} with status code {response.status_code}"
)
return []
else:
content = response.content
else:
content = html
# Parse the content using BeautifulSoup
soup = BeautifulSoup(content, "html.parser")
# Extract all anchor tags and join the URLs
links = [urljoin(url, a.get("href")) for a in soup.find_all("a", href=True)]
# Remove fragments and filter valid links
links = [
urldefrag(link)[0]
for link in links
if self.is_valid_link(urldefrag(link)[0])
]
# Log the number of valid links found
logger.debug(f"Found {len(links)} valid links on {url}")
return set(links)
except requests.RequestException as e:
logger.error(f"Error fetching {url}: {e}")
return []
def scrape_page(self, html, url):
"""
Scrape the content and metadata from the given URL.
Log the scraping process and outcome.
Args:
html (str): The HTML content of the page.
url (str): The URL to scrape.
Returns:
tuple: A tuple containing the extracted content and metadata of the page.
"""
logger.info(f"Scraping page {url}")
try:
metadata = trafilatura.metadata.extract_metadata(filecontent=html, default_url=url).as_dict()
if "body" in metadata:
metadata.pop("body")
if "commentsbody" in metadata:
metadata.pop("commentsbody")
markdown = (
trafilatura.extract(
html,
output_format="markdown",
include_formatting=True,
include_links=True,
include_tables=True,
)
or ""
)
logger.debug(f"Successfully scraped content and metadata from {url}")
return markdown, metadata
except Exception as e:
logger.error(f"Error scraping {url}: {e}")
return None, None
def start_scraping(self, url=None, urls_list=[]):
"""
Initiates the scraping process for a single URL or a list of URLs. It validates URLs,
logs the scraping process, and manages the progress of scraping through the database.
Args:
url (str, optional): A single URL to start scraping from. Defaults to None.
urls_list (list, optional): A list of URLs to scrape. Defaults to an empty list.
"""
# Validate and insert the provided URLs into the database
if urls_list:
# Iterate through the list to check for valid URLs
for url in urls_list:
if not self.is_valid_link(url):
logger.warning(f"Skipping invalid URL: {url}")
urls_list.remove(url) # Remove invalid URLs from the list
# Insert the validated list of URLs into the database
self.db_manager.insert_link(urls_list)
elif url:
# Insert a single URL if provided and valid
self.db_manager.insert_link(url)
# Log the start of the scraping process
logger.info("Starting scraping process")
# Initialize a progress bar to track scraping progress
pbar = tqdm(
total=self.db_manager.get_links_count(),
initial=self.db_manager.get_visited_links_count(),
desc="Scraping",
unit="link",
)
# Initialize rate limit tracking variables
request_count = 0
start_time = time.time()
# Begin the scraping loop
while True:
# Fetch a list of unvisited links from the database
unvisited_links = self.db_manager.get_unvisited_links()
# Exit the loop if there are no more links to visit
if not unvisited_links:
logger.info("No more links to visit. Exiting.")
break
# Process each unvisited link
for link in unvisited_links:
# Check rate limit
if self.rate_limit > 0:
current_time = time.time()
elapsed_time = current_time - start_time
if request_count >= self.rate_limit:
sleep_time = 60 - elapsed_time
if sleep_time > 0:
logger.debug(f"Rate limit reached, sleeping for {sleep_time} seconds")
time.sleep(sleep_time)
# Reset the rate limit tracker
request_count = 0
start_time = time.time()
# Wait for the specified self.delay before making the next request
if self.delay > 0:
logger.debug(f"self.delaying for {self.delay} seconds before next request")
time.sleep(self.delay)
pbar.update(1) # Update the progress bar
url = link[0] # Extract the URL from the link tuple
# Attempt to fetch the page content
response = requests.get(url)
# Increment request count for rate limiting
request_count += 1
# Check for a successful response and correct content type
if response.status_code != 200 or not response.headers.get(
"content-type", ""
).startswith("text/html"):
# Mark the link as visited and log the reason for skipping
self.db_manager.mark_link_visited(url)
logger.info(
f"Skipping link {url} due to invalid status code or content type"
)
continue
# Extract the HTML content from the response
html = response.content
# Scrape the page for content and metadata
content, metadata = self.scrape_page(html, url)
# Insert the scraped data into the database
self.db_manager.insert_page(url, content, json.dumps(metadata))
# Fetch and insert new links found on the page, if not working from a predefined list
if not urls_list:
new_links = self.fetch_links(html=html, url=url)
# Count and insert new links into the database
real_new_links_count = 0
for new_url in new_links:
if self.db_manager.insert_link(new_url):
real_new_links_count += 1
logger.debug(f"Inserted new link {new_url} into the database")
# Update the progress bar total with the count of new links
if real_new_links_count:
pbar.total += real_new_links_count
pbar.refresh()
# Mark the current link as visited in the database
self.db_manager.mark_link_visited(url)
# Close the progress bar upon completion of the scraping process
pbar.close()