-
Notifications
You must be signed in to change notification settings - Fork 1
/
crypto_webscraper_local.py
275 lines (206 loc) · 8.52 KB
/
crypto_webscraper_local.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
import uuid
import json
import urllib.request
import boto3
import logging
import os
import pandas as pd
import tempfile
from selenium import webdriver
from time import sleep
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC
from selenium.common.exceptions import NoSuchElementException
from sqlalchemy import create_engine
class Webscraper:
def __init__(self):
"""
Initaliser:
defines driver local location, creates
empty lists to be used in webscraper &
defines website to be webscraped, initilises
sqlalchemy engine for RDS transfer & sets up
friendly_id to ensure no rescraping of data,
also clicks cookie banner
"""
self.link_list = []
self.coin_image_completed = []
self.friendly_id_list = []
self.final_coin_details=[]
self.driver = webdriver.Chrome('/Users/paddy/Desktop/AiCore/scraper_project/chromedriver')
self.driver.get('https://coinmarketcap.com/')
self.url = 'https://coinmarketcap.com/'
self.next_page_string = '?page='
DATABASE_TYPE = 'postgresql'
DBAPI = 'psycopg2'
ENDPOINT = 'patrickcryptodbfinal.cycquey1wjft.us-east-1.rds.amazonaws.com'
USER = 'postgres'
PASSWORD = 'postgres'
PORT = 5432
DATABASE = 'postgres'
self.engine = create_engine(f"{DATABASE_TYPE}+{DBAPI}://{USER}:{PASSWORD}@{ENDPOINT}:{PORT}/{DATABASE}")
self.client = boto3.client('s3')
self.df = pd.read_sql_table('coin_data', self.engine)
self.id_checker = list(self.df['friendly_id'])
WebDriverWait(self.driver, 20).until(EC.element_to_be_clickable((By.CSS_SELECTOR, "div.cmc-cookie-policy-banner__close"))).click()
sleep(10)
def individual_coin_path(self):
"""
Individual_coin_path:
defines table on webpage which is
to be iterated through to scrape
properties of various coins
"""
sleep(2)
coin_container = self.driver.find_element_by_xpath('//table[@class="h7vnx2-2 czTsgW cmc-table "]')
coin_list = coin_container.find_elements_by_xpath('./tbody/tr')
return coin_list
def crypto_properties(self):
"""
crypto_properties:
this sets up the dictionary with all of
the coins attributes, also is responsible for
setting the friendly id to stop rescraping
& scrolls the page as the scraper collects
results
"""
coin_list = self.individual_coin_path()
self.image_path = '/Users/paddy/Desktop/AiCore/Scraper_Project/Coin_Images'
sleep(3)
i = 1
self.driver.execute_script("window.scrollBy(0, 50)")
self.driver.execute_script("document.body.style.zoom='50%'")
for i in range(len(coin_list)):
try:
full_coin_list= {
'uuid' : str(uuid.uuid4()),
'Name': coin_list[i].find_element_by_xpath('.//td[3]//a//p').text,
'Symbol' : coin_list[i].find_element_by_xpath('.//td[3]/div/a/div/div/div/p').text,
'Price' : coin_list[i].find_element_by_xpath('.//td[4]/div/a/span').text,
'Volume' :coin_list[i].find_element_by_xpath('.//td[8]/div/a/p').text,
'Market_cap' : coin_list[i].find_element_by_xpath('.//td//p/span[2]').text,
'Circulating_Supply' : coin_list[i].find_element_by_xpath('.//td[9]//div/div[1]/p').text
}
except NoSuchElementException:
continue
friendly_id_con = coin_list[i].find_element_by_xpath('.//td[3]/div/a')
friendly_id = friendly_id_con.get_attribute('href')
friendly_id_split = friendly_id.split('/')[-2]
self.friendly_id_list.append(friendly_id_split)
img = coin_list[i].find_element_by_class_name('coin-logo')
src = img.get_attribute('src')
full_coin_list['image'] = src
full_coin_list['friendly_id'] = friendly_id_split
coin_list = self.individual_coin_path()
print(full_coin_list)
self.driver.execute_script("window.scrollBy(0, 50)")
if friendly_id in self.friendly_id_list:
continue
else:
self.final_coin_details.append(full_coin_list)
if i == 100:
self.save_to_json_final()
return full_coin_list()
def save_to_json_final(self):
"""
save_to_json:
responsible for ensuring no duplicate
results saved & all coins saved to
easy to read json format
"""
final_coin_list = []
for coin in self.final_coin_details:
if coin not in final_coin_list:
final_coin_list.append(coin)
complete_full_coin_list = final_coin_list
with open('coins_data.json', encoding='utf-8', mode='w') as file:
json.dump(complete_full_coin_list, file, ensure_ascii=False, indent=4)
def page_iterator(self, no_of_pages):
"""
page_iterator:
responsible for iterating
webscraper through first 10
pages of website, alongside
saving results once webscraper
finishes
input: no_of_pages = int
"""
sleep(5)
element = self.driver.find_elements_by_xpath("//a[@aria-label='Next page']")
page = 1
while page <= no_of_pages:
self.crypto_properties()
self.driver.execute_script("document.body.style.zoom='100%'")
next_page_button = element[1]
sleep(3)
next_page_button.click()
page += 1
if page == no_of_pages:
return
def split_image_url(self):
'''
Split_image_url:
This method collects the image
url from the saved json object &
formats it to be uploaded to the s3 bucket
'''
self.image_link = []
self.image_uuid = []
for j in self.final_coin_details:
dict_values = j.values()
v_image_link = list(dict_values)[7]
v_uuid = list(dict_values)[0]
self.image_link.append(v_image_link)
self.image_uuid.append(v_uuid)
print(self.image_link)
print(self.image_uuid)
def upload_image_jpeg(self):
'''
upload_image_jpeg:
Using a temporary dictionary,
iterates throught the image url's
sending them to the AWS S3 bucket in
jpeg format labelled with the same
uuid as the json data belonging to that coin
'''
self.split_image_url()
v_image_link = self.image_link
v_uuid= self.image_uuid
with tempfile.TemporaryDirectory() as tmpdict:
for i in range(len(self.image_link)):
urllib.request.urlretrieve(v_image_link[i], tmpdict + f'{v_uuid[i]}.jpg')
self.client.upload_file(tmpdict + f'{v_uuid[i]}.jpg', 'patrickcryptobucketfinal', f'{v_uuid[i]}.jpg')
return print('images have been uploaded to S3')
def data_to_sql(self):
'''
data_to_sql:
turns json into pandas df &
imports that to postresql, checks
with postresql that no duplicates exist using DISTINCT
SQL function
'''
with open('./coins_data.json', 'r') as filename:
df_coins = json.load(filename)
df = pd.DataFrame(df_coins)
df.columns = df.columns.str.lower()
self.engine.connect()
df.to_sql('coin_data', con=self.engine, if_exists='replace')
pd.read_sql_query('''SELECT DISTINCT coin_data.Name From coin_data;''', self.engine)
print('duplicate values have been removed')
if __name__ == '__main__':
'''
Main initiliser:
the if statement infers
that this file is the main file
being run instead of importing from
a seperate module as this is true it
will initilise both our classes & run them
sequentially
'''
public_webscraper = Webscraper()
public_webscraper.page_iterator(2)
public_webscraper.upload_image_jpeg()
public_webscraper.save_to_json_final()
public_webscraper.data_to_sql()
print('scraper has finished')