-
-
Notifications
You must be signed in to change notification settings - Fork 4
/
core.py
978 lines (757 loc) Β· 44.3 KB
/
core.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
# Python shipped packages imports
import re
import sys
import time
import json
import base64
import logging
from decimal import Decimal
from pathlib import Path
# Dependencies imports
import requests
from bs4 import BeautifulSoup
import browser_cookie3
# Global variables
SCRIPT_NAME = 'CW Wizard'
VERSION = '1.0.6'
EXIT_ERROR_MSG = 'The Wizard encountered issue(s) please check previous logs.\n'
EXIT_SUCCESS_MSG = 'The Wizard has finish is work, have a great day!\n'
# These two variables are extracted and set from the first wantlist url given to the script
CURR_GAME = 'Magic'
CARDMARKET_TOP_DOMAIN = '.cardmarket.com'
CARDMARKET_BASE_URL = 'https://www' + CARDMARKET_TOP_DOMAIN
CARDMARKET_BASE_URL_REGEX = r'^https:\/\/www\.cardmarket\.com'
# This value can be override via script arguments or via GUI
MAXIMUM_SELLERS = 20
# This value cannot be currently override (this represent a max of 350 offers per card)
MAXIMUM_NB_REQUESTS_PER_CARD = 6
# The delay is now mandatory seen Cardmarket block for 2h to 24h users that makes too much requests in a small amount of time.
DELAY_BETWEEN_REQUEST = 0.170
CARD_LANGUAGES = { 'English': 1, 'French': 2, 'German': 3, 'Spanish': 4,
'Italian': 5, 'S-Chinese': 6, 'Japanese': 7,
'Portuguese': 8, 'Russian': 9, 'Korean': 10,
'T-Chinese': 11, 'Dutch': 12, 'Polish': 13, 'Czech': 14, 'Hungarian': 15 }
CARD_CONDITIONS = { 'Mint': 1, 'Near Mint': 2, 'Excellent': 3,
'Good': 4, 'Light Played': 5, 'Played': 6, 'Poor': 7, }
CARD_CONDITIONS_SHORTNAMES = { 'Mint': 'MT', 'Near Mint': 'NM', 'Excellent': 'EX',
'Good': 'GD', 'Light Played': 'LP', 'Played': 'PL', 'Poor': 'PO', }
# These request errors description are from https://api.cardmarket.com/ws/documentation/API_2.0:Response_Codes
REQUEST_ERRORS = { 307: ['Temporary Redirect', 'Particular requests can deliver thousands of entities (e.g. a large stock or requesting articles for a specified product, and many more). Generally all these request allow you to paginate the results - either returning a 206 or 204 HTTP status code. Nevertheless, all these requests can also be done without specifying a pagination. As general values for force-paginated requests please use maxResults=100 in order to avoid being 307\'d again. The first result is indexed with 0: start=0.'],
400: ['Bad Request', 'Whenever something goes wrong with your request, e.g. your POST data and/or structure is wrong, or you want to access an article in your stock by providing an invalid ArticleID, a 400 Bad Request HTTP status is returned, describing the error within the content.'],
401: ['Unauthorized', 'You get a 401 Unauthorized HTTP status, when authentication or authorization fails during your request, e.g. your Authorization (signature) is not correct.'],
403: ['Forbidden', 'You get a 403 Forbidden HTTP status, when you try to access valid resources, but don\'t have access to it, i. e. you try to access /authenticate with a dedicated or widget app, or resources specifically written for widget apps with a dedicated app.'],
404: ['Invalid request', 'The request isn\'t formatted correctly, request data and/or structure is wrong.'],
405: ['Not Allowed', 'You get a 405 Not Allowed HTTP status, every time you want to access a valid resource with a wrong HTTP method. While OPTIONS requests are now possible on all of the API\'s resources, most resources are limited to one or more other HTTP methods. These are always specified in the Access-Control-Allow-Methods header coming with each response. Please refer to CRUD Operations Documentation to learn more about the different HTTP methods and which purposes they fulfill in a RESTful API.'],
412: ['Precondition Failed', 'When you want to perform an invalid state change on one of your orders, e.g. confirm reception on an order, that\'s still not flagged as sent, you get a 412 Precondition Failed HTTP status.'],
417: ['Expectation Failed', 'Typically you get a 417 Expectation Failed HTTP status code, when your request has an XML body without the corresponding header and/or the body not sent as text, but its byte representation. Please also don\'t send any Expect: header with your request.'],
429: ['Too Many Requests', 'Our API has the following request limits which reset every midnight at 12am (0:00) CET/CEST: - Dedicated App (private users): 5.000 | - Dedicated App (commercial users): 100.000 | - Dedicated App (powerseller users): 1.000.000 | - Widget and 3rd-Party Apps don\'t have any request limits. If your has a request limit, additional response headers are sent by the API: - X-Request-Limit-Max, which contains your request limit, - X-Request-Limit-Count, which contains the actual number of requests you made after the last request limit reset. Once your request limit is reached the API will answer with a 429 Too Many Requests until the next request limit reset.']}
CREDENTIALS_PATH = Path.cwd().joinpath('credentials.json')
REQUEST_HEADERS = {
"accept": "*/*",
"accept-language": "*",
"dnt": "1"
}
COMPATIBLE_BROWSERS = ["chrome", "firefox", "opera", "opera_gx", "edge", "chromium", "brave", "vivaldi", "safari"]
class FunctResult():
def __init__(self):
self.status = 'valid'
self.messages = []
self.result = None
def addWarning(self, message):
if self.status == 'valid':
self.status = 'warning'
self.messages.append({ 'type': 'warning', 'content': message, 'logged': False })
def addError(self, message):
self.status = 'error'
self.messages.append({ 'type': 'error', 'content': message, 'logged': False })
def addDetailedRequestError(self, task, response, as_warning=False):
status_code = response.status_code
message = 'Unable to {}.\nRequest status code "{}":'.format(task, str(status_code))
# Check if we have more info on the request error
status_code_info = REQUEST_ERRORS.get(status_code)
detailed_message = 'Unknown status code!'
if status_code_info:
# Log the additional info
detailed_message = '{} -- {}'.format(status_code_info[0], status_code_info[1])
if as_warning:
self.addWarning(message)
self.addWarning(detailed_message)
else:
self.addError(message)
self.addError(detailed_message)
def append(self, result):
if result.status == 'error':
self.status = 'error'
if result.status == 'warning' and self.status == 'valid':
self.status = 'warning'
self.messages = [*self.messages, *result.messages]
def setResult(self, element):
self.result = element
def addResult(self, element):
if not self.result:
self.result = element
elif isinstance(self.result, list):
self.result.append(element)
else:
self.result = [self.result, element]
def getResult(self):
return self.result
def isWarning(self):
return self.status == 'warning'
def isValid(self):
return self.status == 'valid'
def logMessages(self):
for message in self.messages:
if not message['logged']:
method = getattr(LOG, message['type'])
method(message['content'])
message['logged'] = True
def getMessages(self, message_type='all'):
messages = []
for message in self.messages:
if message_type == 'all' or (message_type != 'all' and message['type'] == message_type):
messages.append(message['content'])
return messages
class ColoredFormatter(logging.Formatter):
"""Custom formatter handling color"""
cyan = '\x1b[36;20m'
green = '\x1b[32;20m'
yellow = '\x1b[33;20m'
red = '\x1b[31;20m'
bold_red = '\x1b[31;1m'
reset = '\x1b[0m'
message_format = '%(levelname)-8s - %(message)s'
FORMATS = {
logging.DEBUG: cyan + message_format + reset,
logging.INFO: green + message_format + reset,
logging.WARNING: yellow + message_format + reset,
logging.ERROR: red + message_format + reset,
logging.CRITICAL: bold_red + message_format + reset
}
def format(self, record):
log_format = self.FORMATS.get(record.levelno)
formatter = logging.Formatter(log_format)
return formatter.format(record)
def init_logger():
"""Initialize script logger"""
logger_name = Path(__file__).stem
logger = logging.getLogger(logger_name)
logger.setLevel(level = logging.DEBUG)
console_handler = logging.StreamHandler(sys.stdout)
console_handler.setLevel(level=logging.DEBUG)
console_handler.setFormatter(ColoredFormatter())
logger.addHandler(console_handler)
return logger
LOG = init_logger()
def extract_log_in_error_msg(response):
error_msg = ''
# Step 1: Convert response to BeautifulSoup object
soup = BeautifulSoup(response.text, 'html.parser')
# Step 2: Try to find the error message in the HTML
error_msg_container = soup.find('h4', class_='alert-heading')
if error_msg_container:
error_msg = error_msg_container.get_text('. ')
return error_msg
def cardmarket_log_in(session, credentials, silently=False):
funct_result = FunctResult()
if not silently:
LOG.debug('------- The Wizard start for you a temporary session on Cardmarket...')
LOG.debug(' |____ This temp session isn\'t related to your browser, it\'s like private navigation session.')
LOG.debug(' |____ No cookies will be stored/remains at the end.\n')
# Step 1: Get the login page html (to retrieve the login token)
try:
response_get_login_page = session.get(CARDMARKET_BASE_URL + '/Login')
except Exception as e:
funct_result.addError('Unable to connect to Cardmarket.\nReason: {}'.format(e))
return funct_result
if response_get_login_page.status_code != 200:
# Issue with the request
funct_result.addDetailedRequestError('access to Cardmarket', response_get_login_page)
return funct_result
# Step 2: Extract the token from the html string
regex_match = re.search(r'name="__cmtkn" value="(?P<token>\w+)"', response_get_login_page.text)
if regex_match.lastindex is None:
# Cannot retrieve the login token
funct_result.addError('Cannot retrieve the login token.')
return funct_result
# Step 3: Prepare payload
token = regex_match.group('token')
referal_page_path = '/en/{}'.format(CURR_GAME)
payload = {'__cmtkn': token, 'referalPage': referal_page_path, 'username': credentials['login'], 'userPassword': credentials['password']}
# Step 4: Do the log-in POST request to Cardmarket with the payload
time.sleep(DELAY_BETWEEN_REQUEST)
response_post_login = session.post('{}/en/{}/PostGetAction/User_Login'.format(CARDMARKET_BASE_URL, CURR_GAME), data=payload)
if response_post_login.status_code != 200:
# Issue with the request
funct_result.addDetailedRequestError('log-in to Cardmarket', response_post_login)
return funct_result
# Step 5: Check in the response HTML if there is a log-in error
log_in_error_msg = extract_log_in_error_msg(response_post_login)
if log_in_error_msg:
# It's most likely an issue with the payload (wrong username and/or password)
funct_result.addError('Unable to log-in to Cardmarket.\nMessage: {}.'.format(log_in_error_msg))
return funct_result
if not silently:
LOG.info('Successfully logged in !\n')
return funct_result
def cardmarket_log_out(session, silently=False):
funct_result = FunctResult()
if not silently:
LOG.debug('------- The Wizard log out of the temporary session on Cardmarket...\n')
time.sleep(DELAY_BETWEEN_REQUEST)
response_get_logout = session.get('{}/en/{}/PostGetAction/User_Logout'.format(CARDMARKET_BASE_URL, CURR_GAME))
if response_get_logout.status_code != 200:
# Issue with the request
funct_result.addDetailedRequestError('logout of Cardmarket', response_get_logout)
return funct_result
if not silently:
LOG.info('Successfully logout!')
return funct_result
def rename_dict_key(dictionnary, old_key_name, new_key_name):
# No operation performed iff key not in dictionnary
if old_key_name in dictionnary:
dictionnary[new_key_name] = dictionnary[old_key_name]
del dictionnary[old_key_name]
return dictionnary
def get_wantlist_columns_correlation_table(wantlist_table_head):
correlation_table = {}
for index, th_column in enumerate(wantlist_table_head.find_all('th')):
if th_column.has_attr('class'):
column_name = th_column['class'][0]
if column_name == 'min-size':
# Special case: This is a boolean column, let's find the correct name
# Get span title in title format (capitalize first letter of every words)
column_name = th_column.find('span')['title'].title()
# Removing all non alphanumerical characters
column_name = ''.join(e for e in column_name if e.isalnum())
# Lastly add a prefix, so we know it's a True/False column
column_name = 'is' + column_name
# Add to correlation dict
correlation_table[column_name] = index
# Rename some keys to match requirements
correlation_table = rename_dict_key(correlation_table, old_key_name='condition', new_key_name='minCondition')
correlation_table = rename_dict_key(correlation_table, old_key_name='buyPrice', new_key_name='maxPrice')
correlation_table = rename_dict_key(correlation_table, old_key_name='isReverseHolo', new_key_name='isReverse')
correlation_table = rename_dict_key(correlation_table, old_key_name='isFirstEdition', new_key_name='isFirstEd')
# Remove useless entries
# (not checking first if they exists, since they always SHOULD exists, common columns)
del correlation_table['select']
del correlation_table['preview']
del correlation_table['mailAlert']
del correlation_table['action']
return correlation_table
def retrieve_wantlist(session, wantlist_url, continue_on_warning=False):
funct_result = FunctResult()
LOG.debug(' |____ The Wizard is retrieving the wantlist ("{}")...'.format(wantlist_url))
wantlist = None
# Get Information from url to recreate it (changing language to english)
wantlist_url_regex = CARDMARKET_BASE_URL_REGEX + r"\/[a-z]{2}\/(?P<game>\w+)\/Wants/(?P<wantlist_id>\d+)$"
match = re.match(wantlist_url_regex, wantlist_url)
if not match:
funct_result.addError('The wantlist url ("{}") seems invalid'.format(wantlist_url))
return funct_result
wantlist_url = '{}/en/{}/Wants/{}'.format(CARDMARKET_BASE_URL, match.group('game'), match.group('wantlist_id'))
# Step 1: Get the desired wantlist page
time.sleep(DELAY_BETWEEN_REQUEST)
response_get_wantlist = session.get(wantlist_url)
if response_get_wantlist.status_code != 200:
# Issue with the request
funct_result.addDetailedRequestError('access to the wantlist ("{}")'.format(wantlist_url), response_get_wantlist, continue_on_warning)
return funct_result
# Step 2: Convert response to BeautifulSoup object
soup = BeautifulSoup(response_get_wantlist.text, 'html.parser')
# Step 3: Retrieve the wantlist title
wantlist_title = soup.find('h1').string
# Step 4: Retrieve the wantlist table (BeautifulSoup object)
wantlist_section = soup.find(id='WantsListTable')
if not wantlist_section:
funct_result.addWarning('The wantlist seems to be empty, or the cards list could\'nt be retrieved')
return funct_result
wantlist_table = wantlist_section.table.tbody
# Get all the columns (attributes) available for this wantlist
# Create a correlation table: column name -> index in article row
wantlist_table_head = wantlist_section.table.thead
column_correlation_table = get_wantlist_columns_correlation_table(wantlist_table_head)
# Step 5: Convert the wantlist table to python list
card_count = len(wantlist_table.contents)
card_count_str = '{} card{}'.format(str(card_count), 's' if card_count > 1 else '')
wantlist = [ '{} ({})'.format(wantlist_title, card_count_str), wantlist_url ]
for row in wantlist_table.children:
card = {}
# Step 5.A: Retrieve and add attributes to the card dict
name_link_tag = row.contents[column_correlation_table['name']].a
card['url'] = CARDMARKET_BASE_URL + name_link_tag['href'].split('?')[0]
card['title'] = name_link_tag.contents[0]
card['languages'] = []
languages_link_tags = row.contents[column_correlation_table['languages']].find_all('a')
for languages_link_tag in languages_link_tags:
card['languages'].append(languages_link_tag.span['data-original-title'])
card['quantity'] = int(row.contents[column_correlation_table['amount']].string)
for attribute_name, column_index in column_correlation_table.items():
if attribute_name not in ['amount', 'name', 'languages', 'maxPrice']:
span_child = row.contents[column_index].find('span', class_='sr-only')
if span_child is not None:
card[attribute_name] = span_child.contents[0]
card['maxPrice'] = row.contents[column_correlation_table['maxPrice']].span.contents[0]
# If maxPrice is a value we convert it into a Decimal
if card['maxPrice'] != 'N/A':
card['maxPrice'] = Decimal(card['maxPrice'].split(' ')[0].replace('.','').replace(',','.'))
# Step 5.B: Add newly created card dict to the wishlish array
wantlist.append(card)
funct_result.addResult(wantlist)
return funct_result
def _get_load_more_args(card, card_id, is_metacard):
args_dict = { 'page': '0' }
filter_settings = {}
filter_settings['idLanguage'] = {str(CARD_LANGUAGES[language]): CARD_LANGUAGES[language] for language in card['languages']}
# The names of attributes in the list below are the ones accepted as filters by Cardmarket
for attribute in ['isReverse', 'isSigned', 'isFirstEd', 'isAltered']:
if attribute in card and card[attribute] != 'Any':
filter_settings[attribute] = card[attribute]
condition = [shortname for shortname in CARD_CONDITIONS_SHORTNAMES.values()]
if card['minCondition'] != 'Poor':
condition = condition[:condition.index(CARD_CONDITIONS_SHORTNAMES[card['minCondition']]) + 1]
filter_settings['condition'] = condition
args_dict['filterSettings'] = json.dumps(filter_settings, separators=('\\,', ':'))
if not is_metacard:
args_dict['idProduct'] = card_id
else:
args_dict['idMetacard'] = card_id
return args_dict
def _get_load_more_card_id(load_more_btn):
onclick_str = load_more_btn['onclick']
is_metacard = False
# Looking for default Product ID
product_id_match = re.search(r'\'idProduct\'\:\'(?P<product_id>\d+)\'', onclick_str)
if product_id_match:
return is_metacard, product_id_match.group('product_id')
# Looking for Metacard ID
metacard_id_match = re.search(r'\'idMetacard\'\:\'(?P<metacard_id>\d+)\'', onclick_str)
if metacard_id_match:
return is_metacard, metacard_id_match.group('metacard_id')
def _get_load_more_request_token(load_more_btn):
onclick_str = load_more_btn['onclick']
return re.match(r"jcp\('(?P<token>[^']+)'", onclick_str).group('token')
def load_more_articles(session, funct_result, soup, card, articles_table):
nb_of_requests = 0
# Step 1: Check if there isn't a load more articles button, in this case we stop
load_more_btn = soup.find(id='loadMoreButton')
if not load_more_btn:
return None
# Step 2: Initialize variables
active = True
card_curr_game = re.match(CARDMARKET_BASE_URL_REGEX + r'\/\w{2}\/(\w+)', card['url']).group(1)
is_metacard, card_id = _get_load_more_card_id(load_more_btn)
load_more_args = _get_load_more_args(card, card_id, is_metacard=is_metacard)
request_token = _get_load_more_request_token(load_more_btn)
# Step 3: Retrieve more article until card['maxPrice'] is reached or there is no more article to load
while active and nb_of_requests < MAXIMUM_NB_REQUESTS_PER_CARD:
# Step 3.A: Get the price of the last card currently displayed
last_article = articles_table.contents[-1]
last_article_price_str = last_article.find('div', class_='price-container').find('span', class_='text-nowrap').contents[0]
last_article_price = Decimal(last_article_price_str.split(' ')[0].replace('.','').replace(',','.'))
# Step 3.B: Check if we need to load more articles, if yes send a new request to get more articles.
if last_article_price <= card['maxPrice']:
# Step 3.B.I: Initialize a payload and do a POST request
args_base64 = base64.b64encode(bytes(json.dumps(load_more_args, separators=(',', ':')), 'utf-8'))
payload = {'args': request_token + args_base64.decode("utf-8")}
LOG.debug(' |_ Loading 50 more offers for the card...')
response_post_load_article = session.post('{}/en/{}/AjaxAction'.format(CARDMARKET_BASE_URL, card_curr_game), data=payload)
if response_post_load_article.status_code != 200:
# Issue with the request
funct_result.addWarning('Failed to load more articles for card page ("{}")'.format(card['title']))
# We cannot retrieve more articles, so break the loop
break
# Step 3.B.II: Handle the request result containing the new articles and the new page_index value
more_article_soup = BeautifulSoup(response_post_load_article.text, 'html.parser')
new_page_index_soup = more_article_soup.find('newpage')
if new_page_index_soup:
load_more_args['page'] = int(new_page_index_soup.contents[0])
new_articles_rows_soup = more_article_soup.find('rows')
if new_articles_rows_soup:
articles_rows_html_str = base64.b64decode(new_articles_rows_soup.contents[0]).decode("utf-8")
articles_table.append(BeautifulSoup(articles_rows_html_str, 'html.parser'))
else:
active = False
# TODO: All these if else active = False should be refactored
if load_more_args['page'] < 0:
# There is no more article available, stop the process
active = False
else:
active = False
else:
active = False
nb_of_requests += 1
def populate_sellers_dict(session, sellers, wantlist, articles_comment=False, continue_on_warning=False):
funct_result = FunctResult()
wantlist_url = wantlist.pop(0)
LOG.debug(' |____ The Wizard is aggregating sellers and articles data for the wantlist ("{}")...'.format(wantlist_url))
for card in wantlist:
# If multiple languages selected, do one request per language
for card_language in card['languages']:
# Save a sellers list for the current card,
# to avoid adding multiple time the same article for a seller
# in case a seller sell the card multiple times in different conditions
card_sellers_names = []
# Step 1: Create the get params for the request according to card attributes
params = {}
params['language'] = CARD_LANGUAGES[card_language]
params['minCondition'] = CARD_CONDITIONS[card['minCondition']]
for attribute in ['isReverse', 'isSigned', 'isFirstEd', 'isAltered']:
if attribute in card and card[attribute] != 'Any':
params[attribute] = card[attribute]
# Step 2: Get the card page
time.sleep(DELAY_BETWEEN_REQUEST)
LOG.debug(' |__ Retrieving offers for card: {} ...'.format(card['title']))
response_get_card_articles = session.get(card['url'], params=params)
if response_get_card_articles.status_code != 200:
# Issue with the request
funct_result.addDetailedRequestError('access the card page ("{}")'.format(card['title']), response_get_card_articles, continue_on_warning)
if continue_on_warning:
continue
return funct_result
card_full_url = response_get_card_articles.url
# Step 3.A: Retrieve the articles table (BeautifulSoup object)
soup = BeautifulSoup(response_get_card_articles.text, 'html.parser')
articles_table = soup.find('div', class_='table-body')
# Step 3.B: Load more articles if necessary
if isinstance(card['maxPrice'], Decimal):
load_more_articles(session, funct_result, soup, card, articles_table)
# Step 4: Iterate over articles (a.k.a offers)
for article_row in articles_table.children:
# Step 4.A: Check if this is a proper article
if 'article-row' not in article_row.attrs['class']:
funct_result.addWarning('No offers found for card ("{}") with parameters: {} {}.'.format(card['title'], params, card['maxPrice']))
break
# Step 4.B: Check if the user can purchase the item, ship to is address available from the seller.
action_container = article_row.find('div', class_='actions-container')
if not action_container.find('div', class_='input-group'):
# Cannot purchase, skip this article_row
continue
# Step 4.C: Retrieve Seller info
seller_name_tag = article_row.find('span', class_='seller-name').find('a')
seller_name = seller_name_tag.contents[0]
seller_sales_info_wrapper = article_row.find('span', class_='seller-extended').contents[1]
seller_sales_number = re.match(r'^\d+', seller_sales_info_wrapper['title']).group(0)
# Step 4.D Skip if we already added an article (of this card) for this seller
if seller_name in card_sellers_names:
# Skip this article_row
continue
seller_profile_url = CARDMARKET_BASE_URL + seller_name_tag['href']
price_str = article_row.find('div', class_='price-container').find('span', class_='text-nowrap').contents[0]
price = Decimal(price_str.split(' ')[0].replace('.','').replace(',','.'))
# Step 4.E: Check if price isn't above maxPrice
if isinstance(card['maxPrice'], Decimal) and price > card['maxPrice']:
# The current article price is superior than the max price
# we stop iterate over article (article are listed according to price)
break
# Step 4.F: Create the new article
article_attributes = article_row.find('div', class_='product-attributes')
article_condition = article_attributes.a.span.contents[0]
article_language = article_attributes.find('span', class_='icon')['data-original-title']
article_dict = { 'name': card['title'], 'url': card_full_url, 'language': article_language, 'condition': article_condition, 'price': price }
# Step 4.G: Handle seller comment for the article
if articles_comment:
article_comment = ""
article_comment_wrapper = article_row.find('div', class_='product-comments')
if article_comment_wrapper:
article_comment = article_comment_wrapper.find('span', class_='text-truncate').contents[0]
article_dict['comment'] = article_comment
# Step 4.H: Add this article on the seller key in the dict
if seller_name in sellers:
sellers[seller_name]['cards'].append(article_dict)
else:
sellers[seller_name] = {'url': seller_profile_url, 'sales': seller_sales_number, 'cards': [ article_dict ] }
# Step 4.I: Add seller name in the corresponding list
card_sellers_names.append(seller_name)
return funct_result
def determine_relevant_sellers(sellers, max_sellers):
LOG.debug('------- The Wizard is sorting sellers to find relevant ones...')
if not sellers:
return LOG.warning(' |____ It\'s seems there is no seller.')
sorted_list = []
# Step 1: Sort sellers dict according to number of articles
for seller_name, data in sellers.items():
total_price = Decimal(0.0)
for article in data['cards']:
total_price += article['price']
sorted_list.append((seller_name, len(data['cards']), str(total_price)))
from operator import itemgetter
sorted_list = sorted(sorted_list,key=itemgetter(1), reverse=True)
# Step 2: reduce sellers list
relevant_sellers = sorted_list[:max_sellers]
return relevant_sellers
def delete_previous_result_page(path_file):
funct_result = FunctResult()
"""Delete a file"""
if path_file and path_file.exists():
try:
if path_file.is_file():
LOG.debug(' |____ Deleting the previous result page...')
import os
os.remove(path_file)
except Exception as e:
funct_result.addError('Failed to delete "{}". Reason: {}'.format(path_file, e))
return funct_result
return funct_result
def build_result_page(wantlists_info, max_sellers, sellers, relevant_sellers):
funct_result = FunctResult()
LOG.debug('------- The Wizard is creating a beautiful result page...')
# Step 1: Retrieve content of template.html
soup = None
template_path = Path.cwd().joinpath('assets', 'template.html')
try:
with open(template_path, 'r', encoding='utf-8') as template_file:
template_contents = template_file.read()
soup = BeautifulSoup(template_contents, 'html.parser')
except IOError as err:
funct_result.addError('Template file "{}" cannot be properly opened.'.format(template_path))
funct_result.addError(err)
return funct_result
# Step 2: Edit the template
# Step 2.A: Add "max_sellers" value
max_sellers_parent_tag = soup.find('span', id='max-sellers-value')
max_sellers_parent_tag.string = str(max_sellers)
# Step 2.B: Add wantlists
wantlists_parent_tag = soup.find('div', id='wantlist-links')
wantlists_links_html_str = ''
for wantlist_info in wantlists_info:
wantlists_links_html_str += '<a class="wantlist-item button" href="{}" target="_blank" rel="noopener noreferrer">{}</a>'.format(wantlist_info[0], wantlist_info[1])
wantlists_parent_tag.append(BeautifulSoup(wantlists_links_html_str, 'html.parser'))
# Step 2.C: Retrieve containers tags and declare variables for relevant sellers
sellers_parent_tag = soup.find('div', id='relevant-sellers-items')
sellers_html_str = ''
sellers_cards_lists_parent_tag = soup.find('div', id='sellers-cards-lists')
sellers_cards_lists_html_str = ''
# Step 2.D: Create HTML contents for relevant sellers
seller_index = 0
for relevant_seller in relevant_sellers:
index_5_digits_str = '{:05}'.format(seller_index)
sellers_html_str += '<div class="seller-item"><a href="{}" id="seller-{}" class="seller-name" target="_blank" rel="noopener noreferrer">{}</a><span id="seller-{}-sales-number" class="hidden">{}</span><hr><span class="number-cards">Cards: <b>{}</b></span><span class="total-price">Total: {} β¬</span><a href="#" onclick="showCardsList(\'{}\'); return false;" class="link-cards-list button">See Cards ></a></div>'.format(sellers[relevant_seller[0]]['url'], index_5_digits_str, relevant_seller[0], index_5_digits_str, sellers[relevant_seller[0]]['sales'], str(relevant_seller[1]), str(relevant_seller[2]), index_5_digits_str)
# Concatenate cards list
sellers_cards_lists_html_str += '<div id="seller-{}-cards" class="cards-list">'.format(index_5_digits_str)
for card in sellers[relevant_seller[0]]['cards']:
article_comment = ""
if 'comment' in card:
article_comment = card['comment']
sellers_cards_lists_html_str += '<div class="card-item"><span class="card-condition {}">{}</span><span class="card-language {}"></span><a href="{}" class="card-title" target="_blank" rel="noopener noreferrer">{}</a><span class="card-comment" title="{}">{}</span><span class="card-price">{} β¬</span></div>'.format(card['condition'], card['condition'], card['language'], card['url'], card['name'], article_comment, article_comment, card['price'])
sellers_cards_lists_html_str += '</div>'
seller_index += 1
# Step 2.E: Append HTML for relevant sellers
sellers_parent_tag.append(BeautifulSoup(sellers_html_str, 'html.parser'))
sellers_cards_lists_parent_tag.append(BeautifulSoup(sellers_cards_lists_html_str, 'html.parser'))
# Step 3: Save as "result.html" in current project directory
result_path = Path.cwd().joinpath('result.html')
# Step 3.A: if "result.html" exists delete it
delete_result = delete_previous_result_page(result_path)
funct_result.append(delete_result)
if not delete_result.isValid():
return funct_result
LOG.debug(' |____ Saving the result page...\n')
# Step 3.B: Create and write to the file
try:
with open(result_path, 'w+', encoding='utf-8') as result_file:
result_file.write(str(soup))
except IOError as err:
funct_result.addError('Error while creating the result file ("{}").'.format(result_path))
funct_result.addError(err)
return funct_result
LOG.info('The result page has been created here: "{}"\n'.format(result_path.as_uri()))
# Step 4: Open the result page
import webbrowser
try:
webbrowser.open_new_tab(result_path.as_uri())
except webbrowser.Error:
# Since it's not critical at all only display a warning.
funct_result.addWarning('Failed to automatically open the result page for you.')
funct_result.addResult(result_path.as_uri())
return funct_result
def get_cardmarket_cookies(browser_name):
# Caution: browser_name should be in snake case
# Retrieve Cardmarket Cookies
if browser_name is not None:
funct_retrieve_cookies = getattr(browser_cookie3, browser_name, None)
if funct_retrieve_cookies is not None:
return funct_retrieve_cookies(domain_name=CARDMARKET_TOP_DOMAIN)
# Fallback get Cardmarket cookies in every browsers installed
# (can fail for example if users don't give access rights for Safari cookies)
return browser_cookie3.load(domain_name=CARDMARKET_TOP_DOMAIN)
def cardmarket_wantlist_wizard(browser_name, user_agent, credentials, wantlist_urls, continue_on_warning, max_sellers, articles_comment=False):
funct_result = FunctResult()
LOG.debug('------- Calling the Wizard...\r\n')
end_intro_message = 'these whislists urls!' if len(wantlist_urls) > 1 else 'this wantlist url!'
LOG.info('Hi there... You are here to find great card deals, right? Humm... okay... Give me {}'.format(end_intro_message))
LOG.warning('Be aware that he\'s very old and can be sometimes grumpy and slow to perform all these tasks.\r\n')
# Step 1: Create a web session (to be able to stay connected)
with requests.Session() as session:
# Setting cookies and headers to bypass / skip Cloudflare protection
session.cookies = get_cardmarket_cookies(browser_name)
headers = REQUEST_HEADERS.copy()
headers["user-agent"] = str(user_agent)
session.headers.update(headers)
# Step 2: Log-in to Cardmarket
funct_result = cardmarket_log_in(session, credentials)
funct_result.logMessages()
if not funct_result.isValid():
# FATAL error we cannot perform anything without being log-in
return funct_result
sellers = {}
wantlists_info = []
LOG.debug('------- Handling the wantlist(s)')
for wantlist_url in wantlist_urls:
# Step 2: Retrieve wantlist in JSON
retrieve_result = retrieve_wantlist(session, wantlist_url, continue_on_warning)
retrieve_result.logMessages()
funct_result.append(retrieve_result)
if retrieve_result.isWarning():
# Empty wantlist, or unable to retrieve wantlist but continue_on_warning is True
continue
elif not retrieve_result.isValid():
# Only break, not return so we can still log out of Cardmarket
break
wantlist = retrieve_result.getResult()
# Step 2.A: Add info to the wantlists_info list
wantlists_info.append((wantlist_url, wantlist.pop(0)))
# Step 3: Populate the sellers dictionary
populate_result = populate_sellers_dict(session, sellers, wantlist, continue_on_warning=continue_on_warning, articles_comment=articles_comment)
populate_result.logMessages()
funct_result.append(populate_result)
# Step 4: Display most relevant sellers
if funct_result.isValid() or (funct_result.isWarning() and continue_on_warning):
relevant_sellers = determine_relevant_sellers(sellers, max_sellers)
if relevant_sellers:
build_result = build_result_page(wantlists_info, max_sellers, sellers, relevant_sellers)
build_result.logMessages()
funct_result.append(build_result)
result_path = build_result.getResult()
if result_path != None:
funct_result.setResult(result_path)
# Step 5: Logout from Cardmarket, simply a safety mesure.
logout_result = cardmarket_log_out(session)
logout_result.logMessages()
funct_result.append(logout_result)
# Avoid issue(s) by ensuring we have a result
if funct_result.getResult() == None:
funct_result.addResult('')
return funct_result
def get_credentials_from_file():
funct_result = FunctResult()
LOG.debug('------- Retrieving credentials...')
credentials = None
# Step 1: Check if the path point to an existing file
if not CREDENTIALS_PATH.exists():
# File not found
funct_result.addWarning('file \'credentials.json\' doesn\'t exists in the current directory.')
return funct_result
# Step 2: Open the file as JSON
try:
with open(CREDENTIALS_PATH, 'r', encoding='utf-8') as json_file:
credentials = json.load(json_file)
except json.JSONDecodeError as err:
funct_result.addError('Issue(s) in the JSON file: {} (line {} pos {}).'.format(err.msg, err.lineno, err.pos))
return funct_result
except IOError as err:
funct_result.addError('Credentials file "{}" cannot be properly opened.'.format(CREDENTIALS_PATH))
funct_result.addError(err)
return funct_result
# Step 3: Check if the JSON object contains the necessary keys and values
# And the values are strings
keys_to_check = ['login', 'password']
for key in keys_to_check:
if key not in credentials:
funct_result.addError('Key "{}" not present in the credentials JSON file "{}".'.format(key, CREDENTIALS_PATH))
return funct_result
if not credentials[key] or not isinstance(credentials[key], str):
funct_result.addError('Value of key "{}" isn\'t of type string in the credentials JSON file "{}".'.format(key, CREDENTIALS_PATH))
return funct_result
funct_result.addResult(credentials)
return funct_result
def check_wantlists_and_max_sellers(wantlist_urls, max_sellers, silently=False):
funct_result = FunctResult()
if not silently:
LOG.debug('------- Checking inputs parameters...')
# Declare variables to store values that will be at the end set to GLOBAL variables
global_language = None
global_game = None
# Step 1: Check the wantlist urls
for wantlist_url in wantlist_urls:
if not isinstance(wantlist_url, str):
funct_result.addError('wantlist url need to be of type string.')
continue
matched = re.match(CARDMARKET_BASE_URL_REGEX + r'\/(\w{2})\/(\w+)\/Wants\/\d+$', wantlist_url)
if not matched:
funct_result.addError('Invalid wantlist url ("{}"), valid pattern is:'.format(wantlist_url))
funct_result.addError(CARDMARKET_BASE_URL + '/<LANGUAGE_CODE>/<GAME_NAME>/Wants/<wantlist_CODE>')
continue
if not global_game:
(global_language, global_game) = matched.groups()
if not funct_result.isValid():
return funct_result
# Step 2: Check value of "max_sellers"
if max_sellers < 0:
funct_result.addError('Input parameter "max_sellers" need to be an int superior to 0.')
return funct_result
elif max_sellers == 0 or max_sellers > 300:
funct_result.addWarning('Value of input parameter "max_sellers" is set to "{}", caution this means the process can be long and the result page can be heavy.'.format(max_sellers))
# Step 3: Check validity of global info
if not global_language or len(global_language) != 2 or not global_language.islower():
funct_result.addError('The language code in one of the wantlist url isn\'t formated correctly, need to be 2 character long, only lowercase.')
return funct_result
# Since new games can be added to Cardmarket, checking "global_game" is not worth it
# Step 4: Assign info to global variables
global CURR_GAME
CURR_GAME = global_game
return funct_result
def create_credentials_file(credentials, silently=False):
# Normaly we write to file only if we checked the credentials so :
# Step 1: Add a special key to skip checking the credentials in the future
credentials['skip-check'] = 'YES'
# Step 2: Create the JSON file and dump credentials
with open(CREDENTIALS_PATH, 'w+', encoding='utf-8') as json_file:
json.dump(credentials, json_file, indent=4, ensure_ascii=False)
if not silently:
LOG.debug('Credentials file successfully created ("{}")'.format(CREDENTIALS_PATH))
return True
def check_credentials_validity(browser_name, user_agent, credentials, silently=False):
funct_result = FunctResult()
# Step 1: Create a web session
with requests.Session() as session:
# Setting cookies and headers to bypass / skip Cloudflare protection
session.cookies = get_cardmarket_cookies(browser_name)
headers = REQUEST_HEADERS.copy()
headers["user-agent"] = str(user_agent)
session.headers.update(headers)
# Step 2: Log-in to Cardmarket
funct_result = cardmarket_log_in(session, credentials, silently=silently)
if not funct_result.isValid():
return funct_result
logout_result = cardmarket_log_out(session, silently=silently)
funct_result.append(logout_result)
# This will update the file and add the 'skip-check' property
create_credentials_file(credentials, silently=True)
funct_result.addResult(credentials)
return funct_result
def get_formatted_browser_name(browser_name):
funct_result = FunctResult()
original_browser_name = browser_name
# Properly format the browser name, than check validity
# First put name in lowercase and remove leading and trailing space(s)
browser_name = browser_name.lower().strip()
# Replace remaining space(s) to underscore(s)
browser_name = browser_name.replace(' ', '_')
# Remove non alphanumarical characters (preserve underscore(s))
browser_name = ''.join(c for c in browser_name if c.isalnum() or c == '_')
# Remove company name if present
if browser_name.startswith('google_'):
browser_name = 'chrome'
elif browser_name.startswith('mozilla_'):
browser_name = 'firefox'
elif browser_name.startswith('microsoft_'):
browser_name = 'edge'
elif browser_name.startswith('apple_'):
browser_name = 'safari'
if browser_name not in COMPATIBLE_BROWSERS:
funct_result.addError("Browser name '{}' not compatible or unrecognized.".format(original_browser_name))
# Set the reformatted browser_name
funct_result.setResult(browser_name)
return funct_result