-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathmain.py
More file actions
611 lines (520 loc) · 24.8 KB
/
main.py
File metadata and controls
611 lines (520 loc) · 24.8 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
import asyncio
import aiohttp
import csv
import os
import html
import time
import re
from datetime import datetime, timedelta, timezone
from bs4 import BeautifulSoup
from aiogram import Bot
from aiogram.enums import ParseMode
from aiogram.client.default import DefaultBotProperties
from selenium import webdriver
from selenium.webdriver.common.by import By
from selenium.common.exceptions import NoSuchElementException
from selenium.webdriver.chrome.options import Options
from selenium.webdriver.chrome.service import Service
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions as EC
# Конфигурация (остается без изменений)
REVIEW_ID_FILE = "review_ids.txt"
SENT_IDS_FILE = "sent_ids.txt"
CSV_FILE = "reviews.csv"
YANDEX_REVIEW_ID_FILE = "yandex_review_ids.txt"
YANDEX_SENT_IDS_FILE = "yandex_sent_ids.txt"
YANDEX_CSV_FILE = "yandex_reviews.csv"
API_KEY = "6e7e1929-4ea9-4a5d-8c05-d601860389bd"
BRANCH_URLS = ["https://2gis.ru/vladivostok/firm/70000001102820158"]
YANDEX_IDS = [163055877176]
API_TOKEN = ""
CHANNEL_ID = -1002998637584
# Инициализация бота
bot = Bot(
token=API_TOKEN,
default=DefaultBotProperties(parse_mode=ParseMode.HTML)
)
# Классы для парсинга Яндекс (с изменениями для уменьшения ошибок)
class ParserHelper:
@staticmethod
def get_count_star(stars_elements):
count = 0
for star in stars_elements:
classes = star.get_attribute('class')
if '._full' in classes or '_full' in classes:
count += 1
elif '._half' in classes or '_half' in classes:
count += 0.5
return count
@staticmethod
def form_date(date_str):
try:
dt = datetime.strptime(date_str, '%Y-%m-%dT%H:%M:%S.%fZ')
return dt.timestamp()
except:
return time.time()
@staticmethod
def format_rating(rating_elements):
if rating_elements:
try:
return float(rating_elements[0].text.replace(',', '.'))
except:
return 0.0
return 0.0
@staticmethod
def list_to_num(text):
try:
return int(re.sub(r'\D', '', text))
except:
return 0
class Parser:
def __init__(self, driver):
self.driver = driver
self.wait = WebDriverWait(driver, 10)
def __scroll_to_bottom(self, elem):
self.driver.execute_script(
"arguments[0].scrollIntoView();",
elem
)
time.sleep(1)
new_elems = self.driver.find_elements(By.CLASS_NAME, "business-reviews-card-view__review")
if new_elems and new_elems[-1] != elem:
self.__scroll_to_bottom(new_elems[-1])
def __get_data_item(self, elem):
try:
name = elem.find_element(By.XPATH, ".//span[@itemprop='name']").text
except NoSuchElementException:
name = None
try:
icon_style = elem.find_element(By.XPATH, ".//div[contains(@class, 'user-icon-view__icon')]").get_attribute('style')
icon_href = re.findall(r'url\("?(.*?)"?\)', icon_style)[0] if icon_style else None
except NoSuchElementException:
icon_href = None
try:
date = elem.find_element(By.XPATH, ".//meta[@itemprop='datePublished']").get_attribute('content')
except NoSuchElementException:
date = None
text = None
try:
review_body = elem.find_element(By.CLASS_NAME, "business-review-view__body")
text_elements = review_body.find_elements(By.CLASS_NAME, "spoiler-view__text-container")
if text_elements:
text = text_elements[0].text
if not text:
text_elements = review_body.find_elements(By.XPATH, ".//span[contains(@class, 'business-review-view__body-text')]")
if text_elements:
text = text_elements[0].text
if not text:
text = review_body.text
except NoSuchElementException:
text = None
try:
stars = elem.find_elements(By.XPATH, ".//div[contains(@class, 'business-rating-badge-view__stars')]/span")
stars_count = ParserHelper.get_count_star(stars)
except NoSuchElementException:
stars_count = 0
answer = None
try:
# Пытаемся найти кнопку раскрытия ответа компании
answer_btn = elem.find_elements(By.CLASS_NAME, "business-review-view__comment-expand")
if answer_btn:
self.driver.execute_script("arguments[0].click()", answer_btn[0])
time.sleep(0.5)
answer = elem.find_element(By.CLASS_NAME, "business-review-comment-content__bubble").text
except NoSuchElementException:
answer = None
return {
'name': name,
'icon_href': icon_href,
'date': ParserHelper.form_date(date) if date else None,
'text': text,
'stars': stars_count,
'answer': answer
}
def __get_data_company(self):
try:
name = self.driver.find_element(By.XPATH, ".//h1[contains(@class, 'orgpage-header-view__header')]").text
except NoSuchElementException:
name = None
try:
rating_block = self.driver.find_element(By.XPATH, ".//div[contains(@class, 'business-summary-rating-badge-view__rating-and-stars')]")
rating_text = rating_block.find_elements(By.XPATH, ".//span[contains(@class, 'business-summary-rating-badge-view__rating-text')]")
rating = ParserHelper.format_rating(rating_text)
count_rating = rating_block.find_element(By.XPATH, ".//span[contains(@class, 'business-rating-amount-view')]").text
count_rating = ParserHelper.list_to_num(count_rating)
stars = rating_block.find_elements(By.XPATH, ".//div[contains(@class, 'business-rating-badge-view__stars')]/span")
stars_count = ParserHelper.get_count_star(stars)
except NoSuchElementException:
rating, count_rating, stars_count = 0, 0, 0
return {
'name': name,
'rating': rating,
'count_rating': count_rating,
'stars': stars_count
}
def __get_data_reviews(self):
reviews = []
try:
self.wait.until(EC.presence_of_element_located((By.CLASS_NAME, "business-reviews-card-view__review")))
elements = self.driver.find_elements(By.CLASS_NAME, "business-reviews-card-view__review")
if elements:
self.__scroll_to_bottom(elements[-1])
elements = self.driver.find_elements(By.CLASS_NAME, "business-reviews-card-view__review")
for elem in elements:
reviews.append(self.__get_data_item(elem))
except NoSuchElementException:
pass
return reviews
def __is_valid_page(self):
try:
return bool(self.driver.find_element(By.XPATH, ".//h1[contains(@class, 'orgpage-header-view__header')]"))
except NoSuchElementException:
return False
def parse_all_data(self):
if not self.__is_valid_page():
return {'error': 'Страница не найдена'}
return {
'company_info': self.__get_data_company(),
'company_reviews': self.__get_data_reviews()
}
def parse_reviews(self):
if not self.__is_valid_page():
return {'error': 'Страница не найдена'}
return {'company_reviews': self.__get_data_reviews()}
def parse_company_info(self):
if not self.__is_valid_page():
return {'error': 'Страница не найдена'}
return {'company_info': self.__get_data_company()}
class YandexParser:
def __init__(self, id_yandex: int):
self.id_yandex = id_yandex
def __open_page(self):
url = f'https://yandex.ru/maps/org/{self.id_yandex}/reviews/'
chrome_options = Options()
chrome_options.add_argument('--no-sandbox')
chrome_options.add_argument('--disable-dev-shm-usage')
chrome_options.add_argument('--disable-gpu')
chrome_options.add_argument('--disable-blink-features=AutomationControlled')
chrome_options.add_experimental_option('excludeSwitches', ['enable-automation'])
chrome_options.add_experimental_option('useAutomationExtension', False)
chrome_options.add_argument('--disable-extensions')
chrome_options.add_argument('--disable-default-apps')
chrome_options.add_argument('--disable-component-extensions-with-background-pages')
chrome_options.add_argument('--disable-background-timer-throttling')
chrome_options.add_argument('--disable-backgrounding-occluded-windows')
chrome_options.add_argument('--disable-renderer-backgrounding')
chrome_options.add_argument('--headless')
chrome_options.add_argument('--log-level=3') # Уменьшаем уровень логов
chrome_options.add_argument('--disable-logging') # Отключаем логи
# Добавляем опции для отключения сообщений DevTools
chrome_options.add_experimental_option("excludeSwitches", ["enable-logging"])
chrome_options.add_argument("--disable-dev-tools") # Отключаем DevTools
# Отключаем различные службы и функции
prefs = {
'profile.default_content_setting_values.notifications': 2,
'profile.default_content_settings.popups': 0,
'profile.managed_default_content_settings.durable_storage': 2,
'credentials_enable_service': False,
'password_manager_enabled': False,
'profile.default_content_setting_values.geolocation': 2,
}
chrome_options.add_experimental_option('prefs', prefs)
try:
service = Service(log_path='NUL')
driver = webdriver.Chrome(service=service, options=chrome_options)
except:
driver = webdriver.Chrome(options=chrome_options)
driver.execute_script("Object.defineProperty(navigator, 'webdriver', {get: () => undefined})")
parser = Parser(driver)
driver.get(url)
return parser
def parse(self, parse_type: str = 'default'):
result = {}
parser = self.__open_page()
time.sleep(5)
try:
if parse_type == 'default':
result = parser.parse_all_data()
elif parse_type == 'company':
result = parser.parse_company_info()
elif parse_type == 'reviews':
result = parser.parse_reviews()
except Exception as e:
print(f"Ошибка при парсинге: {e}")
result = {'error': str(e)}
finally:
parser.driver.quit()
return result
# Общие функции для бота
async def send_message(text: str):
try:
await bot.send_message(CHANNEL_ID, text)
return True
except Exception as e:
if "Flood control exceeded" in str(e):
try:
retry_after = int(str(e).split("retry after ")[1].split(" ")[0])
print(f"Обнаружено ограничение частоты. Ждем {retry_after} секунд...")
await asyncio.sleep(retry_after + 2)
await bot.send_message(CHANNEL_ID, text)
return True
except (IndexError, ValueError):
print(f"Не удалось извлечь время ожидания: {e}")
return False
else:
print(f"Ошибка отправки сообщения: {e}")
return False
def load_existing_ids(filename):
if not os.path.exists(filename):
return set()
with open(filename, "r", encoding="utf-8") as f:
return set(line.strip() for line in f)
def save_ids(filename, new_ids):
with open(filename, "a", encoding="utf-8") as f:
for rid in new_ids:
f.write(rid + "\n")
def save_reviews_to_csv(reviews, csv_file):
write_header = not os.path.exists(csv_file)
# Добавляем поля для ответа компании и даты ответа
fieldnames = ["address", "date", "author", "rating", "text", "answer", "answer_date"] if "yandex" not in csv_file else ["address", "date", "author", "rating", "text", "answer"]
with open(csv_file, "a", encoding="utf-8-sig", newline="") as f:
writer = csv.DictWriter(f, fieldnames=fieldnames, delimiter=";")
if write_header:
writer.writeheader()
writer.writerows(reviews)
def format_datetime(dt_value):
"""Форматирует дату-время из timestamp или строки в ЧЧ:ММ:СС-ДД.ММ.ГГГГ"""
try:
if isinstance(dt_value, (int, float)):
dt = datetime.fromtimestamp(dt_value, timezone.utc)
else:
# Обрабатываем разные форматы дат
if 'T' in dt_value:
if '+' in dt_value:
dt = datetime.fromisoformat(dt_value.replace('Z', '+00:00'))
else:
dt = datetime.fromisoformat(dt_value.replace('Z', ''))
else:
dt = datetime.strptime(dt_value, '%Y-%m-%d %H:%M:%S')
dt = dt.astimezone(timezone(timedelta(hours=7)))
return dt.strftime("%H:%M:%S %d.%m.%Y")
except (ValueError, AttributeError):
return str(dt_value)
def format_review(review, source="2GIS"):
escaped_text = html.escape(review['text'])
if len(escaped_text) > 4000:
escaped_text = escaped_text[:4000] + "..."
review_date = format_datetime(review['date'])
tz = timezone(timedelta(hours=7))
current_time = datetime.now(tz).strftime("%H:%M:%S %d.%m.%Y")
if source == "2GIS":
message = (f"<b>📝 Отзыв с 2ГИС</b>\n\n"
f"<b>🏢 Адрес:</b> {html.escape(review['address'])}\n"
f"<b>📅 Дата отзыва:</b> {review_date}\n"
f"<b>👤 Автор:</b> {html.escape(review['author'])}\n"
f"<b>⭐ Рейтинг:</b> {review['rating']}/5\n\n"
f"<b>💬 Текст:</b>\n{escaped_text}\n\n")
# Добавляем ответ компании, если он есть
if review.get('answer'):
escaped_answer = html.escape(review['answer'])
if len(escaped_answer) > 4000:
escaped_answer = escaped_answer[:4000] + "..."
answer_date = format_datetime(review.get('answer_date', '')) if review.get('answer_date') else "не указана"
message += (f"<b>📣 Ответ компании:</b>\n"
f"{escaped_answer}\n"
f"<b>📅 Дата ответа:</b> {answer_date}\n\n")
#message += f"<i>🕐 Проверено: {current_time}</i>"
return message
else:
# Форматирование для Яндекс с ответом компании
message = (f"<b>📝 Отзыв с Яндекс Карт</b>\n\n"
f"<b>🏢 Организация:</b> {html.escape(review['address'])}\n"
f"<b>📅 Дата отзыва:</b> {review_date}\n"
f"<b>👤 Автор:</b> {html.escape(review['author'])}\n"
f"<b>⭐ Рейтинг:</b> {review['rating']}/5\n\n"
f"<b>💬 Текст:</b>\n{escaped_text}\n\n")
# Добавляем ответ компании, если он есть
if review.get('answer'):
escaped_answer = html.escape(review['answer'])
if len(escaped_answer) > 4000:
escaped_answer = escaped_answer[:4000] + "..."
message += f"<b>📣 Ответ компании:</b>\n{escaped_answer}\n\n"
#message += f"<i>🕐 Проверено: {current_time}</i>"
return message
# Функции для 2GIS
async def get_page_title(session, url):
try:
async with session.get(url, headers={"User-Agent": "Mozilla/5.0"}) as resp:
html_content = await resp.text()
soup = BeautifulSoup(html_content, "html.parser")
return soup.title.string.strip().removesuffix(" — 2ГИС")
except Exception:
return "Ошибка получения адреса"
async def fetch_reviews(session, url, existing_ids):
firm_id = url.split("/")[-1]
address = await get_page_title(session, url)
print(f"Ищем отзывы предприятия: {address}")
new_reviews = []
new_ids = []
limit = 50
offset = 0
while True:
api_url = f"https://public-api.reviews.2gis.com/2.0/branches/{firm_id}/reviews"
params = {
"limit": limit,
"offset": offset,
"is_advertiser": "false",
"fields": "meta.providers,meta.branch_rating,meta.branch_reviews_count,meta.total_count,reviews.hiding_reason,reviews.is_verified,reviews.emojis",
"without_my_first_review": "false",
"rated": "true",
"sort_by": "date_edited",
"key": API_KEY,
"locale": "ru_RU"
}
try:
async with session.get(api_url, params=params) as resp:
data = await resp.json()
reviews = data.get("reviews", [])
if not reviews:
break
for review in reviews:
rid = review["id"]
if rid in existing_ids:
continue
# Добавляем обработку ответа компании
official_answer = review.get("official_answer")
answer_text = ""
answer_date = ""
if official_answer:
answer_text = official_answer.get("text", "")
answer_date = official_answer.get("date_created", "")
new_reviews.append({
"address": address,
"date": review.get("date_edited") or review.get("date_created"),
"author": review["user"]["name"],
"rating": review.get("rating", ""),
"text": review.get("text", "").replace("\n", " ").strip(),
"answer": answer_text,
"answer_date": answer_date
})
new_ids.append(rid)
if len(reviews) < limit:
break
offset += limit
await asyncio.sleep(1)
except Exception as e:
print(f"Ошибка при запросе {url}: {e}")
break
# Сортируем отзывы от старых к новым
if new_reviews:
try:
# Создаем временный список с датами для сортировки
sorted_reviews = sorted(
new_reviews,
key=lambda x: x['date'] if x['date'] else '0'
)
# Сохраняем порядок ID в соответствии с отсортированными отзывами
sorted_ids = []
for review in sorted_reviews:
index = new_reviews.index(review)
sorted_ids.append(new_ids[index])
return sorted_reviews, sorted_ids
except Exception as e:
print(f"Ошибка при сортировке отзывов: {e}")
return new_reviews, new_ids
return new_reviews, new_ids
# Функции для Яндекс
def parse_yandex_reviews(company_id, existing_ids):
parser = YandexParser(company_id)
data = parser.parse('reviews')
if 'error' in data:
print(f"Ошибка парсинга Яндекс: {data['error']}")
return [], []
reviews_data = data.get('company_reviews', [])
new_reviews = []
new_ids = []
for review in reviews_data:
# Создаем уникальный ID на основе содержимого отзыва
review_id = f"yandex_{company_id}_{review['date']}_{review['name']}"
if review_id in existing_ids:
continue
new_reviews.append({
"address": f"Яндекс ID: {company_id}",
"date": review['date'],
"author": review['name'],
"rating": review['stars'],
"text": review['text'] or "Нет текста",
"answer": review.get('answer', '') # Добавляем ответ компании
})
new_ids.append(review_id)
return new_reviews, new_ids
async def check_2gis_reviews():
existing_ids = load_existing_ids(REVIEW_ID_FILE)
sent_ids = load_existing_ids(SENT_IDS_FILE)
all_new_reviews = []
all_new_ids = []
async with aiohttp.ClientSession() as session:
tasks = [fetch_reviews(session, link, existing_ids) for link in BRANCH_URLS if link.strip()]
results = await asyncio.gather(*tasks, return_exceptions=True)
for result in results:
if isinstance(result, Exception):
print(f"Ошибка при сборе отзывов 2GIS: {result}")
continue
reviews, ids = result
all_new_reviews.extend(reviews)
all_new_ids.extend(ids)
if all_new_reviews:
save_reviews_to_csv(all_new_reviews, CSV_FILE)
save_ids(REVIEW_ID_FILE, all_new_ids)
for review in all_new_reviews:
review_text = format_review(review, "2GIS")
success = await send_message(review_text)
if success:
review_id = all_new_ids[all_new_reviews.index(review)]
save_ids(SENT_IDS_FILE, [review_id])
await asyncio.sleep(3)
print(f"2GIS: Найдено и отправлено новых отзывов: {len(all_new_reviews)}")
return True
else:
print("2GIS: Новых отзывов не найдено")
return False
async def check_yandex_reviews():
existing_ids = load_existing_ids(YANDEX_REVIEW_ID_FILE)
sent_ids = load_existing_ids(YANDEX_SENT_IDS_FILE)
all_new_reviews = []
all_new_ids = []
for company_id in YANDEX_IDS:
reviews, ids = await asyncio.to_thread(parse_yandex_reviews, company_id, existing_ids)
all_new_reviews.extend(reviews)
all_new_ids.extend(ids)
if all_new_reviews:
save_reviews_to_csv(all_new_reviews, YANDEX_CSV_FILE)
save_ids(YANDEX_REVIEW_ID_FILE, all_new_ids)
for review in all_new_reviews:
review_text = format_review(review, "Yandex")
success = await send_message(review_text)
if success:
review_id = all_new_ids[all_new_reviews.index(review)]
save_ids(YANDEX_SENT_IDS_FILE, [review_id])
await asyncio.sleep(3)
print(f"Яндекс: Найдено и отправлено новых отзывов: {len(all_new_reviews)}")
return True
else:
print("Яндекс: Новых отзывов не найдено")
return False
async def main():
print("Запуск бота для мониторинга отзывов...")
print("Первоначальная проверка отзывов...")
await check_2gis_reviews()
await check_yandex_reviews()
while True:
print("Ожидание следующей проверки (5 минут)...")
await asyncio.sleep(5 * 60)
print("Начинаем очередную проверку отзывов...")
await check_2gis_reviews()
await check_yandex_reviews()
if __name__ == "__main__":
try:
asyncio.run(main())
except KeyboardInterrupt:
print("Бот остановлен")