Fixed bug where duplicate threads would not get responded
This commit is contained in:
		| @@ -13,8 +13,9 @@ from utils_worker.workers import CompressWorker, DownloadWorker, FetchWorker, Up | ||||
|  | ||||
| class ArticleWatcher: | ||||
|     """Wrapper for a newly created article object. Notifies the coordinator upon change/completition""" | ||||
|     def __init__(self, article, **kwargs) -> None: | ||||
|     def __init__(self, article, thread, **kwargs) -> None: | ||||
|         self.article = article | ||||
|         self.thread = thread | ||||
|  | ||||
|         self.completition_notifier = kwargs.get("notifier") | ||||
|  | ||||
| @@ -48,7 +49,7 @@ class ArticleWatcher: | ||||
|         elif completed_action == "download": | ||||
|             self.compress.process(self) | ||||
|         elif completed_action == "compress": # last step | ||||
|             self.completition_notifier(self.article) | ||||
|             self.completition_notifier(self.article, self.thread) | ||||
|             # triggers action in Coordinator | ||||
|         elif completed_action == "upload": | ||||
|             # this case occurs when upload was faster than compression | ||||
| @@ -122,16 +123,18 @@ class Coordinator(Thread): | ||||
|     def incoming_request(self, message): | ||||
|         """This method is passed onto the slack worker. It gets triggered when a new message is received.""" | ||||
|         url = message.urls[0] # ignore all the other ones | ||||
|         a, is_new = models.ArticleDownload.get_or_create(article_url=url) | ||||
|         message.thread.article = a | ||||
|         message.thread.save() | ||||
|         article, is_new = models.ArticleDownload.get_or_create(article_url=url) | ||||
|         thread = message.thread | ||||
|         thread.article = article | ||||
|         thread.save() | ||||
|         self.kwargs.update({"notifier" : self.article_complete_notifier}) | ||||
|  | ||||
|         if is_new or (a.file_name == "" and a.verified == 0): | ||||
|         if is_new or (article.file_name == "" and article.verified == 0): | ||||
|             # check for models that were created but were abandonned. This means they have missing information, most importantly no associated file | ||||
|             # this overwrites previously set information, but that should not be too important | ||||
|             ArticleWatcher( | ||||
|                 a, | ||||
|                 article, | ||||
|                 thread, | ||||
|                 **self.kwargs    | ||||
|             ) | ||||
|  | ||||
| @@ -140,7 +143,8 @@ class Coordinator(Thread): | ||||
|             # the watcher orchestrates the procedure and notifies upon completition | ||||
|             # the watcher will notify once it is sufficiently populated | ||||
|         else: # manually trigger notification immediatly | ||||
|             self.article_complete_notifier(a) | ||||
|             logger.info(f"Found existing article {article}. Now sending") | ||||
|             self.article_complete_notifier(article, thread) | ||||
|  | ||||
|  | ||||
|  | ||||
| @@ -152,8 +156,8 @@ class Coordinator(Thread): | ||||
|             notifier = lambda article: print(f"Completed manual actions for {article}") | ||||
|             ArticleWatcher(article, workers_manual = workers, notifier = notifier) | ||||
|  | ||||
|     def article_complete_notifier(self, article): | ||||
|         self.worker_slack.bot_worker.respond_channel_message(article) | ||||
|     def article_complete_notifier(self, article, thread): | ||||
|         self.worker_slack.bot_worker.respond_channel_message(thread) | ||||
|         self.worker_mail.send(article) | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -3,8 +3,8 @@ import configuration | ||||
| import requests | ||||
| import os | ||||
| import time | ||||
| import asyncio | ||||
| import sys | ||||
| from threading import Thread | ||||
| from slack_sdk.errors import SlackApiError | ||||
|  | ||||
| logger = logging.getLogger(__name__) | ||||
| @@ -18,26 +18,29 @@ def init(client) -> None: | ||||
|     global slack_client | ||||
|     slack_client = client | ||||
|  | ||||
|     # config["archive_id"] = channel_id | ||||
|     global LATEST_RECORDED_REACTION | ||||
|     try: | ||||
|         LATEST_RECORDED_REACTION = models.Reaction.select(models.Reaction.id).order_by("id")[-1] | ||||
|     except IndexError: #query is actually empty, we have never fetched any messages until now | ||||
|         LATEST_RECORDED_REACTION = 0     | ||||
|     # fetch all te messages we could have possibly missed | ||||
|      | ||||
|     # fetch all te messages we could have possibly missed | ||||
|     logger.info("Querying missed messages, threads and reactions. This can take some time.") | ||||
|     fetch_missed_channel_messages() | ||||
|     if "nofetch" in sys.argv: | ||||
|         logger.info("Omitted update of reactions and thread messages because of argument 'nofetch'.") | ||||
|     fetch_missed_channel_messages() # not threaded | ||||
|     t = Thread(target = fetch_missed_channel_reactions) # threaded, runs in background (usually takes a long time) | ||||
|     t.start() | ||||
|  | ||||
|     if "reducedfetch" in sys.argv: | ||||
|         logger.warning("Only fetching empty threads for bot messages because of argument 'reducedfetch'") | ||||
|         fetch_missed_thread_messages(reduced=True) | ||||
|     else:    # perform these two asyncronously | ||||
|         async def run_async(): | ||||
|             await asyncio.gather(fetch_missed_channel_reactions(), fetch_missed_thread_messages()) | ||||
|         asyncio.run(run_async()) | ||||
|         fetch_missed_thread_messages() | ||||
|      | ||||
|  | ||||
|  | ||||
| def get_past_messages(): | ||||
| def get_unhandled_messages(): | ||||
|     """Gets all messages that have not yet been handled, be it by mistake or by downtime | ||||
|     As the message handler mkaes no distinction between channel messages and thread messages, | ||||
|     As the message handler makes no distinction between channel messages and thread messages, | ||||
|     we don't have to worry about them here. | ||||
|     """ | ||||
|  | ||||
| @@ -51,10 +54,11 @@ def get_past_messages(): | ||||
|     logger.info(f"Set {len(threaded_objects)} thread-messages as not yet handled.") | ||||
|  | ||||
|  | ||||
|     channel_objects = [t.initiator_message for t in models.Thread.select() if t.message_count == 1 and not t.is_fully_processed] | ||||
|     channel_objects = [t.initiator_message for t in models.Thread.select() if (t.message_count == 1 and not t.is_fully_processed)] | ||||
|     logger.info(f"Set {len(channel_objects)} channel-messages as not yet handled.") | ||||
|      | ||||
|     reaction_objects = list(models.Reaction.select().where(models.Reaction.id > LATEST_RECORDED_REACTION)) | ||||
|     logger.info(f"Set {len(reaction_objects)} reactions as not yet handled.") | ||||
|     # the ones newer than the last before the fetch | ||||
|      | ||||
|     all_messages = channel_objects + threaded_objects | ||||
| @@ -108,11 +112,17 @@ def fetch_missed_channel_messages(): | ||||
|     logger.info(f"Fetched {new_fetches} new channel messages.") | ||||
|  | ||||
|  | ||||
| async def fetch_missed_thread_messages(): | ||||
| def fetch_missed_thread_messages(reduced=False): | ||||
|     """After having gotten all base-threads, we need to fetch all their replies"""         | ||||
|     # I don't know of a better way: we need to fetch this for each and every thread (except if it is marked as permanently solved) | ||||
|     logger.info("Starting async fetch of thread messages...") | ||||
|     threads = [t for t in models.Thread.select() if not t.is_fully_processed] | ||||
|     logger.info("Starting fetch of thread messages...") | ||||
|     if reduced: | ||||
|         threads = [t for t in models.Thread.select() if (t.message_count == 1 and not t.is_fully_processed)] | ||||
|         # this only fetches completely empty threads, which might be because the bot-message was not yet saved to the db. | ||||
|         # once we got all the bot-messages the remaining empty threads will be the ones we need to process. | ||||
|     else: | ||||
|         threads = [t for t in models.Thread.select() if not t.is_fully_processed] | ||||
|     logger.info(f"Fetching history for {len(threads)} empty threads") | ||||
|     new_messages = [] | ||||
|     for i,t in enumerate(threads): | ||||
|         try: | ||||
| @@ -123,7 +133,7 @@ async def fetch_missed_thread_messages(): | ||||
|             )["messages"] | ||||
|         except SlackApiError: | ||||
|             logger.error("Hit rate limit while querying threaded messages, retrying in {}s ({}/{} queries elapsed)".format(config["api_wait_time"], i, len(threads))) | ||||
|             await asyncio.sleep(int(config["api_wait_time"])) | ||||
|             time.sleep(int(config["api_wait_time"])) | ||||
|             messages = slack_client.conversations_replies( | ||||
|                 channel = config["archive_id"], | ||||
|                 ts = t.slack_ts, | ||||
| @@ -140,8 +150,8 @@ async def fetch_missed_thread_messages(): | ||||
|     logger.info("Fetched {} new threaded messages.".format(len(new_messages))) | ||||
|  | ||||
|  | ||||
| async def fetch_missed_channel_reactions(): | ||||
|     logger.info("Starting async fetch of channel reactions...") | ||||
| def fetch_missed_channel_reactions(): | ||||
|     logger.info("Starting background fetch of channel reactions...") | ||||
|     threads = [t for t in models.Thread.select() if not t.is_fully_processed] | ||||
|     for i,t in enumerate(threads): | ||||
|         try: | ||||
| @@ -152,7 +162,7 @@ async def fetch_missed_channel_reactions(): | ||||
|             reactions = query["message"].get("reactions", []) # default = [] | ||||
|         except SlackApiError: # probably a rate_limit: | ||||
|             logger.error("Hit rate limit while querying reactions. retrying in {}s ({}/{} queries elapsed)".format(config["api_wait_time"], i, len(threads))) | ||||
|             await asyncio.sleep(int(config["api_wait_time"])) | ||||
|             time.sleep(int(config["api_wait_time"])) | ||||
|             reactions = query["message"].get("reactions", []) | ||||
|  | ||||
|         for r in reactions: | ||||
|   | ||||
| @@ -1,9 +1,7 @@ | ||||
| from threading import Thread | ||||
| from slack_bolt import App | ||||
| from slack_bolt.adapter.socket_mode import SocketModeHandler | ||||
|  | ||||
| import logging | ||||
| from rich.rule import Rule | ||||
| import configuration | ||||
|  | ||||
| from . import message_helpers | ||||
| @@ -18,12 +16,11 @@ class BotApp(App): | ||||
|     def __init__(self, callback, *args, **kwargs): | ||||
|  | ||||
|         super().__init__(*args, **kwargs) | ||||
|         # models = models | ||||
|         self.callback = callback | ||||
|  | ||||
|     def start(self): | ||||
|         message_helpers.init(self.client) | ||||
|         missed_messages, missed_reactions = message_helpers.get_past_messages() | ||||
|         missed_messages, missed_reactions = message_helpers.get_unhandled_messages() | ||||
|  | ||||
|         [self.handle_incoming_message(m) for m in missed_messages] | ||||
|         [self.handle_incoming_reaction(r) for r in missed_reactions] | ||||
| @@ -122,10 +119,8 @@ class BotApp(App): | ||||
|                 ) | ||||
|  | ||||
|  | ||||
|     def respond_channel_message(self, article, say=message_helpers.say_substitute): | ||||
|         # extra={"markup": True} | ||||
|         # self.logger.info(Rule(url[:min(len(url), 30)])) | ||||
|         thread = article.slack_thread.execute()[0] | ||||
|     def respond_channel_message(self, thread, say=message_helpers.say_substitute): | ||||
|         article = thread.article | ||||
|         answers = article.slack_info | ||||
|         for a in answers: | ||||
|             if a["file_path"]: | ||||
| @@ -149,7 +144,6 @@ class BotApp(App): | ||||
|                     thread_ts=thread.slack_ts | ||||
|                 ) | ||||
|                 status = True | ||||
|         # self.logger.info(Rule(f"Fully handled (success={status})")) | ||||
|          | ||||
|  | ||||
|     def startup_status(self): | ||||
|   | ||||
| @@ -211,6 +211,7 @@ class Thread(ChatBaseModel): | ||||
|  | ||||
|     @property | ||||
|     def message_count(self): | ||||
|         # logger.warning("message_count was called") | ||||
|         return self.messages.count() | ||||
|  | ||||
|     @property | ||||
|   | ||||
| @@ -136,12 +136,11 @@ class PDFDownloader: | ||||
|             hrefs = [e.get_attribute("href") for e in self.driver.find_elements_by_xpath("//a[@href]")] | ||||
|         except: | ||||
|             hrefs = [] | ||||
|         old = hrefs | ||||
|         len_old = len(hrefs) | ||||
|         hrefs = [h for h in hrefs \ | ||||
|             if not sum([(domain in h) for domain in blacklisted]) # sum([True, False, False, False]) == 1 (esp. not 0) | ||||
|             ] # filter a tiny bit at least | ||||
|         diff = set(old) ^ set(hrefs) | ||||
|         self.logger.info(f"Removed {len(diff)} hrefs: {diff} (before:{len(old)}, after: {len(hrefs)})") | ||||
|         self.logger.info(f"Hrefs result (before:{len_old}, after: {len(hrefs)})") | ||||
|         return hrefs | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -15,6 +15,7 @@ class NewspaperDummy(): | ||||
|     title = "Error while running fetch" | ||||
|     summary = "Error while running fetch" | ||||
|     text = "Error while running fetch" | ||||
|     meta_lang = "" | ||||
|     authors = [] | ||||
|     keywords = [] | ||||
|  | ||||
| @@ -23,6 +24,7 @@ def get_description(article_object): | ||||
|     url = article_object.article_url | ||||
|     website = urlparse(url).netloc | ||||
|     article_object.source_name = website | ||||
|  | ||||
|     try: | ||||
|         pub_date = datetime.datetime.strptime(find_date(url), '%Y-%d-%M') | ||||
|     except: # other file types | ||||
| @@ -50,6 +52,11 @@ def get_description(article_object): | ||||
|     else: | ||||
|         summary = fallback.summary | ||||
|  | ||||
|     try: | ||||
|         print(f"lang: {news_article.meta_lang}") | ||||
|     except: | ||||
|         print("could not access meta_lang") | ||||
|          | ||||
|     if news_article.meta_lang: | ||||
|         lang = news_article.meta_lang | ||||
|     else: | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Remy Moll
					Remy Moll