tubearchivist/tubearchivist/home/tasks.py

272 lines
8.0 KiB
Python
Raw Normal View History

2021-09-05 17:10:14 +00:00
"""
Functionality:
- initiate celery app
- collect tasks
- user config changes won't get applied here
because tasks are initiated at application start
2021-09-05 17:10:14 +00:00
"""
import json
2021-09-05 17:10:14 +00:00
import os
from celery import Celery, shared_task
2022-01-22 15:13:37 +00:00
from home.src.download.queue import PendingList
from home.src.download.subscriptions import (
2023-03-02 05:30:48 +00:00
SubscriptionHandler,
2023-03-01 12:12:27 +00:00
SubscriptionScanner,
)
2022-08-10 14:03:54 +00:00
from home.src.download.thumbnails import ThumbFilesystem, ThumbValidator
2022-01-22 15:13:37 +00:00
from home.src.download.yt_dlp_handler import VideoDownloader
from home.src.es.backup import ElasticBackup
from home.src.es.index_setup import ElasitIndexWrap
2022-01-22 15:13:37 +00:00
from home.src.index.channel import YoutubeChannel
2022-12-11 05:03:21 +00:00
from home.src.index.filesystem import ImportFolderScanner, scan_filesystem
from home.src.index.reindex import Reindex, ReindexManual, ReindexOutdated
from home.src.index.video_constants import VideoTypeEnum
2022-12-21 07:24:24 +00:00
from home.src.ta.config import AppConfig, ReleaseVersion, ScheduleBuilder
from home.src.ta.helper import clear_dl_cache
2022-01-22 15:13:37 +00:00
from home.src.ta.ta_redis import RedisArchivist, RedisQueue
2023-03-01 12:12:27 +00:00
from home.src.ta.task_manager import TaskManager
2021-09-05 17:10:14 +00:00
CONFIG = AppConfig().config
2021-09-30 11:03:23 +00:00
REDIS_HOST = os.environ.get("REDIS_HOST")
REDIS_PORT = os.environ.get("REDIS_PORT") or 6379
2021-09-05 17:10:14 +00:00
2021-12-02 08:54:29 +00:00
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "config.settings")
2023-03-01 08:31:17 +00:00
app = Celery(
"tasks",
broker=f"redis://{REDIS_HOST}:{REDIS_PORT}",
backend=f"redis://{REDIS_HOST}:{REDIS_PORT}",
result_extended=True,
)
2021-10-27 11:07:35 +00:00
app.config_from_object("django.conf:settings", namespace="ta:")
2021-09-05 17:10:14 +00:00
app.autodiscover_tasks()
app.conf.timezone = os.environ.get("TZ") or "UTC"
2021-09-05 17:10:14 +00:00
2023-03-01 12:12:27 +00:00
@shared_task(name="update_subscribed", bind=True)
def update_subscribed(self):
2021-09-21 09:25:22 +00:00
"""look for missing videos and add to pending"""
2023-03-01 12:12:27 +00:00
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] rescan already running")
message = {
"status": "message:rescan",
"level": "error",
"title": "Rescanning channels and playlists.",
"message": "Rescan already in progress.",
}
RedisArchivist().set_message("message:rescan", message, expire=True)
return
manager.init(self)
SubscriptionScanner().scan()
@shared_task(name="download_pending", bind=True)
def download_pending(self):
2021-09-21 09:25:22 +00:00
"""download latest pending videos"""
2023-03-01 12:12:27 +00:00
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] download queue already running")
return
manager.init(self)
downloader = VideoDownloader()
downloader.add_pending()
downloader.run_queue()
2021-09-05 17:10:14 +00:00
2023-03-01 08:31:17 +00:00
@shared_task(name="download_single")
def download_single(pending_video):
2021-09-21 09:25:22 +00:00
"""start download single video now"""
queue = RedisQueue(queue_name="dl_queue")
to_add = {
"youtube_id": pending_video["youtube_id"],
"vid_type": pending_video.get("vid_type", VideoTypeEnum.VIDEOS.value),
}
queue.add_priority(json.dumps(to_add))
print(f"Added to queue with priority: {to_add}")
# start queue if needed
have_lock = False
my_lock = RedisArchivist().get_lock("downloading")
try:
have_lock = my_lock.acquire(blocking=False)
if have_lock:
key = "message:download"
mess_dict = {
"status": key,
"level": "info",
"title": "Download single video",
"message": "processing",
}
RedisArchivist().set_message(key, mess_dict, expire=True)
VideoDownloader().run_queue()
else:
print("Download queue already running.")
finally:
# release if only single run
if have_lock and not queue.get_next():
my_lock.release()
2021-09-05 17:10:14 +00:00
2023-03-01 08:31:17 +00:00
@shared_task(name="extract_download")
2021-09-05 17:10:14 +00:00
def extrac_dl(youtube_ids):
2021-09-21 09:25:22 +00:00
"""parse list passed and add to pending"""
2022-03-18 11:27:25 +00:00
pending_handler = PendingList(youtube_ids=youtube_ids)
pending_handler.parse_url_list()
pending_handler.add_to_pending()
@shared_task(bind=True, name="check_reindex")
def check_reindex(self, data=False, extract_videos=False):
2021-09-21 09:25:22 +00:00
"""run the reindex main command"""
if data:
# started from frontend through API
print(f"[task][{self.name}] reindex {data}")
ReindexManual(extract_videos=extract_videos).extract_data(data)
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] reindex queue is already running")
return
manager.init(self)
if not data:
# started from scheduler
print(f"[task][{self.name}] reindex outdated documents")
ReindexOutdated().add_outdated()
Reindex().reindex_all()
2021-09-13 15:17:36 +00:00
@shared_task(bind=True, name="manual_import")
def run_manual_import(self):
2021-09-21 09:25:22 +00:00
"""called from settings page, to go through import folder"""
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] manual import is already running")
return
manager.init(self)
ImportFolderScanner().scan()
2021-09-18 13:02:54 +00:00
@shared_task(bind=True, name="run_backup")
def run_backup(self, reason="auto"):
2021-09-21 09:25:22 +00:00
"""called from settings page, dump backup to zip file"""
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] backup is already running")
return
2022-08-12 05:03:09 +00:00
manager.init(self)
ElasticBackup(reason=reason).backup_all_indexes()
@shared_task(bind=True, name="restore_backup")
def run_restore_backup(self, filename):
2021-09-21 09:25:22 +00:00
"""called from settings page, dump backup to zip file"""
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] restore is already running")
return
manager.init(self)
ElasitIndexWrap().reset()
ElasticBackup().restore(filename)
2021-09-21 09:25:22 +00:00
print("index restore finished")
2021-09-24 16:37:26 +00:00
2023-03-01 08:31:17 +00:00
@shared_task(name="kill_download")
2021-09-24 16:37:26 +00:00
def kill_dl(task_id):
"""kill download worker task by ID"""
2021-10-11 08:26:31 +00:00
if task_id:
app.control.revoke(task_id, terminate=True)
_ = RedisArchivist().del_message("dl_queue_id")
RedisQueue(queue_name="dl_queue").clear()
2021-09-24 16:37:26 +00:00
_ = clear_dl_cache(CONFIG)
2021-09-24 16:37:26 +00:00
# notify
mess_dict = {
"status": "message:download",
2021-09-24 16:37:26 +00:00
"level": "error",
"title": "Canceling download process",
"message": "Canceling download queue now.",
2021-09-24 16:37:26 +00:00
}
RedisArchivist().set_message("message:download", mess_dict, expire=True)
@shared_task(bind=True, name="rescan_filesystem")
def rescan_filesystem(self):
2021-10-08 08:10:44 +00:00
"""check the media folder for mismatches"""
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] filesystem rescan already running")
return
manager.init(self)
scan_filesystem()
2022-08-10 14:03:54 +00:00
ThumbValidator().download_missing()
@shared_task(bind=True, name="thumbnail_check")
def thumbnail_check(self):
2021-12-02 08:54:29 +00:00
"""validate thumbnails"""
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] thumbnail check is already running")
return
manager.init(self)
2022-08-10 14:03:54 +00:00
ThumbValidator().download_missing()
2021-12-02 08:54:29 +00:00
@shared_task(bind=True, name="resync_thumbs")
def re_sync_thumbs(self):
"""sync thumbnails to mediafiles"""
manager = TaskManager()
if manager.is_pending(self):
print(f"[task][{self.name}] thumb re-embed is already running")
return
manager.init(self)
2022-08-10 14:03:54 +00:00
ThumbFilesystem().sync()
2023-03-01 08:31:17 +00:00
@shared_task(name="subscribe_to")
def subscribe_to(url_str):
"""take a list of urls to subscribe to"""
2023-03-02 05:30:48 +00:00
SubscriptionHandler(url_str).subscribe()
2023-03-01 08:31:17 +00:00
@shared_task(name="index_playlists")
def index_channel_playlists(channel_id):
"""add all playlists of channel to index"""
2022-01-22 10:48:54 +00:00
channel = YoutubeChannel(channel_id)
# notify
key = "message:playlistscan"
mess_dict = {
"status": key,
"level": "info",
"title": "Looking for playlists",
2023-01-16 02:22:47 +00:00
"message": f"{channel_id}: Channel scan in progress",
}
RedisArchivist().set_message(key, mess_dict, expire=True)
channel.index_channel_playlists()
2021-12-02 08:54:29 +00:00
2022-12-21 07:24:24 +00:00
@shared_task(name="version_check")
def version_check():
"""check for new updates"""
ReleaseVersion().check()
# start schedule here
app.conf.beat_schedule = ScheduleBuilder().build_schedule()