mirror of
https://github.com/tubearchivist/tubearchivist-frontend.git
synced 2024-11-22 20:00:15 +00:00
bulk import subtitle lines into es
This commit is contained in:
parent
9f652802ae
commit
a2cae51f48
@ -4,11 +4,13 @@ functionality:
|
|||||||
- index and update in es
|
- index and update in es
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
import json
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
|
from home.src.es.connect import ElasticWrap
|
||||||
from home.src.index import channel as ta_channel
|
from home.src.index import channel as ta_channel
|
||||||
from home.src.index.generic import YouTubeItem
|
from home.src.index.generic import YouTubeItem
|
||||||
from home.src.ta.helper import DurationConverter, clean_string
|
from home.src.ta.helper import DurationConverter, clean_string
|
||||||
@ -110,20 +112,31 @@ class YoutubeSubtitle:
|
|||||||
videos_base = self.video.config["application"]["videos"]
|
videos_base = self.video.config["application"]["videos"]
|
||||||
for subtitle in relevant_subtitles:
|
for subtitle in relevant_subtitles:
|
||||||
dest_path = os.path.join(videos_base, subtitle["media_url"])
|
dest_path = os.path.join(videos_base, subtitle["media_url"])
|
||||||
|
source = subtitle["media_url"]
|
||||||
response = requests.get(subtitle["url"])
|
response = requests.get(subtitle["url"])
|
||||||
if subtitle["source"] == "auto":
|
if not response.ok:
|
||||||
parser = SubtitleParser(response.text)
|
print(f"{self.video.youtube_id}: failed to download subtitle")
|
||||||
|
continue
|
||||||
|
|
||||||
|
parser = SubtitleParser(response.text, subtitle.get("lang"))
|
||||||
parser.process()
|
parser.process()
|
||||||
subtitle_str_clean = parser.get_subtitle_str()
|
subtitle_str = parser.get_subtitle_str()
|
||||||
else:
|
self._write_subtitle_file(dest_path, subtitle_str)
|
||||||
subtitle_str_clean = response.text
|
query_str = parser.create_bulk_import(self.video, source)
|
||||||
if response.ok:
|
self._index_subtitle(query_str)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _write_subtitle_file(dest_path, subtitle_str):
|
||||||
|
"""write subtitle file to disk"""
|
||||||
# create folder here for first video of channel
|
# create folder here for first video of channel
|
||||||
os.makedirs(os.path.split(dest_path)[0], exist_ok=True)
|
os.makedirs(os.path.split(dest_path)[0], exist_ok=True)
|
||||||
with open(dest_path, "w", encoding="utf-8") as subfile:
|
with open(dest_path, "w", encoding="utf-8") as subfile:
|
||||||
subfile.write(subtitle_str_clean)
|
subfile.write(subtitle_str)
|
||||||
else:
|
|
||||||
print(f"{self.video.youtube_id}: failed to download subtitle")
|
@staticmethod
|
||||||
|
def _index_subtitle(query_str):
|
||||||
|
"""send subtitle to es for indexing"""
|
||||||
|
_, _ = ElasticWrap("_bulk").post(data=query_str, ndjson=True)
|
||||||
|
|
||||||
|
|
||||||
class SubtitleParser:
|
class SubtitleParser:
|
||||||
@ -133,8 +146,9 @@ class SubtitleParser:
|
|||||||
stamp_reg = r"<([0-9]{2}:?){3}\.[0-9]{3}>"
|
stamp_reg = r"<([0-9]{2}:?){3}\.[0-9]{3}>"
|
||||||
tag_reg = r"</?c>"
|
tag_reg = r"</?c>"
|
||||||
|
|
||||||
def __init__(self, subtitle_str):
|
def __init__(self, subtitle_str, lang):
|
||||||
self.subtitle_str = subtitle_str
|
self.subtitle_str = subtitle_str
|
||||||
|
self.lang = lang
|
||||||
self.header = False
|
self.header = False
|
||||||
self.parsed_cue_list = False
|
self.parsed_cue_list = False
|
||||||
self.all_text_lines = False
|
self.all_text_lines = False
|
||||||
@ -209,6 +223,42 @@ class SubtitleParser:
|
|||||||
|
|
||||||
return new_subtitle_str
|
return new_subtitle_str
|
||||||
|
|
||||||
|
def create_bulk_import(self, video, source):
|
||||||
|
"""process matched for es import"""
|
||||||
|
bulk_list = []
|
||||||
|
channel = video.json_data.get("channel")
|
||||||
|
|
||||||
|
document = {
|
||||||
|
"youtube_id": video.youtube_id,
|
||||||
|
"title": video.json_data.get("title"),
|
||||||
|
"subtitle_channel": channel.get("channel_name"),
|
||||||
|
"subtitle_channel_id": channel.get("channel_id"),
|
||||||
|
"subtitle_last_refresh": int(datetime.now().strftime("%s")),
|
||||||
|
"subtitle_lang": self.lang,
|
||||||
|
"subtitle_source": source,
|
||||||
|
}
|
||||||
|
|
||||||
|
for match in self.matched:
|
||||||
|
match_id = match.get("id")
|
||||||
|
document_id = f"{video.youtube_id}-{self.lang}-{match_id}"
|
||||||
|
action = {"index": {"_index": "ta_subtitle", "_id": document_id}}
|
||||||
|
document.update(
|
||||||
|
{
|
||||||
|
"subtitle_fragment_id": document_id,
|
||||||
|
"subtitle_start": match.get("start"),
|
||||||
|
"subtitle_end": match.get("end"),
|
||||||
|
"subtitle_index": match_id,
|
||||||
|
"subtitle_line": " ".join(match.get("lines")),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
bulk_list.append(json.dumps(action))
|
||||||
|
bulk_list.append(json.dumps(document))
|
||||||
|
|
||||||
|
bulk_list.append("\n")
|
||||||
|
query_str = "\n".join(bulk_list)
|
||||||
|
|
||||||
|
return query_str
|
||||||
|
|
||||||
|
|
||||||
class YoutubeVideo(YouTubeItem, YoutubeSubtitle):
|
class YoutubeVideo(YouTubeItem, YoutubeSubtitle):
|
||||||
"""represents a single youtube video"""
|
"""represents a single youtube video"""
|
||||||
|
Loading…
Reference in New Issue
Block a user