2022-09-17 18:12:14 +00:00
|
|
|
import base64
|
2022-08-11 09:14:12 +00:00
|
|
|
import os
|
|
|
|
import re
|
2022-09-17 18:12:14 +00:00
|
|
|
|
2022-08-11 10:06:31 +00:00
|
|
|
from .html2text import html2text
|
2022-08-11 09:14:12 +00:00
|
|
|
|
2022-09-03 10:50:14 +00:00
|
|
|
TOOLTIP_REGEX = r"(\/\/\/(.+)\/\/\/)"
|
|
|
|
contentDir = os.path.join(
|
|
|
|
os.path.dirname(os.path.realpath(__file__)), "..", "..", "discoursio-web", "content"
|
|
|
|
)
|
|
|
|
s3 = "https://discours-io.s3.amazonaws.com/"
|
|
|
|
cdn = "https://assets.discours.io"
|
|
|
|
|
|
|
|
|
|
|
|
def replace_tooltips(body):
|
|
|
|
# change if you prefer regexp
|
|
|
|
newbody = body
|
|
|
|
matches = list(re.finditer(TOOLTIP_REGEX, body, re.IGNORECASE | re.MULTILINE))[1:]
|
|
|
|
for match in matches:
|
|
|
|
newbody = body.replace(
|
|
|
|
match.group(1), '<Tooltip text="' + match.group(2) + '" />'
|
|
|
|
) # NOTE: doesn't work
|
|
|
|
if len(matches) > 0:
|
|
|
|
print("[extract] found %d tooltips" % len(matches))
|
|
|
|
return newbody
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
|
|
|
|
def place_tooltips(body):
|
2022-09-03 10:50:14 +00:00
|
|
|
parts = body.split("&&&")
|
2022-09-04 17:20:38 +00:00
|
|
|
lll = len(parts)
|
2022-09-03 10:50:14 +00:00
|
|
|
newparts = list(parts)
|
|
|
|
placed = False
|
2022-09-04 17:20:38 +00:00
|
|
|
if lll & 1:
|
|
|
|
if lll > 1:
|
2022-09-03 10:50:14 +00:00
|
|
|
i = 1
|
2022-09-04 17:20:38 +00:00
|
|
|
print("[extract] found %d tooltips" % (lll - 1))
|
2022-09-03 10:50:14 +00:00
|
|
|
for part in parts[1:]:
|
|
|
|
if i & 1:
|
|
|
|
placed = True
|
|
|
|
if 'a class="footnote-url" href=' in part:
|
|
|
|
print("[extract] footnote: " + part)
|
|
|
|
fn = 'a class="footnote-url" href="'
|
|
|
|
link = part.split(fn, 1)[1].split('"', 1)[0]
|
|
|
|
extracted_part = (
|
|
|
|
part.split(fn, 1)[0] + " " + part.split("/", 1)[-1]
|
|
|
|
)
|
|
|
|
newparts[i] = (
|
|
|
|
"<Tooltip"
|
|
|
|
+ (' link="' + link + '" ' if link else "")
|
|
|
|
+ ">"
|
|
|
|
+ extracted_part
|
|
|
|
+ "</Tooltip>"
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
newparts[i] = "<Tooltip>%s</Tooltip>" % part
|
|
|
|
# print('[extract] ' + newparts[i])
|
|
|
|
else:
|
|
|
|
# print('[extract] ' + part[:10] + '..')
|
|
|
|
newparts[i] = part
|
|
|
|
i += 1
|
|
|
|
return ("".join(newparts), placed)
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
2022-09-04 17:20:38 +00:00
|
|
|
IMG_REGEX = r"\!\[(.*?)\]\((data\:image\/(png|jpeg|jpg);base64\,((?:[A-Za-z\d+\/]{4})*(?:[A-Za-z\d+\/]{3}="
|
|
|
|
IMG_REGEX += r"|[A-Za-z\d+\/]{2}==)))\)"
|
2022-08-11 09:14:12 +00:00
|
|
|
|
2022-09-03 10:50:14 +00:00
|
|
|
parentDir = "/".join(os.getcwd().split("/")[:-1])
|
|
|
|
public = parentDir + "/discoursio-web/public"
|
2022-08-11 09:14:12 +00:00
|
|
|
cache = {}
|
|
|
|
|
|
|
|
|
2022-09-03 10:50:14 +00:00
|
|
|
def reextract_images(body, oid):
|
|
|
|
# change if you prefer regexp
|
|
|
|
matches = list(re.finditer(IMG_REGEX, body, re.IGNORECASE | re.MULTILINE))[1:]
|
|
|
|
i = 0
|
|
|
|
for match in matches:
|
|
|
|
print("[extract] image " + match.group(1))
|
|
|
|
ext = match.group(3)
|
|
|
|
name = oid + str(i)
|
|
|
|
link = public + "/upload/image-" + name + "." + ext
|
|
|
|
img = match.group(4)
|
|
|
|
title = match.group(1) # NOTE: this is not the title
|
|
|
|
if img not in cache:
|
|
|
|
content = base64.b64decode(img + "==")
|
|
|
|
print(str(len(img)) + " image bytes been written")
|
|
|
|
open("../" + link, "wb").write(content)
|
|
|
|
cache[img] = name
|
|
|
|
i += 1
|
|
|
|
else:
|
|
|
|
print("[extract] image cached " + cache[img])
|
|
|
|
body.replace(
|
|
|
|
str(match), ""
|
|
|
|
) # WARNING: this does not work
|
|
|
|
return body
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
IMAGES = {
|
2022-09-03 10:50:14 +00:00
|
|
|
"data:image/png": "png",
|
|
|
|
"data:image/jpg": "jpg",
|
|
|
|
"data:image/jpeg": "jpg",
|
2022-08-11 09:14:12 +00:00
|
|
|
}
|
|
|
|
|
2022-09-03 10:50:14 +00:00
|
|
|
b64 = ";base64,"
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
def extract_imageparts(bodyparts, prefix):
|
2022-09-03 10:50:14 +00:00
|
|
|
# recursive loop
|
|
|
|
newparts = list(bodyparts)
|
|
|
|
for current in bodyparts:
|
|
|
|
i = bodyparts.index(current)
|
|
|
|
for mime in IMAGES.keys():
|
|
|
|
if mime == current[-len(mime) :] and (i + 1 < len(bodyparts)):
|
|
|
|
print("[extract] " + mime)
|
|
|
|
next = bodyparts[i + 1]
|
|
|
|
ext = IMAGES[mime]
|
|
|
|
b64end = next.index(")")
|
|
|
|
b64encoded = next[:b64end]
|
|
|
|
name = prefix + "-" + str(len(cache))
|
|
|
|
link = "/upload/image-" + name + "." + ext
|
|
|
|
print("[extract] name: " + name)
|
|
|
|
print("[extract] link: " + link)
|
|
|
|
print("[extract] %d bytes" % len(b64encoded))
|
|
|
|
if b64encoded not in cache:
|
|
|
|
try:
|
|
|
|
content = base64.b64decode(b64encoded + "==")
|
|
|
|
open(public + link, "wb").write(content)
|
|
|
|
print(
|
|
|
|
"[extract] "
|
|
|
|
+ str(len(content))
|
|
|
|
+ " image bytes been written"
|
|
|
|
)
|
|
|
|
cache[b64encoded] = name
|
2022-09-04 17:20:38 +00:00
|
|
|
except Exception:
|
2022-09-03 10:50:14 +00:00
|
|
|
raise Exception
|
|
|
|
# raise Exception('[extract] error decoding image %r' %b64encoded)
|
|
|
|
else:
|
|
|
|
print("[extract] cached link " + cache[b64encoded])
|
|
|
|
name = cache[b64encoded]
|
|
|
|
link = cdn + "/upload/image-" + name + "." + ext
|
|
|
|
newparts[i] = (
|
|
|
|
current[: -len(mime)]
|
|
|
|
+ current[-len(mime) :]
|
|
|
|
+ link
|
|
|
|
+ next[-b64end:]
|
|
|
|
)
|
|
|
|
newparts[i + 1] = next[:-b64end]
|
|
|
|
break
|
|
|
|
return (
|
|
|
|
extract_imageparts(
|
2022-09-04 17:20:38 +00:00
|
|
|
newparts[i] + newparts[i + 1] + b64.join(bodyparts[(i + 2) :]), prefix
|
2022-09-03 10:50:14 +00:00
|
|
|
)
|
|
|
|
if len(bodyparts) > (i + 1)
|
|
|
|
else "".join(newparts)
|
|
|
|
)
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
def extract_dataimages(parts, prefix):
|
2022-09-03 10:50:14 +00:00
|
|
|
newparts = list(parts)
|
|
|
|
for part in parts:
|
|
|
|
i = parts.index(part)
|
|
|
|
if part.endswith("]("):
|
|
|
|
[ext, rest] = parts[i + 1].split(b64)
|
|
|
|
name = prefix + "-" + str(len(cache))
|
|
|
|
if ext == "/jpeg":
|
|
|
|
ext = "jpg"
|
|
|
|
else:
|
|
|
|
ext = ext.replace("/", "")
|
|
|
|
link = "/upload/image-" + name + "." + ext
|
|
|
|
print("[extract] filename: " + link)
|
|
|
|
b64end = rest.find(")")
|
|
|
|
if b64end != -1:
|
|
|
|
b64encoded = rest[:b64end]
|
|
|
|
print("[extract] %d text bytes" % len(b64encoded))
|
|
|
|
# write if not cached
|
|
|
|
if b64encoded not in cache:
|
|
|
|
try:
|
|
|
|
content = base64.b64decode(b64encoded + "==")
|
|
|
|
open(public + link, "wb").write(content)
|
|
|
|
print("[extract] " + str(len(content)) + " image bytes")
|
|
|
|
cache[b64encoded] = name
|
2022-09-04 17:20:38 +00:00
|
|
|
except Exception:
|
2022-09-03 10:50:14 +00:00
|
|
|
raise Exception
|
|
|
|
# raise Exception('[extract] error decoding image %r' %b64encoded)
|
|
|
|
else:
|
|
|
|
print("[extract] 0 image bytes, cached for " + cache[b64encoded])
|
|
|
|
name = cache[b64encoded]
|
|
|
|
|
|
|
|
# update link with CDN
|
|
|
|
link = cdn + "/upload/image-" + name + "." + ext
|
|
|
|
|
|
|
|
# patch newparts
|
|
|
|
newparts[i + 1] = link + rest[b64end:]
|
|
|
|
else:
|
|
|
|
raise Exception("cannot find the end of base64 encoded string")
|
|
|
|
else:
|
|
|
|
print("[extract] dataimage skipping part " + str(i))
|
|
|
|
continue
|
|
|
|
return "".join(newparts)
|
|
|
|
|
|
|
|
|
|
|
|
di = "data:image"
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
def extract_md_images(body, oid):
|
2022-09-03 10:50:14 +00:00
|
|
|
newbody = ""
|
|
|
|
body = (
|
|
|
|
body.replace("\n! [](" + di, "\n 
|
|
|
|
.replace("\n[](" + di, "\n
|
|
|
|
.replace(" [](" + di, " 
|
|
|
|
)
|
|
|
|
parts = body.split(di)
|
|
|
|
if len(parts) > 1:
|
|
|
|
newbody = extract_dataimages(parts, oid)
|
|
|
|
else:
|
|
|
|
newbody = body
|
|
|
|
return newbody
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
|
|
|
|
def cleanup(body):
|
2022-09-03 10:50:14 +00:00
|
|
|
newbody = (
|
|
|
|
body.replace("<", "")
|
|
|
|
.replace(">", "")
|
|
|
|
.replace("{", "(")
|
|
|
|
.replace("}", ")")
|
|
|
|
.replace("…", "...")
|
|
|
|
.replace(" __ ", " ")
|
|
|
|
.replace("_ _", " ")
|
|
|
|
.replace("****", "")
|
|
|
|
.replace("\u00a0", " ")
|
|
|
|
.replace("\u02c6", "^")
|
|
|
|
.replace("\u00a0", " ")
|
|
|
|
.replace("\ufeff", "")
|
|
|
|
.replace("\u200b", "")
|
|
|
|
.replace("\u200c", "")
|
|
|
|
) # .replace('\u2212', '-')
|
|
|
|
return newbody
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
def extract_md(body, oid):
|
2022-09-03 10:50:14 +00:00
|
|
|
newbody = body
|
|
|
|
if newbody:
|
|
|
|
newbody = extract_md_images(newbody, oid)
|
|
|
|
if not newbody:
|
|
|
|
raise Exception("extract_images error")
|
|
|
|
newbody = cleanup(newbody)
|
|
|
|
if not newbody:
|
|
|
|
raise Exception("cleanup error")
|
|
|
|
newbody, placed = place_tooltips(newbody)
|
|
|
|
if not newbody:
|
|
|
|
raise Exception("place_tooltips error")
|
|
|
|
if placed:
|
|
|
|
newbody = "import Tooltip from '$/components/Article/Tooltip'\n\n" + newbody
|
|
|
|
return newbody
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
def prepare_md_body(entry):
|
2022-09-03 10:50:14 +00:00
|
|
|
# body modifications
|
|
|
|
body = ""
|
|
|
|
kind = entry.get("type")
|
|
|
|
addon = ""
|
|
|
|
if kind == "Video":
|
|
|
|
addon = ""
|
|
|
|
for m in entry.get("media", []):
|
|
|
|
if "youtubeId" in m:
|
|
|
|
addon += "<VideoPlayer youtubeId='" + m["youtubeId"] + "' />\n"
|
|
|
|
elif "vimeoId" in m:
|
|
|
|
addon += "<VideoPlayer vimeoId='" + m["vimeoId"] + "' />\n"
|
|
|
|
else:
|
|
|
|
print("[extract] media is not supported")
|
|
|
|
print(m)
|
|
|
|
body = "import VideoPlayer from '$/components/Article/VideoPlayer'\n\n" + addon
|
|
|
|
|
|
|
|
elif kind == "Music":
|
|
|
|
addon = ""
|
|
|
|
for m in entry.get("media", []):
|
|
|
|
artist = m.get("performer")
|
|
|
|
trackname = ""
|
|
|
|
if artist:
|
|
|
|
trackname += artist + " - "
|
|
|
|
if "title" in m:
|
|
|
|
trackname += m.get("title", "")
|
|
|
|
addon += (
|
|
|
|
'<MusicPlayer src="'
|
|
|
|
+ m.get("fileUrl", "")
|
|
|
|
+ '" title="'
|
|
|
|
+ trackname
|
|
|
|
+ '" />\n'
|
|
|
|
)
|
|
|
|
body = "import MusicPlayer from '$/components/Article/MusicPlayer'\n\n" + addon
|
|
|
|
|
|
|
|
body_orig = extract_html(entry)
|
|
|
|
if body_orig:
|
|
|
|
body += extract_md(html2text(body_orig), entry["_id"])
|
|
|
|
if not body:
|
|
|
|
print("[extract] empty MDX body")
|
|
|
|
return body
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
def prepare_html_body(entry):
|
2022-09-03 10:50:14 +00:00
|
|
|
# body modifications
|
|
|
|
body = ""
|
|
|
|
kind = entry.get("type")
|
|
|
|
addon = ""
|
|
|
|
if kind == "Video":
|
|
|
|
addon = ""
|
|
|
|
for m in entry.get("media", []):
|
|
|
|
if "youtubeId" in m:
|
|
|
|
addon += '<iframe width="420" height="345" src="http://www.youtube.com/embed/'
|
|
|
|
addon += m["youtubeId"]
|
|
|
|
addon += '?autoplay=1" frameborder="0" allowfullscreen></iframe>\n'
|
|
|
|
elif "vimeoId" in m:
|
|
|
|
addon += '<iframe src="https://player.vimeo.com/video/'
|
|
|
|
addon += m["vimeoId"]
|
2022-09-04 17:20:38 +00:00
|
|
|
addon += ' width="420" height="345" frameborder="0" allow="autoplay; fullscreen"'
|
|
|
|
addon += " allowfullscreen></iframe>"
|
2022-09-03 10:50:14 +00:00
|
|
|
else:
|
|
|
|
print("[extract] media is not supported")
|
|
|
|
print(m)
|
|
|
|
body += addon
|
|
|
|
|
|
|
|
elif kind == "Music":
|
|
|
|
addon = ""
|
|
|
|
for m in entry.get("media", []):
|
|
|
|
artist = m.get("performer")
|
|
|
|
trackname = ""
|
|
|
|
if artist:
|
|
|
|
trackname += artist + " - "
|
|
|
|
if "title" in m:
|
|
|
|
trackname += m.get("title", "")
|
|
|
|
addon += "<figure><figcaption>"
|
|
|
|
addon += trackname
|
|
|
|
addon += '</figcaption><audio controls src="'
|
|
|
|
addon += m.get("fileUrl", "")
|
|
|
|
addon += '"></audio></figure>'
|
|
|
|
body += addon
|
|
|
|
|
|
|
|
body = extract_html(entry)
|
|
|
|
# if body_orig: body += extract_md(html2text(body_orig), entry['_id'])
|
|
|
|
if not body:
|
|
|
|
print("[extract] empty HTML body")
|
|
|
|
return body
|
|
|
|
|
2022-08-11 09:14:12 +00:00
|
|
|
|
|
|
|
def extract_html(entry):
|
2022-09-03 10:50:14 +00:00
|
|
|
body_orig = entry.get("body") or ""
|
|
|
|
media = entry.get("media", [])
|
|
|
|
kind = entry.get("type") or ""
|
|
|
|
print("[extract] kind: " + kind)
|
|
|
|
mbodies = set([])
|
|
|
|
if media:
|
|
|
|
# print('[extract] media is found')
|
|
|
|
for m in media:
|
|
|
|
mbody = m.get("body", "")
|
|
|
|
addon = ""
|
|
|
|
if kind == "Literature":
|
|
|
|
mbody = m.get("literatureBody") or m.get("body", "")
|
|
|
|
elif kind == "Image":
|
|
|
|
cover = ""
|
|
|
|
if "thumborId" in entry:
|
|
|
|
cover = cdn + "/unsafe/1600x/" + entry["thumborId"]
|
|
|
|
if not cover:
|
|
|
|
if "image" in entry:
|
|
|
|
cover = entry["image"].get("url", "")
|
|
|
|
if "cloudinary" in cover:
|
|
|
|
cover = ""
|
|
|
|
# else: print('[extract] cover: ' + cover)
|
|
|
|
title = m.get("title", "").replace("\n", " ").replace(" ", " ")
|
|
|
|
u = m.get("thumborId") or cover or ""
|
|
|
|
if title:
|
|
|
|
addon += "<h4>" + title + "</h4>\n"
|
|
|
|
if not u.startswith("http"):
|
|
|
|
u = s3 + u
|
|
|
|
if not u:
|
|
|
|
print("[extract] no image url for " + str(m))
|
|
|
|
if "cloudinary" in u:
|
|
|
|
u = "img/lost.svg"
|
|
|
|
if u != cover or (u == cover and media.index(m) == 0):
|
|
|
|
addon += '<img src="' + u + '" alt="' + title + '" />\n'
|
|
|
|
if addon:
|
|
|
|
body_orig += addon
|
|
|
|
# print('[extract] item addon: ' + addon)
|
|
|
|
# if addon: print('[extract] addon: %s' % addon)
|
|
|
|
if mbody and mbody not in mbodies:
|
|
|
|
mbodies.add(mbody)
|
|
|
|
body_orig += mbody
|
|
|
|
if len(list(mbodies)) != len(media):
|
|
|
|
print(
|
|
|
|
"[extract] %d/%d media item bodies appended"
|
|
|
|
% (len(list(mbodies)), len(media))
|
|
|
|
)
|
|
|
|
# print('[extract] media items body: \n' + body_orig)
|
|
|
|
if not body_orig:
|
|
|
|
for up in entry.get("bodyHistory", []) or []:
|
|
|
|
body_orig = up.get("text", "") or ""
|
|
|
|
if body_orig:
|
|
|
|
print("[extract] got html body from history")
|
|
|
|
break
|
|
|
|
if not body_orig:
|
|
|
|
print("[extract] empty HTML body")
|
|
|
|
# body_html = str(BeautifulSoup(body_orig, features="html.parser"))
|
|
|
|
return body_orig
|