This commit is contained in:
parent
564a8c10b7
commit
2e68128dfc
|
@ -1,7 +1,7 @@
|
||||||
import json
|
import json
|
||||||
import time
|
import time
|
||||||
|
|
||||||
from sqlalchemy import select, or_, and_, text, desc
|
from sqlalchemy import select, or_, and_, text, desc, cast, Integer
|
||||||
from sqlalchemy.orm import aliased
|
from sqlalchemy.orm import aliased
|
||||||
from sqlalchemy_searchable import search
|
from sqlalchemy_searchable import search
|
||||||
|
|
||||||
|
@ -9,7 +9,7 @@ from orm.author import Author, AuthorFollower
|
||||||
from orm.shout import ShoutAuthor, ShoutTopic
|
from orm.shout import ShoutAuthor, ShoutTopic
|
||||||
from orm.topic import Topic
|
from orm.topic import Topic
|
||||||
from resolvers.stat import get_with_stat, author_follows_authors, author_follows_topics
|
from resolvers.stat import get_with_stat, author_follows_authors, author_follows_topics
|
||||||
from services.event_listeners import update_author_cache
|
from services.cache import update_author_cache
|
||||||
from services.auth import login_required
|
from services.auth import login_required
|
||||||
from services.db import local_session
|
from services.db import local_session
|
||||||
from services.rediscache import redis
|
from services.rediscache import redis
|
||||||
|
@ -214,26 +214,27 @@ def create_author(user_id: str, slug: str, name: str = ''):
|
||||||
|
|
||||||
|
|
||||||
@query.field('get_author_followers')
|
@query.field('get_author_followers')
|
||||||
def get_author_followers(_, _info, slug: str):
|
async def get_author_followers(_, _info, slug: str):
|
||||||
logger.debug(f'getting followers for @{slug}')
|
logger.debug(f'getting followers for @{slug}')
|
||||||
try:
|
try:
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
author_alias = aliased(Author)
|
author_alias = aliased(Author)
|
||||||
author_id_result = (
|
author_id_result = (
|
||||||
session.query(author_alias.id).filter(author_alias.slug == slug).first()
|
session.query(author_alias).filter(author_alias.slug == slug).first()
|
||||||
)
|
)
|
||||||
author_id = author_id_result[0] if author_id_result else None
|
author = author_id_result[0] if author_id_result else None
|
||||||
|
author_id = cast(author.id, Integer)
|
||||||
author_follower_alias = aliased(AuthorFollower, name='af')
|
cached = await redis.execute('GET', f'id:{author_id}:followers')
|
||||||
q = select(Author).join(
|
if not cached:
|
||||||
author_follower_alias,
|
author_follower_alias = aliased(AuthorFollower, name='af')
|
||||||
and_(
|
q = select(Author).join(
|
||||||
author_follower_alias.author == author_id,
|
author_follower_alias,
|
||||||
author_follower_alias.follower == Author.id,
|
and_(
|
||||||
),
|
author_follower_alias.author == author_id,
|
||||||
)
|
author_follower_alias.follower == Author.id,
|
||||||
|
),
|
||||||
return get_with_stat(q)
|
)
|
||||||
|
return json.loads(cached) if cached else get_with_stat(q)
|
||||||
except Exception as exc:
|
except Exception as exc:
|
||||||
logger.error(exc)
|
logger.error(exc)
|
||||||
return []
|
return []
|
||||||
|
|
|
@ -16,7 +16,7 @@ from resolvers.topic import topic_unfollow
|
||||||
from resolvers.stat import get_with_stat, author_follows_topics, author_follows_authors
|
from resolvers.stat import get_with_stat, author_follows_topics, author_follows_authors
|
||||||
from services.auth import login_required
|
from services.auth import login_required
|
||||||
from services.db import local_session
|
from services.db import local_session
|
||||||
from services.event_listeners import DEFAULT_FOLLOWS, update_follows_for_author
|
from services.cache import DEFAULT_FOLLOWS, update_follows_for_author, update_followers_for_author
|
||||||
from services.notify import notify_follower
|
from services.notify import notify_follower
|
||||||
from services.schema import mutation, query
|
from services.schema import mutation, query
|
||||||
from services.logger import root_logger as logger
|
from services.logger import root_logger as logger
|
||||||
|
@ -28,24 +28,27 @@ from services.rediscache import redis
|
||||||
async def follow(_, info, what, slug):
|
async def follow(_, info, what, slug):
|
||||||
try:
|
try:
|
||||||
user_id = info.context['user_id']
|
user_id = info.context['user_id']
|
||||||
with local_session() as session:
|
follower_query = select(Author).select_from(Author).filter(Author.user == user_id)
|
||||||
follower = session.query(Author).filter(Author.user == user_id).first()
|
[follower] = get_with_stat(follower_query)
|
||||||
if follower:
|
if follower:
|
||||||
if what == 'AUTHOR':
|
if what == 'AUTHOR':
|
||||||
if author_unfollow(follower.id, slug):
|
if author_unfollow(follower.id, slug):
|
||||||
author = session.query(Author).where(Author.slug == slug).first()
|
author_query = select(Author).select_from(Author).where(Author.slug == slug)
|
||||||
if author:
|
[author] = get_with_stat(author_query)
|
||||||
await update_follows_for_author(session, follower, 'author', author, True)
|
if author:
|
||||||
await notify_follower(follower.dict(), author.id, 'unfollow')
|
await update_follows_for_author(follower, 'author', author, True)
|
||||||
elif what == 'TOPIC':
|
await update_followers_for_author(follower, author, True)
|
||||||
topic = session.query(Topic).where(Topic.slug == slug).first()
|
await notify_follower(follower.dict(), author.id, 'unfollow')
|
||||||
if topic:
|
elif what == 'TOPIC':
|
||||||
await update_follows_for_author(session, follower, 'topic', topic, True)
|
topic_query = select(Topic).where(Topic.slug == slug)
|
||||||
topic_unfollow(follower.id, slug)
|
[topic] = get_with_stat(topic_query)
|
||||||
elif what == 'COMMUNITY':
|
if topic:
|
||||||
community_follow(follower.id, slug)
|
await update_follows_for_author(follower, 'topic', topic, True)
|
||||||
elif what == 'REACTIONS':
|
topic_unfollow(follower.id, slug)
|
||||||
reactions_follow(follower.id, slug)
|
elif what == 'COMMUNITY':
|
||||||
|
community_follow(follower.id, slug)
|
||||||
|
elif what == 'REACTIONS':
|
||||||
|
reactions_follow(follower.id, slug)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.debug(info, what, slug)
|
logger.debug(info, what, slug)
|
||||||
logger.error(e)
|
logger.error(e)
|
||||||
|
@ -59,24 +62,27 @@ async def follow(_, info, what, slug):
|
||||||
async def unfollow(_, info, what, slug):
|
async def unfollow(_, info, what, slug):
|
||||||
user_id = info.context['user_id']
|
user_id = info.context['user_id']
|
||||||
try:
|
try:
|
||||||
with local_session() as session:
|
follower_query = select(Author).filter(Author.user == user_id)
|
||||||
follower = session.query(Author).filter(Author.user == user_id).first()
|
[follower] = get_with_stat(follower_query)
|
||||||
if follower:
|
if follower:
|
||||||
if what == 'AUTHOR':
|
if what == 'AUTHOR':
|
||||||
if author_unfollow(follower.id, slug):
|
if author_unfollow(follower.id, slug):
|
||||||
author = session.query(Author).where(Author.slug == slug).first()
|
author_query = select(Author).where(Author.slug == slug)
|
||||||
if author:
|
[author] = get_with_stat(author_query)
|
||||||
await update_follows_for_author(session, follower, 'author', author, False)
|
if author:
|
||||||
await notify_follower(follower.dict(), author.id, 'unfollow')
|
await update_follows_for_author(follower, 'author', author, False)
|
||||||
elif what == 'TOPIC':
|
await update_followers_for_author(follower, author, False)
|
||||||
topic = session.query(Topic).where(Topic.slug == slug).first()
|
await notify_follower(follower.dict(), author.id, 'unfollow')
|
||||||
if topic:
|
elif what == 'TOPIC':
|
||||||
await update_follows_for_author(session, follower, 'topic', topic, False)
|
topic_query = select(Topic).where(Topic.slug == slug)
|
||||||
topic_unfollow(follower.id, slug)
|
[topic] = get_with_stat(topic_query)
|
||||||
elif what == 'COMMUNITY':
|
if topic:
|
||||||
community_unfollow(follower.id, slug)
|
await update_follows_for_author(follower, 'topic', topic, False)
|
||||||
elif what == 'REACTIONS':
|
topic_unfollow(follower.id, slug)
|
||||||
reactions_unfollow(follower.id, slug)
|
elif what == 'COMMUNITY':
|
||||||
|
community_unfollow(follower.id, slug)
|
||||||
|
elif what == 'REACTIONS':
|
||||||
|
reactions_unfollow(follower.id, slug)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
return {'error': str(e)}
|
return {'error': str(e)}
|
||||||
|
|
||||||
|
|
|
@ -1,26 +1,219 @@
|
||||||
from functools import wraps
|
import asyncio
|
||||||
|
|
||||||
from dogpile.cache import make_region
|
from sqlalchemy import select, event
|
||||||
|
import json
|
||||||
|
|
||||||
# Создание региона кэша с TTL 300 секунд
|
from orm.author import Author, AuthorFollower
|
||||||
cache_region = make_region().configure('dogpile.cache.memory', expiration_time=300)
|
from orm.reaction import Reaction
|
||||||
|
from orm.shout import ShoutAuthor, Shout
|
||||||
|
from orm.topic import Topic, TopicFollower
|
||||||
|
from resolvers.stat import get_with_stat
|
||||||
|
from services.rediscache import redis
|
||||||
|
from services.logger import root_logger as logger
|
||||||
|
|
||||||
|
|
||||||
# Декоратор для кэширования методов
|
DEFAULT_FOLLOWS = {
|
||||||
def cache_method(cache_key: str):
|
'topics': [],
|
||||||
def decorator(f):
|
'authors': [],
|
||||||
@wraps(f)
|
'communities': [{'id': 1, 'name': 'Дискурс', 'slug': 'discours', 'pic': ''}],
|
||||||
def decorated_function(*args, **kwargs):
|
}
|
||||||
# Генерация ключа для кэширования
|
|
||||||
key = cache_key.format(*args, **kwargs)
|
|
||||||
# Получение значения из кэша
|
|
||||||
result = cache_region.get(key)
|
|
||||||
if result is None:
|
|
||||||
# Если значение отсутствует в кэше, вызываем функцию и кэшируем результат
|
|
||||||
result = f(*args, **kwargs)
|
|
||||||
cache_region.set(key, result)
|
|
||||||
return result
|
|
||||||
|
|
||||||
return decorated_function
|
|
||||||
|
|
||||||
return decorator
|
async def update_author_cache(author: dict, ttl=25 * 60 * 60):
|
||||||
|
payload = json.dumps(author)
|
||||||
|
await redis.execute('SETEX', f'user:{author.get("user")}:author', ttl, payload)
|
||||||
|
await redis.execute('SETEX', f'id:{author.get("id")}:author', ttl, payload)
|
||||||
|
|
||||||
|
|
||||||
|
async def update_follows_topics_cache(follows, author_id: int, ttl=25 * 60 * 60):
|
||||||
|
try:
|
||||||
|
payload = json.dumps(follows)
|
||||||
|
await redis.execute('SETEX', f'author:{author_id}:follows-topics', ttl, payload)
|
||||||
|
except Exception as exc:
|
||||||
|
logger.error(exc)
|
||||||
|
import traceback
|
||||||
|
|
||||||
|
exc = traceback.format_exc()
|
||||||
|
logger.error(exc)
|
||||||
|
|
||||||
|
|
||||||
|
async def update_follows_authors_cache(follows, author_id: int, ttl=25 * 60 * 60):
|
||||||
|
try:
|
||||||
|
payload = json.dumps(follows)
|
||||||
|
await redis.execute('SETEX', f'author:{author_id}:follows-authors', ttl, payload)
|
||||||
|
except Exception:
|
||||||
|
import traceback
|
||||||
|
|
||||||
|
exc = traceback.format_exc()
|
||||||
|
logger.error(exc)
|
||||||
|
|
||||||
|
|
||||||
|
@event.listens_for(Shout, 'after_insert')
|
||||||
|
@event.listens_for(Shout, 'after_update')
|
||||||
|
def after_shouts_update(mapper, connection, shout: Shout):
|
||||||
|
# Main query to get authors associated with the shout through ShoutAuthor
|
||||||
|
authors_query = (
|
||||||
|
select(Author)
|
||||||
|
.select_from(ShoutAuthor) # Select from ShoutAuthor
|
||||||
|
.join(Author, Author.id == ShoutAuthor.author) # Join with Author
|
||||||
|
.where(ShoutAuthor.shout == shout.id) # Filter by shout.id
|
||||||
|
)
|
||||||
|
|
||||||
|
for author_with_stat in get_with_stat(authors_query):
|
||||||
|
asyncio.create_task(update_author_cache(author_with_stat.dict()))
|
||||||
|
|
||||||
|
|
||||||
|
@event.listens_for(Reaction, 'after_insert')
|
||||||
|
def after_reaction_insert(mapper, connection, reaction: Reaction):
|
||||||
|
try:
|
||||||
|
author_subquery = select(Author).where(Author.id == reaction.created_by)
|
||||||
|
replied_author_subquery = (
|
||||||
|
select(Author)
|
||||||
|
.join(Reaction, Author.id == Reaction.created_by)
|
||||||
|
.where(Reaction.id == reaction.reply_to)
|
||||||
|
)
|
||||||
|
|
||||||
|
author_query = select(
|
||||||
|
author_subquery.subquery().c.id,
|
||||||
|
author_subquery.subquery().c.slug,
|
||||||
|
author_subquery.subquery().c.created_at,
|
||||||
|
author_subquery.subquery().c.name,
|
||||||
|
).select_from(author_subquery.subquery()).union(
|
||||||
|
select(
|
||||||
|
replied_author_subquery.subquery().c.id,
|
||||||
|
)
|
||||||
|
.select_from(replied_author_subquery.subquery())
|
||||||
|
)
|
||||||
|
|
||||||
|
for author_with_stat in get_with_stat(author_query):
|
||||||
|
asyncio.create_task(update_author_cache(author_with_stat.dict()))
|
||||||
|
|
||||||
|
shout = connection.execute(select(Shout).select_from(Shout).where(Shout.id == reaction.shout)).first()
|
||||||
|
if shout:
|
||||||
|
after_shouts_update(mapper, connection, shout)
|
||||||
|
except Exception as exc:
|
||||||
|
logger.error(exc)
|
||||||
|
|
||||||
|
|
||||||
|
@event.listens_for(Author, 'after_insert')
|
||||||
|
@event.listens_for(Author, 'after_update')
|
||||||
|
def after_author_update(mapper, connection, author: Author):
|
||||||
|
q = select(Author).where(Author.id == author.id)
|
||||||
|
[author_with_stat] = get_with_stat(q)
|
||||||
|
asyncio.create_task(update_author_cache(author_with_stat.dict()))
|
||||||
|
|
||||||
|
|
||||||
|
@event.listens_for(TopicFollower, 'after_insert')
|
||||||
|
def after_topic_follower_insert(mapper, connection, target: TopicFollower):
|
||||||
|
asyncio.create_task(
|
||||||
|
handle_topic_follower_change(connection, target.topic, target.follower, True)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@event.listens_for(TopicFollower, 'after_delete')
|
||||||
|
def after_topic_follower_delete(mapper, connection, target: TopicFollower):
|
||||||
|
asyncio.create_task(
|
||||||
|
handle_topic_follower_change(connection, target.topic, target.follower, False)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@event.listens_for(AuthorFollower, 'after_insert')
|
||||||
|
def after_author_follower_insert(mapper, connection, target: AuthorFollower):
|
||||||
|
asyncio.create_task(
|
||||||
|
handle_author_follower_change(connection, target.author, target.follower, True)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@event.listens_for(AuthorFollower, 'after_delete')
|
||||||
|
def after_author_follower_delete(mapper, connection, target: AuthorFollower):
|
||||||
|
asyncio.create_task(
|
||||||
|
handle_author_follower_change(connection, target.author, target.follower, False)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
async def update_follows_for_author(follower: Author, entity_type: str, entity: dict, is_insert: bool):
|
||||||
|
ttl = 25 * 60 * 60
|
||||||
|
redis_key = f'id:{follower.id}:follows-{entity_type}s'
|
||||||
|
follows_str = await redis.get(redis_key)
|
||||||
|
follows = json.loads(follows_str) if follows_str else []
|
||||||
|
if is_insert:
|
||||||
|
follows.append(entity)
|
||||||
|
else:
|
||||||
|
# Remove the entity from follows
|
||||||
|
follows = [e for e in follows if e['id'] != entity['id']]
|
||||||
|
await redis.execute('SETEX', redis_key, ttl, json.dumps(follows))
|
||||||
|
|
||||||
|
|
||||||
|
async def update_followers_for_author(follower: Author, author: Author, is_insert: bool):
|
||||||
|
ttl = 25 * 60 * 60
|
||||||
|
redis_key = f'id:{author.id}:followers'
|
||||||
|
followers_str = await redis.get(redis_key)
|
||||||
|
followers = json.loads(followers_str) if followers_str else []
|
||||||
|
if is_insert:
|
||||||
|
followers.append(follower)
|
||||||
|
else:
|
||||||
|
# Remove the entity from follows
|
||||||
|
follows = [e for e in followers if e['id'] != author.id]
|
||||||
|
await redis.execute('SETEX', redis_key, ttl, json.dumps(follows))
|
||||||
|
|
||||||
|
|
||||||
|
async def handle_author_follower_change(
|
||||||
|
connection, author_id: int, follower_id: int, is_insert: bool
|
||||||
|
):
|
||||||
|
author_query = select(Author).select_from(Author).filter(Author.id == author_id)
|
||||||
|
[author] = get_with_stat(author_query)
|
||||||
|
follower_query = select(Author).select_from(Author).filter(Author.id == follower_id)
|
||||||
|
follower = get_with_stat(follower_query)
|
||||||
|
if follower and author:
|
||||||
|
_ = asyncio.create_task(update_author_cache(author.dict()))
|
||||||
|
follows_authors = await redis.execute('GET', f'author:{follower_id}:follows-authors')
|
||||||
|
if follows_authors:
|
||||||
|
follows_authors = json.loads(follows_authors)
|
||||||
|
if not any(x.get('id') == author.id for x in follows_authors):
|
||||||
|
follows_authors.append(author.dict())
|
||||||
|
_ = asyncio.create_task(update_follows_authors_cache(follows_authors, follower_id))
|
||||||
|
_ = asyncio.create_task(update_author_cache(follower.dict()))
|
||||||
|
await update_follows_for_author(
|
||||||
|
connection,
|
||||||
|
follower,
|
||||||
|
'author',
|
||||||
|
{
|
||||||
|
'id': author.id,
|
||||||
|
'name': author.name,
|
||||||
|
'slug': author.slug,
|
||||||
|
'pic': author.pic,
|
||||||
|
'bio': author.bio,
|
||||||
|
'stat': author.stat,
|
||||||
|
},
|
||||||
|
is_insert,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
async def handle_topic_follower_change(
|
||||||
|
connection, topic_id: int, follower_id: int, is_insert: bool
|
||||||
|
):
|
||||||
|
q = select(Topic).filter(Topic.id == topic_id)
|
||||||
|
topics = get_with_stat(q)
|
||||||
|
topic = topics[0]
|
||||||
|
follower_query = select(Author).filter(Author.id == follower_id)
|
||||||
|
follower = get_with_stat(follower_query)
|
||||||
|
if follower and topic:
|
||||||
|
_ = asyncio.create_task(update_author_cache(follower.dict()))
|
||||||
|
follows_topics = await redis.execute('GET', f'author:{follower_id}:follows-topics')
|
||||||
|
if follows_topics:
|
||||||
|
follows_topics = json.loads(follows_topics)
|
||||||
|
if not any(x.get('id') == topic.id for x in follows_topics):
|
||||||
|
follows_topics.append(topic)
|
||||||
|
_ = asyncio.create_task(update_follows_topics_cache(follows_topics, follower_id))
|
||||||
|
await update_follows_for_author(
|
||||||
|
follower,
|
||||||
|
'topic',
|
||||||
|
{
|
||||||
|
'id': topic.id,
|
||||||
|
'title': topic.title,
|
||||||
|
'slug': topic.slug,
|
||||||
|
'body': topic.body,
|
||||||
|
'stat': topic.stat,
|
||||||
|
},
|
||||||
|
is_insert,
|
||||||
|
)
|
||||||
|
|
|
@ -1,210 +0,0 @@
|
||||||
import asyncio
|
|
||||||
|
|
||||||
from sqlalchemy import select, event
|
|
||||||
import json
|
|
||||||
|
|
||||||
from orm.author import Author, AuthorFollower
|
|
||||||
from orm.reaction import Reaction
|
|
||||||
from orm.shout import ShoutAuthor, Shout
|
|
||||||
from orm.topic import Topic, TopicFollower
|
|
||||||
from resolvers.stat import get_with_stat
|
|
||||||
from services.rediscache import redis
|
|
||||||
from services.logger import root_logger as logger
|
|
||||||
|
|
||||||
|
|
||||||
DEFAULT_FOLLOWS = {
|
|
||||||
'topics': [],
|
|
||||||
'authors': [],
|
|
||||||
'communities': [{'id': 1, 'name': 'Дискурс', 'slug': 'discours', 'pic': ''}],
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
async def update_author_cache(author: dict, ttl=25 * 60 * 60):
|
|
||||||
payload = json.dumps(author)
|
|
||||||
await redis.execute('SETEX', f'user:{author.get("user")}:author', ttl, payload)
|
|
||||||
await redis.execute('SETEX', f'id:{author.get("id")}:author', ttl, payload)
|
|
||||||
|
|
||||||
|
|
||||||
async def update_follows_topics_cache(follows, author_id: int, ttl=25 * 60 * 60):
|
|
||||||
try:
|
|
||||||
payload = json.dumps(follows)
|
|
||||||
await redis.execute('SETEX', f'author:{author_id}:follows-topics', ttl, payload)
|
|
||||||
except Exception:
|
|
||||||
import traceback
|
|
||||||
|
|
||||||
exc = traceback.format_exc()
|
|
||||||
logger.error(exc)
|
|
||||||
|
|
||||||
|
|
||||||
async def update_follows_authors_cache(follows, author_id: int, ttl=25 * 60 * 60):
|
|
||||||
try:
|
|
||||||
payload = json.dumps(follows)
|
|
||||||
await redis.execute('SETEX', f'author:{author_id}:follows-authors', ttl, payload)
|
|
||||||
except Exception:
|
|
||||||
import traceback
|
|
||||||
|
|
||||||
exc = traceback.format_exc()
|
|
||||||
logger.error(exc)
|
|
||||||
|
|
||||||
|
|
||||||
@event.listens_for(Shout, 'after_insert')
|
|
||||||
@event.listens_for(Shout, 'after_update')
|
|
||||||
def after_shouts_update(mapper, connection, shout: Shout):
|
|
||||||
# Main query to get authors associated with the shout through ShoutAuthor
|
|
||||||
authors_query = (
|
|
||||||
select(Author)
|
|
||||||
.select_from(ShoutAuthor) # Select from ShoutAuthor
|
|
||||||
.join(Author, Author.id == ShoutAuthor.author) # Join with Author
|
|
||||||
.where(ShoutAuthor.shout == shout.id) # Filter by shout.id
|
|
||||||
)
|
|
||||||
|
|
||||||
for author_with_stat in get_with_stat(authors_query):
|
|
||||||
asyncio.create_task(update_author_cache(author_with_stat.dict()))
|
|
||||||
|
|
||||||
|
|
||||||
@event.listens_for(Reaction, 'after_insert')
|
|
||||||
def after_reaction_insert(mapper, connection, reaction: Reaction):
|
|
||||||
try:
|
|
||||||
author_subquery = select(Author).where(Author.id == reaction.created_by)
|
|
||||||
replied_author_subquery = (
|
|
||||||
select(Author)
|
|
||||||
.join(Reaction, Author.id == Reaction.created_by)
|
|
||||||
.where(Reaction.id == reaction.reply_to)
|
|
||||||
)
|
|
||||||
|
|
||||||
author_query = select(
|
|
||||||
author_subquery.subquery().c.id,
|
|
||||||
author_subquery.subquery().c.slug,
|
|
||||||
author_subquery.subquery().c.created_at,
|
|
||||||
author_subquery.subquery().c.name,
|
|
||||||
).select_from(author_subquery.subquery()).union(
|
|
||||||
select(
|
|
||||||
replied_author_subquery.subquery().c.id,
|
|
||||||
)
|
|
||||||
.select_from(replied_author_subquery.subquery())
|
|
||||||
)
|
|
||||||
|
|
||||||
for author_with_stat in get_with_stat(author_query):
|
|
||||||
asyncio.create_task(update_author_cache(author_with_stat.dict()))
|
|
||||||
|
|
||||||
shout = connection.execute(select(Shout).select_from(Shout).where(Shout.id == reaction.shout)).first()
|
|
||||||
if shout:
|
|
||||||
after_shouts_update(mapper, connection, shout)
|
|
||||||
except Exception as exc:
|
|
||||||
logger.error(exc)
|
|
||||||
|
|
||||||
|
|
||||||
@event.listens_for(Author, 'after_insert')
|
|
||||||
@event.listens_for(Author, 'after_update')
|
|
||||||
def after_author_update(mapper, connection, author: Author):
|
|
||||||
q = select(Author).where(Author.id == author.id)
|
|
||||||
[author_with_stat] = get_with_stat(q)
|
|
||||||
asyncio.create_task(update_author_cache(author_with_stat.dict()))
|
|
||||||
|
|
||||||
|
|
||||||
@event.listens_for(TopicFollower, 'after_insert')
|
|
||||||
def after_topic_follower_insert(mapper, connection, target: TopicFollower):
|
|
||||||
asyncio.create_task(
|
|
||||||
handle_topic_follower_change(connection, target.topic, target.follower, True)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@event.listens_for(TopicFollower, 'after_delete')
|
|
||||||
def after_topic_follower_delete(mapper, connection, target: TopicFollower):
|
|
||||||
asyncio.create_task(
|
|
||||||
handle_topic_follower_change(connection, target.topic, target.follower, False)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@event.listens_for(AuthorFollower, 'after_insert')
|
|
||||||
def after_author_follower_insert(mapper, connection, target: AuthorFollower):
|
|
||||||
asyncio.create_task(
|
|
||||||
handle_author_follower_change(connection, target.author, target.follower, True)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@event.listens_for(AuthorFollower, 'after_delete')
|
|
||||||
def after_author_follower_delete(mapper, connection, target: AuthorFollower):
|
|
||||||
asyncio.create_task(
|
|
||||||
handle_author_follower_change(connection, target.author, target.follower, False)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
async def update_follows_for_author(
|
|
||||||
connection, follower, entity_type, entity: dict, is_insert
|
|
||||||
):
|
|
||||||
ttl = 25 * 60 * 60
|
|
||||||
redis_key = f'id:{follower.id}:follows-{entity_type}s'
|
|
||||||
follows_str = await redis.get(redis_key)
|
|
||||||
follows = json.loads(follows_str) if follows_str else []
|
|
||||||
if is_insert:
|
|
||||||
follows[f'{entity_type}s'].append(entity)
|
|
||||||
else:
|
|
||||||
# Remove the entity from follows
|
|
||||||
follows[f'{entity_type}s'] = [
|
|
||||||
e for e in follows[f'{entity_type}s'] if e['id'] != entity['id']
|
|
||||||
]
|
|
||||||
await redis.execute('SETEX', redis_key, ttl, json.dumps(follows))
|
|
||||||
|
|
||||||
|
|
||||||
async def handle_author_follower_change(
|
|
||||||
connection, author_id: int, follower_id: int, is_insert: bool
|
|
||||||
):
|
|
||||||
author_query = select(Author).select_from(Author).filter(Author.id == author_id)
|
|
||||||
[author] = get_with_stat(author_query)
|
|
||||||
follower_query = select(Author).select_from(Author).filter(Author.id == follower_id)
|
|
||||||
follower = get_with_stat(follower_query)
|
|
||||||
if follower and author:
|
|
||||||
_ = asyncio.create_task(update_author_cache(author.dict()))
|
|
||||||
follows_authors = await redis.execute('GET', f'author:{follower_id}:follows-authors')
|
|
||||||
if follows_authors:
|
|
||||||
follows_authors = json.loads(follows_authors)
|
|
||||||
if not any(x.get('id') == author.id for x in follows_authors):
|
|
||||||
follows_authors.append(author.dict())
|
|
||||||
_ = asyncio.create_task(update_follows_authors_cache(follows_authors, follower_id))
|
|
||||||
_ = asyncio.create_task(update_author_cache(follower.dict()))
|
|
||||||
await update_follows_for_author(
|
|
||||||
connection,
|
|
||||||
follower,
|
|
||||||
'author',
|
|
||||||
{
|
|
||||||
'id': author.id,
|
|
||||||
'name': author.name,
|
|
||||||
'slug': author.slug,
|
|
||||||
'pic': author.pic,
|
|
||||||
'bio': author.bio,
|
|
||||||
'stat': author.stat,
|
|
||||||
},
|
|
||||||
is_insert,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
async def handle_topic_follower_change(
|
|
||||||
connection, topic_id: int, follower_id: int, is_insert: bool
|
|
||||||
):
|
|
||||||
q = select(Topic).filter(Topic.id == topic_id)
|
|
||||||
topics = get_with_stat(q)
|
|
||||||
topic = topics[0]
|
|
||||||
follower_query = select(Author).filter(Author.id == follower_id)
|
|
||||||
follower = get_with_stat(follower_query)
|
|
||||||
if follower and topic:
|
|
||||||
_ = asyncio.create_task(update_author_cache(follower.dict()))
|
|
||||||
follows_topics = await redis.execute('GET', f'author:{follower_id}:follows-topics')
|
|
||||||
if follows_topics:
|
|
||||||
follows_topics = json.loads(follows_topics)
|
|
||||||
if not any(x.get('id') == topic.id for x in follows_topics):
|
|
||||||
follows_topics.append(topic)
|
|
||||||
_ = asyncio.create_task(update_follows_topics_cache(follows_topics, follower_id))
|
|
||||||
await update_follows_for_author(
|
|
||||||
connection,
|
|
||||||
follower,
|
|
||||||
'topic',
|
|
||||||
{
|
|
||||||
'id': topic.id,
|
|
||||||
'title': topic.title,
|
|
||||||
'slug': topic.slug,
|
|
||||||
'body': topic.body,
|
|
||||||
'stat': topic.stat,
|
|
||||||
},
|
|
||||||
is_insert,
|
|
||||||
)
|
|
26
services/memorycache.py
Normal file
26
services/memorycache.py
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
from functools import wraps
|
||||||
|
|
||||||
|
from dogpile.cache import make_region
|
||||||
|
|
||||||
|
# Создание региона кэша с TTL 300 секунд
|
||||||
|
cache_region = make_region().configure('dogpile.cache.memory', expiration_time=300)
|
||||||
|
|
||||||
|
|
||||||
|
# Декоратор для кэширования методов
|
||||||
|
def cache_method(cache_key: str):
|
||||||
|
def decorator(f):
|
||||||
|
@wraps(f)
|
||||||
|
def decorated_function(*args, **kwargs):
|
||||||
|
# Генерация ключа для кэширования
|
||||||
|
key = cache_key.format(*args, **kwargs)
|
||||||
|
# Получение значения из кэша
|
||||||
|
result = cache_region.get(key)
|
||||||
|
if result is None:
|
||||||
|
# Если значение отсутствует в кэше, вызываем функцию и кэшируем результат
|
||||||
|
result = f(*args, **kwargs)
|
||||||
|
cache_region.set(key, result)
|
||||||
|
return result
|
||||||
|
|
||||||
|
return decorated_function
|
||||||
|
|
||||||
|
return decorator
|
Loading…
Reference in New Issue
Block a user