Deal with locked databases, tweaks
This commit is contained in:
parent
fb9bfe0084
commit
6b02fb66bc
2 changed files with 172 additions and 151 deletions
24
cry/cli.py
24
cry/cli.py
|
|
@ -199,22 +199,34 @@ def serve():
|
|||
def do_GET(self):
|
||||
db = database.Database.local()
|
||||
feeds = db.load_all(feed_limit=10)
|
||||
del db
|
||||
|
||||
feeds.sort(key=feed.sort_key, reverse=True)
|
||||
|
||||
buffer = io.StringIO()
|
||||
buffer.write("<head>")
|
||||
buffer.write('<meta charset="utf-8"><title>Subscribed Feeds</title>')
|
||||
buffer.write("</head>")
|
||||
buffer.write("<h1>Feeds</h1>")
|
||||
buffer.write(
|
||||
"""
|
||||
<!doctype html>
|
||||
<head>
|
||||
<meta charset="utf8">
|
||||
<title>Subscribed Feeds</title>
|
||||
</head>
|
||||
<h1>Feeds</h1>
|
||||
"""
|
||||
)
|
||||
for f in feeds:
|
||||
feed_title = html.escape(f.title)
|
||||
buffer.write(f'<h2><a href="{f.link}">{feed_title}</a></h2>')
|
||||
if len(f.entries) > 0:
|
||||
ago = f" ({f.entries[0].time_ago()})"
|
||||
else:
|
||||
ago = ""
|
||||
buffer.write(f'<h2><a href="{f.link}">{feed_title}</a>{ago}</h2>')
|
||||
buffer.write(f"<div>")
|
||||
if len(f.entries) > 0:
|
||||
for entry in f.entries:
|
||||
title = html.escape(entry.title)
|
||||
buffer.write(
|
||||
f'<span>• <a href="{entry.link}">{title}</a> ({entry.time_ago()})</span> '
|
||||
f'<span class="entry">• <a href="{entry.link}">{title}</a> ({entry.time_ago()})</span> '
|
||||
)
|
||||
else:
|
||||
buffer.write("<i>No entries...</i>")
|
||||
|
|
|
|||
299
cry/database.py
299
cry/database.py
|
|
@ -102,20 +102,24 @@ class Database:
|
|||
return db
|
||||
|
||||
def get_property(self, prop: str, default=None) -> typing.Any:
|
||||
cursor = self.db.execute("SELECT value FROM properties WHERE name=?", (prop,))
|
||||
result = cursor.fetchone()
|
||||
if result is None:
|
||||
return default
|
||||
return result[0]
|
||||
with self.db:
|
||||
cursor = self.db.execute(
|
||||
"SELECT value FROM properties WHERE name=?", (prop,)
|
||||
)
|
||||
result = cursor.fetchone()
|
||||
if result is None:
|
||||
return default
|
||||
return result[0]
|
||||
|
||||
def set_property(self, prop: str, value):
|
||||
self.db.execute(
|
||||
"""
|
||||
INSERT INTO properties (name, value) VALUES (?, ?)
|
||||
ON CONFLICT DO UPDATE SET value=excluded.value
|
||||
""",
|
||||
(prop, value),
|
||||
)
|
||||
with self.db:
|
||||
self.db.execute(
|
||||
"""
|
||||
INSERT INTO properties (name, value) VALUES (?, ?)
|
||||
ON CONFLICT DO UPDATE SET value=excluded.value
|
||||
""",
|
||||
(prop, value),
|
||||
)
|
||||
|
||||
def ensure_database_schema(self):
|
||||
with self.db:
|
||||
|
|
@ -138,67 +142,135 @@ class Database:
|
|||
self.set_property("origin", self.origin)
|
||||
|
||||
def load_all_meta(self) -> list[feed.FeedMeta]:
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
url,
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified
|
||||
FROM feeds
|
||||
"""
|
||||
)
|
||||
rows = cursor.fetchall()
|
||||
return [
|
||||
feed.FeedMeta(
|
||||
url=url,
|
||||
last_fetched_ts=int(last_fetched_ts),
|
||||
retry_after_ts=int(retry_after_ts),
|
||||
status=int(status),
|
||||
etag=etag,
|
||||
modified=modified,
|
||||
origin=self.origin,
|
||||
with self.db:
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
url,
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified
|
||||
FROM feeds
|
||||
"""
|
||||
)
|
||||
for url, last_fetched_ts, retry_after_ts, status, etag, modified in rows
|
||||
]
|
||||
rows = cursor.fetchall()
|
||||
return [
|
||||
feed.FeedMeta(
|
||||
url=url,
|
||||
last_fetched_ts=int(last_fetched_ts),
|
||||
retry_after_ts=int(retry_after_ts),
|
||||
status=int(status),
|
||||
etag=etag,
|
||||
modified=modified,
|
||||
origin=self.origin,
|
||||
)
|
||||
for url, last_fetched_ts, retry_after_ts, status, etag, modified in rows
|
||||
]
|
||||
|
||||
def load_all(self, feed_limit: int = 20, pattern: str = "") -> list[feed.Feed]:
|
||||
pattern = pattern.replace("\\", "\\\\").replace("%", "\\%").replace("_", "\\_")
|
||||
sql_pattern = f"%{pattern}%"
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
url,
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified,
|
||||
title,
|
||||
link
|
||||
FROM feeds
|
||||
WHERE (title LIKE :sql_pattern ESCAPE '\\'
|
||||
OR link LIKE :sql_pattern ESCAPE '\\')
|
||||
AND status != 2 -- UNSUBSCRIBED
|
||||
""",
|
||||
{"sql_pattern": sql_pattern},
|
||||
)
|
||||
rows = cursor.fetchall()
|
||||
with self.db:
|
||||
pattern = (
|
||||
pattern.replace("\\", "\\\\").replace("%", "\\%").replace("_", "\\_")
|
||||
)
|
||||
sql_pattern = f"%{pattern}%"
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
url,
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified,
|
||||
title,
|
||||
link
|
||||
FROM feeds
|
||||
WHERE (title LIKE :sql_pattern ESCAPE '\\'
|
||||
OR link LIKE :sql_pattern ESCAPE '\\')
|
||||
AND status != 2 -- UNSUBSCRIBED
|
||||
""",
|
||||
{"sql_pattern": sql_pattern},
|
||||
)
|
||||
rows = cursor.fetchall()
|
||||
|
||||
almost_feeds = []
|
||||
for row in rows:
|
||||
(
|
||||
url,
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified,
|
||||
title,
|
||||
link,
|
||||
) = row
|
||||
almost_feeds = []
|
||||
for row in rows:
|
||||
(
|
||||
url,
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified,
|
||||
title,
|
||||
link,
|
||||
) = row
|
||||
meta = feed.FeedMeta(
|
||||
url=url,
|
||||
last_fetched_ts=last_fetched_ts,
|
||||
retry_after_ts=retry_after_ts,
|
||||
status=status,
|
||||
etag=etag,
|
||||
modified=modified,
|
||||
origin=self.origin,
|
||||
)
|
||||
almost_feeds.append((meta, title, link))
|
||||
|
||||
feeds = []
|
||||
for meta, title, link in almost_feeds:
|
||||
if feed_limit > 0:
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
id,
|
||||
inserted_at,
|
||||
title,
|
||||
link
|
||||
FROM entries
|
||||
WHERE feed_url=?
|
||||
ORDER BY inserted_at DESC
|
||||
LIMIT ?
|
||||
""",
|
||||
[meta.url, feed_limit],
|
||||
)
|
||||
rows = cursor.fetchall()
|
||||
else:
|
||||
rows = []
|
||||
|
||||
entries = [
|
||||
feed.Entry(id=id, inserted_at=inserted_at, title=title, link=link)
|
||||
for id, inserted_at, title, link in rows
|
||||
]
|
||||
f = feed.Feed(meta=meta, title=title, link=link, entries=entries)
|
||||
feeds.append(f)
|
||||
|
||||
return feeds
|
||||
|
||||
def load_feed(self, url: str) -> feed.Feed | None:
|
||||
with self.db:
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified,
|
||||
title,
|
||||
link
|
||||
FROM feeds
|
||||
WHERE url=?
|
||||
""",
|
||||
[url],
|
||||
)
|
||||
|
||||
row = cursor.fetchone()
|
||||
if row is None:
|
||||
return None
|
||||
|
||||
last_fetched_ts, retry_after_ts, status, etag, modified, title, link = row
|
||||
meta = feed.FeedMeta(
|
||||
url=url,
|
||||
last_fetched_ts=last_fetched_ts,
|
||||
|
|
@ -208,88 +280,25 @@ class Database:
|
|||
modified=modified,
|
||||
origin=self.origin,
|
||||
)
|
||||
almost_feeds.append((meta, title, link))
|
||||
|
||||
feeds = []
|
||||
for meta, title, link in almost_feeds:
|
||||
if feed_limit > 0:
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
id,
|
||||
inserted_at,
|
||||
title,
|
||||
link
|
||||
FROM entries
|
||||
WHERE feed_url=?
|
||||
ORDER BY inserted_at DESC
|
||||
LIMIT ?
|
||||
""",
|
||||
[meta.url, feed_limit],
|
||||
)
|
||||
rows = cursor.fetchall()
|
||||
else:
|
||||
rows = []
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
id,
|
||||
inserted_at,
|
||||
title,
|
||||
link
|
||||
FROM entries
|
||||
WHERE feed_url=?
|
||||
""",
|
||||
[url],
|
||||
)
|
||||
|
||||
rows = cursor.fetchall()
|
||||
entries = [
|
||||
feed.Entry(id=id, inserted_at=inserted_at, title=title, link=link)
|
||||
for id, inserted_at, title, link in rows
|
||||
]
|
||||
f = feed.Feed(meta=meta, title=title, link=link, entries=entries)
|
||||
feeds.append(f)
|
||||
|
||||
return feeds
|
||||
|
||||
def load_feed(self, url: str) -> feed.Feed | None:
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
last_fetched_ts,
|
||||
retry_after_ts,
|
||||
status,
|
||||
etag,
|
||||
modified,
|
||||
title,
|
||||
link
|
||||
FROM feeds
|
||||
WHERE url=?
|
||||
""",
|
||||
[url],
|
||||
)
|
||||
|
||||
row = cursor.fetchone()
|
||||
if row is None:
|
||||
return None
|
||||
|
||||
last_fetched_ts, retry_after_ts, status, etag, modified, title, link = row
|
||||
meta = feed.FeedMeta(
|
||||
url=url,
|
||||
last_fetched_ts=last_fetched_ts,
|
||||
retry_after_ts=retry_after_ts,
|
||||
status=status,
|
||||
etag=etag,
|
||||
modified=modified,
|
||||
origin=self.origin,
|
||||
)
|
||||
|
||||
cursor = self.db.execute(
|
||||
"""
|
||||
SELECT
|
||||
id,
|
||||
inserted_at,
|
||||
title,
|
||||
link
|
||||
FROM entries
|
||||
WHERE feed_url=?
|
||||
""",
|
||||
[url],
|
||||
)
|
||||
|
||||
rows = cursor.fetchall()
|
||||
entries = [
|
||||
feed.Entry(id=id, inserted_at=inserted_at, title=title, link=link)
|
||||
for id, inserted_at, title, link in rows
|
||||
]
|
||||
|
||||
return feed.Feed(meta=meta, title=title, link=link, entries=entries)
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue