mirror of
https://github.com/qbittorrent/qBittorrent
synced 2025-07-06 13:11:25 -07:00
Store all RSS Article data in auto downloading job
This commit is contained in:
parent
475348595c
commit
080dd79875
3 changed files with 48 additions and 42 deletions
|
@ -36,47 +36,47 @@
|
|||
|
||||
#include "rss_feed.h"
|
||||
|
||||
const QString Str_Id(QStringLiteral("id"));
|
||||
const QString Str_Date(QStringLiteral("date"));
|
||||
const QString Str_Title(QStringLiteral("title"));
|
||||
const QString Str_Author(QStringLiteral("author"));
|
||||
const QString Str_Description(QStringLiteral("description"));
|
||||
const QString Str_TorrentURL(QStringLiteral("torrentURL"));
|
||||
const QString Str_Link(QStringLiteral("link"));
|
||||
const QString Str_IsRead(QStringLiteral("isRead"));
|
||||
|
||||
using namespace RSS;
|
||||
|
||||
const QString Article::KeyId(QStringLiteral("id"));
|
||||
const QString Article::KeyDate(QStringLiteral("date"));
|
||||
const QString Article::KeyTitle(QStringLiteral("title"));
|
||||
const QString Article::KeyAuthor(QStringLiteral("author"));
|
||||
const QString Article::KeyDescription(QStringLiteral("description"));
|
||||
const QString Article::KeyTorrentURL(QStringLiteral("torrentURL"));
|
||||
const QString Article::KeyLink(QStringLiteral("link"));
|
||||
const QString Article::KeyIsRead(QStringLiteral("isRead"));
|
||||
|
||||
Article::Article(Feed *feed, const QVariantHash &varHash)
|
||||
: QObject(feed)
|
||||
, m_feed(feed)
|
||||
, m_guid(varHash.value(Str_Id).toString())
|
||||
, m_date(varHash.value(Str_Date).toDateTime())
|
||||
, m_title(varHash.value(Str_Title).toString())
|
||||
, m_author(varHash.value(Str_Author).toString())
|
||||
, m_description(varHash.value(Str_Description).toString())
|
||||
, m_torrentURL(varHash.value(Str_TorrentURL).toString())
|
||||
, m_link(varHash.value(Str_Link).toString())
|
||||
, m_isRead(varHash.value(Str_IsRead, false).toBool())
|
||||
, m_guid(varHash.value(KeyId).toString())
|
||||
, m_date(varHash.value(KeyDate).toDateTime())
|
||||
, m_title(varHash.value(KeyTitle).toString())
|
||||
, m_author(varHash.value(KeyAuthor).toString())
|
||||
, m_description(varHash.value(KeyDescription).toString())
|
||||
, m_torrentURL(varHash.value(KeyTorrentURL).toString())
|
||||
, m_link(varHash.value(KeyLink).toString())
|
||||
, m_isRead(varHash.value(KeyIsRead, false).toBool())
|
||||
, m_data(varHash)
|
||||
{
|
||||
// If item does not have a guid, fall back to some other identifier
|
||||
if (m_guid.isEmpty())
|
||||
m_guid = varHash.value(Str_TorrentURL).toString();
|
||||
m_guid = varHash.value(KeyTorrentURL).toString();
|
||||
if (m_guid.isEmpty())
|
||||
m_guid = varHash.value(Str_Title).toString();
|
||||
m_guid = varHash.value(KeyTitle).toString();
|
||||
if (m_guid.isEmpty())
|
||||
throw std::runtime_error("Bad RSS Article data");
|
||||
|
||||
m_data[Str_Id] = m_guid;
|
||||
m_data[KeyId] = m_guid;
|
||||
}
|
||||
|
||||
Article::Article(Feed *feed, const QJsonObject &jsonObj)
|
||||
: Article(feed, jsonObj.toVariantHash())
|
||||
{
|
||||
// JSON object store DateTime as string so we need to convert it
|
||||
m_date = QDateTime::fromString(jsonObj.value(Str_Date).toString(), Qt::RFC2822Date);
|
||||
m_data[Str_Date] = m_date;
|
||||
m_date = QDateTime::fromString(jsonObj.value(KeyDate).toString(), Qt::RFC2822Date);
|
||||
m_data[KeyDate] = m_date;
|
||||
}
|
||||
|
||||
QString Article::guid() const
|
||||
|
@ -128,7 +128,7 @@ void Article::markAsRead()
|
|||
{
|
||||
if (!m_isRead) {
|
||||
m_isRead = true;
|
||||
m_data[Str_IsRead] = m_isRead;
|
||||
m_data[KeyIsRead] = m_isRead;
|
||||
emit read(this);
|
||||
}
|
||||
}
|
||||
|
@ -137,7 +137,7 @@ QJsonObject Article::toJsonObject() const
|
|||
{
|
||||
auto jsonObj = QJsonObject::fromVariantHash(m_data);
|
||||
// JSON object doesn't support DateTime so we need to convert it
|
||||
jsonObj[Str_Date] = m_date.toString(Qt::RFC2822Date);
|
||||
jsonObj[KeyDate] = m_date.toString(Qt::RFC2822Date);
|
||||
|
||||
return jsonObj;
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue