fix: fix mikan rss extractors
This commit is contained in:
@@ -4,21 +4,25 @@ use thiserror::Error;
|
||||
|
||||
#[derive(Error, Debug)]
|
||||
pub enum ExtractError {
|
||||
#[error("Parse bangumi season error: {0}")]
|
||||
#[error("Extract bangumi season error: {0}")]
|
||||
BangumiSeasonError(#[from] std::num::ParseIntError),
|
||||
#[error("Parse file url error: {0}")]
|
||||
#[error("Extract file url error: {0}")]
|
||||
FileUrlError(#[from] url::ParseError),
|
||||
#[error("Parse {desc} with mime error, expected {expected}, but got {found}")]
|
||||
#[error("Extract {desc} with mime error, expected {expected}, but got {found}")]
|
||||
MimeError {
|
||||
desc: String,
|
||||
expected: String,
|
||||
found: String,
|
||||
},
|
||||
#[error("Parse mikan rss {url} format error")]
|
||||
MikanRssFormatError { url: String },
|
||||
#[error("Parse mikan rss item format error, {reason}")]
|
||||
MikanRssItemFormatError { reason: String },
|
||||
#[error("Missing field {field} in extracting meta")]
|
||||
#[error("Invalid or unknown format in extracting mikan rss")]
|
||||
MikanRssInvalidFormatError,
|
||||
#[error("Invalid field {field} in extracting mikan rss")]
|
||||
MikanRssInvalidFieldError {
|
||||
field: Cow<'static, str>,
|
||||
#[source]
|
||||
source: Option<Box<dyn StdError + Send + Sync>>,
|
||||
},
|
||||
#[error("Missing field {field} in extracting mikan meta")]
|
||||
MikanMetaMissingFieldError {
|
||||
field: Cow<'static, str>,
|
||||
#[source]
|
||||
@@ -33,4 +37,21 @@ impl ExtractError {
|
||||
source: None,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn from_mikan_rss_invalid_field(field: Cow<'static, str>) -> Self {
|
||||
Self::MikanRssInvalidFieldError {
|
||||
field,
|
||||
source: None,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn from_mikan_rss_invalid_field_and_source(
|
||||
field: Cow<'static, str>,
|
||||
source: Box<dyn StdError + Send + Sync>,
|
||||
) -> Self {
|
||||
Self::MikanRssInvalidFieldError {
|
||||
field,
|
||||
source: Some(source),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3,10 +3,14 @@ use std::ops::Deref;
|
||||
use async_trait::async_trait;
|
||||
use loco_rs::app::{AppContext, Initializer};
|
||||
use once_cell::sync::OnceCell;
|
||||
use reqwest_middleware::ClientWithMiddleware;
|
||||
use url::Url;
|
||||
|
||||
use super::AppMikanConfig;
|
||||
use crate::{config::AppConfigExt, fetch::HttpClient};
|
||||
use crate::{
|
||||
config::AppConfigExt,
|
||||
fetch::{HttpClient, HttpClientTrait},
|
||||
};
|
||||
|
||||
static APP_MIKAN_CLIENT: OnceCell<AppMikanClient> = OnceCell::new();
|
||||
|
||||
@@ -39,13 +43,15 @@ impl AppMikanClient {
|
||||
}
|
||||
|
||||
impl Deref for AppMikanClient {
|
||||
type Target = HttpClient;
|
||||
type Target = ClientWithMiddleware;
|
||||
|
||||
fn deref(&self) -> &Self::Target {
|
||||
&self.http_client
|
||||
self.http_client.deref()
|
||||
}
|
||||
}
|
||||
|
||||
impl HttpClientTrait for AppMikanClient {}
|
||||
|
||||
pub struct AppMikanClientInitializer;
|
||||
|
||||
#[async_trait]
|
||||
|
||||
@@ -12,8 +12,7 @@ pub use rss_extract::{
|
||||
MikanRssChannel, MikanRssItem, MikanSubscriberAggregationRssChannel,
|
||||
MikanSubscriberAggregationRssLink, build_mikan_bangumi_rss_link,
|
||||
build_mikan_subscriber_aggregation_rss_link, extract_mikan_bangumi_id_from_rss_link,
|
||||
extract_mikan_subscriber_aggregation_id_from_rss_link, parse_mikan_rss_channel_from_rss_link,
|
||||
parse_mikan_rss_items_from_rss_link,
|
||||
extract_mikan_rss_channel_from_rss_link, extract_mikan_subscriber_aggregation_id_from_rss_link,
|
||||
};
|
||||
pub use web_extract::{
|
||||
MikanBangumiMeta, MikanEpisodeMeta, build_mikan_bangumi_homepage, build_mikan_episode_homepage,
|
||||
|
||||
@@ -1,10 +1,11 @@
|
||||
use std::ops::Deref;
|
||||
use std::borrow::Cow;
|
||||
|
||||
use chrono::DateTime;
|
||||
use color_eyre::eyre;
|
||||
use itertools::Itertools;
|
||||
use reqwest::IntoUrl;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use tracing::instrument;
|
||||
use url::Url;
|
||||
|
||||
use crate::{
|
||||
@@ -105,50 +106,55 @@ impl TryFrom<rss::Item> for MikanRssItem {
|
||||
type Error = ExtractError;
|
||||
|
||||
fn try_from(item: rss::Item) -> Result<Self, Self::Error> {
|
||||
let mime_type = item
|
||||
.enclosure()
|
||||
.map(|x| x.mime_type.to_string())
|
||||
.unwrap_or_default();
|
||||
if mime_type == BITTORRENT_MIME_TYPE {
|
||||
let enclosure = item.enclosure.unwrap();
|
||||
let enclosure = item.enclosure.ok_or_else(|| {
|
||||
ExtractError::from_mikan_rss_invalid_field(Cow::Borrowed("enclosure"))
|
||||
})?;
|
||||
|
||||
let homepage = item
|
||||
.link
|
||||
.ok_or_else(|| ExtractError::MikanRssItemFormatError {
|
||||
reason: String::from("must to have link for homepage"),
|
||||
})?;
|
||||
let mime_type = enclosure.mime_type;
|
||||
if mime_type != BITTORRENT_MIME_TYPE {
|
||||
return Err(ExtractError::MimeError {
|
||||
expected: String::from(BITTORRENT_MIME_TYPE),
|
||||
found: mime_type.to_string(),
|
||||
desc: String::from("MikanRssItem"),
|
||||
});
|
||||
}
|
||||
|
||||
let homepage = Url::parse(&homepage)?;
|
||||
let title = item.title.ok_or_else(|| {
|
||||
ExtractError::from_mikan_rss_invalid_field(Cow::Borrowed("title:title"))
|
||||
})?;
|
||||
|
||||
let enclosure_url = Url::parse(&enclosure.url)?;
|
||||
let enclosure_url = Url::parse(&enclosure.url).map_err(|inner| {
|
||||
ExtractError::from_mikan_rss_invalid_field_and_source(
|
||||
Cow::Borrowed("enclosure_url:enclosure.link"),
|
||||
Box::new(inner),
|
||||
)
|
||||
})?;
|
||||
|
||||
let MikanEpisodeHomepage {
|
||||
mikan_episode_id, ..
|
||||
} = parse_mikan_episode_id_from_homepage(&homepage).ok_or_else(|| {
|
||||
ExtractError::MikanRssItemFormatError {
|
||||
reason: String::from("homepage link format invalid"),
|
||||
}
|
||||
let homepage = item
|
||||
.link
|
||||
.and_then(|link| Url::parse(&link).ok())
|
||||
.ok_or_else(|| {
|
||||
ExtractError::from_mikan_rss_invalid_field(Cow::Borrowed("homepage:link"))
|
||||
})?;
|
||||
|
||||
Ok(MikanRssItem {
|
||||
title: item.title.unwrap_or_default(),
|
||||
homepage,
|
||||
url: enclosure_url,
|
||||
content_length: enclosure.length.parse().ok(),
|
||||
mime: enclosure.mime_type,
|
||||
pub_date: item
|
||||
.pub_date
|
||||
.and_then(|s| DateTime::parse_from_rfc2822(&s).ok())
|
||||
.map(|s| s.timestamp_millis()),
|
||||
mikan_episode_id,
|
||||
})
|
||||
} else {
|
||||
Err(ExtractError::MimeError {
|
||||
expected: String::from(BITTORRENT_MIME_TYPE),
|
||||
found: mime_type,
|
||||
desc: String::from("MikanRssItem"),
|
||||
})
|
||||
}
|
||||
let MikanEpisodeHomepage {
|
||||
mikan_episode_id, ..
|
||||
} = parse_mikan_episode_id_from_homepage(&homepage).ok_or_else(|| {
|
||||
ExtractError::from_mikan_rss_invalid_field(Cow::Borrowed("mikan_episode_id"))
|
||||
})?;
|
||||
|
||||
Ok(MikanRssItem {
|
||||
title,
|
||||
homepage,
|
||||
url: enclosure_url,
|
||||
content_length: enclosure.length.parse().ok(),
|
||||
mime: mime_type,
|
||||
pub_date: item
|
||||
.pub_date
|
||||
.and_then(|s| DateTime::parse_from_rfc2822(&s).ok())
|
||||
.map(|s| s.timestamp_millis()),
|
||||
mikan_episode_id,
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
@@ -220,21 +226,12 @@ pub fn extract_mikan_subscriber_aggregation_id_from_rss_link(
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn parse_mikan_rss_items_from_rss_link(
|
||||
client: Option<&AppMikanClient>,
|
||||
url: impl IntoUrl,
|
||||
) -> eyre::Result<Vec<MikanRssItem>> {
|
||||
let channel = parse_mikan_rss_channel_from_rss_link(client, url).await?;
|
||||
|
||||
Ok(channel.into_items())
|
||||
}
|
||||
|
||||
pub async fn parse_mikan_rss_channel_from_rss_link(
|
||||
client: Option<&AppMikanClient>,
|
||||
url: impl IntoUrl,
|
||||
#[instrument(skip_all, fields(channel_rss_link = channel_rss_link.as_str()))]
|
||||
pub async fn extract_mikan_rss_channel_from_rss_link(
|
||||
http_client: &AppMikanClient,
|
||||
channel_rss_link: impl IntoUrl,
|
||||
) -> eyre::Result<MikanRssChannel> {
|
||||
let http_client = client.map(|s| s.deref());
|
||||
let bytes = fetch_bytes(http_client, url.as_str()).await?;
|
||||
let bytes = fetch_bytes(http_client, channel_rss_link.as_str()).await?;
|
||||
|
||||
let channel = rss::Channel::read_from(&bytes[..])?;
|
||||
|
||||
@@ -245,16 +242,34 @@ pub async fn parse_mikan_rss_channel_from_rss_link(
|
||||
mikan_fansub_id,
|
||||
}) = extract_mikan_bangumi_id_from_rss_link(&channel_link)
|
||||
{
|
||||
tracing::trace!(
|
||||
mikan_bangumi_id,
|
||||
mikan_fansub_id,
|
||||
"MikanBangumiRssLink extracting..."
|
||||
);
|
||||
|
||||
let channel_name = channel.title().replace("Mikan Project - ", "");
|
||||
|
||||
let items = channel
|
||||
.items
|
||||
.into_iter()
|
||||
// @TODO log error
|
||||
.flat_map(MikanRssItem::try_from)
|
||||
.enumerate()
|
||||
.flat_map(|(idx, item)| {
|
||||
MikanRssItem::try_from(item).inspect_err(
|
||||
|error| tracing::warn!(error = %error, "failed to extract rss item idx = {}", idx),
|
||||
)
|
||||
})
|
||||
.collect_vec();
|
||||
|
||||
if let Some(mikan_fansub_id) = mikan_fansub_id {
|
||||
tracing::trace!(
|
||||
channel_name,
|
||||
channel_link = channel_link.as_str(),
|
||||
mikan_bangumi_id,
|
||||
mikan_fansub_id,
|
||||
"MikanBangumiRssChannel extracted"
|
||||
);
|
||||
|
||||
Ok(MikanRssChannel::Bangumi(MikanBangumiRssChannel {
|
||||
name: channel_name,
|
||||
mikan_bangumi_id,
|
||||
@@ -263,6 +278,13 @@ pub async fn parse_mikan_rss_channel_from_rss_link(
|
||||
items,
|
||||
}))
|
||||
} else {
|
||||
tracing::trace!(
|
||||
channel_name,
|
||||
channel_link = channel_link.as_str(),
|
||||
mikan_bangumi_id,
|
||||
"MikanBangumiAggregationRssChannel extracted"
|
||||
);
|
||||
|
||||
Ok(MikanRssChannel::BangumiAggregation(
|
||||
MikanBangumiAggregationRssChannel {
|
||||
name: channel_name,
|
||||
@@ -277,25 +299,41 @@ pub async fn parse_mikan_rss_channel_from_rss_link(
|
||||
..
|
||||
}) = extract_mikan_subscriber_aggregation_id_from_rss_link(&channel_link)
|
||||
{
|
||||
tracing::trace!(
|
||||
mikan_aggregation_id,
|
||||
"MikanSubscriberAggregationRssLink extracting..."
|
||||
);
|
||||
|
||||
let items = channel
|
||||
.items
|
||||
.into_iter()
|
||||
// @TODO log error
|
||||
.flat_map(MikanRssItem::try_from)
|
||||
.enumerate()
|
||||
.flat_map(|(idx, item)| {
|
||||
MikanRssItem::try_from(item).inspect_err(
|
||||
|error| tracing::warn!(error = %error, "failed to extract rss item idx = {}", idx),
|
||||
)
|
||||
})
|
||||
.collect_vec();
|
||||
|
||||
return Ok(MikanRssChannel::SubscriberAggregation(
|
||||
tracing::trace!(
|
||||
channel_link = channel_link.as_str(),
|
||||
mikan_aggregation_id,
|
||||
"MikanSubscriberAggregationRssChannel extracted"
|
||||
);
|
||||
|
||||
Ok(MikanRssChannel::SubscriberAggregation(
|
||||
MikanSubscriberAggregationRssChannel {
|
||||
mikan_aggregation_id,
|
||||
items,
|
||||
url: channel_link,
|
||||
},
|
||||
));
|
||||
))
|
||||
} else {
|
||||
return Err(ExtractError::MikanRssFormatError {
|
||||
url: url.as_str().into(),
|
||||
}
|
||||
.into());
|
||||
Err(ExtractError::MikanRssInvalidFormatError)
|
||||
.inspect_err(|error| {
|
||||
tracing::warn!(error = %error);
|
||||
})
|
||||
.map_err(|error| error.into())
|
||||
}
|
||||
}
|
||||
|
||||
@@ -303,20 +341,39 @@ pub async fn parse_mikan_rss_channel_from_rss_link(
|
||||
mod tests {
|
||||
use std::assert_matches::assert_matches;
|
||||
|
||||
use color_eyre::eyre;
|
||||
use rstest::rstest;
|
||||
use url::Url;
|
||||
|
||||
use crate::{
|
||||
extract::mikan::{
|
||||
MikanBangumiAggregationRssChannel, MikanBangumiRssChannel, MikanRssChannel,
|
||||
parse_mikan_rss_channel_from_rss_link,
|
||||
extract_mikan_rss_channel_from_rss_link,
|
||||
},
|
||||
sync::core::BITTORRENT_MIME_TYPE,
|
||||
test_utils::mikan::build_testing_mikan_client,
|
||||
};
|
||||
|
||||
#[rstest]
|
||||
#[tokio::test]
|
||||
pub async fn test_parse_mikan_rss_channel_from_rss_link() {
|
||||
{
|
||||
let bangumi_url = "https://mikanani.me/RSS/Bangumi?bangumiId=3141&subgroupid=370";
|
||||
async fn test_parse_mikan_rss_channel_from_rss_link() -> eyre::Result<()> {
|
||||
let mut mikan_server = mockito::Server::new_async().await;
|
||||
|
||||
let channel = parse_mikan_rss_channel_from_rss_link(None, bangumi_url)
|
||||
let mikan_base_url = Url::parse(&mikan_server.url())?;
|
||||
|
||||
let mikan_client = build_testing_mikan_client(mikan_base_url.clone())?;
|
||||
|
||||
{
|
||||
let bangumi_rss_url =
|
||||
mikan_base_url.join("/RSS/Bangumi?bangumiId=3141&subgroupid=370")?;
|
||||
let bangumi_rss_mock = mikan_server
|
||||
.mock("GET", bangumi_rss_url.path())
|
||||
.with_body_from_file("tests/resources/mikan/Bangumi-3141-370.rss")
|
||||
.match_query(mockito::Matcher::Any)
|
||||
.create_async()
|
||||
.await;
|
||||
|
||||
let channel = extract_mikan_rss_channel_from_rss_link(&mikan_client, bangumi_rss_url)
|
||||
.await
|
||||
.expect("should get mikan channel from rss url");
|
||||
|
||||
@@ -343,11 +400,20 @@ mod tests {
|
||||
|
||||
let name = first_sub_item.title.as_str();
|
||||
assert!(name.contains("葬送的芙莉莲"));
|
||||
|
||||
bangumi_rss_mock.expect(1);
|
||||
}
|
||||
{
|
||||
let bangumi_url = "https://mikanani.me/RSS/Bangumi?bangumiId=3416";
|
||||
let bangumi_rss_url = mikan_base_url.join("/RSS/Bangumi?bangumiId=3416")?;
|
||||
|
||||
let channel = parse_mikan_rss_channel_from_rss_link(None, bangumi_url)
|
||||
let bangumi_rss_mock = mikan_server
|
||||
.mock("GET", bangumi_rss_url.path())
|
||||
.match_query(mockito::Matcher::Any)
|
||||
.with_body_from_file("tests/resources/mikan/Bangumi-3416.rss")
|
||||
.create_async()
|
||||
.await;
|
||||
|
||||
let channel = extract_mikan_rss_channel_from_rss_link(&mikan_client, bangumi_rss_url)
|
||||
.await
|
||||
.expect("should get mikan channel from rss url");
|
||||
|
||||
@@ -357,6 +423,9 @@ mod tests {
|
||||
);
|
||||
|
||||
assert_matches!(&channel.name(), Some("叹气的亡灵想隐退"));
|
||||
|
||||
bangumi_rss_mock.expect(1);
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
use std::{borrow::Cow, ops::Deref};
|
||||
use std::borrow::Cow;
|
||||
|
||||
use bytes::Bytes;
|
||||
use color_eyre::eyre;
|
||||
@@ -117,10 +117,9 @@ pub fn parse_mikan_episode_id_from_homepage(url: &Url) -> Option<MikanEpisodeHom
|
||||
}
|
||||
|
||||
pub async fn extract_mikan_poster_meta_from_src(
|
||||
client: Option<&AppMikanClient>,
|
||||
http_client: &AppMikanClient,
|
||||
origin_poster_src_url: Url,
|
||||
) -> eyre::Result<MikanBangumiPosterMeta> {
|
||||
let http_client = client.map(|s| s.deref());
|
||||
let poster_data = fetch_image(http_client, origin_poster_src_url.clone()).await?;
|
||||
Ok(MikanBangumiPosterMeta {
|
||||
origin_poster_src: origin_poster_src_url,
|
||||
@@ -152,8 +151,7 @@ pub async fn extract_mikan_bangumi_poster_meta_from_src_with_cache(
|
||||
});
|
||||
}
|
||||
|
||||
let poster_data =
|
||||
fetch_image(Some(mikan_client.deref()), origin_poster_src_url.clone()).await?;
|
||||
let poster_data = fetch_image(mikan_client, origin_poster_src_url.clone()).await?;
|
||||
|
||||
let poster_str = dal_client
|
||||
.store_object(
|
||||
@@ -174,10 +172,9 @@ pub async fn extract_mikan_bangumi_poster_meta_from_src_with_cache(
|
||||
|
||||
#[instrument(skip_all, fields(mikan_episode_homepage_url = mikan_episode_homepage_url.as_str()))]
|
||||
pub async fn extract_mikan_episode_meta_from_episode_homepage(
|
||||
client: Option<&AppMikanClient>,
|
||||
http_client: &AppMikanClient,
|
||||
mikan_episode_homepage_url: Url,
|
||||
) -> eyre::Result<MikanEpisodeMeta> {
|
||||
let http_client = client.map(|s| s.deref());
|
||||
let mikan_base_url = Url::parse(&mikan_episode_homepage_url.origin().unicode_serialization())?;
|
||||
let content = fetch_html(http_client, mikan_episode_homepage_url.as_str()).await?;
|
||||
|
||||
@@ -286,10 +283,9 @@ pub async fn extract_mikan_episode_meta_from_episode_homepage(
|
||||
|
||||
#[instrument(skip_all, fields(mikan_bangumi_homepage_url = mikan_bangumi_homepage_url.as_str()))]
|
||||
pub async fn extract_mikan_bangumi_meta_from_bangumi_homepage(
|
||||
client: Option<&AppMikanClient>,
|
||||
http_client: &AppMikanClient,
|
||||
mikan_bangumi_homepage_url: Url,
|
||||
) -> eyre::Result<MikanBangumiMeta> {
|
||||
let http_client = client.map(|s| s.deref());
|
||||
let mikan_base_url = Url::parse(&mikan_bangumi_homepage_url.origin().unicode_serialization())?;
|
||||
let content = fetch_html(http_client, mikan_bangumi_homepage_url.as_str()).await?;
|
||||
let html = Html::parse_document(&content);
|
||||
@@ -369,10 +365,9 @@ pub async fn extract_mikan_bangumi_meta_from_bangumi_homepage(
|
||||
*/
|
||||
#[instrument(skip_all, fields(my_bangumi_page_url = my_bangumi_page_url.as_str()))]
|
||||
pub async fn extract_mikan_bangumis_meta_from_my_bangumi_page(
|
||||
client: Option<&AppMikanClient>,
|
||||
http_client: &AppMikanClient,
|
||||
my_bangumi_page_url: Url,
|
||||
) -> eyre::Result<Vec<MikanBangumiMeta>> {
|
||||
let http_client = client.map(|c| c.deref());
|
||||
let mikan_base_url = Url::parse(&my_bangumi_page_url.origin().unicode_serialization())?;
|
||||
|
||||
let content = fetch_html(http_client, my_bangumi_page_url.clone()).await?;
|
||||
@@ -506,7 +501,7 @@ mod test {
|
||||
.await;
|
||||
|
||||
let bgm_poster =
|
||||
extract_mikan_poster_meta_from_src(Some(&mikan_client), bangumi_poster_url).await?;
|
||||
extract_mikan_poster_meta_from_src(&mikan_client, bangumi_poster_url).await?;
|
||||
bangumi_poster_mock.expect(1);
|
||||
let u8_data = bgm_poster.poster_data.expect("should have poster data");
|
||||
let image = Image::read(u8_data.to_vec(), Default::default());
|
||||
@@ -540,7 +535,7 @@ mod test {
|
||||
.await;
|
||||
|
||||
let ep_meta = extract_mikan_episode_meta_from_episode_homepage(
|
||||
Some(&mikan_client),
|
||||
&mikan_client,
|
||||
episode_homepage_url.clone(),
|
||||
)
|
||||
.await?;
|
||||
@@ -579,7 +574,7 @@ mod test {
|
||||
.await;
|
||||
|
||||
let bgm_meta = extract_mikan_bangumi_meta_from_bangumi_homepage(
|
||||
Some(&mikan_client),
|
||||
&mikan_client,
|
||||
bangumi_homepage_url.clone(),
|
||||
)
|
||||
.await?;
|
||||
@@ -613,31 +608,29 @@ mod test {
|
||||
|
||||
let my_bangumi_page_url = mikan_base_url.join("/Home/MyBangumi")?;
|
||||
|
||||
let mock_my_bangumi = mikan_server
|
||||
let my_bangumi_mock = mikan_server
|
||||
.mock("GET", my_bangumi_page_url.path())
|
||||
.with_body_from_file("tests/resources/mikan/MyBangumi.htm")
|
||||
.create_async()
|
||||
.await;
|
||||
|
||||
let mock_expand_bangumi = mikan_server
|
||||
let expand_bangumi_mock = mikan_server
|
||||
.mock("GET", "/ExpandBangumi")
|
||||
.match_query(mockito::Matcher::Any)
|
||||
.with_body_from_file("tests/resources/mikan/ExpandBangumi.htm")
|
||||
.create_async()
|
||||
.await;
|
||||
|
||||
let bangumi_metas = extract_mikan_bangumis_meta_from_my_bangumi_page(
|
||||
Some(&mikan_client),
|
||||
my_bangumi_page_url,
|
||||
)
|
||||
.await?;
|
||||
let bangumi_metas =
|
||||
extract_mikan_bangumis_meta_from_my_bangumi_page(&mikan_client, my_bangumi_page_url)
|
||||
.await?;
|
||||
|
||||
assert!(!bangumi_metas.is_empty());
|
||||
|
||||
assert!(bangumi_metas[0].origin_poster_src.is_some());
|
||||
|
||||
mock_my_bangumi.expect(1);
|
||||
mock_expand_bangumi.expect(bangumi_metas.len());
|
||||
my_bangumi_mock.expect(1);
|
||||
expand_bangumi_mock.expect(bangumi_metas.len());
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@@ -1,11 +1,12 @@
|
||||
use bytes::Bytes;
|
||||
use reqwest::IntoUrl;
|
||||
|
||||
use super::HttpClient;
|
||||
|
||||
pub async fn fetch_bytes<T: IntoUrl>(client: Option<&HttpClient>, url: T) -> color_eyre::eyre::Result<Bytes> {
|
||||
let client = client.unwrap_or_default();
|
||||
use super::client::HttpClientTrait;
|
||||
|
||||
pub async fn fetch_bytes<T: IntoUrl, H: HttpClientTrait>(
|
||||
client: &H,
|
||||
url: T,
|
||||
) -> color_eyre::eyre::Result<Bytes> {
|
||||
let bytes = client
|
||||
.get(url)
|
||||
.send()
|
||||
|
||||
@@ -6,7 +6,6 @@ use http_cache_reqwest::{
|
||||
CACacheManager, Cache, CacheManager, CacheMode, HttpCache, HttpCacheOptions, MokaManager,
|
||||
};
|
||||
use leaky_bucket::RateLimiter;
|
||||
use once_cell::sync::OnceCell;
|
||||
use reqwest::{ClientBuilder, Request, Response};
|
||||
use reqwest_middleware::{
|
||||
ClientBuilder as ClientWithMiddlewareBuilder, ClientWithMiddleware, Next,
|
||||
@@ -20,6 +19,23 @@ use thiserror::Error;
|
||||
use super::get_random_mobile_ua;
|
||||
use crate::app::App;
|
||||
|
||||
pub struct RateLimiterMiddleware {
|
||||
rate_limiter: RateLimiter,
|
||||
}
|
||||
|
||||
#[async_trait]
|
||||
impl reqwest_middleware::Middleware for RateLimiterMiddleware {
|
||||
async fn handle(
|
||||
&self,
|
||||
req: Request,
|
||||
extensions: &'_ mut Extensions,
|
||||
next: Next<'_>,
|
||||
) -> reqwest_middleware::Result<Response> {
|
||||
self.rate_limiter.acquire_one().await;
|
||||
next.run(req, extensions).await
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, PartialEq, Eq, Serialize, Deserialize)]
|
||||
#[serde(rename_all = "snake_case", tag = "type")]
|
||||
pub enum HttpClientCacheBackendConfig {
|
||||
@@ -96,6 +112,8 @@ pub enum HttpClientError {
|
||||
HttpError(#[from] http::Error),
|
||||
}
|
||||
|
||||
pub trait HttpClientTrait: Deref<Target = ClientWithMiddleware> {}
|
||||
|
||||
pub struct HttpClient {
|
||||
client: ClientWithMiddleware,
|
||||
pub config: HttpClientConfig,
|
||||
@@ -123,23 +141,6 @@ impl Deref for HttpClient {
|
||||
}
|
||||
}
|
||||
|
||||
pub struct RateLimiterMiddleware {
|
||||
rate_limiter: RateLimiter,
|
||||
}
|
||||
|
||||
#[async_trait]
|
||||
impl reqwest_middleware::Middleware for RateLimiterMiddleware {
|
||||
async fn handle(
|
||||
&self,
|
||||
req: Request,
|
||||
extensions: &'_ mut Extensions,
|
||||
next: Next<'_>,
|
||||
) -> reqwest_middleware::Result<Response> {
|
||||
self.rate_limiter.acquire_one().await;
|
||||
next.run(req, extensions).await
|
||||
}
|
||||
}
|
||||
|
||||
impl HttpClient {
|
||||
pub fn from_config(config: HttpClientConfig) -> Result<Self, HttpClientError> {
|
||||
let reqwest_client_builder = ClientBuilder::new().user_agent(
|
||||
@@ -234,16 +235,10 @@ impl HttpClient {
|
||||
}
|
||||
}
|
||||
|
||||
static DEFAULT_HTTP_CLIENT: OnceCell<HttpClient> = OnceCell::new();
|
||||
|
||||
impl Default for HttpClient {
|
||||
fn default() -> Self {
|
||||
HttpClient::from_config(Default::default()).expect("Failed to create default HttpClient")
|
||||
}
|
||||
}
|
||||
|
||||
impl Default for &HttpClient {
|
||||
fn default() -> Self {
|
||||
DEFAULT_HTTP_CLIENT.get_or_init(HttpClient::default)
|
||||
}
|
||||
}
|
||||
impl HttpClientTrait for HttpClient {}
|
||||
|
||||
@@ -1,12 +1,11 @@
|
||||
use reqwest::IntoUrl;
|
||||
|
||||
use super::HttpClient;
|
||||
use super::client::HttpClientTrait;
|
||||
|
||||
pub async fn fetch_html<T: IntoUrl>(
|
||||
client: Option<&HttpClient>,
|
||||
pub async fn fetch_html<T: IntoUrl, H: HttpClientTrait>(
|
||||
client: &H,
|
||||
url: T,
|
||||
) -> color_eyre::eyre::Result<String> {
|
||||
let client = client.unwrap_or_default();
|
||||
let content = client
|
||||
.get(url)
|
||||
.send()
|
||||
|
||||
@@ -1,8 +1,11 @@
|
||||
use bytes::Bytes;
|
||||
use reqwest::IntoUrl;
|
||||
|
||||
use super::{bytes::fetch_bytes, HttpClient};
|
||||
use super::{bytes::fetch_bytes, client::HttpClientTrait};
|
||||
|
||||
pub async fn fetch_image<T: IntoUrl>(client: Option<&HttpClient>, url: T) -> color_eyre::eyre::Result<Bytes> {
|
||||
pub async fn fetch_image<T: IntoUrl, H: HttpClientTrait>(
|
||||
client: &H,
|
||||
url: T,
|
||||
) -> color_eyre::eyre::Result<Bytes> {
|
||||
fetch_bytes(client, url).await
|
||||
}
|
||||
|
||||
@@ -8,6 +8,6 @@ pub mod oidc;
|
||||
pub use core::get_random_mobile_ua;
|
||||
|
||||
pub use bytes::fetch_bytes;
|
||||
pub use client::{HttpClient, HttpClientConfig, HttpClientError};
|
||||
pub use client::{HttpClient, HttpClientConfig, HttpClientError, HttpClientTrait};
|
||||
pub use html::fetch_html;
|
||||
pub use image::fetch_image;
|
||||
|
||||
@@ -14,7 +14,7 @@ use crate::{
|
||||
build_mikan_bangumi_homepage, build_mikan_bangumi_rss_link,
|
||||
extract_mikan_bangumi_meta_from_bangumi_homepage,
|
||||
extract_mikan_episode_meta_from_episode_homepage,
|
||||
parse_mikan_rss_channel_from_rss_link,
|
||||
extract_mikan_rss_channel_from_rss_link,
|
||||
web_extract::{
|
||||
MikanBangumiPosterMeta, extract_mikan_bangumi_poster_meta_from_src_with_cache,
|
||||
},
|
||||
@@ -220,8 +220,7 @@ impl Model {
|
||||
SubscriptionCategory::Mikan => {
|
||||
let mikan_client = ctx.get_mikan_client();
|
||||
let channel =
|
||||
parse_mikan_rss_channel_from_rss_link(Some(mikan_client), &self.source_url)
|
||||
.await?;
|
||||
extract_mikan_rss_channel_from_rss_link(mikan_client, &self.source_url).await?;
|
||||
|
||||
let items = channel.into_items();
|
||||
|
||||
@@ -257,7 +256,7 @@ impl Model {
|
||||
for new_rss_item in new_rss_items.iter() {
|
||||
new_metas.push(
|
||||
extract_mikan_episode_meta_from_episode_homepage(
|
||||
Some(mikan_client),
|
||||
mikan_client,
|
||||
new_rss_item.homepage.clone(),
|
||||
)
|
||||
.await?,
|
||||
@@ -290,7 +289,7 @@ impl Model {
|
||||
mikan_fansub_id.to_string(),
|
||||
async |am| -> color_eyre::eyre::Result<()> {
|
||||
let bgm_meta = extract_mikan_bangumi_meta_from_bangumi_homepage(
|
||||
Some(mikan_client),
|
||||
mikan_client,
|
||||
bgm_homepage.clone(),
|
||||
)
|
||||
.await?;
|
||||
|
||||
@@ -5,7 +5,7 @@ use itertools::Itertools;
|
||||
use lazy_static::lazy_static;
|
||||
use librqbit_core::{
|
||||
magnet::Magnet,
|
||||
torrent_metainfo::{torrent_from_bytes, TorrentMetaV1Owned},
|
||||
torrent_metainfo::{TorrentMetaV1Owned, torrent_from_bytes},
|
||||
};
|
||||
use quirks_path::{Path, PathBuf};
|
||||
use regex::Regex;
|
||||
@@ -13,7 +13,7 @@ use serde::{Deserialize, Serialize};
|
||||
use url::Url;
|
||||
|
||||
use super::{QbitTorrent, QbitTorrentContent, TorrentDownloadError};
|
||||
use crate::fetch::{fetch_bytes, HttpClient};
|
||||
use crate::fetch::{HttpClientTrait, fetch_bytes};
|
||||
|
||||
pub const BITTORRENT_MIME_TYPE: &str = "application/x-bittorrent";
|
||||
pub const MAGNET_SCHEMA: &str = "magnet";
|
||||
@@ -57,7 +57,10 @@ pub enum TorrentSource {
|
||||
}
|
||||
|
||||
impl TorrentSource {
|
||||
pub async fn parse(client: Option<&HttpClient>, url: &str) -> color_eyre::eyre::Result<Self> {
|
||||
pub async fn parse<H: HttpClientTrait>(
|
||||
client: &H,
|
||||
url: &str,
|
||||
) -> color_eyre::eyre::Result<Self> {
|
||||
let url = Url::parse(url)?;
|
||||
let source = if url.scheme() == MAGNET_SCHEMA {
|
||||
TorrentSource::from_magnet_url(url)?
|
||||
|
||||
@@ -10,8 +10,8 @@ pub use qbit_rs::model::{
|
||||
TorrentFilter as QbitTorrentFilter, TorrentSource as QbitTorrentSource,
|
||||
};
|
||||
use qbit_rs::{
|
||||
model::{AddTorrentArg, Credential, GetTorrentListArg, NonEmptyStr, SyncData},
|
||||
Qbit,
|
||||
model::{AddTorrentArg, Credential, GetTorrentListArg, NonEmptyStr, SyncData},
|
||||
};
|
||||
use quirks_path::{Path, PathBuf};
|
||||
use tokio::time::sleep;
|
||||
@@ -19,8 +19,8 @@ use tracing::instrument;
|
||||
use url::Url;
|
||||
|
||||
use super::{
|
||||
utils::path_equals_as_file_url, Torrent, TorrentDownloadError, TorrentDownloader,
|
||||
TorrentFilter, TorrentSource,
|
||||
Torrent, TorrentDownloadError, TorrentDownloader, TorrentFilter, TorrentSource,
|
||||
utils::path_equals_as_file_url,
|
||||
};
|
||||
|
||||
impl From<TorrentSource> for QbitTorrentSource {
|
||||
@@ -490,6 +490,7 @@ pub mod tests {
|
||||
use itertools::Itertools;
|
||||
|
||||
use super::*;
|
||||
use crate::test_utils::fetch::build_testing_http_client;
|
||||
|
||||
fn get_tmp_qbit_test_folder() -> &'static str {
|
||||
if cfg!(all(windows, not(feature = "testcontainers"))) {
|
||||
@@ -500,16 +501,16 @@ pub mod tests {
|
||||
}
|
||||
|
||||
#[cfg(feature = "testcontainers")]
|
||||
pub async fn create_qbit_testcontainer(
|
||||
) -> color_eyre::eyre::Result<testcontainers::ContainerRequest<testcontainers::GenericImage>>
|
||||
pub async fn create_qbit_testcontainer()
|
||||
-> color_eyre::eyre::Result<testcontainers::ContainerRequest<testcontainers::GenericImage>>
|
||||
{
|
||||
use testcontainers::{
|
||||
GenericImage,
|
||||
core::{
|
||||
ContainerPort,
|
||||
// ReuseDirective,
|
||||
WaitFor,
|
||||
},
|
||||
GenericImage,
|
||||
};
|
||||
use testcontainers_modules::testcontainers::ImageExt;
|
||||
|
||||
@@ -590,6 +591,7 @@ pub mod tests {
|
||||
username: Option<&str>,
|
||||
password: Option<&str>,
|
||||
) -> color_eyre::eyre::Result<()> {
|
||||
let http_client = build_testing_http_client()?;
|
||||
let base_save_path = Path::new(get_tmp_qbit_test_folder());
|
||||
|
||||
let mut downloader = QBittorrentDownloader::from_creation(QBittorrentDownloaderCreation {
|
||||
@@ -610,7 +612,7 @@ pub mod tests {
|
||||
.await?;
|
||||
|
||||
let torrent_source = TorrentSource::parse(
|
||||
None,
|
||||
&http_client,
|
||||
"https://mikanani.me/Download/20240301/47ee2d69e7f19af783ad896541a07b012676f858.torrent"
|
||||
).await?;
|
||||
|
||||
|
||||
@@ -0,0 +1,26 @@
|
||||
use loco_rs::prelude::*;
|
||||
use secrecy::SecretString;
|
||||
|
||||
#[derive(Clone, Debug)]
|
||||
pub struct CreateMikanRSSFromMyBangumiTask {
|
||||
subscriber_id: i32,
|
||||
task_id: String,
|
||||
cookie: SecretString,
|
||||
}
|
||||
|
||||
#[async_trait::async_trait]
|
||||
impl Task for CreateMikanRSSFromMyBangumiTask {
|
||||
fn task(&self) -> TaskInfo {
|
||||
TaskInfo {
|
||||
name: format!(
|
||||
"create-mikan-rss-from-my-bangumi-{}-{}",
|
||||
self.subscriber_id, self.task_id
|
||||
),
|
||||
detail: "create mikan rss from my bangumi page for {} {}".to_string(),
|
||||
}
|
||||
}
|
||||
|
||||
async fn run(&self, _app_context: &AppContext, _vars: &task::Vars) -> Result<()> {
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
@@ -1 +1 @@
|
||||
|
||||
pub mod create_mikan_bangumi_subscriptions_from_my_bangumi_page;
|
||||
|
||||
8
apps/recorder/src/test_utils/fetch.rs
Normal file
8
apps/recorder/src/test_utils/fetch.rs
Normal file
@@ -0,0 +1,8 @@
|
||||
use color_eyre::eyre;
|
||||
|
||||
use crate::fetch::HttpClient;
|
||||
|
||||
pub fn build_testing_http_client() -> eyre::Result<HttpClient> {
|
||||
let mikan_client = HttpClient::default();
|
||||
Ok(mikan_client)
|
||||
}
|
||||
@@ -1,3 +1,4 @@
|
||||
pub mod fetch;
|
||||
pub mod mikan;
|
||||
#[cfg(feature = "testcontainers")]
|
||||
pub mod testcontainers;
|
||||
|
||||
Reference in New Issue
Block a user