Skip to content

Commit

Permalink
Fix retry infinite loops. Fixes #1964 (#1967)
Browse files Browse the repository at this point in the history
* Fix retry infinite loops. Fixes #1964

* Moving retry_limit to settings
  • Loading branch information
dessalines authored Dec 5, 2021
1 parent 8bf0f31 commit 040770d
Show file tree
Hide file tree
Showing 5 changed files with 16 additions and 10 deletions.
2 changes: 2 additions & 0 deletions config/defaults.hjson
Original file line number Diff line number Diff line change
Expand Up @@ -111,4 +111,6 @@
slur_filter: "(\bThis\b)|(\bis\b)|(\bsample\b)"
# Maximum length of local community and user names
actor_name_max_length: 20
# Maximum number of HTTP requests allowed to handle a single incoming activity (or a single object fetch through the search).
http_fetch_retry_limit: 25
}
5 changes: 5 additions & 0 deletions crates/apub/src/fetcher/webfinger.rs
Original file line number Diff line number Diff line change
Expand Up @@ -81,6 +81,11 @@ where
);
debug!("Fetching webfinger url: {}", &fetch_url);

*request_counter += 1;
if *request_counter > context.settings().http_fetch_retry_limit {
return Err(LemmyError::from(anyhow!("Request retry limit reached")));
}

let response = retry(|| context.client().get(&fetch_url).send()).await?;

let res: WebfingerResponse = response
Expand Down
8 changes: 2 additions & 6 deletions crates/apub_lib/src/object_id.rs
Original file line number Diff line number Diff line change
Expand Up @@ -18,10 +18,6 @@ use std::{
use tracing::info;
use url::Url;

/// Maximum number of HTTP requests allowed to handle a single incoming activity (or a single object
/// fetch through the search). This should be configurable.
static REQUEST_LIMIT: i32 = 25;

static CLIENT: Lazy<Client> = Lazy::new(|| {
Client::builder()
.user_agent(build_user_agent(&Settings::get()))
Expand Down Expand Up @@ -119,8 +115,8 @@ where
info!("Fetching remote object {}", self.to_string());

*request_counter += 1;
if *request_counter > REQUEST_LIMIT {
return Err(LemmyError::from(anyhow!("Request limit reached")));
if *request_counter > Settings::get().http_fetch_retry_limit {
return Err(LemmyError::from(anyhow!("Request retry limit reached")));
}

let res = retry(|| {
Expand Down
8 changes: 4 additions & 4 deletions crates/utils/src/request.rs
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ pub struct SiteMetadata {

/// Fetches the post link html tags (like title, description, image, etc)
pub async fn fetch_site_metadata(client: &Client, url: &Url) -> Result<SiteMetadata, LemmyError> {
let response = retry(|| client.get(url.as_str()).send()).await?;
let response = client.get(url.as_str()).send().await?;

let html = response
.text()
Expand Down Expand Up @@ -132,7 +132,7 @@ pub(crate) async fn fetch_pictrs(
utf8_percent_encode(image_url.as_str(), NON_ALPHANUMERIC) // TODO this might not be needed
);

let response = retry(|| client.get(&fetch_url).send()).await?;
let response = client.get(&fetch_url).send().await?;

let response: PictrsResponse = response
.json()
Expand Down Expand Up @@ -201,8 +201,8 @@ pub async fn fetch_site_data(
}
}

async fn is_image_content_type(client: &Client, test: &Url) -> Result<(), LemmyError> {
let response = retry(|| client.get(test.to_owned()).send()).await?;
async fn is_image_content_type(client: &Client, url: &Url) -> Result<(), LemmyError> {
let response = client.get(url.as_str()).send().await?;
if response
.headers()
.get("Content-Type")
Expand Down
3 changes: 3 additions & 0 deletions crates/utils/src/settings/structs.rs
Original file line number Diff line number Diff line change
Expand Up @@ -46,6 +46,9 @@ pub struct Settings {
/// Maximum length of local community and user names
#[default(20)]
pub actor_name_max_length: usize,
/// Maximum number of HTTP requests allowed to handle a single incoming activity (or a single object fetch through the search).
#[default(25)]
pub http_fetch_retry_limit: i32,
}

#[derive(Debug, Deserialize, Serialize, Clone, SmartDefault, Document)]
Expand Down

0 comments on commit 040770d

Please sign in to comment.