From 7df8e7b4c63478b0cb129638b67310cb53e1573b Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Sun, 21 Nov 2021 19:27:33 +0000 Subject: [PATCH 01/27] Tweak feature request template --- .github/ISSUE_TEMPLATE/feature_request.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/ISSUE_TEMPLATE/feature_request.md b/.github/ISSUE_TEMPLATE/feature_request.md index 3770798..5d21fed 100644 --- a/.github/ISSUE_TEMPLATE/feature_request.md +++ b/.github/ISSUE_TEMPLATE/feature_request.md @@ -1,6 +1,6 @@ --- name: 💡 Feature request -about: Suggest an idea for this project +about: Suggest a feature for Libreddit that is not found in Reddit title: '' labels: enhancement assignees: '' From 2ef7957a66ba412fb1d51d8bd8fda0575c948bfe Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Sun, 21 Nov 2021 19:48:48 +0000 Subject: [PATCH 02/27] Create feature parity issue template --- .github/ISSUE_TEMPLATE/feature_parity.md | 28 ++++++++++++++++++++++++ 1 file changed, 28 insertions(+) create mode 100644 .github/ISSUE_TEMPLATE/feature_parity.md diff --git a/.github/ISSUE_TEMPLATE/feature_parity.md b/.github/ISSUE_TEMPLATE/feature_parity.md new file mode 100644 index 0000000..95bc729 --- /dev/null +++ b/.github/ISSUE_TEMPLATE/feature_parity.md @@ -0,0 +1,28 @@ +--- +name: ✨ Feature parity +about: Suggest implementing a feature into Libreddit that is found in Reddit.com +title: '' +labels: feature parity +assignees: '' + +--- + +## How does this feature work on Reddit? + + +## Describe the implementation into Libreddit + + +## Describe alternatives you've considered + + +## Additional context + From 768820cd4c4f8c81910b4bc35c35787d824dacd1 Mon Sep 17 00:00:00 2001 From: mikupls <93015331+mikupls@users.noreply.github.com> Date: Mon, 22 Nov 2021 00:17:52 +0100 Subject: [PATCH 03/27] Render markdown correctly in text post previews by using selftext_html. (#335) * Render markdown correctly in text post previews by using selftext_html. I was mistakenly under the impression that we somehow render markdown ourselves, but turns out we just take whatever HTML reddit gives us, and we also need to do this for text previews. Use CSS to limit the size of the previews instead of truncating in the template. Fix table CSS. * Fix post_body padding and trim post_previews Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com> --- src/utils.rs | 7 ++----- static/style.css | 5 +++-- templates/utils.html | 2 +- 3 files changed, 6 insertions(+), 8 deletions(-) diff --git a/src/utils.rs b/src/utils.rs index 2c74783..0bcad3c 100644 --- a/src/utils.rs +++ b/src/utils.rs @@ -251,11 +251,8 @@ impl Post { // Determine the type of media along with the media URL let (post_type, media, gallery) = Media::parse(data).await; - // selftext is set for text posts when browsing a (sub)reddit. - // Do NOT use selftext_html, because we truncate this content - // in the template code, and using selftext_html might result - // in truncated html. - let mut body = rewrite_urls(&val(post, "selftext")); + // selftext_html is set for text posts when browsing. + let mut body = rewrite_urls(&val(post, "selftext_html")); if body == "" { body = rewrite_urls(&val(post, "body_html")) } diff --git a/static/style.css b/static/style.css index 2d3bf24..67a694d 100644 --- a/static/style.css +++ b/static/style.css @@ -838,14 +838,16 @@ a.search_subreddit:hover { .post_body { opacity: 0.9; font-weight: normal; - padding: 5px 15px; + padding: 5px 15px 5px 12px; grid-area: post_body; width: calc(100% - 30px); } +/* Used only for text post preview */ .post_preview { mask-image: linear-gradient(180deg,#000 60%,transparent); opacity: 0.8; + max-height: 250px; } .post_footer { @@ -1223,7 +1225,6 @@ input[type="submit"] { .md table { margin: 5px; - display: block; overflow-x: auto; } diff --git a/templates/utils.html b/templates/utils.html index e5ba8f8..efd591e 100644 --- a/templates/utils.html +++ b/templates/utils.html @@ -132,7 +132,7 @@
Interface
From bd413060c6aa687b4fa0c65d48ff63424d58bdc8 Mon Sep 17 00:00:00 2001 From: Diego Magdaleno <38844659+DiegoMagdaleno@users.noreply.github.com> Date: Wed, 24 Nov 2021 20:08:27 -0600 Subject: [PATCH 18/27] Support displaying awards (#168) * Initial implementation of award parsing * Posts: Implement awards as part of post * Posts: remove parse_awards dead code * Posts: initial implementation of displaying Awards at the post title * Posts: Proxy static award images * Client: i.redd.it should take path as argument not ID * Posts: Just like Reddit make award size 16px * Templates: limit the awards to 4 awards to increase performance * Comments: Make awards a property of comments and display them * Format and correct /img/:id * Update comment.html * [Optimization] Awards is not longer async * [Revert] Posts can now display more than 4 awards again * [Implementation] Awards not display on the frontpage * [Implementation] Display count on awards * Post: Start working on awards css * Awards: Move the image size to css * Awards: Start implementing tooltips * Refactor awards code and tweak CSS indentation * Unify Awards::new and Awards::parse * Use native tooltips and brighten awards background Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com> --- CODEOWNERS | 1 - src/main.rs | 2 +- src/post.rs | 11 ++++++-- src/utils.rs | 60 +++++++++++++++++++++++++++++++++++++++++- static/style.css | 25 ++++++++++++++++-- templates/comment.html | 8 ++++++ templates/post.html | 11 ++++++++ templates/utils.html | 7 +++++ 8 files changed, 118 insertions(+), 7 deletions(-) delete mode 100644 CODEOWNERS diff --git a/CODEOWNERS b/CODEOWNERS deleted file mode 100644 index 8d1ba42..0000000 --- a/CODEOWNERS +++ /dev/null @@ -1 +0,0 @@ -* @spikecodes diff --git a/src/main.rs b/src/main.rs index f311d3e..611c0c9 100644 --- a/src/main.rs +++ b/src/main.rs @@ -180,7 +180,7 @@ async fn main() { // Proxy media through Libreddit app.at("/vid/:id/:size").get(|r| proxy(r, "https://v.redd.it/{id}/DASH_{size}").boxed()); app.at("/hls/:id/*path").get(|r| proxy(r, "https://v.redd.it/{id}/{path}").boxed()); - app.at("/img/:id").get(|r| proxy(r, "https://i.redd.it/{id}").boxed()); + app.at("/img/*path").get(|r| proxy(r, "https://i.redd.it/{path}").boxed()); app.at("/thumb/:point/:id").get(|r| proxy(r, "https://{point}.thumbs.redditmedia.com/{id}").boxed()); app.at("/emoji/:id/:name").get(|r| proxy(r, "https://emoji.redditmedia.com/{id}/{name}").boxed()); app.at("/preview/:loc/:id").get(|r| proxy(r, "https://{loc}view.redd.it/{id}").boxed()); diff --git a/src/post.rs b/src/post.rs index bde5c5b..6bdf409 100644 --- a/src/post.rs +++ b/src/post.rs @@ -3,8 +3,9 @@ use crate::client::json; use crate::esc; use crate::server::RequestExt; use crate::subreddit::{can_access_quarantine, quarantine}; -use crate::utils::{error, format_num, format_url, param, rewrite_urls, setting, template, time, val, Author, Comment, Flags, Flair, FlairPart, Media, Post, Preferences}; - +use crate::utils::{ + error, format_num, format_url, param, rewrite_urls, setting, template, time, val, Author, Awards, Comment, Flags, Flair, FlairPart, Media, Post, Preferences, +}; use hyper::{Body, Request, Response}; use askama::Template; @@ -93,6 +94,8 @@ async fn parse_post(json: &serde_json::Value) -> Post { // Determine the type of media along with the media URL let (post_type, media, gallery) = Media::parse(&post["data"]).await; + let awards: Awards = Awards::parse(&post["data"]["all_awardings"]); + // Build a post using data parsed from Reddit post API Post { id: val(post, "id"), @@ -148,6 +151,7 @@ async fn parse_post(json: &serde_json::Value) -> Post { created, comments: format_num(post["data"]["num_comments"].as_i64().unwrap_or_default()), gallery, + awards, } } @@ -178,6 +182,8 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, Vec::new() }; + let awards: Awards = Awards::parse(&data["all_awardings"]); + let parent_kind_and_id = val(&comment, "parent_id"); let parent_info = parent_kind_and_id.split('_').collect::{% for c in replies -%}{{ c.render().unwrap() }}{%- endfor %} diff --git a/templates/post.html b/templates/post.html index 60d37aa..99a4065 100644 --- a/templates/post.html +++ b/templates/post.html @@ -43,6 +43,17 @@ {% endif %} • {{ post.rel_time }} + {% if !post.awards.is_empty() %} + • + + {% for award in post.awards.clone() %} + + + {{ award.count }} + + {% endfor %} + + {% endif %}{{ post.title }} diff --git a/templates/utils.html b/templates/utils.html index 3745dea..e50b785 100644 --- a/templates/utils.html +++ b/templates/utils.html @@ -75,6 +75,13 @@ u/{{ post.author.name }} • {{ post.rel_time }} + {% if !post.awards.is_empty() %} + {% for award in post.awards.clone() %} + + + + {% endfor %} + {% endif %}
{% if post.flair.flair_parts.len() > 0 %} From beada1f2b2a63aefbcdcd87f0aca0544a3d1fe84 Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Thu, 25 Nov 2021 05:38:09 +0000 Subject: [PATCH 19/27] Update privacy policy --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index a3f0879..f95f857 100644 --- a/README.md +++ b/README.md @@ -145,13 +145,13 @@ Results from Google Lighthouse ([Libreddit Report](https://lighthouse-dot-webdot For transparency, I hope to describe all the ways Libreddit handles user privacy. -**Logging:** In production (when running the binary, hosting with docker, or using the official instances), Libreddit logs when Reddit is ratelimiting Libreddit and when Reddit's JSON responses can't be parsed. When debugging (running from source without `--release`), Libreddit logs post IDs and URL paths fetched to aid with troubleshooting. +**Logging:** In production (when running the binary, hosting with docker, or using the official instances), Libreddit logs nothing. When debugging (running from source without `--release`), Libreddit logs post IDs fetched to aid with troubleshooting. **DNS:** Both official domains (`libredd.it` and `libreddit.spike.codes`) use Cloudflare as the DNS resolver. Though, the sites are not proxied through Cloudflare meaning Cloudflare doesn't have access to user traffic. -**Cookies:** Libreddit uses optional cookies to store any configured settings in [the settings menu](https://libreddit.spike.codes/settings). This is not a cross-site cookie and the cookie holds no personal data, only a value of the possible layout. +**Cookies:** Libreddit uses optional cookies to store any configured settings in [the settings menu](https://libreddit.spike.codes/settings). These are not cross-site cookies and the cookies hold no personal data. -**Hosting:** The official instances are hosted on [Replit](https://replit.com/) which monitors usage to prevent abuse. I can understand if this invalidates certain users' threat models and therefore, selfhosting and browsing through Tor are welcomed. +**Hosting:** The official instances are hosted on [Replit](https://replit.com/) which monitors usage to prevent abuse. I can understand if this invalidates certain users' threat models and therefore, selfhosting, using unofficial instances and browsing through Tor are welcomed. --- From 888e7b302deadeafb2138d6b5a538c020bd3523f Mon Sep 17 00:00:00 2001 From: Nick Lowery
Date: Thu, 25 Nov 2021 21:02:04 -0700 Subject: [PATCH 20/27] Filter subreddits and users (#317) * Initial work on filtering subreddits and users * Fix doubly-prefixed subreddit name in search alt text (e.g. r/r/pics) * Don't set post title to "Comment" if empty - this could throw off actual posts with the title "Comment" * Filter search results * Fix filtering to differentiate between "this subject itself is filtered" vs "all posts on this current page have been filtered" * Remove unnecessary check * Clean up * Cargo format * Collapse comments from filtered users Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com> --- src/main.rs | 6 +- src/post.rs | 43 +++++----- src/search.rs | 68 ++++++++++++---- src/settings.rs | 2 +- src/subreddit.rs | 169 ++++++++++++++++++++++++++------------- src/user.rs | 57 ++++++++----- src/utils.rs | 28 ++++++- static/style.css | 29 +++++-- templates/comment.html | 6 +- templates/search.html | 14 +++- templates/settings.html | 17 +++- templates/subreddit.html | 44 +++++++--- templates/user.html | 34 ++++++-- 13 files changed, 374 insertions(+), 143 deletions(-) diff --git a/src/main.rs b/src/main.rs index 611c0c9..fa772be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -215,8 +215,10 @@ async fn main() { .at("/r/u_:name") .get(|r| async move { Ok(redirect(format!("/user/{}", r.param("name").unwrap_or_default()))) }.boxed()); - app.at("/r/:sub/subscribe").post(|r| subreddit::subscriptions(r).boxed()); - app.at("/r/:sub/unsubscribe").post(|r| subreddit::subscriptions(r).boxed()); + app.at("/r/:sub/subscribe").post(|r| subreddit::subscriptions_filters(r).boxed()); + app.at("/r/:sub/unsubscribe").post(|r| subreddit::subscriptions_filters(r).boxed()); + app.at("/r/:sub/filter").post(|r| subreddit::subscriptions_filters(r).boxed()); + app.at("/r/:sub/unfilter").post(|r| subreddit::subscriptions_filters(r).boxed()); app.at("/r/:sub/comments/:id").get(|r| post::item(r).boxed()); app.at("/r/:sub/comments/:id/:title").get(|r| post::item(r).boxed()); diff --git a/src/post.rs b/src/post.rs index 6bdf409..ff430fc 100644 --- a/src/post.rs +++ b/src/post.rs @@ -4,11 +4,12 @@ use crate::esc; use crate::server::RequestExt; use crate::subreddit::{can_access_quarantine, quarantine}; use crate::utils::{ - error, format_num, format_url, param, rewrite_urls, setting, template, time, val, Author, Awards, Comment, Flags, Flair, FlairPart, Media, Post, Preferences, + error, format_num, format_url, get_filters, param, rewrite_urls, setting, template, time, val, Author, Awards, Comment, Flags, Flair, FlairPart, Media, Post, Preferences, }; use hyper::{Body, Request, Response}; use askama::Template; +use std::collections::HashSet; // STRUCTS #[derive(Template)] @@ -55,7 +56,7 @@ pub async fn item(req: Request) -> Result , String> { Ok(response) => { // Parse the JSON into Post and Comment structs let post = parse_post(&response[0]).await; - let comments = parse_comments(&response[1], &post.permalink, &post.author.name, highlighted_comment); + let comments = parse_comments(&response[1], &post.permalink, &post.author.name, highlighted_comment, &get_filters(&req)); let url = req.uri().to_string(); // Use the Post and Comment structs to generate a website to show users @@ -156,7 +157,7 @@ async fn parse_post(json: &serde_json::Value) -> Post { } // COMMENTS -fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, highlighted_comment: &str) -> Vec { +fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, highlighted_comment: &str, filters: &HashSet ) -> Vec { // Parse the comment JSON into a Vector of Comments let comments = json["data"]["children"].as_array().map_or(Vec::new(), std::borrow::ToOwned::to_owned); @@ -177,7 +178,7 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, // If this comment contains replies, handle those too let replies: Vec = if data["replies"].is_object() { - parse_comments(&data["replies"], post_link, post_author, highlighted_comment) + parse_comments(&data["replies"], post_link, post_author, highlighted_comment, filters) } else { Vec::new() }; @@ -190,13 +191,29 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, let id = val(&comment, "id"); let highlighted = id == highlighted_comment; + let author = Author { + name: val(&comment, "author"), + flair: Flair { + flair_parts: FlairPart::parse( + data["author_flair_type"].as_str().unwrap_or_default(), + data["author_flair_richtext"].as_array(), + data["author_flair_text"].as_str(), + ), + text: esc!(&comment, "link_flair_text"), + background_color: val(&comment, "author_flair_background_color"), + foreground_color: val(&comment, "author_flair_text_color"), + }, + distinguished: val(&comment, "distinguished"), + }; + let is_filtered = filters.contains(&["u_", author.name.as_str()].concat()); + // Many subreddits have a default comment posted about the sub's rules etc. // Many libreddit users do not wish to see this kind of comment by default. // Reddit does not tell us which users are "bots", so a good heuristic is to // collapse stickied moderator comments. let is_moderator_comment = data["distinguished"].as_str().unwrap_or_default() == "moderator"; let is_stickied = data["stickied"].as_bool().unwrap_or_default(); - let collapsed = is_moderator_comment && is_stickied; + let collapsed = (is_moderator_comment && is_stickied) || is_filtered; Comment { id, @@ -206,20 +223,7 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, post_link: post_link.to_string(), post_author: post_author.to_string(), body, - author: Author { - name: val(&comment, "author"), - flair: Flair { - flair_parts: FlairPart::parse( - data["author_flair_type"].as_str().unwrap_or_default(), - data["author_flair_richtext"].as_array(), - data["author_flair_text"].as_str(), - ), - text: esc!(&comment, "link_flair_text"), - background_color: val(&comment, "author_flair_background_color"), - foreground_color: val(&comment, "author_flair_text_color"), - }, - distinguished: val(&comment, "distinguished"), - }, + author, score: if data["score_hidden"].as_bool().unwrap_or_default() { ("\u{2022}".to_string(), "Hidden".to_string()) } else { @@ -232,6 +236,7 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, highlighted, awards, collapsed, + is_filtered, } }) .collect() diff --git a/src/search.rs b/src/search.rs index aff7173..0eef077 100644 --- a/src/search.rs +++ b/src/search.rs @@ -1,5 +1,5 @@ // CRATES -use crate::utils::{catch_random, error, format_num, format_url, param, redirect, setting, template, val, Post, Preferences}; +use crate::utils::{catch_random, error, filter_posts, format_num, format_url, get_filters, param, redirect, setting, template, val, Post, Preferences}; use crate::{ client::json, subreddit::{can_access_quarantine, quarantine}, @@ -37,6 +37,11 @@ struct SearchTemplate { params: SearchParams, prefs: Preferences, url: String, + /// Whether the subreddit itself is filtered. + is_filtered: bool, + /// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place, + /// and all fetched posts being filtered). + all_posts_filtered: bool, } // SERVICES @@ -59,15 +64,23 @@ pub async fn find(req: Request) -> Result , String> { let typed = param(&path, "type").unwrap_or_default(); let sort = param(&path, "sort").unwrap_or_else(|| "relevance".to_string()); + let filters = get_filters(&req); // If search is not restricted to this subreddit, show other subreddits in search results - let subreddits = param(&path, "restrict_sr").map_or(search_subreddits(&query, &typed).await, |_| Vec::new()); + let subreddits = if param(&path, "restrict_sr").is_none() { + let mut subreddits = search_subreddits(&query, &typed).await; + subreddits.retain(|s| !filters.contains(s.name.as_str())); + subreddits + } else { + Vec::new() + }; let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); - match Post::fetch(&path, String::new(), quarantined).await { - Ok((posts, after)) => template(SearchTemplate { - posts, + // If all requested subs are filtered, we don't need to fetch posts. + if sub.split("+").all(|s| filters.contains(s)) { + template(SearchTemplate { + posts: Vec::new(), subreddits, sub, params: SearchParams { @@ -75,19 +88,46 @@ pub async fn find(req: Request) -> Result , String> { sort, t: param(&path, "t").unwrap_or_default(), before: param(&path, "after").unwrap_or_default(), - after, + after: "".to_string(), restrict_sr: param(&path, "restrict_sr").unwrap_or_default(), typed, }, prefs: Preferences::new(req), url, - }), - Err(msg) => { - if msg == "quarantined" { - let sub = req.param("sub").unwrap_or_default(); - quarantine(req, sub) - } else { - error(req, msg).await + is_filtered: true, + all_posts_filtered: false, + }) + } else { + match Post::fetch(&path, quarantined).await { + Ok((mut posts, after)) => { + let all_posts_filtered = filter_posts(&mut posts, &filters); + + template(SearchTemplate { + posts, + subreddits, + sub, + params: SearchParams { + q: query.replace('"', """), + sort, + t: param(&path, "t").unwrap_or_default(), + before: param(&path, "after").unwrap_or_default(), + after, + restrict_sr: param(&path, "restrict_sr").unwrap_or_default(), + typed, + }, + prefs: Preferences::new(req), + url, + is_filtered: false, + all_posts_filtered, + }) + } + Err(msg) => { + if msg == "quarantined" { + let sub = req.param("sub").unwrap_or_default(); + quarantine(req, sub) + } else { + error(req, msg).await + } } } } @@ -109,7 +149,7 @@ async fn search_subreddits(q: &str, typed: &str) -> Vec { let icon = subreddit["data"]["community_icon"].as_str().map_or_else(|| val(subreddit, "icon_img"), ToString::to_string); Subreddit { - name: val(subreddit, "display_name_prefixed"), + name: val(subreddit, "display_name"), url: val(subreddit, "url"), icon: format_url(&icon), description: val(subreddit, "public_description"), diff --git a/src/settings.rs b/src/settings.rs index efa4708..9cdd266 100644 --- a/src/settings.rs +++ b/src/settings.rs @@ -109,7 +109,7 @@ fn set_cookies_method(req: Request, remove_cookies: bool) -> Response response.insert_cookie( Cookie::build(name.to_owned(), value.clone()) diff --git a/src/subreddit.rs b/src/subreddit.rs index 66938f5..f94a583 100644 --- a/src/subreddit.rs +++ b/src/subreddit.rs @@ -1,6 +1,8 @@ // CRATES use crate::esc; -use crate::utils::{catch_random, error, format_num, format_url, param, redirect, rewrite_urls, setting, template, val, Post, Preferences, Subreddit}; +use crate::utils::{ + catch_random, error, filter_posts, format_num, format_url, get_filters, param, redirect, rewrite_urls, setting, template, val, Post, Preferences, Subreddit, +}; use crate::{client::json, server::ResponseExt, RequestExt}; use askama::Template; use cookie::Cookie; @@ -17,6 +19,11 @@ struct SubredditTemplate { ends: (String, String), prefs: Preferences, url: String, + /// Whether the subreddit itself is filtered. + is_filtered: bool, + /// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place, + /// and all fetched posts being filtered). + all_posts_filtered: bool, } #[derive(Template)] @@ -48,7 +55,7 @@ pub async fn community(req: Request) -> Result , String> { let post_sort = req.cookie("post_sort").map_or_else(|| "hot".to_string(), |c| c.value().to_string()); let sort = req.param("sort").unwrap_or_else(|| req.param("id").unwrap_or(post_sort)); - let sub = req.param("sub").unwrap_or(if front_page == "default" || front_page.is_empty() { + let sub_name = req.param("sub").unwrap_or(if front_page == "default" || front_page.is_empty() { if subscribed.is_empty() { "popular".to_string() } else { @@ -57,59 +64,77 @@ pub async fn community(req: Request) -> Result , String> { } else { front_page.clone() }); - let quarantined = can_access_quarantine(&req, &sub) || root; + let quarantined = can_access_quarantine(&req, &sub_name) || root; // Handle random subreddits - if let Ok(random) = catch_random(&sub, "").await { + if let Ok(random) = catch_random(&sub_name, "").await { return Ok(random); } - if req.param("sub").is_some() && sub.starts_with("u_") { - return Ok(redirect(["/user/", &sub[2..]].concat())); + if req.param("sub").is_some() && sub_name.starts_with("u_") { + return Ok(redirect(["/user/", &sub_name[2..]].concat())); } - let path = format!("/r/{}/{}.json?{}&raw_json=1", sub, sort, req.uri().query().unwrap_or_default()); - - match Post::fetch(&path, String::new(), quarantined).await { - Ok((posts, after)) => { - // If you can get subreddit posts, also request subreddit metadata - let sub = if !sub.contains('+') && sub != subscribed && sub != "popular" && sub != "all" { - // Regular subreddit - subreddit(&sub, quarantined).await.unwrap_or_default() - } else if sub == subscribed { - // Subscription feed - if req.uri().path().starts_with("/r/") { - subreddit(&sub, quarantined).await.unwrap_or_default() - } else { - Subreddit::default() - } - } else if sub.contains('+') { - // Multireddit - Subreddit { - name: sub, - ..Subreddit::default() - } - } else { - Subreddit::default() - }; - - let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); - - template(SubredditTemplate { - sub, - posts, - sort: (sort, param(&path, "t").unwrap_or_default()), - ends: (param(&path, "after").unwrap_or_default(), after), - prefs: Preferences::new(req), - url, - }) + // Request subreddit metadata + let sub = if !sub_name.contains('+') && sub_name != subscribed && sub_name != "popular" && sub_name != "all" { + // Regular subreddit + subreddit(&sub_name, quarantined).await.unwrap_or_default() + } else if sub_name == subscribed { + // Subscription feed + if req.uri().path().starts_with("/r/") { + subreddit(&sub_name, quarantined).await.unwrap_or_default() + } else { + Subreddit::default() + } + } else if sub_name.contains('+') { + // Multireddit + Subreddit { + name: sub_name.clone(), + ..Subreddit::default() + } + } else { + Subreddit::default() + }; + + let path = format!("/r/{}/{}.json?{}&raw_json=1", sub_name.clone(), sort, req.uri().query().unwrap_or_default()); + let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); + let filters = get_filters(&req); + + // If all requested subs are filtered, we don't need to fetch posts. + if sub_name.split("+").all(|s| filters.contains(s)) { + template(SubredditTemplate { + sub, + posts: Vec::new(), + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), "".to_string()), + prefs: Preferences::new(req), + url, + is_filtered: true, + all_posts_filtered: false, + }) + } else { + match Post::fetch(&path, quarantined).await { + Ok((mut posts, after)) => { + let all_posts_filtered = filter_posts(&mut posts, &filters); + + template(SubredditTemplate { + sub, + posts, + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), after), + prefs: Preferences::new(req), + url, + is_filtered: false, + all_posts_filtered, + }) + } + Err(msg) => match msg.as_str() { + "quarantined" => quarantine(req, sub_name), + "private" => error(req, format!("r/{} is a private community", sub_name)).await, + "banned" => error(req, format!("r/{} has been banned from Reddit", sub_name)).await, + _ => error(req, msg).await, + }, } - Err(msg) => match msg.as_str() { - "quarantined" => quarantine(req, sub), - "private" => error(req, format!("r/{} is a private community", sub)).await, - "banned" => error(req, format!("r/{} has been banned from Reddit", sub)).await, - _ => error(req, msg).await, - }, } } @@ -150,18 +175,25 @@ pub fn can_access_quarantine(req: &Request, sub: &str) -> bool { setting(req, &format!("allow_quaran_{}", sub.to_lowercase())).parse().unwrap_or_default() } -// Sub or unsub by setting subscription cookie using response "Set-Cookie" header -pub async fn subscriptions(req: Request) -> Result , String> { +// Sub, filter, unfilter, or unsub by setting subscription cookie using response "Set-Cookie" header +pub async fn subscriptions_filters(req: Request) -> Result , String> { let sub = req.param("sub").unwrap_or_default(); + let action: Vec = req.uri().path().split('/').map(String::from).collect(); + // Handle random subreddits if sub == "random" || sub == "randnsfw" { - return Err("Can't subscribe to random subreddit!".to_string()); + if action.contains(&"filter".to_string()) || action.contains(&"unfilter".to_string()) { + return Err("Can't filter random subreddit!".to_string()); + } else { + return Err("Can't subscribe to random subreddit!".to_string()); + } } let query = req.uri().query().unwrap_or_default().to_string(); - let action: Vec = req.uri().path().split('/').map(String::from).collect(); - let mut sub_list = Preferences::new(req).subscriptions; + let preferences = Preferences::new(req); + let mut sub_list = preferences.subscriptions; + let mut filters = preferences.filters; // Retrieve list of posts for these subreddits to extract display names let posts = json(format!("/r/{}/hot.json?raw_json=1", sub), true).await?; @@ -182,8 +214,10 @@ pub async fn subscriptions(req: Request) -> Result , String> for part in sub.split('+') { // Retrieve display name for the subreddit let display; - let part = if let Some(&(_, display)) = display_lookup.iter().find(|x| x.0 == part.to_lowercase()) { - // This is already known, doesn't require seperate request + let part = if part.starts_with("u_") { + part + } else if let Some(&(_, display)) = display_lookup.iter().find(|x| x.0 == part.to_lowercase()) { + // This is already known, doesn't require separate request display } else { // This subreddit display name isn't known, retrieve it @@ -196,16 +230,28 @@ pub async fn subscriptions(req: Request) -> Result , String> if action.contains(&"subscribe".to_string()) && !sub_list.contains(&part.to_owned()) { // Add each sub name to the subscribed list sub_list.push(part.to_owned()); - // Reorder sub names alphabettically + filters.retain(|s| s.to_lowercase() != part.to_lowercase()); + // Reorder sub names alphabetically sub_list.sort_by_key(|a| a.to_lowercase()); + filters.sort_by_key(|a| a.to_lowercase()); } else if action.contains(&"unsubscribe".to_string()) { // Remove sub name from subscribed list sub_list.retain(|s| s.to_lowercase() != part.to_lowercase()); + } else if action.contains(&"filter".to_string()) && !filters.contains(&part.to_owned()) { + // Add each sub name to the filtered list + filters.push(part.to_owned()); + sub_list.retain(|s| s.to_lowercase() != part.to_lowercase()); + // Reorder sub names alphabetically + filters.sort_by_key(|a| a.to_lowercase()); + sub_list.sort_by_key(|a| a.to_lowercase()); + } else if action.contains(&"unfilter".to_string()) { + // Remove sub name from filtered list + filters.retain(|s| s.to_lowercase() != part.to_lowercase()); } } // Redirect back to subreddit - // check for redirect parameter if unsubscribing from outside sidebar + // check for redirect parameter if unsubscribing/unfiltering from outside sidebar let path = if let Some(redirect_path) = param(&format!("?{}", query), "redirect") { format!("/{}/", redirect_path) } else { @@ -226,6 +272,17 @@ pub async fn subscriptions(req: Request) -> Result , String> .finish(), ); } + if filters.is_empty() { + response.remove_cookie("filters".to_string()); + } else { + response.insert_cookie( + Cookie::build("filters", filters.join("+")) + .path("/") + .http_only(true) + .expires(OffsetDateTime::now_utc() + Duration::weeks(52)) + .finish(), + ); + } Ok(response) } diff --git a/src/user.rs b/src/user.rs index 9179551..61772e5 100644 --- a/src/user.rs +++ b/src/user.rs @@ -2,7 +2,7 @@ use crate::client::json; use crate::esc; use crate::server::RequestExt; -use crate::utils::{error, format_url, param, template, Post, Preferences, User}; +use crate::utils::{error, filter_posts, format_url, get_filters, param, template, Post, Preferences, User}; use askama::Template; use hyper::{Body, Request, Response}; use time::OffsetDateTime; @@ -17,6 +17,11 @@ struct UserTemplate { ends: (String, String), prefs: Preferences, url: String, + /// Whether the user themself is filtered. + is_filtered: bool, + /// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place, + /// and all fetched posts being filtered). + all_posts_filtered: bool, } // FUNCTIONS @@ -27,31 +32,45 @@ pub async fn profile(req: Request) -> Result , String> { req.param("name").unwrap_or_else(|| "reddit".to_string()), req.uri().query().unwrap_or_default() ); + let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); // Retrieve other variables from Libreddit request let sort = param(&path, "sort").unwrap_or_default(); let username = req.param("name").unwrap_or_default(); + let user = user(&username).await.unwrap_or_default(); - // Request user posts/comments from Reddit - let posts = Post::fetch(&path, "Comment".to_string(), false).await; - let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); + let filters = get_filters(&req); + if filters.contains(&["u_", &username].concat()) { + template(UserTemplate { + user, + posts: Vec::new(), + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), "".to_string()), + prefs: Preferences::new(req), + url, + is_filtered: true, + all_posts_filtered: false, + }) + } else { + // Request user posts/comments from Reddit + match Post::fetch(&path, false).await { + Ok((mut posts, after)) => { + let all_posts_filtered = filter_posts(&mut posts, &filters); - match posts { - Ok((posts, after)) => { - // If you can get user posts, also request user data - let user = user(&username).await.unwrap_or_default(); - - template(UserTemplate { - user, - posts, - sort: (sort, param(&path, "t").unwrap_or_default()), - ends: (param(&path, "after").unwrap_or_default(), after), - prefs: Preferences::new(req), - url, - }) + template(UserTemplate { + user, + posts, + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), after), + prefs: Preferences::new(req), + url, + is_filtered: false, + all_posts_filtered, + }) + } + // If there is an error show error page + Err(msg) => error(req, msg).await, } - // If there is an error show error page - Err(msg) => error(req, msg).await, } } diff --git a/src/utils.rs b/src/utils.rs index d6961ec..dad2e99 100644 --- a/src/utils.rs +++ b/src/utils.rs @@ -7,7 +7,7 @@ use cookie::Cookie; use hyper::{Body, Request, Response}; use regex::Regex; use serde_json::Value; -use std::collections::HashMap; +use std::collections::{HashMap, HashSet}; use std::str::FromStr; use time::{Duration, OffsetDateTime}; use url::Url; @@ -219,7 +219,7 @@ pub struct Post { impl Post { // Fetch posts of a user or subreddit and return a vector of posts and the "after" value - pub async fn fetch(path: &str, fallback_title: String, quarantine: bool) -> Result<(Vec , String), String> { + pub async fn fetch(path: &str, quarantine: bool) -> Result<(Vec , String), String> { let res; let post_list; @@ -262,7 +262,7 @@ impl Post { posts.push(Self { id: val(post, "id"), - title: esc!(if title.is_empty() { fallback_title.clone() } else { title }), + title, community: val(post, "subreddit"), body, author: Author { @@ -346,6 +346,7 @@ pub struct Comment { pub highlighted: bool, pub awards: Awards, pub collapsed: bool, + pub is_filtered: bool, } #[derive(Default, Clone)] @@ -458,6 +459,7 @@ pub struct Preferences { pub comment_sort: String, pub post_sort: String, pub subscriptions: Vec , + pub filters: Vec , } impl Preferences { @@ -475,10 +477,28 @@ impl Preferences { comment_sort: setting(&req, "comment_sort"), post_sort: setting(&req, "post_sort"), subscriptions: setting(&req, "subscriptions").split('+').map(String::from).filter(|s| !s.is_empty()).collect(), + filters: setting(&req, "filters").split('+').map(String::from).filter(|s| !s.is_empty()).collect(), } } } +/// Gets a `HashSet` of filters from the cookie in the given `Request`. +pub fn get_filters(req: &Request) -> HashSet { + setting(&req, "filters").split('+').map(String::from).filter(|s| !s.is_empty()).collect:: >() +} + +/// Filters a `Vec ` by the given `HashSet` of filters (each filter being a subreddit name or a user name). If a +/// `Post`'s subreddit or author is found in the filters, it is removed. Returns `true` if _all_ posts were filtered +/// out, or `false` otherwise. +pub fn filter_posts(posts: &mut Vec , filters: &HashSet ) -> bool { + if posts.is_empty() { + false + } else { + posts.retain(|p| !filters.contains(&p.community) && !filters.contains(&["u_", &p.author.name].concat())); + posts.is_empty() + } +} + // // FORMATTING // @@ -515,7 +535,7 @@ pub fn setting(req: &Request, name: &str) -> String { // Detect and redirect in the event of a random subreddit pub async fn catch_random(sub: &str, additional: &str) -> Result , String> { - if (sub == "random" || sub == "randnsfw") && !sub.contains('+') { + if sub == "random" || sub == "randnsfw" { let new_sub = json(format!("/r/{}/about.json?raw_json=1", sub), false).await?["data"]["display_name"] .as_str() .unwrap_or_default() diff --git a/static/style.css b/static/style.css index fefba1a..514b642 100644 --- a/static/style.css +++ b/static/style.css @@ -372,7 +372,7 @@ aside { margin-bottom: 20px; } -#user_details, #sub_details { +#user_details, #sub_details, #sub_actions, #user_actions { display: grid; grid-template-columns: repeat(2, 1fr); grid-column-gap: 20px; @@ -384,7 +384,7 @@ aside { /* Subscriptions */ -#sub_subscription, #user_subscription { +#sub_subscription, #user_subscription, #user_filter, #sub_filter { margin-top: 20px; } @@ -392,18 +392,18 @@ aside { margin-bottom: 20px; } -.subscribe, .unsubscribe { +.subscribe, .unsubscribe, .filter, .unfilter { padding: 10px 20px; border-radius: 5px; cursor: pointer; } -.subscribe { +.subscribe, .filter { color: var(--foreground); background-color: var(--accent); } -.unsubscribe { +.unsubscribe, .unfilter { color: var(--text); background-color: var(--highlighted); } @@ -1042,7 +1042,7 @@ a.search_subreddit:hover { overflow: auto; } -.comment_body.highlighted { +.comment_body.highlighted, .comment_body_filtered.highlighted { background: var(--highlighted); } @@ -1055,6 +1055,15 @@ a.search_subreddit:hover { color: var(--accent); } +.comment_body_filtered { + opacity: 0.4; + font-weight: normal; + font-style: italic; + padding: 5px 5px; + margin: 5px 0; + overflow: auto; +} + .deeper_replies { color: var(--accent); margin-left: 15px; @@ -1226,6 +1235,14 @@ input[type="submit"] { color: var(--accent); } +#settings_filters .unsubscribe { + margin-left: 30px; +} + +#settings_filters a { + color: var(--accent); +} + .helper { padding: 10px; width: 250px; diff --git a/templates/comment.html b/templates/comment.html index 8734e2a..7090251 100644 --- a/templates/comment.html +++ b/templates/comment.html @@ -8,7 +8,7 @@ {{ score.0 }}
-++ {% endif %} {% endif %} diff --git a/templates/user.html b/templates/user.html index bfcef08..8095d06 100644 --- a/templates/user.html +++ b/templates/user.html @@ -13,11 +13,12 @@ {% block body %}diff --git a/templates/search.html b/templates/search.html index b7b99c5..0218c35 100644 --- a/templates/search.html +++ b/templates/search.html @@ -30,7 +30,8 @@ - + + {% if !is_filtered %} {% if subreddits.len() > 0 || params.typed == "sr_user" %}u/{{ author.name }} {% if author.flair.flair_parts.len() > 0 %} @@ -25,7 +25,11 @@ {% endfor %} {% endif %}
+ {% if is_filtered %} +(Filtered content)+ {% else %}{{ body }}+ {% endif %}{% for c in replies -%}{{ c.render().unwrap() }}{%- endfor %}{% if params.typed == "sr_user" %} @@ -41,7 +42,7 @@diff --git a/templates/subreddit.html b/templates/subreddit.html index 9038440..bdbc48d 100644 --- a/templates/subreddit.html +++ b/templates/subreddit.html @@ -17,6 +17,7 @@ {% block body %}{% if subreddit.icon != "" %}{% endif %}{% endif %} - {% if params.typed != "sr_user" %} + {% endif %} + {% if all_posts_filtered %} +- {{ subreddit.name }} + r/{{ subreddit.name }} • {{ subreddit.subscribers.0 }} Members
@@ -54,10 +55,15 @@ {% endif %}(All content on this page has been filtered) + {% else if is_filtered %} +(Content from r/{{ sub }} has been filtered) + {% else if params.typed != "sr_user" %} {% for post in posts %} {% if post.flags.nsfw && prefs.show_nsfw != "on" %} - {% else if post.title != "Comment" %} + {% else if !post.title.is_empty() %} {% call utils::post_in_list(post) %} {% else %}diff --git a/templates/settings.html b/templates/settings.html index 2538e09..2072614 100644 --- a/templates/settings.html +++ b/templates/settings.html @@ -92,10 +92,25 @@ {% endfor %}{% endif %} + {% if !prefs.filters.is_empty() %} +++ {% endif %}Filtered Feeds
+ {% for sub in prefs.filters %} ++ {% let feed -%} + {% if sub.starts_with("u_") -%}{% let feed = format!("u/{}", &sub[2..]) -%}{% else -%}{% let feed = format!("r/{}", sub) -%}{% endif -%} + {{ feed }} + ++ {% endfor %} +Note: settings and subscriptions are saved in browser cookies. Clearing your cookies will reset them.
-You can restore your current settings and subscriptions after clearing your cookies using this link.
+You can restore your current settings and subscriptions after clearing your cookies using this link.
+ {% if !is_filtered %} + {% if !is_filtered %} + {% if all_posts_filtered %} +From 99a83ea11b99663080d73101a03ecb2938c6c8c6 Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Sat, 27 Nov 2021 19:13:15 +0000 Subject: [PATCH 21/27] Add northboot.xyz instance --- README.md | 2 ++ 1 file changed, 2 insertions(+) diff --git a/README.md b/README.md index f95f857..207c488 100644 --- a/README.md +++ b/README.md @@ -56,12 +56,14 @@ Feel free to [open an issue](https://github.com/spikecodes/libreddit/issues/new) | [libreddit.de](https://libreddit.de) | 🇩🇪 DE | | | [libreddit.pussthecat.org](https://libreddit.pussthecat.org) | 🇩🇪 DE | | | [libreddit.mutahar.rocks](https://libreddit.mutahar.rocks) | 🇫🇷 FR | | +| [libreddit.northboot.xyz](https://libreddit.northboot.xyz) | 🇩🇪 DE | | | [spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion](http://spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion) | 🇮🇳 IN | | | [fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion](http://fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion) | 🇩🇪 DE | | | [kphht2jcflojtqte4b4kyx7p2ahagv4debjj32nre67dxz7y57seqwyd.onion](http://kphht2jcflojtqte4b4kyx7p2ahagv4debjj32nre67dxz7y57seqwyd.onion) | 🇳🇱 NL | | | [inytumdgnri7xsqtvpntjevaelxtgbjqkuqhtf6txxhwbll2fwqtakqd.onion](http://inytumdgnri7xsqtvpntjevaelxtgbjqkuqhtf6txxhwbll2fwqtakqd.onion) | 🇨🇭 CH | | | [liredejj74h5xjqr2dylnl5howb2bpikfowqoveub55ru27x43357iid.onion](http://liredejj74h5xjqr2dylnl5howb2bpikfowqoveub55ru27x43357iid.onion) | 🇩🇪 DE | | | [kzhfp3nvb4qp575vy23ccbrgfocezjtl5dx66uthgrhu7nscu6rcwjyd.onion](http://kzhfp3nvb4qp575vy23ccbrgfocezjtl5dx66uthgrhu7nscu6rcwjyd.onion) | 🇺🇸 US | | +| [ecue64ybzvn6vjzl37kcsnwt4ycmbsyf74nbttyg7rkc3t3qwnj7mcyd.onion](http://ecue64ybzvn6vjzl37kcsnwt4ycmbsyf74nbttyg7rkc3t3qwnj7mcyd.onion) | 🇩🇪 DE | | A checkmark in the "Cloudflare" category here refers to the use of the reverse proxy, [Cloudflare](https://cloudflare). The checkmark will not be listed for a site which uses Cloudflare DNS but rather the proxying service which grants Cloudflare the ability to monitor traffic to the website. From 401ee2ee418aa865a6da037737fce8e28528a3d9 Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Sat, 27 Nov 2021 22:27:39 +0000 Subject: [PATCH 22/27] Create FUNDING.yml --- .github/FUNDING.yml | 12 ++++++++++++ 1 file changed, 12 insertions(+) create mode 100644 .github/FUNDING.yml diff --git a/.github/FUNDING.yml b/.github/FUNDING.yml new file mode 100644 index 0000000..a765c29 --- /dev/null +++ b/.github/FUNDING.yml @@ -0,0 +1,12 @@ +# These are supported funding model platforms + +github: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2] +patreon: # Replace with a single Patreon username +open_collective: # Replace with a single Open Collective username +ko_fi: # Replace with a single Ko-fi username +tidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel +community_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry +liberapay: spike +issuehunt: # Replace with a single IssueHunt username +otechie: # Replace with a single Otechie username +custom: # Replace with up to 4 custom sponsorship URLs e.g., ['link1', 'link2'] From b63000a93fc07160d91feb7313cf608f2a58182f Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Sat, 27 Nov 2021 22:27:47 +0000 Subject: [PATCH 23/27] Create FUNDING.yml --- FUNDING.yml | 12 ++++++++++++ 1 file changed, 12 insertions(+) create mode 100644 FUNDING.yml diff --git a/FUNDING.yml b/FUNDING.yml new file mode 100644 index 0000000..a765c29 --- /dev/null +++ b/FUNDING.yml @@ -0,0 +1,12 @@ +# These are supported funding model platforms + +github: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2] +patreon: # Replace with a single Patreon username +open_collective: # Replace with a single Open Collective username +ko_fi: # Replace with a single Ko-fi username +tidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel +community_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry +liberapay: spike +issuehunt: # Replace with a single IssueHunt username +otechie: # Replace with a single Otechie username +custom: # Replace with up to 4 custom sponsorship URLs e.g., ['link1', 'link2'] From 1a1dee36b8b9c6779ccc13f7c9b6ab5c6411281d Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Sat, 27 Nov 2021 22:55:27 +0000 Subject: [PATCH 24/27] Update FUNDING.yml --- .github/FUNDING.yml | 11 ----------- 1 file changed, 11 deletions(-) diff --git a/.github/FUNDING.yml b/.github/FUNDING.yml index a765c29..5346ceb 100644 --- a/.github/FUNDING.yml +++ b/.github/FUNDING.yml @@ -1,12 +1 @@ -# These are supported funding model platforms - -github: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2] -patreon: # Replace with a single Patreon username -open_collective: # Replace with a single Open Collective username -ko_fi: # Replace with a single Ko-fi username -tidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel -community_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry liberapay: spike -issuehunt: # Replace with a single IssueHunt username -otechie: # Replace with a single Otechie username -custom: # Replace with up to 4 custom sponsorship URLs e.g., ['link1', 'link2'] From 67e26479ae5f441964e0096334229b83013fd6aa Mon Sep 17 00:00:00 2001 From: Spike <19519553+spikecodes@users.noreply.github.com> Date: Sat, 27 Nov 2021 22:57:56 +0000 Subject: [PATCH 25/27] Add leddit.xyz instance. Closes #344 --- README.md | 1 + 1 file changed, 1 insertion(+) diff --git a/README.md b/README.md index 207c488..c37af39 100644 --- a/README.md +++ b/README.md @@ -57,6 +57,7 @@ Feel free to [open an issue](https://github.com/spikecodes/libreddit/issues/new) | [libreddit.pussthecat.org](https://libreddit.pussthecat.org) | 🇩🇪 DE | | | [libreddit.mutahar.rocks](https://libreddit.mutahar.rocks) | 🇫🇷 FR | | | [libreddit.northboot.xyz](https://libreddit.northboot.xyz) | 🇩🇪 DE | | +| [leddit.xyz](https://www.leddit.xyz) | 🇩🇪 DE | | | [spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion](http://spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion) | 🇮🇳 IN | | | [fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion](http://fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion) | 🇩🇪 DE | | | [kphht2jcflojtqte4b4kyx7p2ahagv4debjj32nre67dxz7y57seqwyd.onion](http://kphht2jcflojtqte4b4kyx7p2ahagv4debjj32nre67dxz7y57seqwyd.onion) | 🇳🇱 NL | | From 6f29d9433772e4b941216f6481d8818b9475e0c6 Mon Sep 17 00:00:00 2001 From: spikecodes <19519553+spikecodes@users.noreply.github.com> Date: Sat, 27 Nov 2021 15:07:44 -0800 Subject: [PATCH 26/27] List Liberapay as donation method --- README.md | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index c37af39..c6862b0 100644 --- a/README.md +++ b/README.md @@ -15,9 +15,13 @@ --- -**BTC:** bc1qwyxjnafpu3gypcpgs025cw9wa7ryudtecmwa6y +I appreciate any donations! Your support allows me to continue developing Libreddit. -**XMR:** 45FJrEuFPtG2o7QZz2Nps77TbHD4sPqxViwbdyV9A6ktfHiWs47UngG5zXPcLoDXAc8taeuBgeNjfeprwgeXYXhN3C9tVSR +**Liberapay:** + +**Bitcoin:** [bc1qwyxjnafpu3gypcpgs025cw9wa7ryudtecmwa6y](bitcoin:bc1qwyxjnafpu3gypcpgs025cw9wa7ryudtecmwa6y) + +**Monero:** [45FJrEuFPtG2o7QZz2Nps77TbHD4sPqxViwbdyV9A6ktfHiWs47UngG5zXPcLoDXAc8taeuBgeNjfeprwgeXYXhN3C9tVSR](monero:45FJrEuFPtG2o7QZz2Nps77TbHD4sPqxViwbdyV9A6ktfHiWs47UngG5zXPcLoDXAc8taeuBgeNjfeprwgeXYXhN3C9tVSR) --- From 91cc140091cb6994430eb8cc307e14cb9b067302 Mon Sep 17 00:00:00 2001 From: Nick Lowery(All content on this page has been filtered) + {% else %}{% for post in posts %} {% if post.flags.nsfw && prefs.show_nsfw != "on" %} - {% else if post.title != "Comment" %} + {% else if !post.title.is_empty() %} {% call utils::post_in_list(post) %} {% else %}+ {% endif %}@@ -55,6 +59,7 @@ {% endif %}+ {% endif %}Date: Sat, 27 Nov 2021 19:49:41 -0700 Subject: [PATCH 27/27] Set sub and user descriptions to `overflow-wrap: anywhere` (#345) --- static/style.css | 1 + 1 file changed, 1 insertion(+) diff --git a/static/style.css b/static/style.css index 514b642..3b6d20b 100644 --- a/static/style.css +++ b/static/style.css @@ -365,6 +365,7 @@ aside { #user_description, #sub_description { margin: 0 15px; text-align: left; + overflow-wrap: anywhere; } #user_name, #user_description:not(:empty), #user_icon,