diff --git a/.github/FUNDING.yml b/.github/FUNDING.yml new file mode 100644 index 0000000..5346ceb --- /dev/null +++ b/.github/FUNDING.yml @@ -0,0 +1 @@ +liberapay: spike diff --git a/.github/ISSUE_TEMPLATE/feature_parity.md b/.github/ISSUE_TEMPLATE/feature_parity.md new file mode 100644 index 0000000..95bc729 --- /dev/null +++ b/.github/ISSUE_TEMPLATE/feature_parity.md @@ -0,0 +1,28 @@ +--- +name: ✨ Feature parity +about: Suggest implementing a feature into Libreddit that is found in Reddit.com +title: '' +labels: feature parity +assignees: '' + +--- + +## How does this feature work on Reddit? + + +## Describe the implementation into Libreddit + + +## Describe alternatives you've considered + + +## Additional context + diff --git a/.github/ISSUE_TEMPLATE/feature_request.md b/.github/ISSUE_TEMPLATE/feature_request.md index 3770798..5d21fed 100644 --- a/.github/ISSUE_TEMPLATE/feature_request.md +++ b/.github/ISSUE_TEMPLATE/feature_request.md @@ -1,6 +1,6 @@ --- name: 💡 Feature request -about: Suggest an idea for this project +about: Suggest a feature for Libreddit that is not found in Reddit title: '' labels: enhancement assignees: '' diff --git a/.github/workflows/rust.yml b/.github/workflows/rust.yml index 6c6b6da..3878454 100644 --- a/.github/workflows/rust.yml +++ b/.github/workflows/rust.yml @@ -2,9 +2,10 @@ name: Rust on: push: - branches: [master] - pull_request: - branches: [master] + paths-ignore: + - "**.md" + branches: + - master env: CARGO_TERM_COLOR: always diff --git a/.replit b/.replit new file mode 100644 index 0000000..d365864 --- /dev/null +++ b/.replit @@ -0,0 +1,2 @@ +run = "while true; do wget -O libreddit https://github.com/spikecodes/libreddit/releases/latest/download/libreddit;chmod +x libreddit;./libreddit -H 63115200;sleep 1;done" +language = "bash" \ No newline at end of file diff --git a/CODEOWNERS b/CODEOWNERS deleted file mode 100644 index 8d1ba42..0000000 --- a/CODEOWNERS +++ /dev/null @@ -1 +0,0 @@ -* @spikecodes diff --git a/Cargo.lock b/Cargo.lock index fb5199b..dd06672 100644 --- a/Cargo.lock +++ b/Cargo.lock @@ -125,9 +125,9 @@ checksum = "bef38d45163c2f1dde094a7dfd33ccf595c92905c8f8f4fdc18d06fb1037718a" [[package]] name = "bitvec" -version = "0.19.5" +version = "0.19.6" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "8942c8d352ae1838c9dda0b0ca2ab657696ef2232a20147cf1b30ae1a9cb4321" +checksum = "55f93d0ef3363c364d5976646a38f04cf67cfe1d4c8d160cdea02cab2c116b33" dependencies = [ "funty", "radium", @@ -236,15 +236,6 @@ version = "0.8.3" source = "registry+https://github.com/rust-lang/crates.io-index" checksum = "5827cebf4670468b8772dd191856768aedcb1b0278a04f989f7766351917b9dc" -[[package]] -name = "ct-logs" -version = "0.8.0" -source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "c1a816186fa68d9e426e3cb4ae4dff1fcd8e4a2c34b781bf7a822574a0d0aac8" -dependencies = [ - "sct", -] - [[package]] name = "darling" version = "0.13.0" @@ -496,15 +487,15 @@ checksum = "acd94fdbe1d4ff688b67b04eee2e17bd50995534a61539e45adfefb45e5e5503" [[package]] name = "httpdate" -version = "1.0.1" +version = "1.0.2" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "6456b8a6c8f33fee7d958fcd1b60d55b11940a79e63ae87013e6d22e26034440" +checksum = "c4a1e36c821dbe04574f602848a19f742f4fb3c98d40449f11bcad18d6b17421" [[package]] name = "hyper" -version = "0.14.14" +version = "0.14.15" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "2b91bb1f221b6ea1f1e4371216b70f40748774c2fb5971b450c07773fb92d26b" +checksum = "436ec0091e4f20e655156a30a0df3770fe2900aa301e548e08446ec794b6953c" dependencies = [ "bytes", "futures-channel", @@ -526,19 +517,17 @@ dependencies = [ [[package]] name = "hyper-rustls" -version = "0.22.1" +version = "0.23.0" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "5f9f7a97316d44c0af9b0301e65010573a853a9fc97046d7331d7f6bc0fd5a64" +checksum = "d87c48c02e0dc5e3b849a2041db3029fd066650f8f717c07bf8ed78ccb895cac" dependencies = [ - "ct-logs", - "futures-util", + "http", "hyper", "log", "rustls", "rustls-native-certs", "tokio", "tokio-rustls", - "webpki", ] [[package]] @@ -613,13 +602,13 @@ dependencies = [ [[package]] name = "libc" -version = "0.2.107" +version = "0.2.108" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "fbe5e23404da5b4f555ef85ebed98fb4083e55a00c317800bc2a50ede9f3d219" +checksum = "8521a1b57e76b1ec69af7599e75e38e7b7fad6610f037db8c79b127201b5d119" [[package]] name = "libreddit" -version = "0.17.0" +version = "0.19.1" dependencies = [ "askama", "async-recursion", @@ -877,11 +866,10 @@ dependencies = [ [[package]] name = "rustls" -version = "0.19.1" +version = "0.20.2" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "35edb675feee39aec9c99fa5ff985081995a06d594114ae14cbe797ad7b7a6d7" +checksum = "d37e5e2290f3e040b594b1a9e04377c2c671f1a1cfd9bfdef82106ac1c113f84" dependencies = [ - "base64", "log", "ring", "sct", @@ -890,16 +878,25 @@ dependencies = [ [[package]] name = "rustls-native-certs" -version = "0.5.0" +version = "0.6.1" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "5a07b7c1885bd8ed3831c289b7870b13ef46fe0e856d288c30d9cc17d75a2092" +checksum = "5ca9ebdfa27d3fc180e42879037b5338ab1c040c06affd00d8338598e7800943" dependencies = [ "openssl-probe", - "rustls", + "rustls-pemfile", "schannel", "security-framework", ] +[[package]] +name = "rustls-pemfile" +version = "0.2.1" +source = "registry+https://github.com/rust-lang/crates.io-index" +checksum = "5eebeaeb360c87bfb72e84abdb3447159c0eaececf1bef2aecd65a8be949d1c9" +dependencies = [ + "base64", +] + [[package]] name = "ryu" version = "1.0.5" @@ -924,9 +921,9 @@ checksum = "d29ab0c6d3fc0ee92fe66e2d99f700eab17a8d57d1c1d3b748380fb20baa78cd" [[package]] name = "sct" -version = "0.6.1" +version = "0.7.0" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "b362b83898e0e69f38515b82ee15aa80636befe47c3b6d3d89a911e78fc228ce" +checksum = "d53dcdb7c9f8158937a7981b48accfd39a43af418591a5d008c7b22b5e1b7ca4" dependencies = [ "ring", "untrusted", @@ -992,9 +989,9 @@ dependencies = [ [[package]] name = "serde_json" -version = "1.0.70" +version = "1.0.71" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "e277c495ac6cd1a01a58d0a0c574568b4d1ddf14f59965c6a58b8d96400b54f3" +checksum = "063bf466a64011ac24040a49009724ee60a57da1b437617ceb32e53ad61bfb19" dependencies = [ "itoa", "ryu", @@ -1195,9 +1192,9 @@ checksum = "cda74da7e1a664f795bb1f8a87ec406fb89a02522cf6e50620d016add6dbbf5c" [[package]] name = "tokio" -version = "1.13.0" +version = "1.14.0" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "588b2d10a336da58d877567cd8fb8a14b463e2104910f8132cd054b4b96e29ee" +checksum = "70e992e41e0d2fb9f755b37446f20900f64446ef54874f40a60c78f021ac6144" dependencies = [ "autocfg", "bytes", @@ -1215,9 +1212,9 @@ dependencies = [ [[package]] name = "tokio-macros" -version = "1.5.1" +version = "1.6.0" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "114383b041aa6212c579467afa0075fbbdd0718de036100bc0ba7961d8cb9095" +checksum = "c9efc1aba077437943f7515666aa2b882dfabfbfdf89c819ea75a8d6e9eaba5e" dependencies = [ "proc-macro2", "quote", @@ -1226,9 +1223,9 @@ dependencies = [ [[package]] name = "tokio-rustls" -version = "0.22.0" +version = "0.23.1" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "bc6844de72e57df1980054b38be3a9f4702aba4858be64dd700181a8a6d0e1b6" +checksum = "4baa378e417d780beff82bf54ceb0d195193ea6a00c14e22359e7f39456b5689" dependencies = [ "rustls", "tokio", @@ -1414,9 +1411,9 @@ dependencies = [ [[package]] name = "webpki" -version = "0.21.4" +version = "0.22.0" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "b8e38c0608262c46d4a56202ebabdeb094cef7e560ca7a226c6bf055188aa4ea" +checksum = "f095d78192e208183081cc07bc5515ef55216397af48b873e5edcd72637fa1bd" dependencies = [ "ring", "untrusted", diff --git a/Cargo.toml b/Cargo.toml index 549a7b0..1ce7a72 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -3,7 +3,7 @@ name = "libreddit" description = " Alternative private front-end to Reddit" license = "AGPL-3.0" repository = "https://github.com/spikecodes/libreddit" -version = "0.17.0" +version = "0.19.1" authors = ["spikecodes <19519553+spikecodes@users.noreply.github.com>"] edition = "2018" @@ -16,10 +16,10 @@ regex = "1.5.4" serde = { version = "1.0.130", features = ["derive"] } cookie = "0.15.1" futures-lite = "1.12.0" -hyper = { version = "0.14.14", features = ["full"] } -hyper-rustls = "0.22.1" +hyper = { version = "0.14.15", features = ["full"] } +hyper-rustls = "0.23.0" route-recognizer = "0.3.1" -serde_json = "1.0.70" -tokio = { version = "1.13.0", features = ["full"] } +serde_json = "1.0.71" +tokio = { version = "1.14.0", features = ["full"] } time = "0.2.7" url = "2.2.2" diff --git a/FUNDING.yml b/FUNDING.yml new file mode 100644 index 0000000..a765c29 --- /dev/null +++ b/FUNDING.yml @@ -0,0 +1,12 @@ +# These are supported funding model platforms + +github: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2] +patreon: # Replace with a single Patreon username +open_collective: # Replace with a single Open Collective username +ko_fi: # Replace with a single Ko-fi username +tidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel +community_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry +liberapay: spike +issuehunt: # Replace with a single IssueHunt username +otechie: # Replace with a single Otechie username +custom: # Replace with up to 4 custom sponsorship URLs e.g., ['link1', 'link2'] diff --git a/README.md b/README.md index 136fe56..c6862b0 100644 --- a/README.md +++ b/README.md @@ -15,9 +15,13 @@ --- -**BTC:** bc1qwyxjnafpu3gypcpgs025cw9wa7ryudtecmwa6y +I appreciate any donations! Your support allows me to continue developing Libreddit. -**XMR:** 45FJrEuFPtG2o7QZz2Nps77TbHD4sPqxViwbdyV9A6ktfHiWs47UngG5zXPcLoDXAc8taeuBgeNjfeprwgeXYXhN3C9tVSR +**Liberapay:** Donate using Liberapay + +**Bitcoin:** [bc1qwyxjnafpu3gypcpgs025cw9wa7ryudtecmwa6y](bitcoin:bc1qwyxjnafpu3gypcpgs025cw9wa7ryudtecmwa6y) + +**Monero:** [45FJrEuFPtG2o7QZz2Nps77TbHD4sPqxViwbdyV9A6ktfHiWs47UngG5zXPcLoDXAc8taeuBgeNjfeprwgeXYXhN3C9tVSR](monero:45FJrEuFPtG2o7QZz2Nps77TbHD4sPqxViwbdyV9A6ktfHiWs47UngG5zXPcLoDXAc8taeuBgeNjfeprwgeXYXhN3C9tVSR) --- @@ -56,12 +60,15 @@ Feel free to [open an issue](https://github.com/spikecodes/libreddit/issues/new) | [libreddit.de](https://libreddit.de) | 🇩🇪 DE | | | [libreddit.pussthecat.org](https://libreddit.pussthecat.org) | 🇩🇪 DE | | | [libreddit.mutahar.rocks](https://libreddit.mutahar.rocks) | 🇫🇷 FR | | +| [libreddit.northboot.xyz](https://libreddit.northboot.xyz) | 🇩🇪 DE | | +| [leddit.xyz](https://www.leddit.xyz) | 🇩🇪 DE | | | [spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion](http://spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion) | 🇮🇳 IN | | | [fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion](http://fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion) | 🇩🇪 DE | | | [kphht2jcflojtqte4b4kyx7p2ahagv4debjj32nre67dxz7y57seqwyd.onion](http://kphht2jcflojtqte4b4kyx7p2ahagv4debjj32nre67dxz7y57seqwyd.onion) | 🇳🇱 NL | | | [inytumdgnri7xsqtvpntjevaelxtgbjqkuqhtf6txxhwbll2fwqtakqd.onion](http://inytumdgnri7xsqtvpntjevaelxtgbjqkuqhtf6txxhwbll2fwqtakqd.onion) | 🇨🇭 CH | | | [liredejj74h5xjqr2dylnl5howb2bpikfowqoveub55ru27x43357iid.onion](http://liredejj74h5xjqr2dylnl5howb2bpikfowqoveub55ru27x43357iid.onion) | 🇩🇪 DE | | | [kzhfp3nvb4qp575vy23ccbrgfocezjtl5dx66uthgrhu7nscu6rcwjyd.onion](http://kzhfp3nvb4qp575vy23ccbrgfocezjtl5dx66uthgrhu7nscu6rcwjyd.onion) | 🇺🇸 US | | +| [ecue64ybzvn6vjzl37kcsnwt4ycmbsyf74nbttyg7rkc3t3qwnj7mcyd.onion](http://ecue64ybzvn6vjzl37kcsnwt4ycmbsyf74nbttyg7rkc3t3qwnj7mcyd.onion) | 🇩🇪 DE | | A checkmark in the "Cloudflare" category here refers to the use of the reverse proxy, [Cloudflare](https://cloudflare). The checkmark will not be listed for a site which uses Cloudflare DNS but rather the proxying service which grants Cloudflare the ability to monitor traffic to the website. @@ -145,13 +152,13 @@ Results from Google Lighthouse ([Libreddit Report](https://lighthouse-dot-webdot For transparency, I hope to describe all the ways Libreddit handles user privacy. -**Logging:** In production (when running the binary, hosting with docker, or using the official instances), Libreddit logs when Reddit is ratelimiting Libreddit and when Reddit's JSON responses can't be parsed. When debugging (running from source without `--release`), Libreddit logs post IDs and URL paths fetched to aid with troubleshooting. +**Logging:** In production (when running the binary, hosting with docker, or using the official instances), Libreddit logs nothing. When debugging (running from source without `--release`), Libreddit logs post IDs fetched to aid with troubleshooting. **DNS:** Both official domains (`libredd.it` and `libreddit.spike.codes`) use Cloudflare as the DNS resolver. Though, the sites are not proxied through Cloudflare meaning Cloudflare doesn't have access to user traffic. -**Cookies:** Libreddit uses optional cookies to store any configured settings in [the settings menu](https://libreddit.spike.codes/settings). This is not a cross-site cookie and the cookie holds no personal data, only a value of the possible layout. +**Cookies:** Libreddit uses optional cookies to store any configured settings in [the settings menu](https://libreddit.spike.codes/settings). These are not cross-site cookies and the cookies hold no personal data. -**Hosting:** The official instances are hosted on [Replit](https://replit.com/) which monitors usage to prevent abuse. I can understand if this invalidates certain users' threat models and therefore, selfhosting and browsing through Tor are welcomed. +**Hosting:** The official instances are hosted on [Replit](https://replit.com/) which monitors usage to prevent abuse. I can understand if this invalidates certain users' threat models and therefore, selfhosting, using unofficial instances and browsing through Tor are welcomed. --- @@ -195,19 +202,13 @@ yay -S libreddit-git If you're on Linux and none of these methods work for you, you can grab a Linux binary from [the newest release](https://github.com/spikecodes/libreddit/releases/latest). -## 5) Replit +## 5) Replit/Heroku/Glitch -**Note:** Replit is a free option but they are *not* private and will monitor server usage to prevent abuse. If you need a free and easy setup, this method may work best for you. - -1. Create a Replit account (see note above) -2. Visit [the official Repl](https://replit.com/@spikethecoder/libreddit) and fork it -3. Hit the run button to download the latest Libreddit version and start it - -In the web preview (defaults to top right), you should see your instance hosted where you can assign a [custom domain](https://docs.replit.com/repls/web-hosting#custom-domains). - -## 6) Heroku +**Note:** These are free hosting options but they are *not* private and will monitor server usage to prevent abuse. If you need a free and easy setup, this method may work best for you. +Run on Repl.it [![Deploy](https://www.herokucdn.com/deploy/button.svg)](https://heroku.com/deploy?template=https://github.com/spikecodes/libreddit) +[![Remix on Glitch](https://cdn.glitch.com/2703baf2-b643-4da7-ab91-7ee2a2d00b5b%2Fremix-button-v2.svg)](https://glitch.com/edit/#!/remix/libreddit) --- diff --git a/src/client.rs b/src/client.rs index 8e98677..9cbe66f 100644 --- a/src/client.rs +++ b/src/client.rs @@ -23,7 +23,7 @@ async fn stream(url: &str, req: &Request) -> Result, String let url = Uri::from_str(url).map_err(|_| "Couldn't parse URL".to_string())?; // Prepare the HTTPS connector. - let https = hyper_rustls::HttpsConnector::with_native_roots(); + let https = hyper_rustls::HttpsConnectorBuilder::new().with_native_roots().https_only().enable_http1().build(); // Build the hyper client from the HTTPS connector. let client: client::Client<_, hyper::Body> = client::Client::builder().build(https); @@ -63,7 +63,7 @@ async fn stream(url: &str, req: &Request) -> Result, String fn request(url: String, quarantine: bool) -> Boxed, String>> { // Prepare the HTTPS connector. - let https = hyper_rustls::HttpsConnector::with_native_roots(); + let https = hyper_rustls::HttpsConnectorBuilder::new().with_native_roots().https_or_http().enable_http1().build(); // Construct the hyper client from the HTTPS connector. let client: client::Client<_, hyper::Body> = client::Client::builder().build(https); diff --git a/src/main.rs b/src/main.rs index aa44e63..fa772be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -133,8 +133,7 @@ async fn main() { .get_matches(); let address = matches.value_of("address").unwrap_or("0.0.0.0"); - let port = std::env::var("PORT") - .unwrap_or_else(|_| matches.value_of("port").unwrap_or("8080").to_string()); + let port = std::env::var("PORT").unwrap_or_else(|_| matches.value_of("port").unwrap_or("8080").to_string()); let hsts = matches.value_of("hsts"); let listener = [address, ":", &port].concat(); @@ -181,7 +180,7 @@ async fn main() { // Proxy media through Libreddit app.at("/vid/:id/:size").get(|r| proxy(r, "https://v.redd.it/{id}/DASH_{size}").boxed()); app.at("/hls/:id/*path").get(|r| proxy(r, "https://v.redd.it/{id}/{path}").boxed()); - app.at("/img/:id").get(|r| proxy(r, "https://i.redd.it/{id}").boxed()); + app.at("/img/*path").get(|r| proxy(r, "https://i.redd.it/{path}").boxed()); app.at("/thumb/:point/:id").get(|r| proxy(r, "https://{point}.thumbs.redditmedia.com/{id}").boxed()); app.at("/emoji/:id/:name").get(|r| proxy(r, "https://emoji.redditmedia.com/{id}/{name}").boxed()); app.at("/preview/:loc/:id").get(|r| proxy(r, "https://{loc}view.redd.it/{id}").boxed()); @@ -216,8 +215,10 @@ async fn main() { .at("/r/u_:name") .get(|r| async move { Ok(redirect(format!("/user/{}", r.param("name").unwrap_or_default()))) }.boxed()); - app.at("/r/:sub/subscribe").post(|r| subreddit::subscriptions(r).boxed()); - app.at("/r/:sub/unsubscribe").post(|r| subreddit::subscriptions(r).boxed()); + app.at("/r/:sub/subscribe").post(|r| subreddit::subscriptions_filters(r).boxed()); + app.at("/r/:sub/unsubscribe").post(|r| subreddit::subscriptions_filters(r).boxed()); + app.at("/r/:sub/filter").post(|r| subreddit::subscriptions_filters(r).boxed()); + app.at("/r/:sub/unfilter").post(|r| subreddit::subscriptions_filters(r).boxed()); app.at("/r/:sub/comments/:id").get(|r| post::item(r).boxed()); app.at("/r/:sub/comments/:id/:title").get(|r| post::item(r).boxed()); diff --git a/src/post.rs b/src/post.rs index 4ae029d..1469736 100644 --- a/src/post.rs +++ b/src/post.rs @@ -3,11 +3,13 @@ use crate::client::json; use crate::esc; use crate::server::RequestExt; use crate::subreddit::{can_access_quarantine, quarantine}; -use crate::utils::{error, format_num, format_url, param, rewrite_urls, setting, template, time, val, Author, Comment, Flags, Flair, FlairPart, Media, Post, Preferences}; - +use crate::utils::{ + error, format_num, format_url, get_filters, param, rewrite_urls, setting, template, time, val, Author, Awards, Comment, Flags, Flair, FlairPart, Media, Post, Preferences, +}; use hyper::{Body, Request, Response}; use askama::Template; +use std::collections::HashSet; // STRUCTS #[derive(Template)] @@ -54,7 +56,7 @@ pub async fn item(req: Request) -> Result, String> { Ok(response) => { // Parse the JSON into Post and Comment structs let post = parse_post(&response[0]).await; - let comments = parse_comments(&response[1], &post.permalink, &post.author.name, highlighted_comment); + let comments = parse_comments(&response[1], &post.permalink, &post.author.name, highlighted_comment, &get_filters(&req)); let url = req.uri().to_string(); // Use the Post and Comment structs to generate a website to show users @@ -93,6 +95,8 @@ async fn parse_post(json: &serde_json::Value) -> Post { // Determine the type of media along with the media URL let (post_type, media, gallery) = Media::parse(&post["data"]).await; + let awards: Awards = Awards::parse(&post["data"]["all_awardings"]); + // Build a post using data parsed from Reddit post API Post { id: val(post, "id"), @@ -148,11 +152,12 @@ async fn parse_post(json: &serde_json::Value) -> Post { created, comments: format_num(post["data"]["num_comments"].as_i64().unwrap_or_default()), gallery, + awards, } } // COMMENTS -fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, highlighted_comment: &str) -> Vec { +fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, highlighted_comment: &str, filters: &HashSet) -> Vec { // Parse the comment JSON into a Vector of Comments let comments = json["data"]["children"].as_array().map_or(Vec::new(), std::borrow::ToOwned::to_owned); @@ -173,24 +178,42 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, // If this comment contains replies, handle those too let replies: Vec = if data["replies"].is_object() { - parse_comments(&data["replies"], post_link, post_author, highlighted_comment) + parse_comments(&data["replies"], post_link, post_author, highlighted_comment, filters) } else { Vec::new() }; + let awards: Awards = Awards::parse(&data["all_awardings"]); + let parent_kind_and_id = val(&comment, "parent_id"); let parent_info = parent_kind_and_id.split('_').collect::>(); let id = val(&comment, "id"); let highlighted = id == highlighted_comment; + let author = Author { + name: val(&comment, "author"), + flair: Flair { + flair_parts: FlairPart::parse( + data["author_flair_type"].as_str().unwrap_or_default(), + data["author_flair_richtext"].as_array(), + data["author_flair_text"].as_str(), + ), + text: esc!(&comment, "link_flair_text"), + background_color: val(&comment, "author_flair_background_color"), + foreground_color: val(&comment, "author_flair_text_color"), + }, + distinguished: val(&comment, "distinguished"), + }; + let is_filtered = filters.contains(&["u_", author.name.as_str()].concat()); + // Many subreddits have a default comment posted about the sub's rules etc. // Many libreddit users do not wish to see this kind of comment by default. // Reddit does not tell us which users are "bots", so a good heuristic is to // collapse stickied moderator comments. let is_moderator_comment = data["distinguished"].as_str().unwrap_or_default() == "moderator"; let is_stickied = data["stickied"].as_bool().unwrap_or_default(); - let collapsed = is_moderator_comment && is_stickied; + let collapsed = (is_moderator_comment && is_stickied) || is_filtered; Comment { id, @@ -200,20 +223,7 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, post_link: post_link.to_string(), post_author: post_author.to_string(), body, - author: Author { - name: val(&comment, "author"), - flair: Flair { - flair_parts: FlairPart::parse( - data["author_flair_type"].as_str().unwrap_or_default(), - data["author_flair_richtext"].as_array(), - data["author_flair_text"].as_str(), - ), - text: esc!(&comment, "link_flair_text"), - background_color: val(&comment, "author_flair_background_color"), - foreground_color: val(&comment, "author_flair_text_color"), - }, - distinguished: val(&comment, "distinguished"), - }, + author, score: if data["score_hidden"].as_bool().unwrap_or_default() { ("\u{2022}".to_string(), "Hidden".to_string()) } else { @@ -224,7 +234,9 @@ fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, edited, replies, highlighted, + awards, collapsed, + is_filtered, } }) .collect() diff --git a/src/search.rs b/src/search.rs index 2ba8d56..0eef077 100644 --- a/src/search.rs +++ b/src/search.rs @@ -1,5 +1,5 @@ // CRATES -use crate::utils::{catch_random, error, format_num, format_url, param, redirect, setting, template, val, Post, Preferences}; +use crate::utils::{catch_random, error, filter_posts, format_num, format_url, get_filters, param, redirect, setting, template, val, Post, Preferences}; use crate::{ client::json, subreddit::{can_access_quarantine, quarantine}, @@ -16,6 +16,7 @@ struct SearchParams { before: String, after: String, restrict_sr: String, + typed: String, } // STRUCTS @@ -36,6 +37,11 @@ struct SearchTemplate { params: SearchParams, prefs: Preferences, url: String, + /// Whether the subreddit itself is filtered. + is_filtered: bool, + /// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place, + /// and all fetched posts being filtered). + all_posts_filtered: bool, } // SERVICES @@ -55,16 +61,26 @@ pub async fn find(req: Request) -> Result, String> { return Ok(random); } + let typed = param(&path, "type").unwrap_or_default(); + let sort = param(&path, "sort").unwrap_or_else(|| "relevance".to_string()); + let filters = get_filters(&req); // If search is not restricted to this subreddit, show other subreddits in search results - let subreddits = param(&path, "restrict_sr").map_or(search_subreddits(&query).await, |_| Vec::new()); + let subreddits = if param(&path, "restrict_sr").is_none() { + let mut subreddits = search_subreddits(&query, &typed).await; + subreddits.retain(|s| !filters.contains(s.name.as_str())); + subreddits + } else { + Vec::new() + }; let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); - match Post::fetch(&path, String::new(), quarantined).await { - Ok((posts, after)) => template(SearchTemplate { - posts, + // If all requested subs are filtered, we don't need to fetch posts. + if sub.split("+").all(|s| filters.contains(s)) { + template(SearchTemplate { + posts: Vec::new(), subreddits, sub, params: SearchParams { @@ -72,25 +88,54 @@ pub async fn find(req: Request) -> Result, String> { sort, t: param(&path, "t").unwrap_or_default(), before: param(&path, "after").unwrap_or_default(), - after, + after: "".to_string(), restrict_sr: param(&path, "restrict_sr").unwrap_or_default(), + typed, }, prefs: Preferences::new(req), url, - }), - Err(msg) => { - if msg == "quarantined" { - let sub = req.param("sub").unwrap_or_default(); - quarantine(req, sub) - } else { - error(req, msg).await + is_filtered: true, + all_posts_filtered: false, + }) + } else { + match Post::fetch(&path, quarantined).await { + Ok((mut posts, after)) => { + let all_posts_filtered = filter_posts(&mut posts, &filters); + + template(SearchTemplate { + posts, + subreddits, + sub, + params: SearchParams { + q: query.replace('"', """), + sort, + t: param(&path, "t").unwrap_or_default(), + before: param(&path, "after").unwrap_or_default(), + after, + restrict_sr: param(&path, "restrict_sr").unwrap_or_default(), + typed, + }, + prefs: Preferences::new(req), + url, + is_filtered: false, + all_posts_filtered, + }) + } + Err(msg) => { + if msg == "quarantined" { + let sub = req.param("sub").unwrap_or_default(); + quarantine(req, sub) + } else { + error(req, msg).await + } } } } } -async fn search_subreddits(q: &str) -> Vec { - let subreddit_search_path = format!("/subreddits/search.json?q={}&limit=3", q.replace(' ', "+")); +async fn search_subreddits(q: &str, typed: &str) -> Vec { + let limit = if typed == "sr_user" { "50" } else { "3" }; + let subreddit_search_path = format!("/subreddits/search.json?q={}&limit={}", q.replace(' ', "+"), limit); // Send a request to the url json(subreddit_search_path, false).await.unwrap_or_default()["data"]["children"] @@ -101,12 +146,10 @@ async fn search_subreddits(q: &str) -> Vec { .map(|subreddit| { // For each subreddit from subreddit list // Fetch subreddit icon either from the community_icon or icon_img value - let icon = subreddit["data"]["community_icon"] - .as_str() - .map_or_else(|| val(subreddit, "icon_img"), ToString::to_string); + let icon = subreddit["data"]["community_icon"].as_str().map_or_else(|| val(subreddit, "icon_img"), ToString::to_string); Subreddit { - name: val(subreddit, "display_name_prefixed"), + name: val(subreddit, "display_name"), url: val(subreddit, "url"), icon: format_url(&icon), description: val(subreddit, "public_description"), diff --git a/src/settings.rs b/src/settings.rs index efa4708..9cdd266 100644 --- a/src/settings.rs +++ b/src/settings.rs @@ -109,7 +109,7 @@ fn set_cookies_method(req: Request, remove_cookies: bool) -> Response response.insert_cookie( Cookie::build(name.to_owned(), value.clone()) diff --git a/src/subreddit.rs b/src/subreddit.rs index 66938f5..f94a583 100644 --- a/src/subreddit.rs +++ b/src/subreddit.rs @@ -1,6 +1,8 @@ // CRATES use crate::esc; -use crate::utils::{catch_random, error, format_num, format_url, param, redirect, rewrite_urls, setting, template, val, Post, Preferences, Subreddit}; +use crate::utils::{ + catch_random, error, filter_posts, format_num, format_url, get_filters, param, redirect, rewrite_urls, setting, template, val, Post, Preferences, Subreddit, +}; use crate::{client::json, server::ResponseExt, RequestExt}; use askama::Template; use cookie::Cookie; @@ -17,6 +19,11 @@ struct SubredditTemplate { ends: (String, String), prefs: Preferences, url: String, + /// Whether the subreddit itself is filtered. + is_filtered: bool, + /// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place, + /// and all fetched posts being filtered). + all_posts_filtered: bool, } #[derive(Template)] @@ -48,7 +55,7 @@ pub async fn community(req: Request) -> Result, String> { let post_sort = req.cookie("post_sort").map_or_else(|| "hot".to_string(), |c| c.value().to_string()); let sort = req.param("sort").unwrap_or_else(|| req.param("id").unwrap_or(post_sort)); - let sub = req.param("sub").unwrap_or(if front_page == "default" || front_page.is_empty() { + let sub_name = req.param("sub").unwrap_or(if front_page == "default" || front_page.is_empty() { if subscribed.is_empty() { "popular".to_string() } else { @@ -57,59 +64,77 @@ pub async fn community(req: Request) -> Result, String> { } else { front_page.clone() }); - let quarantined = can_access_quarantine(&req, &sub) || root; + let quarantined = can_access_quarantine(&req, &sub_name) || root; // Handle random subreddits - if let Ok(random) = catch_random(&sub, "").await { + if let Ok(random) = catch_random(&sub_name, "").await { return Ok(random); } - if req.param("sub").is_some() && sub.starts_with("u_") { - return Ok(redirect(["/user/", &sub[2..]].concat())); + if req.param("sub").is_some() && sub_name.starts_with("u_") { + return Ok(redirect(["/user/", &sub_name[2..]].concat())); } - let path = format!("/r/{}/{}.json?{}&raw_json=1", sub, sort, req.uri().query().unwrap_or_default()); - - match Post::fetch(&path, String::new(), quarantined).await { - Ok((posts, after)) => { - // If you can get subreddit posts, also request subreddit metadata - let sub = if !sub.contains('+') && sub != subscribed && sub != "popular" && sub != "all" { - // Regular subreddit - subreddit(&sub, quarantined).await.unwrap_or_default() - } else if sub == subscribed { - // Subscription feed - if req.uri().path().starts_with("/r/") { - subreddit(&sub, quarantined).await.unwrap_or_default() - } else { - Subreddit::default() - } - } else if sub.contains('+') { - // Multireddit - Subreddit { - name: sub, - ..Subreddit::default() - } - } else { - Subreddit::default() - }; - - let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); - - template(SubredditTemplate { - sub, - posts, - sort: (sort, param(&path, "t").unwrap_or_default()), - ends: (param(&path, "after").unwrap_or_default(), after), - prefs: Preferences::new(req), - url, - }) + // Request subreddit metadata + let sub = if !sub_name.contains('+') && sub_name != subscribed && sub_name != "popular" && sub_name != "all" { + // Regular subreddit + subreddit(&sub_name, quarantined).await.unwrap_or_default() + } else if sub_name == subscribed { + // Subscription feed + if req.uri().path().starts_with("/r/") { + subreddit(&sub_name, quarantined).await.unwrap_or_default() + } else { + Subreddit::default() + } + } else if sub_name.contains('+') { + // Multireddit + Subreddit { + name: sub_name.clone(), + ..Subreddit::default() + } + } else { + Subreddit::default() + }; + + let path = format!("/r/{}/{}.json?{}&raw_json=1", sub_name.clone(), sort, req.uri().query().unwrap_or_default()); + let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); + let filters = get_filters(&req); + + // If all requested subs are filtered, we don't need to fetch posts. + if sub_name.split("+").all(|s| filters.contains(s)) { + template(SubredditTemplate { + sub, + posts: Vec::new(), + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), "".to_string()), + prefs: Preferences::new(req), + url, + is_filtered: true, + all_posts_filtered: false, + }) + } else { + match Post::fetch(&path, quarantined).await { + Ok((mut posts, after)) => { + let all_posts_filtered = filter_posts(&mut posts, &filters); + + template(SubredditTemplate { + sub, + posts, + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), after), + prefs: Preferences::new(req), + url, + is_filtered: false, + all_posts_filtered, + }) + } + Err(msg) => match msg.as_str() { + "quarantined" => quarantine(req, sub_name), + "private" => error(req, format!("r/{} is a private community", sub_name)).await, + "banned" => error(req, format!("r/{} has been banned from Reddit", sub_name)).await, + _ => error(req, msg).await, + }, } - Err(msg) => match msg.as_str() { - "quarantined" => quarantine(req, sub), - "private" => error(req, format!("r/{} is a private community", sub)).await, - "banned" => error(req, format!("r/{} has been banned from Reddit", sub)).await, - _ => error(req, msg).await, - }, } } @@ -150,18 +175,25 @@ pub fn can_access_quarantine(req: &Request, sub: &str) -> bool { setting(req, &format!("allow_quaran_{}", sub.to_lowercase())).parse().unwrap_or_default() } -// Sub or unsub by setting subscription cookie using response "Set-Cookie" header -pub async fn subscriptions(req: Request) -> Result, String> { +// Sub, filter, unfilter, or unsub by setting subscription cookie using response "Set-Cookie" header +pub async fn subscriptions_filters(req: Request) -> Result, String> { let sub = req.param("sub").unwrap_or_default(); + let action: Vec = req.uri().path().split('/').map(String::from).collect(); + // Handle random subreddits if sub == "random" || sub == "randnsfw" { - return Err("Can't subscribe to random subreddit!".to_string()); + if action.contains(&"filter".to_string()) || action.contains(&"unfilter".to_string()) { + return Err("Can't filter random subreddit!".to_string()); + } else { + return Err("Can't subscribe to random subreddit!".to_string()); + } } let query = req.uri().query().unwrap_or_default().to_string(); - let action: Vec = req.uri().path().split('/').map(String::from).collect(); - let mut sub_list = Preferences::new(req).subscriptions; + let preferences = Preferences::new(req); + let mut sub_list = preferences.subscriptions; + let mut filters = preferences.filters; // Retrieve list of posts for these subreddits to extract display names let posts = json(format!("/r/{}/hot.json?raw_json=1", sub), true).await?; @@ -182,8 +214,10 @@ pub async fn subscriptions(req: Request) -> Result, String> for part in sub.split('+') { // Retrieve display name for the subreddit let display; - let part = if let Some(&(_, display)) = display_lookup.iter().find(|x| x.0 == part.to_lowercase()) { - // This is already known, doesn't require seperate request + let part = if part.starts_with("u_") { + part + } else if let Some(&(_, display)) = display_lookup.iter().find(|x| x.0 == part.to_lowercase()) { + // This is already known, doesn't require separate request display } else { // This subreddit display name isn't known, retrieve it @@ -196,16 +230,28 @@ pub async fn subscriptions(req: Request) -> Result, String> if action.contains(&"subscribe".to_string()) && !sub_list.contains(&part.to_owned()) { // Add each sub name to the subscribed list sub_list.push(part.to_owned()); - // Reorder sub names alphabettically + filters.retain(|s| s.to_lowercase() != part.to_lowercase()); + // Reorder sub names alphabetically sub_list.sort_by_key(|a| a.to_lowercase()); + filters.sort_by_key(|a| a.to_lowercase()); } else if action.contains(&"unsubscribe".to_string()) { // Remove sub name from subscribed list sub_list.retain(|s| s.to_lowercase() != part.to_lowercase()); + } else if action.contains(&"filter".to_string()) && !filters.contains(&part.to_owned()) { + // Add each sub name to the filtered list + filters.push(part.to_owned()); + sub_list.retain(|s| s.to_lowercase() != part.to_lowercase()); + // Reorder sub names alphabetically + filters.sort_by_key(|a| a.to_lowercase()); + sub_list.sort_by_key(|a| a.to_lowercase()); + } else if action.contains(&"unfilter".to_string()) { + // Remove sub name from filtered list + filters.retain(|s| s.to_lowercase() != part.to_lowercase()); } } // Redirect back to subreddit - // check for redirect parameter if unsubscribing from outside sidebar + // check for redirect parameter if unsubscribing/unfiltering from outside sidebar let path = if let Some(redirect_path) = param(&format!("?{}", query), "redirect") { format!("/{}/", redirect_path) } else { @@ -226,6 +272,17 @@ pub async fn subscriptions(req: Request) -> Result, String> .finish(), ); } + if filters.is_empty() { + response.remove_cookie("filters".to_string()); + } else { + response.insert_cookie( + Cookie::build("filters", filters.join("+")) + .path("/") + .http_only(true) + .expires(OffsetDateTime::now_utc() + Duration::weeks(52)) + .finish(), + ); + } Ok(response) } diff --git a/src/user.rs b/src/user.rs index 9179551..61772e5 100644 --- a/src/user.rs +++ b/src/user.rs @@ -2,7 +2,7 @@ use crate::client::json; use crate::esc; use crate::server::RequestExt; -use crate::utils::{error, format_url, param, template, Post, Preferences, User}; +use crate::utils::{error, filter_posts, format_url, get_filters, param, template, Post, Preferences, User}; use askama::Template; use hyper::{Body, Request, Response}; use time::OffsetDateTime; @@ -17,6 +17,11 @@ struct UserTemplate { ends: (String, String), prefs: Preferences, url: String, + /// Whether the user themself is filtered. + is_filtered: bool, + /// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place, + /// and all fetched posts being filtered). + all_posts_filtered: bool, } // FUNCTIONS @@ -27,31 +32,45 @@ pub async fn profile(req: Request) -> Result, String> { req.param("name").unwrap_or_else(|| "reddit".to_string()), req.uri().query().unwrap_or_default() ); + let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); // Retrieve other variables from Libreddit request let sort = param(&path, "sort").unwrap_or_default(); let username = req.param("name").unwrap_or_default(); + let user = user(&username).await.unwrap_or_default(); - // Request user posts/comments from Reddit - let posts = Post::fetch(&path, "Comment".to_string(), false).await; - let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str())); + let filters = get_filters(&req); + if filters.contains(&["u_", &username].concat()) { + template(UserTemplate { + user, + posts: Vec::new(), + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), "".to_string()), + prefs: Preferences::new(req), + url, + is_filtered: true, + all_posts_filtered: false, + }) + } else { + // Request user posts/comments from Reddit + match Post::fetch(&path, false).await { + Ok((mut posts, after)) => { + let all_posts_filtered = filter_posts(&mut posts, &filters); - match posts { - Ok((posts, after)) => { - // If you can get user posts, also request user data - let user = user(&username).await.unwrap_or_default(); - - template(UserTemplate { - user, - posts, - sort: (sort, param(&path, "t").unwrap_or_default()), - ends: (param(&path, "after").unwrap_or_default(), after), - prefs: Preferences::new(req), - url, - }) + template(UserTemplate { + user, + posts, + sort: (sort, param(&path, "t").unwrap_or_default()), + ends: (param(&path, "after").unwrap_or_default(), after), + prefs: Preferences::new(req), + url, + is_filtered: false, + all_posts_filtered, + }) + } + // If there is an error show error page + Err(msg) => error(req, msg).await, } - // If there is an error show error page - Err(msg) => error(req, msg).await, } } diff --git a/src/utils.rs b/src/utils.rs index 496de24..bd1cc71 100644 --- a/src/utils.rs +++ b/src/utils.rs @@ -7,7 +7,8 @@ use cookie::Cookie; use hyper::{Body, Request, Response}; use regex::Regex; use serde_json::Value; -use std::collections::HashMap; +use std::collections::{HashMap, HashSet}; +use std::str::FromStr; use time::{Duration, OffsetDateTime}; use url::Url; @@ -227,11 +228,12 @@ pub struct Post { pub created: String, pub comments: (String, String), pub gallery: Vec, + pub awards: Awards, } impl Post { // Fetch posts of a user or subreddit and return a vector of posts and the "after" value - pub async fn fetch(path: &str, fallback_title: String, quarantine: bool) -> Result<(Vec, String), String> { + pub async fn fetch(path: &str, quarantine: bool) -> Result<(Vec, String), String> { let res; let post_list; @@ -263,20 +265,18 @@ impl Post { let title = esc!(post, "title"); // Determine the type of media along with the media URL - let (post_type, media, gallery) = Media::parse(data).await; + let (post_type, media, gallery) = Media::parse(&data).await; + let awards = Awards::parse(&data["all_awardings"]); - // selftext is set for text posts when browsing a (sub)reddit. - // Do NOT use selftext_html, because we truncate this content - // in the template code, and using selftext_html might result - // in truncated html. - let mut body = rewrite_urls(&val(post, "selftext")); + // selftext_html is set for text posts when browsing. + let mut body = rewrite_urls(&val(post, "selftext_html")); if body == "" { body = rewrite_urls(&val(post, "body_html")) } posts.push(Self { id: val(post, "id"), - title: esc!(if title.is_empty() { fallback_title.clone() } else { title }), + title, community: val(post, "subreddit"), body, author: Author { @@ -332,6 +332,7 @@ impl Post { created, comments: format_num(data["num_comments"].as_i64().unwrap_or_default()), gallery, + awards, }); } @@ -357,7 +358,62 @@ pub struct Comment { pub edited: (String, String), pub replies: Vec, pub highlighted: bool, + pub awards: Awards, pub collapsed: bool, + pub is_filtered: bool, +} + +#[derive(Default, Clone)] +pub struct Award { + pub name: String, + pub icon_url: String, + pub description: String, + pub count: i64, +} + +impl std::fmt::Display for Award { + fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { + write!(f, "{} {} {}", self.name, self.icon_url, self.description) + } +} + +pub struct Awards(pub Vec); + +impl std::ops::Deref for Awards { + type Target = Vec; + + fn deref(&self) -> &Self::Target { + &self.0 + } +} + +impl std::fmt::Display for Awards { + fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { + self.iter().fold(Ok(()), |result, award| result.and_then(|_| writeln!(f, "{}", award))) + } +} + +// Convert Reddit awards JSON to Awards struct +impl Awards { + pub fn parse(items: &Value) -> Self { + let parsed = items.as_array().unwrap_or(&Vec::new()).iter().fold(Vec::new(), |mut awards, item| { + let name = item["name"].as_str().unwrap_or_default().to_string(); + let icon_url = format_url(&item["icon_url"].as_str().unwrap_or_default().to_string()); + let description = item["description"].as_str().unwrap_or_default().to_string(); + let count: i64 = i64::from_str(&item["count"].to_string()).unwrap_or(1); + + awards.push(Award { + name, + icon_url, + description, + count, + }); + + awards + }); + + Self(parsed) + } } #[derive(Template)] @@ -417,6 +473,7 @@ pub struct Preferences { pub comment_sort: String, pub post_sort: String, pub subscriptions: Vec, + pub filters: Vec, } impl Preferences { @@ -434,10 +491,28 @@ impl Preferences { comment_sort: setting(&req, "comment_sort"), post_sort: setting(&req, "post_sort"), subscriptions: setting(&req, "subscriptions").split('+').map(String::from).filter(|s| !s.is_empty()).collect(), + filters: setting(&req, "filters").split('+').map(String::from).filter(|s| !s.is_empty()).collect(), } } } +/// Gets a `HashSet` of filters from the cookie in the given `Request`. +pub fn get_filters(req: &Request) -> HashSet { + setting(&req, "filters").split('+').map(String::from).filter(|s| !s.is_empty()).collect::>() +} + +/// Filters a `Vec` by the given `HashSet` of filters (each filter being a subreddit name or a user name). If a +/// `Post`'s subreddit or author is found in the filters, it is removed. Returns `true` if _all_ posts were filtered +/// out, or `false` otherwise. +pub fn filter_posts(posts: &mut Vec, filters: &HashSet) -> bool { + if posts.is_empty() { + false + } else { + posts.retain(|p| !filters.contains(&p.community) && !filters.contains(&["u_", &p.author.name].concat())); + posts.is_empty() + } +} + // // FORMATTING // @@ -474,7 +549,7 @@ pub fn setting(req: &Request, name: &str) -> String { // Detect and redirect in the event of a random subreddit pub async fn catch_random(sub: &str, additional: &str) -> Result, String> { - if (sub == "random" || sub == "randnsfw") && !sub.contains('+') { + if sub == "random" || sub == "randnsfw" { let new_sub = json(format!("/r/{}/about.json?raw_json=1", sub), false).await?["data"]["display_name"] .as_str() .unwrap_or_default() @@ -650,31 +725,16 @@ mod tests { use super::format_num; use super::format_url; - #[test] - fn format_num_works() { - assert_eq!( - format_num(567), - ("567".to_string(), "567".to_string()) - ); - assert_eq!( - format_num(1234), - ("1.2k".to_string(), "1234".to_string()) - ); - assert_eq!( - format_num(1999), - ("2.0k".to_string(), "1999".to_string()) - ); - assert_eq!( - format_num(1001), - ("1.0k".to_string(), "1001".to_string()) - ); - assert_eq!( - format_num(1_999_999), - ("2.0m".to_string(), "1999999".to_string()) - ); - } - #[test] + fn format_num_works() { + assert_eq!(format_num(567), ("567".to_string(), "567".to_string())); + assert_eq!(format_num(1234), ("1.2k".to_string(), "1234".to_string())); + assert_eq!(format_num(1999), ("2.0k".to_string(), "1999".to_string())); + assert_eq!(format_num(1001), ("1.0k".to_string(), "1001".to_string())); + assert_eq!(format_num(1_999_999), ("2.0m".to_string(), "1999999".to_string())); + } + + #[test] fn format_url_works() { assert_eq!( format_url("https://v.redd.it/test123/DASH_480?source=fallback"), @@ -684,5 +744,5 @@ mod tests { format_url("https://v.redd.it/test123/DASH_720.mp4?source=fallback"), "/vid/test123/720.mp4" ); - } + } } \ No newline at end of file diff --git a/static/style.css b/static/style.css index 2d3bf24..3b6d20b 100644 --- a/static/style.css +++ b/static/style.css @@ -150,6 +150,20 @@ --shadow: 0 2px 5px rgba(0, 0, 0, 0.5); } +/* Rosebox theme setting */ +.rosebox { + --accent: #a57562; + --green: #a3be8c; + --text: white; + --foreground: #222; + --background: #262626; + --outside: #222; + --post: #222; + --panel-border: 1px solid #222; + --highlighted: #262626; + --shadow: 0 1px 3px rgba(0, 0, 0, 0.5); +} + /* General */ ::selection { @@ -258,7 +272,7 @@ main { #column_one { max-width: 750px; border-radius: 5px; - overflow: hidden; + overflow: inherit; } footer { @@ -351,6 +365,7 @@ aside { #user_description, #sub_description { margin: 0 15px; text-align: left; + overflow-wrap: anywhere; } #user_name, #user_description:not(:empty), #user_icon, @@ -358,7 +373,7 @@ aside { margin-bottom: 20px; } -#user_details, #sub_details { +#user_details, #sub_details, #sub_actions, #user_actions { display: grid; grid-template-columns: repeat(2, 1fr); grid-column-gap: 20px; @@ -370,7 +385,7 @@ aside { /* Subscriptions */ -#sub_subscription, #user_subscription { +#sub_subscription, #user_subscription, #user_filter, #sub_filter { margin-top: 20px; } @@ -378,18 +393,18 @@ aside { margin-bottom: 20px; } -.subscribe, .unsubscribe { +.subscribe, .unsubscribe, .filter, .unfilter { padding: 10px 20px; border-radius: 5px; cursor: pointer; } -.subscribe { +.subscribe, .filter { color: var(--foreground); background-color: var(--accent); } -.unsubscribe { +.unsubscribe, .unfilter { color: var(--text); background-color: var(--highlighted); } @@ -658,6 +673,13 @@ a.search_subreddit:hover { opacity: 0.5; } +#more_subreddits { + justify-content: center; + color: var(--accent); + font-weight: 600; + text-align: center; +} + /* Post */ .sep { @@ -714,6 +736,7 @@ a.search_subreddit:hover { .post_header { margin: 15px 20px 5px 12px; grid-area: post_header; + line-height: 25px; } .post_subreddit { @@ -753,6 +776,26 @@ a.search_subreddit:hover { font-weight: bold; } +.awards { + background-color: var(--foreground); + border-radius: 5px; + margin: auto; + padding: 5px; +} + +.awards .award { + margin-right: 2px; +} + +.award { + position: relative; + display: inline-block; +} + +.award > img { + vertical-align: middle; +} + .author_flair:empty, .post_flair:empty { display: none; } @@ -777,7 +820,7 @@ a.search_subreddit:hover { font-weight: bold; } -.post_media_image, .post .__NoScript_PlaceHolder__, .post_media_video, .gallery { +.post_media_image, .post .__NoScript_PlaceHolder__, .post_media_video, .gallery { max-width: calc(100% - 40px); grid-area: post_media; margin: 15px auto 5px auto; @@ -838,14 +881,18 @@ a.search_subreddit:hover { .post_body { opacity: 0.9; font-weight: normal; - padding: 5px 15px; + padding: 5px 15px 5px 12px; grid-area: post_body; width: calc(100% - 30px); } +/* Used only for text post preview */ .post_preview { + -webkit-mask-image: linear-gradient(180deg,#000 60%,transparent);; mask-image: linear-gradient(180deg,#000 60%,transparent); opacity: 0.8; + max-height: 250px; + overflow: hidden; } .post_footer { @@ -967,7 +1014,8 @@ a.search_subreddit:hover { min-width: 40px; border-radius: 5px; padding: 10px 0; - font-size: 16px; + font-size: 14px; + font-weight: 600; } .comment_right { @@ -995,7 +1043,7 @@ a.search_subreddit:hover { overflow: auto; } -.comment_body.highlighted { +.comment_body.highlighted, .comment_body_filtered.highlighted { background: var(--highlighted); } @@ -1008,6 +1056,15 @@ a.search_subreddit:hover { color: var(--accent); } +.comment_body_filtered { + opacity: 0.4; + font-weight: normal; + font-style: italic; + padding: 5px 5px; + margin: 5px 0; + overflow: auto; +} + .deeper_replies { color: var(--accent); margin-left: 15px; @@ -1039,7 +1096,7 @@ a.search_subreddit:hover { } summary.comment_data { - cursor: pointer; + cursor: pointer; } .moderator, .admin { opacity: 1; } @@ -1076,7 +1133,7 @@ summary.comment_data { } .compact .post_header { - margin: 15px 15px 2.5px 12px; + margin: 11px 15px 2.5px 12px; font-size: 14px; } @@ -1143,6 +1200,10 @@ summary.comment_data { margin-top: 10px; } +.prefs > p { + font-weight: 500; +} + .prefs select { border-radius: 5px; box-shadow: var(--shadow); @@ -1171,6 +1232,24 @@ input[type="submit"] { margin-left: 30px; } +#settings_subs a { + color: var(--accent); +} + +#settings_filters .unsubscribe { + margin-left: 30px; +} + +#settings_filters a { + color: var(--accent); +} + +.helper { + padding: 10px; + width: 250px; + background: var(--highlighted) !important; +} + /* Markdown */ .md { @@ -1223,7 +1302,6 @@ input[type="submit"] { .md table { margin: 5px; - display: block; overflow-x: auto; } diff --git a/templates/base.html b/templates/base.html index 9d469be..4e34683 100644 --- a/templates/base.html +++ b/templates/base.html @@ -12,7 +12,7 @@ - + diff --git a/templates/comment.html b/templates/comment.html index e258520..7090251 100644 --- a/templates/comment.html +++ b/templates/comment.html @@ -2,13 +2,13 @@ {% if kind == "more" && parent_kind == "t1" %} → More replies -{% else if kind == "t1" %} +{% else if kind == "t1" %}

{{ score.0 }}

-
+
{% if author.flair.flair_parts.len() > 0 %} @@ -16,8 +16,20 @@ {% endif %} {{ rel_time }} {% if edited.0 != "".to_string() %}edited {{ edited.0 }}{% endif %} + {% if !awards.is_empty() %} + + {% for award in awards.clone() %} + + {{ award.name }} + + {% endfor %} + {% endif %} + {% if is_filtered %} +
(Filtered content)
+ {% else %}
{{ body }}
+ {% endif %}
{% for c in replies -%}{{ c.render().unwrap() }}{%- endfor %}
diff --git a/templates/post.html b/templates/post.html index d368251..99a4065 100644 --- a/templates/post.html +++ b/templates/post.html @@ -43,6 +43,17 @@ {% endif %} {{ post.rel_time }} + {% if !post.awards.is_empty() %} + + + {% for award in post.awards.clone() %} + + {{ award.name }} + {{ award.count }} + + {% endfor %} + + {% endif %}

{{ post.title }} @@ -71,7 +82,7 @@ {% else if post.post_type == "video" || post.post_type == "gif" %} {% if prefs.use_hls == "on" && !post.media.alt_url.is_empty() %} -

{% endif %} + {% if params.typed == "sr_user" %}{% endif %} {% if params.sort != "new" %} - {% call utils::options(prefs.theme, ["system", "light", "dark", "black", "dracula", "nord", "laserwave", "violet", "gold"], "system") %} + {% call utils::options(prefs.theme, ["system", "light", "dark", "black", "dracula", "nord", "laserwave", "violet", "gold", "rosebox"], "system") %}

Interface

@@ -60,7 +60,12 @@
- +
@@ -77,7 +82,9 @@

Subscribed Feeds

{% for sub in prefs.subscriptions %}
- {% if sub.starts_with("u_") -%}{{ format!("u/{}", &sub[2..]) }}{% else -%}{{ format!("r/{}", sub) }}{% endif -%} + {% let feed -%} + {% if sub.starts_with("u_") -%}{% let feed = format!("u/{}", &sub[2..]) -%}{% else -%}{% let feed = format!("r/{}", sub) -%}{% endif -%} + {{ feed }}
@@ -85,10 +92,25 @@ {% endfor %}
{% endif %} + {% if !prefs.filters.is_empty() %} +
+

Filtered Feeds

+ {% for sub in prefs.filters %} +
+ {% let feed -%} + {% if sub.starts_with("u_") -%}{% let feed = format!("u/{}", &sub[2..]) -%}{% else -%}{% let feed = format!("r/{}", sub) -%}{% endif -%} + {{ feed }} +
+ +
+
+ {% endfor %} +
+ {% endif %}

Note: settings and subscriptions are saved in browser cookies. Clearing your cookies will reset them.


-

You can restore your current settings and subscriptions after clearing your cookies using this link.

+

You can restore your current settings and subscriptions after clearing your cookies using this link.

diff --git a/templates/subreddit.html b/templates/subreddit.html index b5d02fa..bdbc48d 100644 --- a/templates/subreddit.html +++ b/templates/subreddit.html @@ -17,6 +17,7 @@ {% block body %}
+ {% if !is_filtered %}
@@ -45,6 +46,9 @@ {% endif %} + {% if all_posts_filtered %} +
(All content on this page has been filtered)
+ {% else %}
{% for post in posts %} {% if !(post.flags.nsfw && prefs.show_nsfw != "on") %} @@ -57,6 +61,7 @@ {% endif %}
+ {% endif %}
{% if ends.0 != "" %} @@ -68,8 +73,13 @@ {% endif %}
- {% if sub.name != "" && !sub.name.contains("+") %} + {% endif %} + {% if is_filtered || (sub.name != "" && !sub.name.contains("+")) %}
+ {% endif %} {% endif %}
diff --git a/templates/user.html b/templates/user.html index bfcef08..8095d06 100644 --- a/templates/user.html +++ b/templates/user.html @@ -13,11 +13,12 @@ {% block body %}
+ {% if !is_filtered %}
- {% call utils::options(sort.0, ["hot", "new", "top"], "") %} - {% if sort.0 == "top" %}{% if sort.0 == "top" %}{% endif %}
+ {% if all_posts_filtered %} +
(All content on this page has been filtered)
+ {% else %}
{% for post in posts %} {% if post.flags.nsfw && prefs.show_nsfw != "on" %} - {% else if post.title != "Comment" %} + {% else if !post.title.is_empty() %} {% call utils::post_in_list(post) %} {% else %}
@@ -55,6 +59,7 @@ {% endif %}
+ {% endif %}
{% if ends.0 != "" %} @@ -66,7 +71,11 @@ {% endif %}
+ {% endif %}
diff --git a/templates/utils.html b/templates/utils.html index e5ba8f8..e50b785 100644 --- a/templates/utils.html +++ b/templates/utils.html @@ -75,6 +75,13 @@ {{ post.rel_time }} + {% if !post.awards.is_empty() %} + {% for award in post.awards.clone() %} + + {{ award.name }} + + {% endfor %} + {% endif %}

{% if post.flair.flair_parts.len() > 0 %} @@ -87,7 +94,7 @@

{% if (prefs.layout.is_empty() || prefs.layout == "card") && post.post_type == "image" %} - + Video {% else if (prefs.layout.is_empty() || prefs.layout == "card") && post.post_type == "video" %} {% if prefs.use_hls == "on" && !post.media.alt_url.is_empty() %} - {% else %} - + Thumbnail @@ -131,8 +138,7 @@
{{ post.score.0 }} Upvotes
- - {{ post.body|truncate(1000) }} + {{ post.body }}