Compare commits

...

1248 Commits

Author SHA1 Message Date
1c1e627815 v0.35.3 2024-10-20 19:31:07 +13:00
4b854eb84d fix/change feed list alignment 2024-10-20 19:20:01 +13:00
712790acbe fix footer spacing 2024-10-20 18:08:26 +13:00
062d810aad settings rearrangement 2024-10-20 18:06:26 +13:00
64eec64ebe Merge remote-tracking branch 'upstream/main' 2024-10-20 17:32:13 +13:00
3ff907d6c1 additional new colour tweaks (#285) 2024-10-12 11:42:15 -04:00
f4a457e529 Add additional themes to README (#284) 2024-10-11 15:43:45 -04:00
7dda8d9bbb use better accent colour + add libreddit styles (#281)
* update favicon with new logo

* only have 32x32 .ico file

* use #d74253 as new accent colour + add old libreddit styles + bolden accented buttons

* fix unrenamed libreddit themes
2024-10-10 18:45:39 -04:00
b99412b4a1 feat: update logos and accents (#280)
* feat: update logos, accents

* fix apple-touch-icon.png size

* remove width, length

---------

Co-authored-by: DokterKaj <dokterkaj@gmail.com>
2024-10-06 15:19:37 -04:00
1838fdaea4 Replace askama with rinja (#276) 2024-10-02 17:43:13 -04:00
f71b0cd178 chore(deps): Update brotli from 6.0 to 7.0 (#277)
* chore(deps): Update brotli from 6.0 to 7.0

* Update Cargo.lock for brotli 7.0
2024-10-02 14:18:41 -04:00
604db902e9 Fix systemd service (#275)
This format is not recognized by systemd. As shown in the following log:

/etc/systemd/system/redlib.service:33: System call ~@privileged is not known, ignoring.
/etc/systemd/system/redlib.service:33: System call ~@resources is not known, ignoring.
2024-10-01 15:55:42 -04:00
e57eaa0b78 fix(issue): render checkbox in issue template 2024-09-29 22:31:23 -04:00
31ad8c5f7b fix(issue): add checkbox for latest commit 2024-09-29 16:29:38 -04:00
5aef97410c Update redlib.conf (#271)
This setting expects an array of subs and not a boolean value. 
This confuses new users and also seems to be unintentional. 

Removing the comment character would lead to an error output on the start page. “Couldn't send request to Reddit: Post url contains non-ASCII characters | /r/off (sub1%2Bsub2%2Bsub3)/hot.json?&raw_json=1”
2024-09-29 14:45:03 -04:00
8d0ed4682e feat(search): redirect u/ and user/ to profile (#268) 2024-09-27 08:29:33 -04:00
fe4fed0504 Make jump to comment work from user's page and make last <p> in the post and comment bodies get the margin above it like the others (#219)
* Make last <p> in post body get padding above it

* When going to a comment from a user's page jump to it on page load
2024-09-27 00:44:32 -04:00
6e2e679a0e chore(oauth): add additional logging to login routine 2024-09-26 15:06:39 -04:00
6b44c1abf2 chore(oauth): add additional logging to login routine 2024-09-26 15:04:36 -04:00
a807002ddf Fix #206 and make (most) emotes embed in comments (#209)
* Fix links not being converted when multiple emojis are in one comment

* Make (most) emotes embed within comments

* Restore the behavior that the "rewrite_urls_removes_backslashes_and_rewrites_url" test looks for

* Listen to cargo fmt and cargo clippy's suggestions as well as removing some leftover comments and code

---------

Co-authored-by: Matthew Esposito <matt@matthew.science>
2024-09-25 13:36:23 -04:00
403513ac4c fix(search): handle queries' urlencoding (#264)
* fix(search): handle queries' urlencoding

* fix(search): handle queries' urlencoding
2024-09-24 23:30:06 -04:00
72f7d9d08c fix(search): handle multi-sub search (#263) 2024-09-24 23:20:12 -04:00
e6273e2ed5 fix(client): catch json suspended user error (#262)
* fix(client): catch json suspended user error
2024-09-24 23:13:36 -04:00
f1d4e6a417 fix(client): catch various json errors to properly render error page (#261)
* fix(client): catch various json errors to properly render error page

* fix(client): catch various json errors to properly render error page
2024-09-24 23:01:28 -04:00
e0d7837c02 fix(client): don't catch network policy errors, since they indicate q… (#259) 2024-09-24 21:45:47 -04:00
2d6ac78acf chore(client): update new oauth path (#258) 2024-09-24 21:28:54 -04:00
1e54c639d3 fix(client): add a timeout and retry logic to oauth daemon (#256)
* fix(client): add a timeout and retry logic to oauth daemon

* fix(client): add a timeout and retry logic to oauth daemon
2024-09-24 21:02:12 -04:00
d5f137ce47 fix(funding): add sponsor link 2024-09-21 15:48:19 -04:00
245fd9d408 fix(funding): update funding 2024-09-21 15:47:44 -04:00
b54620b5aa fix(client): use async_recursion crate 2024-09-21 15:44:27 -04:00
69c7a69afd add description for rss item (just like https://news.ycombinator.com/… (#220)
fix #201
2024-09-21 00:05:32 -04:00
2991813c2d Make poll results appear inside of a post (#218) 2024-09-21 00:02:34 -04:00
7156be6ad0 fix(client): fix failing tests, retries for canonical_path 2024-09-20 23:57:18 -04:00
f9b1a832c4 .dockerignore 2024-09-19 14:17:02 +12:00
1ffbce1ad8 Merge remote-tracking branch 'upstream/main' 2024-09-19 14:14:17 +12:00
793047f63f fix(client): revert to hyper-rustls=0.24.2 2024-09-18 11:24:00 -04:00
3625fdfdbe fix(ci): temporarily disable README updates 2024-09-17 14:42:09 -04:00
28f85f2599 Attempting to fix main-docker.yml for downloading digests 2024-09-17 14:39:42 -04:00
7be29f609c Attempting to fix main-docker.yml for downloading digests (#243)
* Update main-docker.yml (digests)

Updated digest name on upload (as per 1187084)

* Update main-docker.yml (digests download)

Trying another fix based on the template provided here https://github.com/actions/download-artifact for downloading multiple (filtered) Artifacts to the same directory
2024-09-17 14:37:39 -04:00
f18d135045 Update main-docker.yml (digests) (#238)
Updated digest name on upload (as per 1187084)
2024-09-17 14:09:32 -04:00
f2c410f9ea Merge remote-tracking branch 'upstream/main' 2024-09-17 12:28:21 +12:00
8ef45456d6 actions trigger 2024-09-16 16:39:07 -04:00
118708400a fix(ci): unique name 2024-09-16 16:36:24 -04:00
28e72c9058 fix(ci): bump versions (#234) 2024-09-16 16:29:52 -04:00
0b15250cc8 fix(oauth): catch network policy violation and rate limit (#233) 2024-09-16 16:16:08 -04:00
408ebe6ef1 feat(post): add archive.is link for link posts (#165) 2024-09-05 12:00:09 -04:00
Pim
7a0ea1fbd3 fix: spoiler hover and video size (#196)
* fix: unblur both media as body when hoevering over either one

* fix: set min size for video

when the video is not loaded, the size is determined by the poster image. But when the poster image returns a 404, then the video had a size of 0x0

* chore: tab/space
2024-09-05 11:59:43 -04:00
db8b92ea55 Fix a whole bunch of styling bugs (#193)
* Fix a whole bunch of mobile styling bugs

* Make searchbox scroll fix only apply in mobile mode to prevent bug

* Remove the min-width requirement for the main column

This was meant to be removed already, this is what fixes posts having an odd right side gap before swapping to the mobile layout

* Make margins consistent between fixed and unfixed navbar settings

* Remove some empty space from deleted option

* Make mobile layout post width fix only apply in mobile mode to prevent bug

* Make sure some options only get applied to the elements that need them, also fix the margins on the settings page

* Move search comments option before it starts touching the sort options and wrapping the x amount of comments text

* Trigger the even further compacted layout a little earlier, right before text begins wrapping in odd ways

* In the extra small mobile layout make give up/downvote numbers enough room so they aren't clipping out of their box

* Fix https://github.com/redlib-org/redlib/issues/172

* Properly center search box instead of having it slightly skewed

* Undo word wrapping since it breaks the sorting options and the only other viable setting has an absolute conniption on Chrome for some reason

* Readd word wrapping and just force it to normal for the sorting section

* Make post flair line up with title

* Make post flair position consistent

* Make footer text properly horizontally centered in mobile mode and fix slight vertical misalignment issues

* Make feeds button appear in settings menu to keep navbar looking consistent

* Fix extra navbar padding on search page

* Reduce gap between navbar and content in mobile mode

* Reduce gap between navbar and content in mobile mode
2024-09-05 11:59:21 -04:00
c494fbec31 Insert noindex meta for ROBOTS_DISABLE_INDEXING (#199) (#207) 2024-09-03 19:44:04 -04:00
438e412be3 Add anchor to comment link (#212) 2024-09-03 19:21:33 -04:00
d0e081e6a0 chore(deps): bump actions/download-artifact from 3 to 4.1.7 in /.github/workflows (#214)
Bumps [actions/download-artifact](https://github.com/actions/download-artifact) from 3 to 4.1.7.
- [Release notes](https://github.com/actions/download-artifact/releases)
- [Commits](https://github.com/actions/download-artifact/compare/v3...v4.1.7)

---
updated-dependencies:
- dependency-name: actions/download-artifact
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-09-03 19:21:11 -04:00
b076b076e3 v0.35.2 2024-08-31 10:20:57 +12:00
fdaba8344d set video download MIME
fixes #11
2024-08-31 10:18:05 +12:00
041ecceeaf Update license tag (#200) 2024-08-07 14:09:54 -04:00
3677ca10e2 Fix sort options overflow on small screens (#192) 2024-07-25 20:26:27 -04:00
21fd34710c Fix font sizes in search bar being incorrect (#190) 2024-07-24 19:56:06 -04:00
4205959ade fix(docs): Add env HIDE_SIDEBAR_AND_SUMMARY (#188) 2024-07-23 21:02:15 -04:00
27b56c1781 fix(client): handle new gated and quarantined error types (#187)
* fix(client): handle new gated and quarantined error types

* test(client): add test for gated and quarantined
2024-07-21 14:22:54 -04:00
374238abc3 Add RSS feeds (fix #57) (#90)
* Add RSS feeds

* feat(rss): feature-ify rss

* feat(rss): config-ify rss

* fix(rss): update info page

* feat(rss): conditionally add RSS feeds to user and sub pages

* feat(rss): implement URLs for RSS
2024-07-21 14:09:34 -04:00
Pim
9bdb5c8966 feat: also blur text in post body for spoilers (#186)
chore: simplify blur classes
2024-07-21 10:22:40 -04:00
410872d988 Make search bar responsive on mobile devices (#178)
* Search: Apply bg on elements rather than container

This changes allows moving the individual elements that composes
the search bar around without losing the background on the elements.

* Update search widget semantic structure

* Make search bar design responsive on small screens

* Fix border color

* Polish
2024-07-17 21:28:50 -04:00
c890e809b7 Improve post information widget of user comments for devices under 480px width (#183)
* Fix user comment post link disappearing when < 480px

* Improve user comment metadata design for mobile

* Remove formerly unused CSS class style

The prior commit introduced the usage of the `comment_subreddit` class
to identify the subreddit that the reddit user posted the comment on.

However, this class came with a legacy style within the CSS file that was
previously not used anywhere within the current day Redlib

As such this style has been removed.
2024-07-17 21:27:13 -04:00
eedbe06af6 Merge branch 'ugly-layouts'
sure why not :)
- last words 2024
2024-07-07 19:34:27 +12:00
6042bca612 mosaic -> waterfall 2024-07-07 18:35:53 +12:00
7e85ae25b1 Merge remote-tracking branch 'upstream/main'
+ repository link update
2024-07-07 17:19:28 +12:00
Pim
4f21388643 fix: also use hls if possible for gifs in post_in_list macro (#177) 2024-07-05 16:33:06 -04:00
Pim
8a917fcde3 feat: add download button on image/gif/video posts (#173)
* feat: add download button on image/gif/video posts

* chore: fix formatting

* chore: dont create reference
2024-07-04 21:32:12 -04:00
67a890cab3 fix(posts): fix sort call on new (#171) 2024-07-02 08:04:27 -04:00
Pim
366bc17f97 feat: show post link title for comments on user page (#169) 2024-07-01 17:15:50 -04:00
d9e7681004 v0.35.1 2024-06-29 13:28:18 -04:00
f74d1affb6 fix(posts): manually sort by flags (#168)
* fix(posts): manually sort by flags

* fix(posts): shorten sort call
2024-06-29 13:26:09 -04:00
f44638a2cb v0.35.0 2024-06-29 12:00:34 -04:00
beb4cf193b fix(posts): manually sort by created date (#166) 2024-06-29 11:48:42 -04:00
c565ebfb01 refactor(log): update some logs 2024-06-29 10:44:33 -04:00
2b8112a3fb mosaic layout 2024-06-30 00:04:46 +12:00
459a8e1245 refactor(log): shorten some logs 2024-06-29 00:20:19 -04:00
0f7eba717e fix(client): Handle invalid reddit response of base URL location 2024-06-28 22:41:36 -04:00
ea87ec33a1 fix(subreddit): handle plus-encoding errors even better (#163)
* fix(subreddit): handle plus-encoding errors even better

* chore(clippy): fix lint
2024-06-28 22:28:58 -04:00
102cd2f23f Merge pull request #162 from redlib-org/oauth_arc_swap
fix(oauth): arc_swap
2024-06-28 18:17:00 -04:00
3b2ad212d5 fix(oauth): arc_swap 2024-06-28 18:14:47 -04:00
4dc7ff8165 Merge pull request #160 from redlib-org/oauth_oppenheimer
fix(oauth): even more atomics to avoid simultaneous token rollover
2024-06-27 23:35:51 -04:00
2f8a38d8c7 chore(clippy): fix lint 2024-06-27 23:34:27 -04:00
13083e999c fix(oauth): handle extremely rare race condition by atomically compare_exchanging 2024-06-27 23:32:17 -04:00
4e2ec3fbc9 fix(oauth): handle case where a rate limit sneaks in 2024-06-27 23:29:55 -04:00
89313f73e6 fix(oauth): atomics to avoid simultaneous token rollover 2024-06-27 23:26:31 -04:00
1644e68e43 v0.35.1 2024-06-28 13:34:12 +12:00
f5d4cc49fb fix some default options #7 2024-06-28 13:33:14 +12:00
f22f7841a8 Merge remote-tracking branch 'upstream/main' 2024-06-28 12:40:15 +12:00
b6c6e64cfe move the restore note
my scroll wheel will fall off if i have to scroll to get down there ohmagaw
2024-06-28 12:31:31 +12:00
6d3db31b11 Merge pull request 'README updates' (#9) from README-update into main
Reviewed-on: #9
2024-06-28 00:34:03 +12:00
f0eb496d6a fix README.md 2024-06-28 00:32:20 +12:00
e984646ae6 Update README.md
add information about the name, a thank you and bits here and there
2024-06-28 00:07:29 +12:00
60c0d63583 nsfw_landing style & function changes
theme the NSFW title and allow the user to directly change their setting
2024-06-27 17:43:43 +12:00
3bd8b511a7 fix(oauth): strengthen sync guarantees 2024-06-26 23:41:26 -04:00
8c5aaaa33d feat(scripts): add load testing 2024-06-26 23:40:31 -04:00
bd0a312487 add disclaimer 2024-06-27 15:04:58 +12:00
023cc8505b Merge pull request #158 from redlib-org/oauth_proper_atomics
fix(oauth): reset rate limit earlier in refresh cycle
2024-06-26 22:20:00 -04:00
2e476dea63 fix(oauth): reset rate limit earlier in refresh cycle 2024-06-26 22:16:41 -04:00
f398d16c22 fix ffmpeg_video_downloads on cookie restore 2024-06-27 14:13:16 +12:00
d045a5760a Merge pull request #156 from redlib-org/fix_oauth_ratelimit
feat(oauth): roll over oauth key on rate limit
2024-06-26 19:33:04 -04:00
07bf20dbc0 feat(oauth): roll over oauth key on rate limit 2024-06-26 19:19:30 -04:00
518bf03e04 fix(client): Add trace logging for ratelimit info, render error page if exceeded 2024-06-26 08:05:22 -04:00
00dee52320 chore(deps): Cargo update 2024-06-25 20:42:06 -04:00
48873c01b9 Merge pull request #154 from redlib-org/fix_multi_sub
fix(subreddit): handle plus-encoding errors
2024-06-25 19:52:33 -04:00
951fe400ae fix(subreddit): handle plus-encoding errors 2024-06-25 19:50:00 -04:00
bacc9e35df refactor(oauth): leave android header unmodified (fixes #131) 2024-06-25 19:28:41 -04:00
724b960112 Merge pull request #149 from pimlie/feat-blur-spoiler-previews
feat: add support to blur spoiler previews
2024-06-23 10:28:46 -04:00
ff2e274e5e v0.35.0 2024-06-23 20:48:26 +12:00
7c821d541a fix PREFS size
whoops
2024-06-23 20:44:39 +12:00
7aca5791b1 Merge branch 'hls-video-downloads' 2024-06-23 20:41:27 +12:00
8b3783096b revert ugly-layouts from hls-video-downloads
This reverts commits
97def1bab1
feda45f803
2024-06-23 20:33:32 +12:00
59b702aa15 kinda horrific download implementation 2024-06-23 19:57:23 +12:00
edf0109095 allow wasm to run 2024-06-23 19:31:30 +12:00
ecf0f91464 remove failed hls.js source map request 2024-06-23 19:30:44 +12:00
69f9d9ff3c feat: also blur spoiler previews on post view 2024-06-22 12:47:33 +02:00
1d44bd180e feat: add spoiler badge to post title 2024-06-22 12:38:13 +02:00
3301da1ef1 feat: add support to blur spoiler previews 2024-06-22 12:16:12 +02:00
138934f30c change resource links and container links 2024-06-21 23:49:37 +12:00
5396c0783e add mascots to user settings readme 2024-06-21 23:11:23 +12:00
213481ef53 Merge pull request #148 from Pix3l01/fix-healthcheck
Make server listen on both IPv6 and IPv4 by default to fix docker healthcheck
2024-06-20 08:06:27 -04:00
2d5cfcb41d Merge pull request #146 from ac615223s5/default-filter
Default filters
2024-06-20 07:58:41 -04:00
f460895cc0 chore(clippy): fix lint 2024-06-20 07:56:43 -04:00
5193164719 Merge branch 'redlib-org:main' into default-filter 2024-06-19 20:10:05 -04:00
9013e589dd chore(clippy): fix lint 2024-06-19 14:45:55 -04:00
997cd8f829 feat(bug): Improve bug reporting while keeping logs private 2024-06-19 14:45:32 -04:00
5a13b9892b chore(clippy): add lint 2024-06-19 14:28:48 -04:00
91975865b8 Make server listen on both IPv6 and IPv4 by default 2024-06-19 00:42:38 +02:00
3491e754ac Update .gitignore 2024-06-18 15:16:41 -04:00
a5a5cbb734 add defaults for mascot fixes some of #7 2024-06-18 21:00:02 +12:00
5117af4137 add hide_sidebar_and_summary to restore link 2024-06-18 20:51:02 +12:00
394e975724 Merge remote-tracking branch 'upstream/main' 2024-06-18 19:57:42 +12:00
a67cbb99a6 sync upstream 2024-06-18 19:56:04 +12:00
1408c32a4d Update .env.example 2024-06-18 00:25:29 -04:00
30944579d7 add default filter config 2024-06-18 00:21:00 -04:00
ededa849f4 Merge pull request 'mascots' (#6) from mascots into main solves #5
Reviewed-on: #6
2024-06-17 22:44:22 +12:00
1d8b6f58fb fix mascot css 2024-06-17 22:40:58 +12:00
4c6a71171b add boymoder 2024-06-17 22:29:58 +12:00
846377b586 implement mascots 2024-06-17 22:29:52 +12:00
2dbcc071a6 update CREDITS 2024-06-17 18:00:58 +12:00
a12596e1a2 remove FFmpeg from subreddit template 2024-06-16 20:02:57 +12:00
df8d36c661 serve ffmpeg.wasm 2024-06-16 19:53:16 +12:00
5018184e7c rename ffmpeg.wasm to FFmpeg 2024-06-16 18:59:47 +12:00
5c73f043cd add preference option 2024-06-16 15:01:17 +12:00
013e805ece Merge branch 'main' into ugly-layouts 2024-06-15 17:32:35 +12:00
9a7da3abce Merge pull request #142 from runofthemillgeek/fix/font-weight-in-font-face
Add font-weight range in @font-face rule
2024-06-13 23:13:27 -04:00
7fa37e48d5 Add font-weight range in @font-face rule
This makes Safari render the bold weights correctly which otherwise
looks wrong.
2024-06-11 21:16:44 +05:30
21da37246e update catppuccin
discernible difference in foreground color
2024-06-06 04:54:03 +12:00
fb0814babd change alignment fix 2024-06-06 04:43:43 +12:00
bae220d4fe fix settings selection alignment 2024-06-06 04:30:32 +12:00
97def1bab1 Disallow wide ui with old layout 2024-06-06 04:26:47 +12:00
a6f901c094 Merge pull request #128 from arulagrawal/nix-flake
add nix flake
2024-06-04 22:49:33 -04:00
afad65f204 Merge pull request #127 from pimlie/patch-1
fix: healthcheck in Dockerfile
2024-06-04 22:48:51 -04:00
c12da45059 Merge pull request #121 from bennettmsherman/patch-1
Properly apply REDLIB_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY
2024-06-04 22:48:33 -04:00
10d5b4a583 fix dockerfile 2024-06-05 10:38:15 +12:00
feda45f803 "old" layout draft 2024-06-05 10:31:41 +12:00
1132d73975 Merge pull request #134 from EMarshal/main 2024-06-02 22:08:38 -04:00
8ece7562ec Remove obsolete references to latest-arm and latest-armv7 images 2024-06-02 18:45:49 -06:00
08e463fd44 Merge pull request #126 from Harm133/main
Create build-artifacts on release
2024-06-02 18:17:05 -04:00
f8f964741a super minimal rebrand 2024-06-02 22:33:45 +12:00
58c736c51b fix build not including git hash 2024-06-02 22:33:45 +12:00
336d5dcaa7 Update screenshot
yeah boymoder let your tits out
2024-06-02 01:04:14 +12:00
4c00e4ac05 Update README.md
fix claustrophobia
2024-06-02 00:37:07 +12:00
92afbe55de half rebrand the README 2024-06-02 00:34:41 +12:00
ec11a5511b add nix flake 2024-05-31 17:10:36 +04:00
9133062e68 fix healthcheck 2024-06-01 00:02:08 +12:00
Pim
3caa0592f3 fix: healthcheck in Dockerfile
Fix typo causing wget to return a warning:

```
# wget --q
wget: option '--q' is ambiguous; possibilities: '--quiet' '--quota'
```
2024-05-31 11:40:01 +02:00
190c92339e Update build-artifacts.yaml
Only build artifacts on published release
2024-05-31 10:19:44 +02:00
83d50316fb moved repo org 2024-05-31 20:00:00 +12:00
e187ccd14a Merge remote-tracking branch 'upstream/main' 2024-05-31 19:54:04 +12:00
284e03ccf9 version 2024-05-31 19:48:23 +12:00
17c7738d6e Properly apply REDLIB_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY
This change fixes the REDLIB_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY environment variable being ignored.
2024-05-30 20:40:16 -07:00
8a3ceaf94a Merge pull request #119 from redlib-org/fix_unauth
fix(oauth): Make Android user-agent patching unconditional
2024-05-30 18:13:43 -04:00
bd47c206a1 fix(oauth): Make Android user-agent patching unconditional 2024-05-30 18:08:45 -04:00
7a099f259f v0.34.0 2024-05-29 20:59:27 -04:00
4ea911e6b2 fix workflow again 2024-05-29 20:51:39 -04:00
31d68afdc9 fix workflow 2024-05-29 20:50:38 -04:00
96a7e155c5 v0.33.3 2024-05-29 20:49:29 -04:00
045a8852ec fix workflow 2024-05-29 20:47:44 -04:00
4cc8bf8318 v0.33.2 2024-05-29 20:05:56 -04:00
1715b36ae9 Update fix for submitted route 2024-05-29 20:02:56 -04:00
6102b08894 v0.33.1 2024-05-29 19:24:09 -04:00
892b0e89c8 Cargo fmt 2024-05-29 19:15:54 -04:00
a64e2143f3 Fix Display impl, resolving clippy 2024-05-29 18:51:20 -04:00
e13d9b7239 Merge pull request #86 from EMarshal/main
Update hide HLS env var to match documentation
2024-05-29 18:49:40 -04:00
e597ed8f06 Merge pull request #87 from ButteredCats/fix_post_footer_text
Stop post footer text from disappearing when screen is exactly 480px wide
2024-05-29 18:49:15 -04:00
8c67d33721 Merge pull request #102 from ismailkarsli/main
added geo_filter query param to /r/popular endpoint
2024-05-29 18:48:48 -04:00
26aa374bbd Merge pull request #114 from ufUNnxagpM/main
feat: add icebergDark theme
2024-05-29 18:47:19 -04:00
5b8b1e36ca Merge pull request #106 from axeII/feat/hide-summary-and-sidebar
Feature: Adds an option in the settings to hide the summary sidebar.
2024-05-29 18:46:50 -04:00
093d240530 Merge branch 'main' into feat/hide-summary-and-sidebar 2024-05-29 18:46:10 -04:00
9048565d48 Merge pull request #112 from Myzel394/fix-docs
Add `PORT` env to docs
2024-05-29 18:45:25 -04:00
6b2aab23c8 Merge pull request #92 from ButteredCats/fix_preview_captions
Update embedding Reddit preview links to include captions and support i.redd.it embeds
2024-05-29 18:44:48 -04:00
6b11d936b3 Fix clippys 2024-05-29 18:44:19 -04:00
34692359cf Merge branch 'redlib-org:main' into fix_preview_captions 2024-05-29 18:42:39 -04:00
33411a7588 Bump version 2024-05-29 18:40:09 -04:00
5a0f0f96f5 Update OAuth resources 2024-05-29 18:37:13 -04:00
273d889f1b Fix retrieval of multi-subs 2024-05-29 18:36:56 -04:00
93594fc642 feat: add icebergDark theme 2024-05-24 22:16:49 -07:00
af6f3d3b3f fix(docs): Add PORT env 2024-05-24 13:21:02 +02:00
62b791bb24 Add in support for embedding i.redd.it images and gifs and remove leftover println 2024-05-23 21:29:36 -04:00
e9af28b6eb Make sure that the extra <p></p> at the bottom of a comment containing only an image doesn't get a margin 2024-05-23 20:31:40 -04:00
2f2cded671 Make sure new system can handle both normal and external previews 2024-05-22 17:22:10 -04:00
b22fb7cd7b Fix embedded preview images having a gap from the top of a comment 2024-05-22 16:40:57 -04:00
75b0149313 Remove useless replace 2024-05-22 16:34:39 -04:00
50fad938dd Fix infinite loop when replacing text that contains dollar signs 2024-05-22 16:31:07 -04:00
b6f5831d10 feat: adds hide summary sidebar option 2024-05-13 23:49:59 +02:00
565b50646f added geo_filter query param to /r/popular endpoint 2024-04-25 14:01:06 +03:00
6484ebf897 Fix failing check 2024-04-14 17:32:10 -04:00
3f863c8991 Prevent panic if image_caption is empty, don't replace <p>'s in case text is inside them with the image, update test to reflect change in image replacing 2024-04-14 17:26:43 -04:00
e581f432dd Use substring instead of .remove and .pop, change image_text to image_caption to better reflect its usage, only replace quotes in image_caption when needed, and add comments for what some of the code does 2024-04-10 10:47:24 -04:00
2c8f5a7ac1 Make figure margins only apply to comments to bring embedded previews more in line with gallery posts 2024-04-09 18:51:32 -04:00
6d83b07aaa Update embedding Reddit preview links to include captions where applicable 2024-04-09 18:33:13 -04:00
5f67eacf93 update build 2024-04-09 22:19:40 +12:00
6684db5e82 sidebar open by default 2024-04-09 06:20:04 +12:00
1760f5a676 pagination arrows 2024-04-09 06:19:44 +12:00
8307aa1e42 show fork info 2024-04-09 06:19:00 +12:00
27f25e0fb1 Merge pull request #88 from ButteredCats/better_preview_handling
Handle preview embedding better
2024-04-08 10:04:23 -04:00
c15f25e27b Handle preview embedding better 2024-04-09 01:27:40 +12:00
20e3129d88 fix catppuccin 2024-04-09 01:13:04 +12:00
54c2ffad95 add catppuccin theme 2024-04-09 01:00:04 +12:00
6991fc6ae3 add build 2024-04-09 01:00:04 +12:00
30b27aea44 Update Dockerfile.old 2024-04-09 01:00:04 +12:00
63e2fadf4a monumental fuckup :D
revert revert
2024-04-09 01:00:04 +12:00
4ba26285d7 revert e79242c9e7
revert v0.31.2
2024-04-09 01:00:04 +12:00
89140c8cf7 Actually fix checks this time 2024-04-07 19:59:54 -04:00
99048c4683 Fix failing checks 2024-04-07 19:51:20 -04:00
ccfe7d0eeb Make embedded images keep aspect ration when shrinking 2024-04-07 19:24:55 -04:00
f7c182dcd8 Fix a couple of edge cases with image embedding and don't check if REDDIT_PREVIEW_REGEX matches before executing loop 2024-04-07 19:22:20 -04:00
9bd540d659 Stop post footer text from disappear at exactly 480px 2024-04-07 17:23:24 -04:00
4f6a14739b Update hide HLS env var to match documentation 2024-04-07 14:43:39 -06:00
7c87d63d34 Merge pull request #85 from ButteredCats/embed_images
Embed Reddit preview links
2024-04-07 14:01:47 -04:00
75b139dff2 Update image link test to account for embedded images 2024-04-07 12:08:53 -04:00
10499df423 Make image preview links embed 2024-04-07 12:07:53 -04:00
c86ca16c1a Merge pull request #80 from ButteredCats/fix_multiple_images
Fix multiple Reddit preview links becoming the same
2024-04-07 11:58:51 -04:00
858299c861 Add test for rewriting multiple preview links 2024-04-07 11:24:18 -04:00
75f5c6668c Merge pull request #77 from EMarshal/main
Update PUSHSHIFT_FRONTEND examples to undelete.pullpush.io
2024-04-07 10:53:40 -04:00
e6b9a2e426 Fix anchor tags scrolling to the wrong place when fixed navbar is enabled (#82) 2024-04-07 10:52:59 -04:00
d4a2b3edc6 Add SVG logo (#84)
* Add SVG version of logo

* Add PNG generated from SVG
2024-04-07 10:51:33 -04:00
4f0b29f930 Fix failing checks 2024-03-30 18:53:46 -04:00
4e2648280d Fix multiple Reddit preview links becoming the same 2024-03-30 18:36:28 -04:00
35ae71302f Update PUSHSHIFT_FRONTEND examples to undelete.pullpush.io
Matches the change in 3592728
2024-03-28 17:09:12 -06:00
e79242c9e7 v0.31.2 2024-03-04 20:20:23 -05:00
da581cb79b Update deps (GHSA-r8w9-5wcg-vfj7) 2024-03-04 20:02:58 -05:00
3f4526debe Container pipeline overhaul (#59)
* Simplified docker image building

* minor name fix

* Optimize container pipeline

* Change config to redlib-org account

* Added README push to Quay.io

* Fixes
2024-03-04 20:02:06 -05:00
5de171b13a Make HLS support checking more robust (#61) 2024-03-04 19:59:54 -05:00
22910956db Overhaul README.md (#56)
* Use GitHub Flavored Markdown for admonitions

* Add syntax highlighting to code blocks

* Minor raw formatting changes

* Add 'Metric' header to PageSpeed Insights table

* Fix typo

* Minor wording changes

* Add link to launchd page on Wikipedia

* Overhaul README.md

* Add table of contents

* Reorganise About section

* Edit PageSpeed table

* Add instructions for streaming container logs

* Rework Instances section

* Add note about Compose plugin installation

* Add link to instance monitoring

* Move info on architecture specific images into an admonition

* Change NGINX reverse proxy admonition to important

* Change Redlib external links admonition to a note

* Edit section headings

* Fix anchor link under Deployment section

* Add comparison with Libreddit

* Update ToC
2024-02-15 15:58:33 -05:00
94ada2b10c Remove seccomp profile from yml (#34) 2024-02-12 14:43:36 -05:00
1f246c956d Rename compose.dev.yml to compose.dev.yaml (#52)
Minor typo
2024-02-12 14:13:48 -05:00
b7778d5f95 Update dev compose yaml 2024-02-12 13:37:11 -05:00
f507fcfcf8 Expand Docker documentation and allow using .env (#49)
Fix seccomp error for Docker Compose
2024-02-12 13:34:02 -05:00
c6030064f1 Added video quality feature (#43)
* Fixed docker compose errors

* Added quality selector

* Remove log statement

Co-authored-by: Matthew Esposito <matt@matthew.science>

* Show kbps in quality

Co-authored-by: Matthew Esposito <matt@matthew.science>

* Make highest quality default

* Add styling, default option to highest

---------

Co-authored-by: Matthew Esposito <matt@matthew.science>
2024-02-06 15:27:23 -05:00
fe6123e05f Update docker-compose.yml to remove duplicate 'security_opt' (#45)
* Update docker-compose.yml to remove duplicate 'security_opt'

Docker compose complains with "mapping key "security_opt" already defined"

* Update and rename docker-compose.yml to compose.yaml

`version:` is deprecated:

https://docs.docker.com/compose/compose-file/04-version-and-name/

compose.yaml is the current standard:

https://docs.docker.com/contribute/style/terminology/#composeyaml
2024-02-06 12:14:37 -05:00
3bb5dc5f3e Update dependencies (Hyper v1.x) (#39)
* Cargo update

* Update major non-breaking changes

* Add deprecation feature-flags to hyper v0.14

* Semi-upgrade hyper-rustls

* Revert deprecated warnings
2024-02-06 08:01:59 -05:00
35927287f1 Replace www.unddit.com with undelete.pullpush.io (#42)
* Replace www.unddit.com with undelete.pullpush.io

* Update comment

---------

Co-authored-by: Matthew Esposito <matt@matthew.science>
2024-02-05 16:50:09 -05:00
469d0994f1 Handle errors from reddit (#35)
* Fix error handling logic

A 401 code is still an Ok(<...>) response

* Fix json key

* Run `cargo fmt`
2024-02-02 14:53:15 -05:00
99097da6b8 Remove pedantic clippy's 2024-01-28 13:36:27 -05:00
3d2c936a9e Refresh OAuth on 401 only (#33) 2024-01-28 09:28:42 -05:00
03e267f02e Fix pedantic clippy 2024-01-27 23:34:23 -05:00
6c2579cda9 Add check for unauthorized - refresh token 2024-01-27 23:31:21 -05:00
d0c5a1d93a Merge pull request #30 from tmak2002/main
compose: use official image instead building image
2024-01-27 13:36:43 -05:00
119b661639 add developmemt docker compose file 2024-01-26 17:06:30 +01:00
a8e2430e34 use official image 2024-01-26 17:03:56 +01:00
e8c257f801 Update README.md (fix #31) 2024-01-24 14:34:02 -05:00
ea3d248766 Update oauth_resources.rs 2024-01-24 14:30:17 -05:00
5604786146 Add new hls.js version, add script for updating (fixes #29) 2024-01-24 14:26:52 -05:00
9f9ae45f6e Add many Clippy's, fix many Clippy's 2024-01-19 20:16:17 -05:00
3e459f5415 Cargo update - fix possible DoS 2024-01-19 19:06:47 -05:00
95373f8261 More succinct fix to header parsing 2024-01-19 19:06:05 -05:00
3609564db0 Add error logging when rendering the Error page 2024-01-19 19:00:13 -05:00
fcde6ff689 Fix client.rs - properly return Err on invalid header (fix #28) 2024-01-19 18:58:08 -05:00
0f148c58d3 Merge pull request #21 from dethos/patch-1
Fix app.json
2024-01-12 16:56:59 -05:00
b1ef598f3c Update README.md to remove old referenced links 2024-01-12 16:53:40 -05:00
825e38b25f fix app.json
Remove a comma that was left behind and made the JSON content invalid.
2024-01-12 18:38:56 +00:00
f50872a88c Merge pull request #18 from ButteredCats/fix_footer
Less buggy solution to solving the floating footer problem
2024-01-07 16:54:50 -05:00
a445759a69 Be slightly smarter with options, fix footer being 30px above or below bottom on short posts 2024-01-06 16:53:58 -05:00
b578b717d7 Less buggy solution to solving the floating footer problem 2024-01-06 03:49:38 +00:00
78e51eb11f Merge pull request #17 from ButteredCats/fix_footer
Fix floating footer with "Keep navbar fixed" off
2024-01-03 21:31:26 -05:00
5d8529d6bb Fix floating footer with "Keep navbar fixed" off 2024-01-03 21:26:56 -05:00
c597a20311 Add websockets URL to parsing 2024-01-03 20:06:08 -05:00
89ba46e15d Redirect and proxy redditstatic gifs in-body (fix #14) 2024-01-03 09:36:19 -05:00
3dee29f3ef Add scrolling to highlighted comment (fix #13) 2024-01-02 19:43:00 -05:00
dea805936c Fix preview URL (fixes libreddit/libreddit/issues/559) 2024-01-02 19:21:24 -05:00
0c79cefed7 Fix publish action 2024-01-02 19:05:36 -05:00
5cb36ee15d v0.31.0 2024-01-02 18:45:19 -05:00
5bdcf64237 Update links and branches 2024-01-02 16:45:31 -05:00
3755f0cb24 README images (fix #9) 2024-01-02 00:39:11 -05:00
3145a6286b Update test runner to *run* cargo nextest run 2023-12-31 13:26:46 -05:00
c2e650b03b Update test runner to cargo nextest run 2023-12-30 21:46:37 -05:00
6d97f4c8dd Change Tokio tests - fix GHA runner (again) 2023-12-30 21:33:27 -05:00
cd836308db Update oauth.rs to use Android client only (fixes #8) 2023-12-30 17:32:54 -05:00
d327ab2c95 Small changes to params generation in subreddit.rs 2023-12-30 17:10:46 -05:00
53e8811f32 Remove all stats tracking (fixes #7) 2023-12-30 10:22:49 -05:00
d86b77ab56 Reset test threads to 1 (should fix test issues in GHA) 2023-12-29 20:33:43 -05:00
90a800ff44 Remove share parameters at canonical_path 2023-12-29 19:34:57 -05:00
45d8f1bbc8 Better handle redirects with new OAuth endpoints 2023-12-29 19:28:41 -05:00
3a4a39f577 Add config tests 2023-12-28 19:15:00 -05:00
ce0c6eca8a Fix obfuscated link handling 2023-12-28 18:21:07 -05:00
878ef8e95e Change formatting of autogenerated script 2023-12-28 17:36:45 -05:00
05f9d4f3bd Add count to update_oauth_resources.sh 2023-12-28 15:53:09 -05:00
0955f902f8 Fix update_oauth_resources.sh 2023-12-28 15:49:29 -05:00
d4c4d61ce8 Merge pull request #5 from redlib-org/improve_spoofing
Improve spoofing
2023-12-28 15:44:52 -05:00
c1214939ef Add requirements to scripts/update_oauth_resources.sh 2023-12-28 15:38:31 -05:00
9f41af6eee Improve spoofing - match headers more closely, pull in real versions/builds 2023-12-28 15:37:02 -05:00
4461a7d172 Only rerun build script if src/ changes 2023-12-28 15:21:06 -05:00
9850109326 Minor stylistic changes 2023-12-28 12:42:06 -05:00
bfe1c3db57 Fix Settings HTML 2023-12-28 11:35:06 -05:00
28f39329ad Add package name to instance_info in order to identify redlib instances 2023-12-28 11:21:56 -05:00
1316c8491c Merge pull request #4 from redlib-org/fix_popular_localization
Fix /r/popular localization - set geo filter to global
2023-12-28 10:43:10 -05:00
42902cc8d0 Add test for popular globalization 2023-12-28 10:40:17 -05:00
b43ed01958 Fix /r/popular localization - set geo filter to global 2023-12-28 10:30:42 -05:00
7d952f7f18 [FIX] Readme fix commands 2023-12-27 00:20:22 -05:00
819be89f84 Update cargo.toml 2023-12-27 00:19:54 -05:00
5f84d12774 0.30.2 (update config handling + test GH release) 2023-12-27 00:11:24 -05:00
5e68a66e40 Accept legacy config files 2023-12-27 00:11:06 -05:00
457b0bd57e Accept legacy environment variables 2023-12-26 23:16:36 -05:00
47822d8d6c Fix clippy warning 2023-12-26 23:15:06 -05:00
e452b8d6b5 Update oauth::choose to use new fastrand::choose_multiple 2023-12-26 20:00:36 -05:00
09df7713b1 Fix more README links 2023-12-26 19:55:37 -05:00
a2ca895e1f Fix docker repo URL - GHA 2023-12-26 19:09:54 -05:00
3fd3d4e145 Update action versions :fingers_crossed: 2023-12-26 19:01:45 -05:00
4c78ab30d3 More README changes, modify contrib/ 2023-12-26 18:48:09 -05:00
c5d11f220e Fix clippy warnings 2023-12-26 18:27:25 -05:00
b0f985c687 Libreddit -> Redlib 2023-12-26 18:25:52 -05:00
dac059573d Update rest of Cargo.toml, excepting askama, futures-lite, hyper 2023-12-26 17:50:56 -05:00
3e3c30d7f1 Update cookie + changes 2023-12-26 16:24:53 -05:00
e82c3fbea0 Update 11-15 of deps 2023-12-26 16:12:42 -05:00
d9f7ebcb79 Update util calls 2023-12-26 16:12:00 -05:00
36357e2609 Cargo update 2023-12-26 15:55:35 -05:00
b7bf9c74be Fix import error 2023-12-26 15:54:43 -05:00
1c36467c9c Merge remote-tracking branch 'origin/pull/867' 2023-12-26 15:52:53 -05:00
d76051302e Merge remote-tracking branch 'origin/pull/738' 2023-12-26 15:51:15 -05:00
90d1831352 Merge remote-tracking branch 'origin/pull/819' 2023-12-26 15:48:27 -05:00
3ac2048247 Merge remote-tracking branch 'origin/pull/861' 2023-12-26 15:47:33 -05:00
902acb257d Merge remote-tracking branch 'origin/pull/176' 2023-12-26 15:47:12 -05:00
28611da602 Add seccomp (merge 441) 2023-12-26 15:46:20 -05:00
f26c8be931 Add TokyoNight - Merge 450 2023-12-26 15:45:12 -05:00
de268314f3 Fix tests 2023-12-26 15:42:41 -05:00
8c9565c57b Readme values 2023-12-26 15:22:34 -05:00
0eb5e18cef Merge remote-tracking branch 'origin/pull/536' 2023-12-26 15:20:21 -05:00
fc4b686607 Merge remote-tracking branch 'origin/pull/746' 2023-12-26 15:18:49 -05:00
c17af7db75 Small changes to table and html 2023-12-26 15:18:29 -05:00
6625d106c3 Merge remote-tracking branch 'origin/pull/753' 2023-12-26 15:17:23 -05:00
82fdcf7443 Merge remote-tracking branch 'origin/pull/768' 2023-12-26 15:15:06 -05:00
2a525b744a Merge remote-tracking branch 'origin/pull/854' 2023-12-26 15:14:46 -05:00
c71a9dddd7 Merge remote-tracking branch 'origin/pull/857' 2023-12-26 15:13:46 -05:00
cc9023dc64 Merge remote-tracking branch 'origin/pull/865' 2023-12-26 15:12:36 -05:00
f5b54197c4 Merge remote-tracking branch 'origin/pull/808' 2023-12-26 15:11:44 -05:00
9b71822be6 Match on both http and https in format_url (414) 2023-12-26 15:11:16 -05:00
9d948abadc Merge pull request #831 from bennettmsherman/header-filters
Remove Reddit's 'Nel' and 'Report-To' (network error logging) response headers
2023-11-29 09:25:16 -05:00
2d64c092ea Fix short links again. Just using a split 2023-11-21 21:34:13 -08:00
2b06d22687 Made userpage posts show subreddit name
Made userpage posts show subreddit name instead of ambigious `COMMENT`
2023-11-08 13:13:46 +02:00
3e236e7ab5 client.rs: remove some String allocations 2023-10-27 09:05:22 -04:00
469aff0689 Handle obfuscated share links 2023-10-04 09:55:33 -07:00
dd611b17ad Add start URL to manifest.json
Adds the last criteria to make the application an installable PWA on mobile.
2023-09-17 23:00:01 -07:00
b39db0fcd4 Make launchd service for macOS 2023-08-23 13:14:32 -04:00
2815dc5209 Correct the shutdown announcement 2023-07-14 12:05:57 -07:00
00697c6ae4 Add shut down announcement 2023-07-14 11:57:25 -07:00
7a14975fb8 Remove 'Nel' and 'Report-To' response headers 2023-07-08 19:20:58 -07:00
ea696687be Merge pull request #821 from fawni/feat/hide-subreddit-panel 2023-06-09 19:01:57 -04:00
136aa0aa7d Format 2023-06-09 17:32:21 -04:00
a39bb9d502 Merge branch 'master' into reddit-stats 2023-06-09 17:31:12 -04:00
5f562876f4 Make stats collection opt-out 2023-06-09 17:26:23 -04:00
f7f1aa4bde Abstract out random choosing 2023-06-08 16:27:36 -04:00
c00beaa5d8 Improve OAuth refresh, logging 2023-06-08 14:33:54 -04:00
49dde7ad72 Improve subreddit test 2023-06-08 14:06:58 -04:00
13394b4a5e Add ability to hide subreddit panel (closes #801) 2023-06-07 13:51:27 +03:00
0ca0eefaa4 Add tests to check fetching sub/user/oauth 2023-06-06 15:28:36 -04:00
6cd53abd42 Documentation 2023-06-06 15:26:31 -04:00
dc7601375e Ignore dotenv failure 2023-06-06 15:07:11 -04:00
659a82bf63 Improve spoofing of devices, handle token refreshes 2023-06-06 15:05:20 -04:00
a5833dc05c Add .env to .gitignore 2023-06-06 15:04:06 -04:00
e94a9c81e2 Add deps - rand, logging 2023-06-06 14:33:01 -04:00
8a23616920 Stray space 2023-06-05 20:57:34 -04:00
00355de727 Set proper headers 2023-06-05 20:39:56 -04:00
383d2789ce Initial PoC of spoofing Android OAuth 2023-06-05 20:31:25 -04:00
ba89b76332 Merge pull request #814 from Tokarak/deps-update 2023-06-04 18:14:27 -04:00
96e9e0ea9f Update .replit to download from nightly build artifacts (#815) 2023-06-03 23:36:39 +00:00
c1dd1a091e Update release binary paths 2023-06-03 16:30:58 -04:00
05ae39f743 Update RUSTFLAGS 2023-06-03 16:15:24 -04:00
221260c282 Remove MUSL, build statically via flags 2023-06-03 16:12:48 -04:00
f3c835bee7 Proof-read README.md 2023-06-03 20:02:02 +01:00
f9fd54aa3c Specify newer dependencies + cargo update 2023-06-03 19:41:32 +01:00
510d967777 Add MUSL target 2023-06-03 14:33:27 -04:00
0bcebff6f2 Fix YAML formatting 2023-06-03 14:24:19 -04:00
0c74305617 Add MUSL builds to GH Actions and fix Release event trigger (#810) 2023-06-03 18:19:20 +00:00
97f0f69059 Rebase #811 (#812)
Co-authored-by: Matthew Esposito <matt@matthew.science>
2023-06-03 17:32:46 +00:00
255307a4f7 Add request stats to instance info HTML 2023-05-31 20:02:00 -04:00
b5fc4bef28 Fix github-actions versioning 2023-05-31 19:50:38 -04:00
81a6e6458c ci: cleanup github actions (#803) 2023-05-31 23:47:58 +00:00
de68409610 Add request stats to instance info page 2023-05-31 19:39:44 -04:00
193a6effbf Merge pull request #792 from beucismis/master 2023-05-31 18:42:39 -04:00
09551fca29 Merge pull request #806 from gmnsii/comment-searchbar-color 2023-05-31 18:40:25 -04:00
38ee0d9428 make comment search bar color change based on theme 2023-05-31 19:41:13 +02:00
ca7ad9f812 Merge pull request #796 from StuffNoOneCaresAbout/lazy-init-regex 2023-05-01 10:09:59 -04:00
98e2833881 Merge pull request #790 from StuffNoOneCaresAbout/allow-disabling-indexing 2023-05-01 10:08:20 -04:00
4d5c52b83b Rename variables to more descriptive names. 2023-05-01 05:00:49 +01:00
6c47ea921b performance: compile regex only once 2023-05-01 04:22:10 +01:00
6c0e5cfe93 Add cursor:pointer for button and select 2023-04-29 21:16:02 +03:00
0c591149d5 Add option to disable all indexing. 2023-04-26 12:52:12 +01:00
8b4b2dd268 Ignore idea files. 2023-04-26 12:52:00 +01:00
ac58bb532a Merge pull request #787 from libreddit/clippy_refactor 2023-04-19 13:08:44 -04:00
af8fe176ea Fix clippy warnings 2023-04-19 10:37:47 -04:00
bfa9c084bb Merge pull request #786 from libreddit/update_deps 2023-04-19 10:32:46 -04:00
3c892d3cfd Update Cargo.lock - h2 moderate 2023-04-19 10:27:50 -04:00
4a1b448abb Merge pull request #776 from iTzBoboCz/polls 2023-04-17 18:12:02 -04:00
991677cd1e Add variable for now_utc, format 2023-04-17 18:00:41 -04:00
3b8a13d050 Merge pull request #773 from libreddit/fmt_clippy 2023-04-15 11:01:19 -04:00
0e90ebc1a1 Merge pull request #769 from gmnsii/bypass-gate 2023-04-15 11:00:20 -04:00
af89d4c88f Merge pull request #778 from Akanksh12/comments-to-contrib-files 2023-04-15 10:59:28 -04:00
5f87875b8e Merge branch 'master' into bypass-gate 2023-04-15 10:56:28 -04:00
aaf05de1a8 Merge pull request #771 from gmnsii/comment-search 2023-04-15 10:55:10 -04:00
17f7f6a9d1 changed default port to 12345 2023-04-08 21:17:19 +05:30
ec226e0cab fix(polls): apply clippy suggestions 2023-04-08 10:41:12 +02:00
2b8931c032 Merge pull request #770 from invakid404/patch-1
fix(style): fit footer width to body size
2023-04-07 12:05:41 -04:00
62771bf4a3 Merge pull request #751 from master-hax/optimize-docker
optimize arm dockerfile
2023-04-07 12:02:03 -04:00
22e3e0eb91 added comments to libreddit.service and .conf 2023-04-06 10:06:37 +05:30
94a781c82c fix(polls): minor improvements 2023-04-01 14:31:39 +02:00
75af984154 fix(polls): apply suggestions and fix id parsing 2023-04-01 14:26:04 +02:00
8bed342a6d fix: print time suffix only for relative dates 2023-04-01 13:21:15 +02:00
de5d8d5f86 Requested code style changes 2023-03-26 11:52:02 -07:00
f465394f93 Address fmt + clippy 2023-03-25 16:32:42 -04:00
1e418619f1 Feat: search for comments within posts
Add the ability to search for specific comments within posts.
Known issues:
  - Just like on reddit, this does not work with comment sorting. The
    sorting order is ignored during the search and changing the sorting
    order after the search does not change anything. I do not think we
    can fix this before reddit does, since in my understanding we rely
    on them for the sorting. However we could implement a default
    sorting method ourselves by taking the vector of comments returned
    from the search and sorting it manually.
  - The UI could be improved on mobile. On screens with a max width
    inferior to 480 pixels, the comment search bar is displayed below
    the comment sorting form. It would be great if we could make the
    search bar have the same width as the whole comment sorting form
    but I do not have the willpower to write any more css.
2023-03-24 17:41:26 -07:00
8be69f6fe5 Checks if the link contains the parameter instead of ends with it
To know if the gate should be bypassed, we check if the link contains
the pasameter instead of checking if the link ends with it. This is
impostant, for example if we were to implement searching for comments
within a post. If we wanted to search for comments within a post that we
have bypassed the gate to view: the link will look like
https://libreddit-instance/r/somesub/comments/post-id/post-title&bypass_nsfw_landing/?q=some-query&type=comment
2023-03-23 12:36:04 -07:00
e3b1c5b587 Use a bullet instead of empty margin when score is hidden
This is prettier and keeps consistency across the app.
2023-03-23 11:29:28 -07:00
a0726c5903 Change the bypass message and format code
The bypass message now indicates that the bypass is only temporary.
2023-03-23 11:09:33 -07:00
c1c867a5ff feat: add polls 2023-03-23 13:21:09 +01:00
5dc3279ac3 fix: make time work with future dates 2023-03-23 13:18:48 +01:00
dead990ba0 fix(style): fit footer width to body size 2023-03-23 13:49:40 +02:00
e046144bf3 Allow bypassing nsfw gate for posts
On instances that are not sfw-only, the nsfw gate for posts can now be
bypassed.
2023-03-22 23:18:35 -07:00
f8ba3cf815 Fix formatting in some places 2023-03-22 20:29:48 -07:00
df3d894947 Add option to hide score
Add the option to hide score for posts and comments in preferences.
There is still however a blank margin where the score is supposed to be.
2023-03-22 20:08:20 -07:00
e25622dac2 harden docker-compose.yml (#760)
`user: nobody`: the least privileged account.
`read_only: true`: this container doesn't write anything to the filesystem, this removes a vector.
`security_opt`: disallows the container to grab more privileges.
`cap_drop`: this container doesn't need any capabilities, drop them.
`networks`: put `libreddit` into its own network so it cannot see other containers by default.
2023-03-17 10:17:01 -06:00
6bcc4aa368 Update version string in Cargo.lock. 2023-03-17 09:36:52 -06:00
eb0928acc3 add link to reddit status page. 2023-03-14 22:21:41 +01:00
6d652fc38c optimize arm dockerfile 2023-03-12 23:36:25 -07:00
f62f7bf200 v0.30.1 2023-03-10 21:34:42 -07:00
aece392a86 Pad bottom of body to prevent footer collision (fixes #747) 2023-03-10 21:33:45 -07:00
aeeb066e47 Update README.md (#748)
* Remove duplicated config

Was accidentally introduced in  412ce8f1f3
2023-03-10 21:04:05 -07:00
741613e27f Ignore errors while fetching subreddit names in subscriptions_filters()
If we can't retrieve subreddit name, just use the user-supplied name.
This fixes banned subreddits being impossible to to unfilter or
unsubscribe from.
A drawback of such approach is that it might be possible to subscribe to
a subreddit twice with different casing, however the chance of this is
extremely low.
2023-03-09 15:18:03 +03:00
51cdf574f7 v0.30.0 2023-03-08 22:15:31 -07:00
af6722c053 Move unimportant links to footer (#728) 2023-03-08 22:14:43 -07:00
412ce8f1f3 Fix default subscriptions (#732)
Co-authored-by: Daniel Valentine <daniel@vielle.ws>
2023-03-08 21:53:23 -07:00
dfa57c890d fix build error on windows (#741) 2023-03-08 21:32:41 -07:00
01f9907aaf show the count of 'more replies'. (#740)
Co-authored-by: Daniel Valentine <daniel@vielle.ws>
Co-authored-by: Matthew Esposito <matt@matthew.science>
2023-03-08 21:30:41 -07:00
bf19ff513f add support for gifs in galleries. (#744) 2023-03-08 21:04:26 -07:00
ffc9ca2e98 use the documented LIBREDDIT_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION config option. (#737) 2023-03-04 13:04:40 -07:00
a7f59ccac1 Display previews and inline images for reddit-hosted images.
This version does not change the svg behaviour in the other cases in an attempt to reduce breakage.
2023-03-03 13:50:05 +01:00
cef9266648 Restructure section on Libreddit user privacy. 2023-02-26 03:35:36 -07:00
d3b4f4e379 Update tempfile to v3.4.0. 2023-02-26 03:11:17 -07:00
b90b41c009 v0.29.4 2023-02-26 03:01:35 -07:00
pin
0eccb9bcf2 Add NetBSD install (#720) 2023-02-26 01:13:56 -07:00
eb07a2ce7c Make gated subreddits accessible by treating them as quarantined (#722)
* Fix gated communities being unviewable by treating them as quarantined

* Show restriction reason in quarantine template

* Add `gated` checks for other requests
2023-02-26 00:40:32 -07:00
0b39d4f059 Mark search query as safe on Prev/Next button (#731)
Fixes: #677 again. Complement to #686.
2023-02-26 00:35:05 -07:00
58fa213be8 Reuse hyper client. (#727)
Making a new connection on every request is very slow and wasteful, espectially on slower network.

Fix this by reuse a hyper client which shares a connection pool.

I'm able to lower /r/popular loading time from 5s to 1.5s on my machine.
2023-02-26 00:33:55 -07:00
5e03d701e4 Revert "Move unimportant links to footer"
This reverts commit e3df3a9470.
2023-02-19 18:03:55 +00:00
e3df3a9470 Move unimportant links to footer 2023-02-19 18:00:56 +00:00
35504eda14 v0.29.3 -- fix layout bugs on mobile
Addresses the following layout bugs in mobile view:

* improper rendering of award images on posts
* upvote ratio no longer appearing on bottom-right corner of post as
  before
* Reddit warning pop-up background cut off at bottom of page

Fixes #713.
2023-02-14 20:19:19 -07:00
a05cfe60fe v0.29.2 2023-02-12 03:36:48 -07:00
2774d15298 Fix bug causing user/sub title to appear off-center. 2023-02-12 01:02:25 -07:00
f544daf8c0 Replace snoo with r/ icon 2023-02-09 21:40:51 -08:00
089315f9bb v0.29.1 (fixes #713) 2023-02-09 22:25:42 -07:00
1f7e14dd4e v0.29.0 2023-02-08 00:33:57 -07:00
37f71c48d1 Reduce size of instance info button in footer. 2023-02-08 00:33:31 -07:00
fa68bf561b added leaving reddit dialog (#643) 2023-02-08 00:24:06 -07:00
a4eecb251e Fix listing_options hidden overflow 2023-02-04 00:02:32 -08:00
9bf6194b09 v0.28.1
Remove font-weight associated with instance info button, which made the
icon look ghastly in Chrome.
2023-01-31 00:14:23 -07:00
f405f509c4 v0.28.0 2023-01-30 02:07:32 -07:00
8be5fdee2d Implement instance info endpoint (JSON, YAML, TXT) (#685)
Co-authored-by: Daniel Valentine <daniel@vielle.ws>
Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2023-01-30 02:02:43 -07:00
7efa26e811 Fix #699 2023-01-21 00:35:49 -08:00
755fff0818 Use Markdown Highlights in README 2023-01-19 18:28:24 -08:00
53e1e302d5 Register Dockerfile.* as Dockerfiles for Linguist (#694)
This allows GitHub Linguist to generate slightly more accurate language
stats for this repository, and also enable syntax highlighting in the
GitHub web UI. Due to caching, it may take a few days for this change to
have a visible effect on github.com.
2023-01-16 21:57:55 -07:00
bb5f2674d1 Merge branch 'master' into feature/fixed-navbar 2023-01-16 19:43:54 -08:00
3d0287f04f Add comment count in post (#659)
* Add comment count in post

* Restyle comment count
2023-01-16 12:05:53 -08:00
7cb132af01 Update packages 2023-01-16 11:09:57 -08:00
63b0b936aa Update CREDITS file. 2023-01-12 02:19:09 -07:00
412122d7d9 v0.27.1 2023-01-12 01:57:03 -07:00
eb9ef9f6d9 added leaving reddit dialog (#643) 2023-01-12 01:46:56 -07:00
27091db53b Create rust-tests.yml (#690)
This will run tests on every push and PR to master.
2023-01-12 01:43:08 -07:00
2a54043afc Simplify listener definition (#681)
This simplifies the logic to build the listener by using more clap
features instead of manually accessing the PORT environment variable.
This also removes unnecessary `unwrap_or` calls that set defaults that
are already set by clap.
2023-01-12 01:41:59 -07:00
e238a7b168 Bump tokio from 1.23.0 to 1.23.1 (#691)
Bumps [tokio](https://github.com/tokio-rs/tokio) from 1.23.0 to 1.23.1.
- [Release notes](https://github.com/tokio-rs/tokio/releases)
- [Commits](https://github.com/tokio-rs/tokio/compare/tokio-1.23.0...tokio-1.23.1)

---
updated-dependencies:
- dependency-name: tokio
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>

Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2023-01-12 01:39:23 -07:00
1e554acd20 Merge pull request #687 from jojosch/fix-cfg-test
Fix tests
2023-01-04 16:03:48 -05:00
dff91da877 config: fix SFW test 2023-01-04 11:12:19 +01:00
f6bb53e388 Mark search query as safe in askama template (#686) 2023-01-03 20:55:17 -08:00
709292339a Merge pull request #674 from spenserblack/codespace 2023-01-03 20:12:36 -05:00
799e5b882b Merge pull request #667 from erdnaxe/scrollbar_theme 2023-01-03 19:34:40 -05:00
0ff92cbfe3 v0.27.0 2023-01-03 11:21:27 -07:00
e9891236cd Remove unnecessary SFW-only disclosure in settings in SFW-only mode. 2023-01-03 11:20:55 -07:00
e2c48c3438 Add hide_awards config option (fixes #442) 2023-01-03 11:16:22 -07:00
9a7b3b29f5 Merge remote-tracking branch 'origin/master' into hide_awards 2023-01-03 11:12:27 -07:00
10add895fb Merge pull request #683 from Tokarak/master 2023-01-03 09:28:13 -07:00
050eaedf15 Remove unused dep "async-recursion"
Found using cargo-udeps. Checked.
2023-01-03 14:56:17 +00:00
5b06a3fc64 Add config system to read from file (#664)
Co-authored-by: Daniel Valentine <daniel@vielle.ws>
2023-01-03 02:55:22 -07:00
4817f51bc0 v0.26.0 2023-01-03 02:40:44 -07:00
c83a4e0cc8 Landing page for NSFW content, SFW-only mode (#656)
Co-authored-by: Matt <matt@matthew.science>
Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2023-01-03 02:39:45 -07:00
c15f305be0 v0.25.3 2023-01-01 23:54:35 -07:00
222d216854 Merge pull request #673 from spenserblack/code-in-new-tab 2023-01-01 23:08:15 -05:00
6a785baa2c Add hide_awards config 2023-01-01 21:39:38 -05:00
6d8aaba8bb Merge pull request #676 from ellieeet123/master 2023-01-01 19:23:30 -05:00
6cf3748642 Fix for #675
/:id route now accepts 7 character post IDs.
2023-01-01 17:06:58 -06:00
9c938c6210 build: enable LTO, set codegen-unit to 1 and strip the binary (#467)
Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2023-01-01 14:33:31 -07:00
b1182e7cf5 Create devcontainer.json 2023-01-01 14:52:33 -05:00
a49d399f72 Link to libreddit/libreddit and open in new tab
This sets the target of the "code" link to `_blank`, so that it will
open in a new tab in browsers. Because the GitHub page is a different
context from libreddit, and accessing the repository doesn't imply that
the user is finished browsing libreddit, this seemed reasonable. This
also changes the link from `spikecodes/libreddit` to
`libreddit/libreddit`.
2023-01-01 13:43:33 -05:00
9178b50b73 fix a11y and HTML issues on settings page (#662)
- connect labels with corresponding form controls
- use fieldsets to group form sections
- don't nest details/summary element into label
2023-01-01 01:56:09 -07:00
b5d04f1a50 v0.25.2 2022-12-31 21:34:15 -07:00
9e434e7db6 Search - add support for raw reddit links (#663)
* Search - add support for raw reddit links

If a search query starts with 'https://www.reddit.com/' or 'https://old.reddit.com/',
this prefix will be truncated and the query will be processed normally.
For example, a search query 'https://www.reddit.com/r/rust' will redirect to
r/rust.

* Search - support a wider variety of reddit links.

Add once cell dependency for static regex support (avoid compiling the
same regex multiple times).
All search queries are now matched against a regex (provided by @Daniel-Valentine)
that determines if it is a reddit link. If it is, the prefix specifying
the reddit instance will be truncated from the query that will then be
processed normally.
For example, the query 'https://www.reddit.com/r/rust' will be treated
the same way as the query 'r/rust'.
2022-12-31 20:57:42 -07:00
ab30b8bbec Bugfix: 'all posts are hidden because NSFW' when no posts where found (#666)
* Fix 'all_posts_hidden_nsfw' when there are no posts.

If a search query yielded no results and the user set nsfw posts to be
hidden, libreddit would show 'All posts are hidden because they are NSFW.
Enable "Show NSFW posts" in settings to view'. This is fixed by
verifying tnat posts.len > 0 before setting 'all_posts_hidden_nsfw' to
true.

* Add a message when no posts were found.

* Delete 2
2022-12-31 19:11:59 -07:00
1fa9f27619 Theme browser scrollbar
Hint current color-scheme to the browser. This enables chromium-based
browsers to change the scrollbar color according to the current theme.
2022-12-26 22:57:04 +01:00
37d1939dc0 Fix #658.
Dimensions for embedded video in post are explicitly set only when defined by Reddit.

c/o: NKIPSC <15067635+NKIPSC@users.noreply.github.com>
2022-12-13 21:15:28 -07:00
08a20b89a6 Merge branch 'cache-determine-compressor' 2022-12-10 18:35:38 -07:00
5d518cfc18 Cache result of server::determine_compressor. 2022-12-04 17:56:02 -07:00
7e752b3d81 Fix Docker credential secrets 2022-12-04 11:07:18 -08:00
87729d0daa Use new libreddit org for GitLab and Docker links 2022-12-04 11:05:19 -08:00
dc06ae3b29 Automatically-update Docker Repo description 2022-12-04 11:01:28 -08:00
225380b7d9 Fix workflow to push to new Libreddit Docker repo 2022-12-04 10:57:19 -08:00
7391a5bc7a v0.25.0 2022-12-03 01:18:23 -07:00
3ff5aff32f Merge branch 'list-post-duplicates' 2022-12-03 01:11:45 -07:00
e579b97442 List post duplicates (resolves #574). 2022-12-03 01:08:36 -07:00
8fa8a449cf Sign release (resolves #651). 2022-12-01 16:42:04 -07:00
473a498bea Update CREDITS file. 2022-11-30 21:08:51 -07:00
92f5286667 Make the column size in posts consistent.
Signed-off-by: Daniel Valentine <Daniel-Valentine@users.noreply.github.com>
2022-11-30 21:06:21 -07:00
0a6bf6bbee Update CREDITS file. 2022-11-27 15:57:31 -07:00
618b074ad5 Fix embeds (#648) 2022-11-27 11:42:34 -07:00
d86cebf975 Request CSS with explicit version.
base.html will now request with a query parameter `v=` whose value is
the current version of Libreddit. This will cause the browser to request
the stylesheet for a specific version of Libreddit, bypassing the cache.
A new version of Libreddit will cause the browser to fetch a new
stylesheet.

Resolves #622. Credit is due to GitHub user @chloekek for offering this
solution in the following post:
        https://github.com/libreddit/libreddit/issues/622#issuecomment-1315961742
2022-11-23 14:43:36 -07:00
ab39b62533 Dockerfile.arm: Add git to builder. 2022-11-22 15:42:10 -07:00
5aee695bae Dockerfile.arm: Force cargo to use git binary.
Hopefully resolves #641.
2022-11-22 15:38:17 -07:00
c9633e1464 Revert "Dockerfile.arm: Verbose cargo install."
This reverts commit 0152752913.
2022-11-22 15:32:45 -07:00
0152752913 Dockerfile.arm: Verbose cargo install.
Temporarily provide `--verbose` to `cargo install` to track when during
the build the process(es) receive SIGKILL.
2022-11-22 15:29:02 -07:00
6912307349 Update version to v0.24.1. 2022-11-22 12:14:12 -07:00
f76243e0af Revert "Dockerfile.arm: disable cargo build parallelization"
This reverts commit f0fa2f2709.

This did not stop the OS from issuing SIGKILL to cargo and/or one of its
child processes.
2022-11-22 00:22:15 -07:00
f0fa2f2709 Dockerfile.arm: disable cargo build parallelization 2022-11-22 00:16:55 -07:00
88bed73e5e Extract Location URL path correctly in client::request. (fixes #645) (#646) 2022-11-21 08:58:40 -07:00
3a33c70e7c Update CREDITS file. 2022-11-20 17:52:28 -07:00
40dfddc44d Added gruvbox-dark and gruvbox-light themes (#490) 2022-11-20 13:49:20 -07:00
3f3d9e9c3b Indicate pinned posts on user profiles (close #606) 2022-11-14 18:08:44 -08:00
501b47894c Add "BLUR_NSFW" to the list of settings in README (#639) 2022-11-12 10:37:58 -08:00
d8c661177b Update Google PageInsights speed comparison 2022-11-11 09:43:18 -08:00
fade305f90 Blur NSFW posts.
Reimplementation of libreddit/libreddit#482.

Co-authored by: Daniel Valentine <Daniel-Valentine@users.noreply.github.com>
2022-11-09 08:49:39 -07:00
e62d33ccae Blur NSFW posts.
Reimplementation of libreddit/libreddit#482.

Co-authored by: Daniel Valentine <Daniel-Valentine@users.noreply.github.com>
2022-11-08 09:01:12 -07:00
465d9b7ba7 Implement 'posts hidden because of NSFW'. (Resolves #159) (#619) 2022-11-07 20:54:49 -07:00
5c366e14a3 Add CREDITS file and script to generate. (Resolves ferritreader/ferrit#33) 2022-11-06 16:04:02 -07:00
d4ca376e8d Add format_url tests (#615) 2022-11-05 23:51:56 -06:00
371b7b2635 Update Libreddit GitHub links. 2022-11-05 21:24:16 -06:00
cc27dc2a26 Update README.md to point to markdown instances list. 2022-11-05 20:50:42 -06:00
bfe03578f0 Update Instances section in README.md. 2022-11-05 13:25:12 -06:00
c6487799ed Redirect /:id to canonical URL for post. (#617)
* Redirect /:id to canonical URL for post.

This implements redirection of `/:id` (a short-form URL to a post) to
the post's canonical URL. Libreddit issues a `HEAD /:id` to Reddit to get
the canonical URL, and on success will send an HTTP 302 to a client with
the canonical URL set in as the value of the `Location:` header.

This also implements support for short IDs for non-ASCII posts, c/o
spikecodes.

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2022-11-05 02:29:04 -06:00
584cd4aac1 Add DoomOne theme, c/o Tildemaster <root@vern.cc> (#611) 2022-11-03 23:08:03 -06:00
377634841c Upgrade to v0.23.2 2022-11-03 21:31:32 -07:00
c0e37443ae Allow the spoilering of links (fixes #610) 2022-11-03 21:30:35 -07:00
8348e20724 Use permalink offered by Reddit (fixes #613). (#614) 2022-11-03 21:08:36 -07:00
ae3ea2da7c HTTP compression (Reddit -> Libreddit -> client) (#612)
Implements HTTP compression, between both Reddit and Libreddit and Libreddit
and a web browser. Compression between Reddit and Libreddit is mandatory,
whereas compression between Libreddit and a client is opt-in (client must
specify a compressor in the Accept-Encoding header).

Supported compressors are gzip and brotli. gzip support is ubiquitous,
whereas brotli is supported by almost all modern browsers except Safari
(iOS, iPhone, macOS), although Safari may support brotli in the future.

Co-authored-by: Matthew E <matt@matthew.science>
2022-11-03 22:04:34 -06:00
8435b8eab9 Update hls.js.min to v1.2.4
Mirrors ferritreader/ferrit#6
2022-11-02 08:46:59 -07:00
510c8679d6 Show full "Submissions" btn on mobile (fixes #548) 2022-11-01 21:59:16 -07:00
98674310bc Remove some-things.org instance (closes #561) 2022-11-01 21:29:50 -07:00
170ea384fb Support /comments endpoint (closes #568)
Code based on @Daniel-Valentine's [implementation](e2c84879d6)
2022-11-01 20:53:42 -07:00
1b5e9a4279 Fix #592 2022-11-01 20:47:47 -07:00
b170a8dd99 Switch Reveddit to Unddit 2022-10-31 22:30:31 -07:00
aa54301054 Upgrade to version 0.23 2022-10-31 20:35:00 -07:00
b4d3f03335 Upgrade dependencies 2022-10-31 20:23:59 -07:00
1a1ff2e600 Use singular form of "comment" for posts with 1 comment (#567)
* Use singular form of "comment" for posts with 1 comment

* Fix incorrect text on comment count tooltip

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-10-31 18:36:24 -07:00
4fc07c02b5 update instance (igna.rocks => intent.cool) (#603)
* update instance (igna.rocks => intent.cool)

* Remove accidentally-added broken instances

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-10-31 18:22:20 -07:00
8d58cf61d2 Removed 8 dead links from instance list (#545)
* Update README.md

Removed the following instances for dead links:

* libreddit.sugoma.tk
* libreddit.jamiethalacker.dev
* libreddit.database.red
* reddit.phii.me
* libreddit.autarkic.org
* lr.oversold.host
* libreddit.datatunnel.xyz
* libreddit.crewz.me

* Fix double pipe on flux.industries instance

Co-authored-by: Mohammed Anas <triallax@tutanota.com>

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
Co-authored-by: Mohammed Anas <triallax@tutanota.com>
2022-10-31 18:06:32 -07:00
711e3c205d Add libreddit.cachyos.org instance (#571)
Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-10-31 17:50:59 -07:00
0704eb10b8 Add new instance (libreddit.oxymagnesium.com) (#591) 2022-10-31 17:49:20 -07:00
ef86c1be86 Remove reddit.artemislena.eu (uses Teddit, not Libreddit now) (#586)
* L: Fixed two swapped config variables in the documentation.

* L: reddit.artemislena.eu is now Teddit, not Libreddit anymore
2022-10-31 17:46:16 -07:00
8141b74817 Update Esmail's onion instance (#593) 2022-10-31 17:45:05 -07:00
57d304161b Add ~vern onion instance (#537)
* Add ~vern onion instance

 Add vern.cc's onion instance of libreddit to instance list

* Remove http:// from link name

* Fix previous commit
2022-06-24 10:52:08 -07:00
b5f21bcb97 Added 3 instances (#531)
* Add Instance - encrypted-data.xyz

* Add Instance - eu.org

* Add Instance - opnxng.com
2022-06-24 10:50:16 -07:00
36c560144a Remove libreddit.awesomehub.io (#535) 2022-06-24 10:48:41 -07:00
6d49858d59 Fixed navbar: Add preference to settings restore link 2022-06-18 23:05:36 +01:00
6c202a59b0 Make the fixed navbar optional
Adds another on/off preference (default: on, keeps same
behaviour) for the fixed navbar.
When off the navbar will not remain at the top of the
page when scrolling.
This is useful for small displays such as phones where
otherwise the navbar takes up a sizeable portion of
the viewport.
2022-06-18 22:53:30 +01:00
2bc714d0c5 Added mha.fi Instance (#519)
Close #518
2022-06-11 20:41:54 +00:00
ff4a515e24 change lr.vern.cc hosting location as we moved vps (#526) 2022-06-11 20:39:29 +00:00
93f089c2cf Add libreddit.foss.wtf (Close #527) 2022-06-11 20:39:11 +00:00
23569206cc L: Fixed two swapped config variables in the documentation. (#524) 2022-06-06 02:09:36 +00:00
5f20e8ee27 Fix dark theme hidden in settings 2022-05-28 19:55:13 -07:00
a8a8980b98 Update README.md (#516) 2022-05-28 05:31:38 +00:00
fd7d977835 Add instance rd.jae.su (close #515) 2022-05-28 05:31:07 +00:00
50f26333cb remove 40two.app - dead/serves ads (#517)
40two.app looks like it serves ads instead of libreddit. Hasn't worked for 1 week+
2022-05-28 05:17:38 +00:00
f5cd48b07f Fix #514 2022-05-21 21:06:03 -07:00
50665bbeb3 Switch titles to <h1>s (Fixes #444) 2022-05-21 15:47:58 -07:00
d558127306 Add keyboard shortcuts to nav buttons (closes #466) 2022-05-20 23:10:11 -07:00
0c757023f9 Correct localhost to 0.0.0.0 in SystemD conf (#498) 2022-05-21 05:53:48 +00:00
90828cc71c Fix "Post url contains non-ASCII characters" error (#479) 2022-05-21 05:48:59 +00:00
7f5bfc04b3 Always show Feeds dropdown (Fixes #408) 2022-05-20 22:42:05 -07:00
322aa97a18 Fix HTML encoding in templating (#404) 2022-05-21 05:28:31 +00:00
7e07ca3df1 Fix #480 2022-05-20 21:26:53 -07:00
428dc58e3c Update to v0.22.8 2022-05-20 19:20:44 -07:00
0ec8e4e9a2 Harden Systemd configuration (#453) 2022-05-21 01:48:32 +00:00
60c7b6b23f Embed css themes to simplify adding and testing new themes (#489) 2022-05-21 01:41:31 +00:00
1c8bcf33c1 Made .onion instance url consistant (#511)
Without `http://`, it may have caused problems for libredirect
2022-05-18 04:30:51 +00:00
3bdc21f90a Remove silkky.cloud instance. Closes #510 2022-05-17 03:12:03 +00:00
c3dade257d Restore post sorting preference by link (#406) 2022-05-17 03:11:01 +00:00
62b2bbb231 Add reddit.dr460nf1r3.org instance (#504)
Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-05-15 21:37:39 +00:00
653aee9294 CI: Add docker build caching (#493) 2022-05-15 21:37:13 +00:00
bb7fb1313d Fix multireddit subscription redirect url 2022-05-15 13:50:17 -07:00
01bc729a80 ✏️ Fix link to Cloudflare in README.md (#506)
Added `.com` to `https://cloudflare` to fix the link in `README.md`.
2022-05-15 19:58:32 +00:00
39e6e6bf81 Add lr.vern.cc to instance list (#505) 2022-05-15 19:58:10 +00:00
8c94c0dd17 Let native elements use theme accent colour (#509) 2022-05-15 19:56:25 +00:00
1c50c8f30d Update instances
Closes #503, #495, #488, #484, #483, #454, and #507
2022-05-15 19:16:15 +00:00
3facaefb53 Add strongthany.cc instance (#478)
* Update to list of public instances

added strongthany.cc instance to the list of public instances

* fixed accidental removal

put the location status back for riverside.
2022-04-06 18:25:11 +00:00
aec45311cc ✍️fix: Tries to add better readability - ➡️ fix: Moves funding links to correct file (#477)
* ✍️ fix(readme): Adds better readability

* ➡️ fix(buymeacoffee): Moves custom link to correct FUNDING.yml

* ➡️ fix(buymeacoffee): Moves custom link to correct FUNDING.yml

* ✍️ fix(bugreport): Adds better readability

* ✍️ fix(featureparity): Adds better readability

* ✍️ fix(featurerequest): Adds better readability

* ✍️ fix(crypto): Removes broken linking, changed to codeblock
2022-04-06 18:24:10 +00:00
47ab857103 Scroll overflowing tables (fixes #469) 2022-04-02 21:24:20 -07:00
a9ef5bc08b Add lunar.icu instance. Closes #460 2022-03-27 00:50:19 +00:00
eb6c5e5e1e Fix backslash url rewriting and add tests for rewrite_urls. (#461)
* Fix backslash url rewriting.

Add test for rewrite_urls.

Fixes #281.

* Update to v0.22.5

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2022-03-26 20:26:30 +00:00
ed11135af8 Update to v0.22.4 2022-03-26 13:09:24 -07:00
3a1af78e26 Wrap long post urls. (#462) 2022-03-26 19:55:53 +00:00
345770c64d Update libreddit.privacy.com.de domain 2022-03-26 19:54:26 +00:00
9eb42932df Hide empty sidebar 2022-03-24 21:19:21 -07:00
f0a6bdc21b Fix sorting buttons on r/all and r/popular (#402)
* Fix sorting buttons on r/all and r/popular

* Bump version to v0.22.2

* Fix empty sidebar in r/all and r/popular

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2022-03-15 03:39:39 +00:00
3eef60d486 Add instances (#432, #433, #436, #438) 2022-03-14 01:01:09 +00:00
59043456ba Wrap long post titles (fixes #435) 2022-03-13 12:59:15 -07:00
90c7088da2 Link Privacy Redirect as well 2022-03-13 19:19:56 +00:00
9e65a65556 Promote LibRedirect in README 2022-03-13 19:15:47 +00:00
8cfbde2710 Add LiberaPay "Donate" button back 2022-03-13 19:11:39 +00:00
70ff150ab4 Add user listing buttons (#400)
* Add user listing buttons

* Update to v0.22

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2022-03-13 19:06:27 +00:00
388779c1f2 Update instances (#421)
close #411 
close #412 
close #417

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-03-13 18:38:35 +00:00
6b605d859f Add German leddit.xyz instance (#429)
* new DE instance

* new hidden service

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-03-13 18:37:01 +00:00
0ae48c400c Add kylrth instances (#446)
Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-03-13 18:35:13 +00:00
a6ed18d674 Changed location of my VPS :) (#415)
* Changed location of my VPS :)

So my tor url changed too along side my VPS's country so yup!

* Update README.md
2022-03-13 18:32:26 +00:00
838cdd95d1 Update libreddit.drivet.xyz (#360)
* Remove Cloudflare Proxy from libreddit.drivet.xyz

I'm testing some stuff and as a result i have disabled proxy for libreddit.drivet.xyz. It exposes my public ip, but also gives more privacy i guess

* Move libreddit.drivet.xyz to Poland

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2022-03-13 18:32:10 +00:00
bc95b08ffd Update libreddit.datatunnel.xyz to Finland 2022-01-21 17:36:34 +00:00
e6190267e4 Add libreddit.datatunnel.xyz instance. Closes #401 2022-01-20 22:34:47 +00:00
3ceeac5fb0 Add lr.rfl890.cf instance. Closes #399 2022-01-19 18:14:51 +00:00
60eb0137c2 Add libreddit.bus-hit.me to instances (#398) 2022-01-19 18:13:46 +00:00
b6bca68d4e Add reddi.tk instance. Closes #397 2022-01-17 20:13:38 +00:00
91bff826f0 Fix and improve admin/mod distinguishers (#386)
* Fix regression with comments from deleted mods

Starting with https://github.com/spikecodes/libreddit/pull/367/files
comments from deleted moderators and admins(?) aren't highlighted.

* Highlight mod and admin usernames in posts

Works like on reddit + shows highlight for mods on the search page.
2022-01-09 02:50:53 +00:00
af6606a855 leddit.xyz instance location change (#387) 2022-01-09 01:14:05 +00:00
977cd0763a Fix #379 2022-01-05 16:46:45 -08:00
fcadd44cb3 Update dependencies 2022-01-05 16:39:56 -08:00
9c325c2cbf Search fixes (#384)
* Default to searching within subreddit

* Redirect to subreddit from search
2022-01-05 14:06:41 -08:00
e9038f4fe2 Add stilic.ml instance. Closes #380 2021-12-31 20:33:53 +00:00
8b8f55e09a Fix sort button scrollbars 2021-12-31 10:42:44 -08:00
f1b3749cf0 Fix #378 — formatting of dates/times 2021-12-29 12:48:57 -08:00
0708fdfb37 Cover more Reddit domains with libreddit link rewrites 2021-12-29 11:38:35 -08:00
cad29e9544 Add libreddit.nl instance. Closes #377 2021-12-28 16:16:53 +00:00
6b59976fcf Fix #376 2021-12-27 23:16:01 -08:00
f9b3981448 Fix debug log in post.rs 2021-12-27 19:56:37 -08:00
db3196df5a Use Reveddit to show removed posts/comments. Closes #299 2021-12-27 19:40:35 -08:00
b3d4f6f91c Show external page links again 2021-12-27 18:00:19 -08:00
45b875b85d Continue Rust workflow if version is already published 2021-12-27 14:21:06 -08:00
992d7889c4 Automatically publish to crates.io 2021-12-27 13:53:58 -08:00
3188f9d8e7 Tweak settings page design 2021-12-27 13:43:44 -08:00
90fa0b5496 Automatically generate release notes 2021-12-27 10:15:25 -08:00
7aeabfc4bc Rewrite Reddit post links to Libreddit equivalents 2021-12-26 21:18:20 -08:00
150ebe38f3 Add Buy Me a Coffee button as donation option 2021-12-27 01:38:14 +00:00
2905d114fa Add Buy Me a Coffee donation option 2021-12-27 00:48:56 +00:00
40e97cc75d Add esmailelbob.xyz instance (#369)
* add new instance

I want to add my own instance - i'm new to selfhosting so i can't guarantee it will be running all the time but i can guarantee whenever it's down i will fix it within hours!

* Fix onion protocol for instance

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2021-12-25 05:11:30 +00:00
7c73e352ce Fix [deleted] user link color 2021-12-19 17:12:33 -08:00
341c623be8 Refactor Media parsing (#334)
* Parse video data from cross_post_parent_list as vanilla Reddit does.

introduce testdata directory for testing JSON parsing functions.

refactor Media::parse for slightly more readability.

Add various test cases.

* Trim down to just refactoring

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2021-12-20 01:07:20 +00:00
4c8b724a9d Merge pull request #367 from alyaeanyx/no-href-to-deleted-user
Don't create hrefs to u/[deleted]
2021-12-19 16:58:04 -08:00
227d74b187 Add totaldarkness.net instance. Closes #366 2021-12-20 00:50:20 +00:00
f05a818edd Don't create hrefs to u/[deleted] 2021-12-19 12:20:37 +01:00
ceee13cfb7 docs: add the missing AUTOPLAY_VIDEOS in README
Merge pull request #361 from xatier/patch-2
2021-12-14 22:03:43 -08:00
a39495b3cb Update README.md
This config was introduced in 1d4ea50a45 , but missed from the documentation.
2021-12-12 14:41:00 -08:00
38cfe4ad71 Add libreddit.hu instsance. Closes #357 2021-12-08 23:13:09 +00:00
0b89539c2b Add lr.cowfee.moe instance. Closes #353 2021-12-01 03:44:44 +00:00
046b8b3edc Add new onion instance. Closes #349 2021-12-01 03:42:54 +00:00
0656756d21 Fix #196 2021-11-29 22:29:41 -08:00
43551f70fd Add leddit onion instance 2021-11-30 04:09:41 +00:00
364c29c4d5 Use resized icons for awards. Fixes #346 2021-11-28 14:47:50 -08:00
e6c978a2f7 Update to v0.20 2021-11-27 20:34:43 -08:00
91cc140091 Set sub and user descriptions to overflow-wrap: anywhere (#345) 2021-11-28 02:49:41 +00:00
6f29d94337 List Liberapay as donation method 2021-11-27 15:07:44 -08:00
67e26479ae Add leddit.xyz instance. Closes #344 2021-11-27 22:57:56 +00:00
1a1dee36b8 Update FUNDING.yml 2021-11-27 22:55:27 +00:00
b63000a93f Create FUNDING.yml 2021-11-27 22:27:47 +00:00
401ee2ee41 Create FUNDING.yml 2021-11-27 22:27:39 +00:00
99a83ea11b Add northboot.xyz instance 2021-11-27 19:13:15 +00:00
888e7b302d Filter subreddits and users (#317)
* Initial work on filtering subreddits and users

* Fix doubly-prefixed subreddit name in search alt text (e.g. r/r/pics)

* Don't set post title to "Comment" if empty - this could throw off actual posts with the title "Comment"

* Filter search results

* Fix filtering to differentiate between "this subject itself is filtered" vs "all posts on this current page have been filtered"

* Remove unnecessary check

* Clean up

* Cargo format

* Collapse comments from filtered users

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-11-26 04:02:04 +00:00
beada1f2b2 Update privacy policy 2021-11-25 05:38:09 +00:00
bd413060c6 Support displaying awards (#168)
* Initial implementation of award parsing

* Posts: Implement awards as part of post

* Posts: remove parse_awards dead code

* Posts: initial implementation of displaying Awards at the post title

* Posts: Proxy static award images

* Client: i.redd.it should take path as argument not ID

* Posts: Just like Reddit make award size 16px

* Templates: limit the awards to 4 awards to increase performance

* Comments: Make awards a property of comments and display them

* Format and correct /img/:id

* Update comment.html

* [Optimization] Awards is not longer async

* [Revert] Posts can now display more than 4 awards again

* [Implementation] Awards not display on the frontpage

* [Implementation] Display count on awards

* Post: Start working on awards css

* Awards: Move the image size to css

* Awards: Start implementing tooltips

* Refactor awards code and tweak CSS indentation

* Unify Awards::new and Awards::parse

* Use native tooltips and brighten awards background

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2021-11-25 02:08:27 +00:00
3054b9f4a0 Add rosebox theme (#237) 2021-11-24 19:31:19 +00:00
1cccef12a4 Add settings helper for HLS toggle 2021-11-23 22:43:25 -08:00
8e332b0630 Show full subreddit results in search 2021-11-23 22:24:23 -08:00
85ae7c1f60 Fix indentation and formatting 2021-11-23 22:23:29 -08:00
6d73024183 Remove obselete HTML comment 2021-11-22 18:16:06 -08:00
923ff776bd Fix HLS + autoplay from breaking videos 2021-11-22 18:07:00 -08:00
e181e3f57d Improve Replit deployment method 2021-11-22 13:01:34 -08:00
79bb913fa6 Condense free hosting options into one deployment method 2021-11-22 13:00:39 -08:00
632b64c98b Add Glitch deployment option. Closes #339 2021-11-22 20:50:12 +00:00
2878d9c799 Shrink and widen comment scores to fit better 2021-11-22 12:19:51 -08:00
9f8d36cb00 Prevent post previews from overflowing on Chromium 2021-11-22 08:40:36 -08:00
25e641e7b3 Don't run GitHub Workflow when README.md changes 2021-11-21 22:46:01 -08:00
4faa9d46d6 Fix HTTPS connector 2021-11-21 22:44:05 -08:00
7220190811 Link subscriptions in settings to their respective feeds 2021-11-21 22:30:44 -08:00
768820cd4c Render markdown correctly in text post previews by using selftext_html. (#335)
* Render markdown correctly in text post previews by using selftext_html.

I was mistakenly under the impression that we somehow render markdown ourselves, but turns out we just take whatever HTML reddit gives us, and we also need to do this for text previews.

Use CSS to limit the size of the previews instead of truncating in the template.

Fix table CSS.

* Fix post_body padding and trim post_previews

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-11-21 23:17:52 +00:00
2ef7957a66 Create feature parity issue template 2021-11-21 19:48:48 +00:00
7df8e7b4c6 Tweak feature request template 2021-11-21 19:27:33 +00:00
67d3be06e1 Fix Jamie's instance address 2021-11-21 04:17:06 +00:00
6be5eb8991 Update README.md 2021-11-21 04:15:35 +00:00
5d9c320a7e Format post and comment votes with a decimal place, like vanilla reddit does. (#324)
* Format post and comment votes with a decimal place, like vanilla reddit does.

Before this change, a vote count of 1999 was displayed as 1k, which is a pretty big gap. The displayed count also differed from what Reddit does. Now, the behaviour is consistent.

Added some tests for format_num.

* Provide more space for post scores

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-11-21 04:07:45 +00:00
f7de5285e4 Hide post preview in compact mode. (#329) 2021-11-21 02:05:37 +00:00
c2053524c7 Add text post previews. (#328)
* Add text post previews.

* Add mask gradient over post preview text

* Increase post title font weight for contrast

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-11-20 21:13:50 +00:00
3a9e6b4ca0 Add mutahar.rocks instance 2021-11-20 17:29:27 +00:00
731a407466 Collapse (sticky) bot comments by default. (#321)
* Collapse bot comments by default.

Comments are considered bot comments if they are posted by a moderator and are stickied. Some false positives are expected.

* Remove unneeded String conversion

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-11-19 05:42:53 +00:00
34ea679519 Update crates and optimize parameters 2021-11-14 18:51:36 -08:00
0f7ba3c61d Add "open in reddit" button to all pages (#304)
* Pass the url parameter to all templates. Add a reddit_link to the navbar, which opens the current url on reddit.

* Add icon for reddit link

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-11-15 02:39:33 +00:00
2486347b14 Fix follows not being case-sensitive (#316) 2021-11-15 00:45:18 +00:00
c298109a7b Change healthcheck script to utilise "wget" (#309)
* Change healthcheck script to utilise "wget"

* add "tries=1" option
2021-11-11 00:47:13 +00:00
a0509890b7 Add Riverside onion (#314) 2021-11-11 00:45:22 +00:00
5644d621f7 Remove cloudflare checkmark from my instance. (#313) 2021-11-11 00:45:13 +00:00
1fc5bda486 Move libreddit.dothq.co to German server (#311) 2021-11-06 02:58:36 +00:00
b3255c22cf Add libreddit.pussthecat.org (#310) 2021-11-05 01:36:39 +00:00
1d4ea50a45 Add setting to autoplay videos 2021-10-25 21:27:55 -07:00
546c8a4cda Add poster attribute and disable autoplay on GIFs
* Add the poster attribute even if a post claims to be type gif. Default to none-preloading for gifs like video-typed posts do.

* Disable autoplay for videos in feeds

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2021-10-23 21:25:51 +00:00
03336ecafd Add libreddit.de and onion host. Closes #300 2021-10-19 13:59:46 -07:00
957e1c7728 Add oversold.host instance. Closes #289 2021-10-08 23:28:21 +00:00
09053ef0ad Add drivet.xyz instance. Closes #287 2021-10-07 00:13:48 +00:00
aff030fc3a Remove bcow instances 2021-10-02 05:12:49 +00:00
97555dbfdd Fix silkky.cloud Cloudflare Status (#284) 2021-09-29 00:56:09 +00:00
32360e5165 Update v0.15.3 2021-09-19 14:37:31 -07:00
350b796571 Support Deployment to Heroku (#280)
* Added heroku.yml

* Added app.json

* PORT as env var
2021-09-19 19:03:01 +00:00
567556711b Update style.css (#282)
Add cursor pointer on hover of summary bar
2021-09-19 19:00:07 +00:00
1ff725ba2e Add flux.industries instance. Closes #279 2021-09-11 19:46:50 +00:00
6a4191f3b5 Fix #272 2021-09-09 17:28:55 -07:00
668493b72c Add autarkic.org instance. Closes #274 2021-09-10 00:18:03 +00:00
db04dcb238 Add igna.rock instance. Closes #276 2021-09-10 00:16:46 +00:00
cc0a1e0324 Use proper spelling of systemd (#275) 2021-09-10 00:15:09 +00:00
e073fc87aa Add alefvanoon.xyz (#271) 2021-09-08 14:28:53 -07:00
982f57efd9 Fix user profiles showing up in search engines 2021-09-06 12:05:03 -07:00
52a1b45014 Lazy load images 2021-09-06 12:02:52 -07:00
6f88fdfc75 Add mint.lgbt instance. Closes #269 2021-09-04 20:07:42 +00:00
015d0b3414 Add stuehieyr.com instance. Closes #267 2021-09-01 00:20:11 +00:00
b41eabecf7 Remove instance (#266) 2021-08-29 18:41:30 +00:00
5cb5f46fa2 Add new instance (libreddit.sugoma.tk) (#262) 2021-08-21 16:52:39 +00:00
a900339529 Add some-things.org instances. Closes #260 2021-08-18 17:06:04 +00:00
41b3dc5739 More apt error messages for Reddit outages 2021-08-11 20:49:42 -07:00
b3b5782373 Handle Docker amd64 builds in GitHub Actions 2021-08-04 12:08:25 -07:00
5c753ee171 Fix #251 2021-08-04 11:52:24 -07:00
229518c40b Remove cyberhost instance. Closes #252 2021-08-01 16:48:26 +00:00
45a5778571 Escape text-only flairs 2021-07-19 10:20:00 -07:00
be253d40dd Escape html characters in post flairs (#247)
* Encode HTML characters in flairs

* Encode HTML characters in flairs

* Use esc! macro for HTML escaping

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-07-19 17:15:15 +00:00
e571cc3b1e Tweak styling of Dracula theme 2021-07-19 10:07:00 -07:00
345f8e7b80 Dampen title color of visited posts. Fixes #222 2021-07-18 14:53:08 -07:00
a190890239 Fix Regex matching of Reddit links 2021-07-17 22:24:28 -07:00
ee51ce1a76 Add instance (#242) 2021-07-07 00:08:02 +00:00
81a2df98cb Add awesomehub.io instance. Closes #239 2021-07-05 00:14:38 +00:00
e79a4b704a Add cyberhost.uk instance. Closes #238 2021-07-05 00:03:38 +00:00
56998b8332 Rewrite redd.it links 2021-06-21 22:51:50 -07:00
5418303b08 ARMV7 Docker Image (#234)
* feat: Push dockerfile.

Progress on trying to get armv7 image.

* feat: Add Dockerfile.

Using rust:slim as builder image.

* refactor: Changes to build for armv7.

* feat: Add .cargo config.

Taken from: https://medium.com/swlh/compiling-rust-for-raspberry-pi-arm-922b55dbb050

* refactor: Add environment variable for linker.

Instead of .cargo config file.

* feat: Working cross compile version.

For Armv7.

* refactor: Clean up dockerfile.

* refactor: Rename to armv7.

Rename Dockerfile.armv7rust to Dockerfile.armv7.

* feat: Add workflow to build ARMv7 docker image.

* docs: Add armv7 deployment instructions.
2021-06-21 23:03:27 +00:00
5ab41c4e6e Add r.nf instance. Closes #233 2021-06-18 23:46:20 +00:00
807b3ffeca Add artemislena.eu instance. Closes #232 2021-06-17 20:44:42 +00:00
85deb4947d Support HLS playback in search and user feeds 2021-06-11 17:38:43 -07:00
d2002c9027 Disable dysfunctional moderator list feature 2021-06-11 11:03:36 -07:00
f84f4c0326 Add Trevor instance 2021-05-31 04:55:39 +00:00
ca3f6c0579 Fix #228 2021-05-28 12:01:20 -07:00
decc9e5139 Include SystemD configuration (#227) 2021-05-28 04:33:14 +00:00
d27bd782ce Specify fallback fonts 2021-05-26 20:30:08 -07:00
4defb58f2a Optimizations and commenting 2021-05-20 12:24:06 -07:00
ba42fc066f Fix two subscription bugs 2021-05-19 20:30:10 -07:00
2cd35fb3b6 Upgrade to v0.14.2 2021-05-19 16:12:21 -07:00
b9af6f47f3 Use Inter font 2021-05-19 16:09:08 -07:00
73732a2a44 Fix subscription clearing when saving settings 2021-05-19 15:59:32 -07:00
43ed9756dc Upgrade to v0.14 2021-05-16 09:11:38 -07:00
8bb247af3b Added support for quarantined subreddits (#219)
* Added support for quarantined subreddits

* Added confirmation wall for quarantined subreddits

* Added quarantine walls to other routes and fixed case issue

* Correct obsolete use of cookie()

* Refactor param() and quarantine()

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2021-05-16 15:53:39 +00:00
ed05f5a092 misc: fix HLS typo (#220) 2021-05-16 15:41:47 +00:00
4f09333cd7 Handle three unwraps 2021-05-15 14:51:57 -07:00
31bf8c802e Document server configuration 2021-05-15 14:32:38 -07:00
e4f9bd7b8d Configure default settings using environment variables 2021-05-15 13:59:42 -07:00
83a667347d Add rel="nofollow" to hardcoded outbound links 2021-05-10 10:31:19 -07:00
499a56aed4 Bump version to v0.12.0 2021-05-09 18:27:25 -07:00
928907086c HLS video playback (#182)
* HLS video playback

Signed-off-by: Adrian Lebioda <adrianlebioda@gmail.com>

* Add LibreJS compliance

* Locally host hls.js

* Notification about HLS under videos that support it

Signed-off-by: Adrian Lebioda <adrianlebioda@gmail.com>

* Use .contains() instead of .find() == None

* Make list of preferences constant

* Change headers_keys from Vector into Array

* Fix incorrect detecting of # in paths

* Remove trailing-slash-appending if statement

* Change HLS notification styling

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-05-10 01:25:52 +00:00
dc9fbc1a05 Don't upload Libreddit releases in pull requests 2021-05-09 18:13:24 -07:00
7ae7a88eed Upgrade to v0.11.2 2021-05-09 08:51:01 -07:00
536a766960 WIP: Various subreddit & post fixes (#215)
* Fixed random subreddit issue

* Fixed large subreddit icon rendering

* Formatting fix

* Fix dodgy HTML rendering issues

* Revert "Fix dodgy HTML rendering issues"

This reverts commit 58be5f449b72f271d2b3c046870b652d1e715289.
2021-05-09 15:40:49 +00:00
e34329cfee Upgrade to v0.11.1 2021-05-08 22:09:47 -07:00
97a0680bd0 Support GIFs in comments (#217)
* Support GIFs in comments

* Fix removing Giphy links so it only removes Giphy links

* Remove removing link to Giphy
2021-05-09 01:22:26 +00:00
c1560f4eba Upgrade to v0.11.0 2021-05-06 12:31:59 -07:00
242ffab0da Fix bug with subreddit subscription case & RTL languages (#214)
* Fixed subreddit subscription case issues

* Fixed formatting

* Fixed flair RTL language issue (#132)

* Convert display_lookup to Vec

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-05-06 19:11:25 +00:00
1211d781d0 Add list of moderators to sidebar (#213)
* Added list of moderators to sidebar & added wiki not found message

* Improved code formatting
2021-05-04 17:30:54 +00:00
9e4066658c Added 2 new themes: violet & gold (#212)
* Added 2 new themes: violet & gold

* Increased contrast in Violet theme

* Changed accent colour on violet theme
2021-05-03 16:48:21 +00:00
560de4e91f removed himiko.cloud instances (#211) 2021-05-02 19:26:13 +00:00
bd1c890961 Update to v0.10.6 2021-04-30 16:35:41 +00:00
6f799b2617 Added laserwave theme (#210) 2021-04-30 16:26:49 +00:00
38e176f59f Add riverside.rocks instance (#209) 2021-04-30 16:25:50 +00:00
8248eca95c Correct Actix → Hyper in Readme 2021-04-27 18:54:38 +00:00
ffc3bfe72d Add libreddit.domain.glass instance 2021-04-22 19:26:42 +00:00
d713746407 doc: add new self-hosted hidden service => http://kphht2jcflojtqte4b4kyx7p2ahagv4debjj32nre67dxz7y57seqwyd.onion/ (#203) 2021-04-22 19:14:11 +00:00
21b45760eb Add exonip.de instance. Closes #200 2021-04-20 16:41:29 +00:00
e3fb93946a Revert ARM builds to only arm64 2021-04-17 21:32:48 -07:00
b6134a39d0 Specify build platform for Docker ARM builds 2021-04-17 21:11:33 -07:00
c844655c98 Use rust:latest as base Docker image for arm tag 2021-04-17 20:56:12 -07:00
cac83493da Add arm/v7 platform for Docker builds 2021-04-17 18:58:18 -07:00
b47cfd1ba5 Only scroll overflowing tables in Wikis 2021-04-16 14:47:42 -07:00
28ca3589ed Add scrollbar to overflowing wikipages. Fixes #192 2021-04-15 15:56:48 -07:00
3cf787cf98 Fix #195 2021-04-14 21:53:17 -07:00
46e22cf74e Fix certificate error on ARM #193 2021-04-15 04:50:03 +00:00
5c2e134924 Include Cargo.lock. Fixes #191 2021-04-14 21:44:16 -07:00
c6244585fa Add database.red instance. Closes #194 2021-04-15 04:35:57 +00:00
9f1ba274eb Document ARM Docker deployment in README 2021-04-10 14:31:26 -07:00
93ed1c6f0c Fix Dockerfile healthchecks 2021-04-09 22:38:13 -07:00
6ce82c36fb Use alpine only for ARM builds 2021-04-09 18:59:04 -07:00
2974d92e30 Set correct Dockerfile for ARM builds 2021-04-09 18:27:12 -07:00
34dfcb2512 Revert ARM Dockerfile 2021-04-09 17:05:20 -07:00
6b42e97bda Test rust:alpine in ARM Dockerfile 2021-04-09 16:38:44 -07:00
49bfe4d27c Create dockerfile for arm64 2021-04-09 15:50:43 -07:00
c8965ae51b Switch Docker base image to "scratch" 2021-04-09 15:24:47 -07:00
0b64a52a63 Switch Docker builds in GitHub Actions to only ARM 2021-04-09 15:07:07 -07:00
a18db1e2b7 Properly pass preview queries to media proxy 2021-04-08 22:26:03 -07:00
3b53e5be4c Fix issue templates (#181)
* Fix comments bug report

* Fix comments feature request
2021-04-06 17:46:23 +00:00
42e8351285 Enhance Issue templates (#177)
* comment out in bugreport

* Comment out in featurerequest
2021-04-06 17:32:44 +00:00
b3e4b7bfae Add user following functionality 2021-04-06 10:23:05 -07:00
c8805f1078 Add opensearch support 2021-04-06 01:23:14 +02:00
4a42a25ed3 Add libreddit.silkky.cloud. Closes #175 2021-04-05 21:57:06 +00:00
2bacaa163f Bump to v0.9 2021-04-01 18:00:18 -07:00
48c3a8c0d0 Added Dracula/Nord theme (#171)
* Added Dracula theme

* Updated accent and added Nord theme

* Updated accent and added Nord theme

* Added official foreground colors
2021-04-02 00:56:28 +00:00
c23d2dc50b Re-add unprivileged user to Dockerfile 2021-04-01 12:09:08 -07:00
46dbd88d91 New alpine-based Dockerfile with healthcheck 2021-03-31 13:05:22 -07:00
f0f484288e Fix server.rs function name 2021-03-31 13:03:44 -07:00
90d39b121f Example docker-compose 2021-03-31 13:03:18 -07:00
44dee302c9 Publish SHA512 checksums for future releases 2021-03-27 20:11:05 +00:00
c7f9386c01 Fix #169 2021-03-27 13:03:13 -07:00
66ac72beab Fix clippy errors 2021-03-26 20:00:47 -07:00
14f9ac4ca7 Automatically draft a release on push 2021-03-26 21:48:00 +00:00
6a7f725c12 Default subreddit post sorting. Closes #166 2021-03-25 21:41:58 -07:00
2533e8cef5 Add phii.me instance. Closes #163 2021-03-23 21:08:00 +00:00
772d20615b Sidebar about page. Closes #162 2021-03-21 19:28:05 -07:00
0bb1677520 Revert client to HTTP/1.1 2021-03-21 13:56:05 -07:00
da4883db29 Upgrade client to HTTP/2 2021-03-21 11:37:03 -07:00
d50b6ca4b3 Add reddit.invak.id instance. Closes #157 2021-03-21 18:23:05 +00:00
4c66e75f6b Add HSTS command line flag 2021-03-20 22:10:31 -07:00
966e0ce921 Expand truncated numbers on mouseover. Close #156 2021-03-20 15:42:47 -07:00
ab886d1e67 Fix #155 2021-03-20 13:03:05 -07:00
dc7e087ed0 Truncate negative scores 2021-03-19 22:04:44 -07:00
0d6e18d97d Fill background of Apple Touch Icon 2021-03-18 21:36:39 -07:00
f872baa1fe Update 0.5.2 2021-03-18 21:35:14 -07:00
9b5176f7b9 Sub icons and truncated subscribers in search results 2021-03-18 21:32:54 -07:00
60c89197e5 Update "Built with" section of Readme 2021-03-18 16:53:03 +00:00
7d94876d90 Update screenshot 2021-03-18 16:47:10 +00:00
467342edf4 Patch broken homepage sorting 2021-03-18 08:51:31 -07:00
3c5b4037e2 Add libreddit.40two.app to instance list 2021-03-18 04:47:16 +00:00
a81502dde1 Fix wiki routing 2021-03-17 21:40:55 -07:00
0ce2d9054e Handle non-ASCII post headers 2021-03-17 21:26:06 -07:00
a5203fe8dd Include docker pull in Docker instructions (#153) 2021-03-18 02:48:38 +00:00
038fafa378 Remove unnecessary trailing slashes 2021-03-17 17:28:05 -07:00
e15c15c390 Fix author flairs not showing 2021-03-17 17:08:43 -07:00
07363e47a9 Request building error handler 2021-03-17 16:53:49 -07:00
fb7faf6477 Handle HeaderValue and Uri parsing errors 2021-03-17 16:32:28 -07:00
b14b4ff551 Rewrite server and client using Hyper 2021-03-17 15:30:33 -07:00
4b1195f221 Update to v0.4.2 2021-03-12 13:48:43 -08:00
a472461ee8 Switch some links in Readme to spike.codes instance 2021-03-12 12:59:50 -08:00
baf5e3d7ee Fix Replit links 2021-03-12 12:55:02 -08:00
f209757ed6 Handle proxy unwraps 2021-03-12 12:21:02 -08:00
4173362ce1 Fix #148 2021-03-11 20:15:26 -08:00
b2ae5e486f Rename subreddit::page to subreddit::community 2021-03-10 21:43:06 -08:00
cda19a1912 Remove duplicate "description" meta tag for posts 2021-03-10 21:41:39 -08:00
f0b69f8a4a Update to v0.4 2021-03-10 20:51:08 -08:00
118ff9485c Document proxy.rs 2021-03-10 19:02:03 -08:00
4a51b7cfb0 Horizontally squish comments 2021-03-10 15:10:59 -08:00
f877face80 Update README.md 2021-03-10 20:56:33 +00:00
f0e8deb000 Add alt attribute to user icon 2021-03-10 11:29:36 -08:00
e70dfe2c0b Fix <video> size attributes 2021-03-10 10:49:18 -08:00
2e89a85858 Handle alternative status codes 2021-03-09 22:23:26 -08:00
e59b2b1346 Custom HTTP client with Rustls 2021-03-09 22:13:46 -08:00
1c36549134 Fix #146 2021-03-09 07:22:17 -08:00
5fb88d4744 Allow certain clippy lints 2021-03-08 19:22:10 -08:00
6c7188a1b9 Prevent pushing of Cargo.lock 2021-03-08 18:50:03 -08:00
84009fbb8e Remove Cargo.lock 2021-03-08 18:49:35 -08:00
bf783c2f3a Optimize type casting 2021-03-08 18:49:06 -08:00
213babb057 Update dependencies 2021-03-08 16:30:34 -08:00
7dbc02d930 Update himiko instances' location 2021-03-05 18:39:31 +00:00
10873dd0c6 Fix #144 2021-03-05 06:24:40 -08:00
c0d1519341 Update screenshot in README.md (#143)
* Update screenshot in README.md

New screenshot for v0.3.1. Also 35% lighter!

* Update screenshot

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2021-03-05 04:30:32 +00:00
8709c49f39 NGINX reverse proxy notice 2021-03-04 03:53:49 +00:00
56cfeba9e5 Update to v0.3.1 2021-03-03 09:35:40 -08:00
890d5ae625 Revert to curl-client 2021-03-03 09:24:31 -08:00
caa8f1d49e Test h1-client 2021-03-03 09:15:19 -08:00
dd51b23dc4 Switch Surf backend to h1-client-rustls 2021-03-02 19:44:51 -08:00
52d9698879 Add british cow instances. Closes #141 2021-03-02 08:56:01 -08:00
20f6945160 Fix #140 2021-02-27 13:34:02 -08:00
10c73fad7f Switch favicon to ico 2021-02-26 12:04:11 -08:00
2bddc952cb Link comment timestamps. Closes #137 2021-02-25 21:53:27 -08:00
1de01d7283 Log errors to io::stderr 2021-02-25 16:21:56 -08:00
9183ce1921 Handle links in Media::parse 2021-02-25 11:43:58 -08:00
a197df89ff Turn off media logging 2021-02-25 11:30:15 -08:00
be2a1d876b Fix url rewrites 2021-02-25 11:01:25 -08:00
686d61801f Fix #110 2021-02-25 10:24:37 -08:00
5d643277bc Geometric logo 2021-02-25 09:07:45 -08:00
a3ec44149c Categorize utilities 2021-02-24 21:29:23 -08:00
83ba0fb913 Reduce dependencies 2021-02-24 20:26:12 -08:00
55e9915bb0 Refactor post_body width 2021-02-24 11:28:26 -08:00
5cd5b553b0 Handle suspended users 2021-02-24 11:26:23 -08:00
2b2bd8421b Added multi-staged container build and a user (#134)
* Added multi-staged container build and a user

This reduces the image size from 2Gb to 92Mb by only put the necessary files into the container.
Container now runs without root priviliges.

* Add EXPOSE to Dockerfile
2021-02-24 19:17:36 +00:00
47d01a0dca Fix logging 2021-02-24 11:00:04 -08:00
0a69937238 Specify Ubuntu version of workflow 2021-02-24 18:44:50 +00:00
6d08f2dd24 Fix post body overflow on mobile 2021-02-24 09:31:58 -08:00
4a06882dc8 Simplify routes in main.rs 2021-02-24 09:26:01 -08:00
3e567d9acf Add new instance 2021-02-23 12:54:39 -08:00
8034594006 Better subreddit error messages. Closes #131 2021-02-22 16:43:32 -08:00
2f3315dcfc Fixes #130 2021-02-22 12:56:23 -08:00
df118764df Update README.md 2021-02-22 05:45:56 +00:00
d78f82649e List other images in manifest 2021-02-21 20:37:53 -08:00
80fb3a5c18 Fix #110 2021-02-21 20:28:04 -08:00
518d5753a7 Handle about pages 2021-02-21 10:13:20 -08:00
de38f7ef18 Fix post flairs 2021-02-21 10:11:17 -08:00
dd67b52199 Fix #126 2021-02-20 18:36:30 -08:00
9cfab348eb Filter by flair. Closes #124 2021-02-20 13:59:16 -08:00
e1f7b6d0c0 Remove defunct Libreddit instance 2021-02-20 20:19:56 +00:00
a606e48435 Handle 4 more unwraps 2021-02-20 12:14:32 -08:00
2091f26bda Fix media previews 2021-02-19 21:49:02 -08:00
b3341b49c0 Individually proxy subreddit and user icons 2021-02-19 21:46:44 -08:00
65e4ceff7b Individually proxy previews 2021-02-19 20:50:55 -08:00
bacb22f7f9 Fix post url indentation 2021-02-19 18:19:04 -08:00
902c9a6e42 Individually proxy custom emojis 2021-02-19 18:18:09 -08:00
c586de66ba Individually proxy images and thumbnails 2021-02-19 12:55:07 -08:00
e466be8946 Fix manifest and update dependencies 2021-02-19 11:10:48 -08:00
bed3465475 Remove "Options" and fix "Built With" 2021-02-19 09:20:09 -08:00
8560e8a37a Add "port" command line argument 2021-02-18 11:49:50 -08:00
3652342f46 Use clap for arg parsing 2021-02-18 11:40:10 -08:00
58127b17d8 Individually proxy videos 2021-02-18 10:04:59 -08:00
2f4deb221a Reduce dependencies 2021-02-16 11:39:56 -08:00
38230ed473 Add more rich meta tags (#121) 2021-02-16 19:16:32 +00:00
71501b064c Update dependencies 2021-02-15 14:09:45 -08:00
47a58ea05c Add CODEOWNERS file 2021-02-15 13:46:07 -08:00
14ecf3cf60 Edit indicator 2021-02-14 14:53:09 -08:00
aa7c8c85df Templatize redirects 2021-02-13 15:02:38 -08:00
0cb7031c36 Fix focus outline 2021-02-13 13:38:12 -08:00
93cfc713c6 Generate URL to restore settings, including subscriptions. Closes #89 (#116)
* Start recursive comments

* Update comment.html

* Fix move error

* Comment improvements

* Fix merge

* Remove extra endif from post.html

* Fix post.html

* Restore setting from link

* Tweak settings page

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-02-13 20:55:23 +00:00
ff8685ae4c Add tooltips for accessibility 2021-02-12 20:53:33 -08:00
f06320a4ae Subscribe to multireddit button. Closes #104 2021-02-12 20:47:54 -08:00
809be42e01 Add "View all comments" and "Show parent comments" buttons when viewing a single thread. Closes #65 (#115)
* Start recursive comments

* Update comment.html

* Fix move error

* Comment improvements

* Fix merge

* Remove extra endif from post.html

* Fix post.html

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-02-12 09:16:59 -08:00
58ca085521 Fix subscription list overflow 2021-02-11 09:18:32 -08:00
4a40e16277 Fix comment structuring (#113)
* Start recursive comments

* Update comment.html

* Fix move error

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-02-10 10:48:51 -08:00
fee2cb1b56 Split subscription names by + 2021-02-09 21:56:38 -08:00
8785bc95f5 Fix extra slashes in post bodies 2021-02-09 21:54:55 -08:00
16454213cf Fix Dockerfile 2021-02-10 01:31:28 +00:00
6feb347c27 Fix post ID parsing 2021-02-09 12:08:38 -08:00
e731cfbac4 Support post links without titles 2021-02-09 10:11:39 -08:00
008924fff8 Fix listen address 2021-02-09 09:54:13 -08:00
ebbdd7185f Move from Actix Web to Tide (#99)
* Initial commit

* Port posts

* Pinpoint Tide Bug

* Revert testing

* Add basic sub support

* Unwrap nested routes

* Front page & sync templates

* Port remaining functions

* Log request errors

* Clean main and settings

* Handle /w/ requests

* Create template() util

* Reduce caching time to 30s

* Fix subscription redirects

* Handle frontpage sorting
2021-02-09 17:38:52 +00:00
402b3149e1 Fix 'no entry found for key' error 2021-02-07 17:56:06 -08:00
ac5ef89dff Fix gallery unwrapping 2021-02-07 17:33:54 -08:00
7edca18f8d Inline videos/gifs for card view (#107)
* Basic gallery support

* Inline videos for card view
2021-02-08 00:22:14 +00:00
cf45d53fdd Basic gallery support (#103) 2021-02-06 20:05:11 +00:00
2a475d127a Added black theme (#101) 2021-02-06 20:04:29 +00:00
3fa523e67b Add a multi-architecture docker build. (#98)
* Add a multi-arch docker build.

* Remove caching
2021-02-04 22:17:04 +00:00
3fbb433e37 Update dependencies 2021-02-03 21:48:56 -08:00
5fbcfd850f Merge pull request #97 from Mennaruuk/master
Fix installation method links
2021-02-03 21:46:28 -08:00
c758db84ec Merge pull request #95 from robrobinbin/patch-5
Remove tiny "padding" below inline images
2021-02-03 21:45:11 -08:00
90d3063f93 Fix #96 2021-02-03 21:42:43 -08:00
82a601d534 Minor update to README.md 2021-02-04 00:39:46 -05:00
12a1b3f459 Update style.css 2021-02-03 22:01:46 +01:00
e23eaf0be0 Merge pull request #92 from robrobinbin/master
Fix layout on browsers with faulty "display: contents" support (Safari)
2021-02-03 11:55:04 -08:00
821709c8d2 Resolve merge conflict 2021-02-03 20:14:37 +01:00
653b0e7024 Don't use display contents and remove duplication 2021-02-03 20:11:04 +01:00
c7a2c43287 Merge pull request #93 from mcrossman/master
Make feed sorting case-insensitive
2021-02-02 19:13:35 -08:00
9824370771 Fix feed sorting to be case-insensitive. 2021-02-03 10:53:09 +11:00
d87b96d0ea Update style.css 2021-02-02 21:33:23 +01:00
6eae4bc47a Update style.css 2021-02-02 21:23:51 +01:00
1bcb070fbb Update user.html 2021-02-02 21:21:29 +01:00
24bc758090 Update subreddit.html 2021-02-02 21:21:11 +01:00
ffbb1cf7cd Update search.html 2021-02-02 21:20:38 +01:00
cbf1f540d6 Update post.html 2021-02-02 21:20:03 +01:00
f8e0d2d4b9 Merge pull request #7 from spikecodes/master
Merge upstream
2021-02-02 21:18:21 +01:00
8a27b2bac8 Support /w/ for Wikis 2021-02-02 08:59:50 -08:00
69941d9efd Implement #88 2021-02-01 17:50:00 -08:00
956de50419 Change Libreddit PWA Theme Color 2021-02-01 16:26:35 -08:00
d790264a62 Merge pull request #87 from robrobinbin/master
Add fallback to thumbnails when SVG is not supported
2021-02-01 12:53:03 -08:00
f4f2d8a377 Update style.css 2021-02-01 21:02:38 +01:00
dd908c9f68 Update user.html 2021-02-01 21:00:03 +01:00
9e1948733d Update subreddit.html 2021-02-01 20:59:31 +01:00
9df1dfae32 Update search.html 2021-02-01 20:58:59 +01:00
cfbee1bb81 Merge pull request #6 from spikecodes/master
Merge upstream
2021-02-01 20:57:06 +01:00
8430cbc6f3 Merge pull request #86 from robrobinbin/patch-4
Place noscript placeholder into grid
2021-02-01 11:45:34 -08:00
a9dd2e6f2c Place noscript placeholder into grid 2021-02-01 20:43:32 +01:00
36964982fb Merge pull request #85 from robrobinbin/master
Improve support for browsers without inline SVG support
2021-02-01 11:39:35 -08:00
0742a33304 Update base.html 2021-02-01 20:32:57 +01:00
7f320b3143 Update style.css 2021-02-01 20:27:56 +01:00
58f4fc4e77 Update user.html 2021-02-01 20:27:08 +01:00
7d8faefad0 Update search.html 2021-02-01 20:26:35 +01:00
ba9b5afd4e Update post.html 2021-02-01 20:25:57 +01:00
ae09f77bf6 Update subreddit.html 2021-02-01 20:25:06 +01:00
5030c418de Merge pull request #5 from spikecodes/master
Merge upstream
2021-02-01 20:23:35 +01:00
4ccd6b1751 Merge pull request #84 from JPyke3/tor-service
Added new Tor Hidden Service
2021-02-01 10:25:52 -08:00
7d17aa0627 Merge pull request #83 from JPyke3/master
iOS "Add to Homescreen" functionality
2021-02-01 10:25:28 -08:00
4b73e2d914 Added personal Hidden Service 2021-02-01 18:44:55 +01:00
0a140a6ffc Merge branch 'master' of github.com:JPyke3/libreddit into master 2021-02-01 11:13:36 +01:00
e837d84105 Add Support for iOS "Add to Homescreen"
* Adds basic Manifest.json
 * Adds Meta Tags for iOS
 * Adds Meta Tags for Android
 * Adds Logo for Manifest.json
 * Adds iOS Logo for homescreen
2021-02-01 11:10:53 +01:00
f6d791ccd9 Style focus outline 2021-01-31 20:56:13 -08:00
effaeb7508 Fix debug logging error 2021-01-31 19:08:50 -08:00
6257faf9dc Update screenshot 2021-01-31 19:05:06 -08:00
ee0da63862 Update version and screenshot 2021-01-31 18:50:31 -08:00
971f14bb55 Fix #66 2021-01-31 14:10:13 -08:00
9a1733ac99 Fix cog class 2021-01-31 13:49:55 -08:00
c32d62fbd5 Shrink links to icons on mobile 2021-01-31 13:22:11 -08:00
1a0d12d2ff Merge pull request #82 from robrobinbin/patch-2
Minor style tweak
2021-01-31 12:40:32 -08:00
2a27850914 Minor style tweak
on small screens, subscriptions are put below the logo, which looks a bit strange and makes it somehow harder to hit
2021-01-31 19:41:32 +01:00
bfcc4c985d Fix subreddit sorting and media 2021-01-30 21:46:35 -08:00
1653d4fb4c Hide NSFW content by default 2021-01-30 21:43:46 -08:00
79027c4c75 Merge pull request #78 from mcrossman/subscriptions
Subscribing to subreddits (favorites)
2021-01-30 21:28:51 -08:00
269bb0bfb6 Convert subscription requests to POST 2021-01-30 21:21:46 -08:00
7933d840b3 Squish navbar further 2021-01-30 20:42:12 -08:00
b875e9377e Merge branch 'master' into subscriptions 2021-01-30 20:25:48 -08:00
8c80946121 Fix multireddit sidebars 2021-01-30 20:24:09 -08:00
21d96e261f Set subscriptions as default front page 2021-01-30 20:18:57 -08:00
9c58d23b41 Unwrap #subscriptions 2021-01-30 18:16:42 -08:00
4ae2191392 Refactor subscriptions 2021-01-30 18:10:38 -08:00
d62a3ab86b Refactor redirect path setting 2021-01-30 17:50:26 -08:00
9b7cd1da5a Prevent adding duplicate subs. 2021-01-31 12:45:19 +11:00
a301f1ecb6 Simplify redirect path 2021-01-30 17:38:47 -08:00
f14639ee00 Simplify sub list editing. 2021-01-31 12:25:07 +11:00
b527735f6f Fix cookie deletion when removing last sub. 2021-01-31 12:17:43 +11:00
8cc01c58f3 Move the subs list into the navbar. 2021-01-31 09:52:18 +11:00
a1d800a0f0 Fix video loading 2021-01-30 11:00:55 -08:00
449899962a Change subscription to get. Add subs to settings. 2021-01-30 22:27:49 +11:00
dc2030e6f3 Vertical list subscriptions. 2021-01-30 21:21:54 +11:00
ef5a1cd66e Rename block sub_list > subscriptions 2021-01-30 19:47:30 +11:00
11e4ff42ed Clean-up and more consistent styles. 2021-01-30 19:33:38 +11:00
c71df35b22 Add basic unsubscribe. 2021-01-30 18:18:53 +11:00
345308a9ac Basic subscribe functionality. 2021-01-30 18:00:00 +11:00
75bbcefbec Display sub list from list in cookie.
Very basic sub list setup. Cookie must be manually added in devtools.
2021-01-30 16:00:55 +11:00
49a6168607 Improve CSP 2021-01-29 14:39:03 -08:00
f55ea5a353 Specify default headers for security 2021-01-28 21:53:10 -08:00
30c33d91e1 Specify video MIME in posts 2021-01-28 20:06:35 -08:00
00b135fb0f Fix #75 2021-01-28 15:50:18 -08:00
5fe9ce8d7b Refactor JSON parsing error logging 2021-01-27 21:48:32 -08:00
8c04365049 Improve error logging 2021-01-27 17:48:58 -08:00
d5b1c3a5bb Fix connection reset error 2021-01-26 11:00:07 -08:00
f038aa61f4 Log JSON parse errors and Reddit request failures 2021-01-26 10:54:57 -08:00
f72c9d39be Simplify proxy 2021-01-25 17:01:02 -08:00
e6c2d08425 Add libreddit.himiko.cloud instance 2021-01-24 23:33:18 +00:00
e901e99278 Merge pull request #71 from robrobinbin/patch-1
Lower number of cached items
2021-01-24 16:22:58 +00:00
acd2cff747 Lower number of cached items 2021-01-24 11:08:24 +01:00
8f913e696c Merge pull request #70 from robrobinbin/master
Experiment with caching requests to Reddit
2021-01-23 21:50:16 +00:00
226d39328c Make it 60 2021-01-23 22:45:41 +01:00
b2ad2f636c Resolve merge conflicts 2021-01-23 10:55:11 +01:00
18fe7ff8cf Merge pull request #4 from spikecodes/master
Merge upstream
2021-01-23 10:48:54 +01:00
077c222a4e Experiment with caching 2021-01-23 10:48:33 +01:00
2270b6cf95 Reduce post padding 2021-01-21 21:25:51 -08:00
758b627660 Merge pull request #68 from robrobinbin/master
Absolutly no jumping when images (fail to) load
2021-01-21 23:32:59 +00:00
baf7272cfd Absolutly no jumping when images (fail to) load 2021-01-21 22:04:06 +01:00
6641e242af Allow NSFW searching 2021-01-20 18:26:58 -08:00
610fcfbf87 Fix #67 2021-01-20 17:55:04 -08:00
dea7f33910 Add code optimizations 2021-01-20 17:38:34 -08:00
c299e128ab Inline images rework 2021-01-21 00:12:54 +01:00
53fa946c75 Merge pull request #64 from robrobinbin/master
Should fix the strange overflowing placeholder when an image fails to load
2021-01-20 21:46:58 +00:00
5d44a071f9 Why?? 2021-01-20 20:59:57 +01:00
e29e203188 Merge pull request #3 from spikecodes/master
Merge upstream
2021-01-20 20:18:08 +01:00
6ead6e08dc Update README.md 2021-01-20 05:55:59 +00:00
7360503234 Make posts single-color 2021-01-18 21:32:25 -08:00
140c1b1bfa Small fixes 2021-01-18 13:12:59 -08:00
040982f1fd Document command line options in Readme 2021-01-18 18:44:03 +00:00
4b0677d10e Merge pull request #2 from spikecodes/master
Merge upstream
2021-01-18 18:48:23 +01:00
616751e054 Embolden markdown blockquotes 2021-01-17 22:32:42 -08:00
5df957f193 Update README.md 2021-01-18 06:30:34 +00:00
7f9cb1b35a Fix post score and thumbnail backgrounds 2021-01-17 20:35:49 -08:00
c030771d36 Refine transitions 2021-01-17 19:16:15 -08:00
a562395c26 Refactor system theme 2021-01-17 18:11:36 -08:00
2bcdf68e40 Merge pull request #62 from mcrossman/master
Add "system" theme that matches browser/OS theme
2021-01-18 01:59:12 +00:00
72eaa685d0 Prevent "system" class from being added to body. 2021-01-18 11:25:39 +11:00
899a414cf6 Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-18 11:24:49 +11:00
524538eeb8 Update Google Lighthouse Report 2021-01-17 16:22:34 -08:00
a184559c21 Clean CSS 2021-01-17 16:20:17 -08:00
1c9fd46e98 Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-18 11:16:02 +11:00
738941d830 Fix arrow alignment 2021-01-17 16:03:14 -08:00
06ab7a4181 Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-18 10:59:26 +11:00
6981d94417 Clean up theme section of CSS. 2021-01-18 10:56:30 +11:00
dd60cb5b2b SVG arrow 2021-01-17 15:51:03 -08:00
1d57e29d56 Add "system" theme to settings. 2021-01-18 10:44:51 +11:00
2d973707f3 Add auto theme selection in style. 2021-01-18 10:42:55 +11:00
cbb937b494 Merge pull request #61 from robrobinbin/master
Improve accessibility rating
2021-01-17 23:00:25 +00:00
d45ee03122 Resolve merge conflicts 2021-01-17 23:54:48 +01:00
162e00b243 Improve accessibility rating 2021-01-17 23:49:36 +01:00
7a32ba087e Merge pull request #1 from spikecodes/master
Merge upstream
2021-01-17 23:47:09 +01:00
801216dfe9 Include code link on mobile 2021-01-17 13:24:44 -08:00
21763c51cd Make number formatting inclusive 2021-01-17 12:59:40 -08:00
138f8320e9 Create media struct 2021-01-17 12:58:12 -08:00
571ba3392c Merge pull request #60 from robrobinbin/master
Add comment counter and other post improvements
2021-01-17 20:05:01 +00:00
090ca1a140 Add comment counter and other post improvements 2021-01-17 20:39:57 +01:00
6127f2a90c Support forced HTTPS redirects #39 2021-01-16 22:04:03 -08:00
ef9bc791e1 Fix inline style tags 2021-01-16 19:21:47 -08:00
894323becf Update Screenshot 2021-01-16 17:47:45 -08:00
4c89d31948 Fix version number 2021-01-16 17:45:45 -08:00
471d181284 Disable production error logging 2021-01-16 15:17:08 -08:00
0e48c66b8c Fix user agent 2021-01-16 15:13:34 -08:00
a0bc1732cf Moderator and admin distinguishers 2021-01-16 15:02:24 -08:00
6d5fd1dbf6 Use main instance more in Readme 2021-01-16 11:56:13 -08:00
0f6e73dd87 Reformat code 2021-01-16 11:50:12 -08:00
151490faf0 Add space next to comment collapse marker 2021-01-16 11:49:49 -08:00
fdf60e7255 Separate datetime into relative and absolute 2021-01-16 11:40:32 -08:00
ab102ca32c Merge pull request #57 from robrobinbin/master
Improve support for text-only browsers
2021-01-16 10:49:41 -08:00
998b301229 Improve support for text-only browsers 2021-01-16 11:00:15 +01:00
d7839899e6 Merge pull request #3 from spikecodes/master
Merge upstream
2021-01-16 09:20:14 +01:00
2385fa33ec Use ureq until AWC IO error is fixed 2021-01-15 21:26:51 -08:00
1fd688eeed Improve awc error log 2021-01-15 20:57:51 -08:00
65543a43b2 Make User-Agent Reddit-compliant 2021-01-15 20:29:34 -08:00
0099021478 Refactor flair spacing 2021-01-15 15:55:10 -08:00
3a9b2dba32 Fix error log 2021-01-15 15:35:09 -08:00
59021b9331 Switch back to ureq temporarily 2021-01-15 15:28:51 -08:00
078d6fe25b Request about pages before posts 2021-01-15 15:05:55 -08:00
373ce55203 Recommend secondary instance 2021-01-15 11:27:06 -08:00
aef0442e9d Add rate-limit warning 2021-01-15 11:24:12 -08:00
21ff8d7b6f Fix #56 2021-01-15 11:21:59 -08:00
bca2a7e540 Error logging 2021-01-15 10:58:53 -08:00
0c014ad41b Comment utils.rs 2021-01-14 15:13:52 -08:00
32b8637c7e Handle failed redirects 2021-01-14 14:56:28 -08:00
5ed122d92c Merge pull request #55 from robrobinbin/master
Add placeholder image for posts without thumbnail
2021-01-14 14:26:40 -08:00
45660816ce Add cardview to search results too 2021-01-14 21:53:07 +01:00
d19e73f059 Add placeholder for posts without thumbnail 2021-01-14 21:45:43 +01:00
18684c934b Refactor subreddit searching 2021-01-14 11:45:04 -08:00
cf4c5e1fe8 Implement #53 2021-01-14 10:57:50 -08:00
7ef4a20aff Merge pull request #54 from robrobinbin/master
Add subreddits to search results, closes #18
2021-01-14 10:54:20 -08:00
292f8fbbb7 remove lines that aren't used yet 2021-01-14 19:33:17 +01:00
735f79d80b Merge pull request #2 from spikecodes/master
Merge upstream into code
2021-01-14 19:29:06 +01:00
a85a4278f6 Add subreddits to search results 2021-01-14 19:22:50 +01:00
dbe617d7eb Switch to awc 2021-01-14 09:53:54 -08:00
842d97e9fa Fix short post IDs 2021-01-13 21:02:48 -08:00
0bf5576427 Categorize routes and refactor error handlers 2021-01-13 19:53:52 -08:00
dd027bff4b Refactor flair parsing 2021-01-13 18:19:40 -08:00
f95ef51017 Add days to time() 2021-01-13 16:31:24 -08:00
740641cb4e Move nested_val() to user.rs 2021-01-13 15:55:10 -08:00
09c98c8da6 Refactor code 2021-01-13 12:52:00 -08:00
33c8bdffb9 Merge pull request #1 from spikecodes/master
Merge upstream
2021-01-13 20:22:43 +01:00
5ab88567de Merge pull request #50 from robrobinbin/rich-flairs
Add support for rich flairs with "Emoji"
2021-01-13 10:48:25 -08:00
c6627ceece Merge branch 'master' into rich-flairs 2021-01-13 08:27:39 +01:00
d9affcdefc Rich flairs 2021-01-13 08:23:48 +01:00
96607256fc Add Favicon 2021-01-12 20:18:20 -08:00
eb9a0dcb4a Fix GIFs 2021-01-12 19:52:02 -08:00
89fa0d5489 Merge pull request #47 from robrobinbin/scrollbar-for-overflowing-code
Add overflow "auto" to "pre"
2021-01-12 15:16:02 -08:00
22589c8296 Merge pull request #48 from robrobinbin/relative_timestamps
Relative timestamps for posts younger than 24h
2021-01-12 15:15:37 -08:00
b0540d2c57 Rich flairs 2021-01-13 00:10:06 +01:00
41c4661bbb Rich flairs 2021-01-12 23:55:35 +01:00
d2314580a9 Rich flairs 2021-01-12 23:50:50 +01:00
a4d77926b6 Rich flairs 2021-01-12 23:34:16 +01:00
bbe7024323 Start richtext flairs 2021-01-12 22:43:03 +01:00
32e1469e11 whitespace 2021-01-12 20:03:54 +01:00
2d4ca2379f whitespace 2021-01-12 20:02:35 +01:00
374f53eb32 Relative timestamps for recent posts 2021-01-12 19:59:32 +01:00
add7efea3c Update style.css 2021-01-12 18:53:10 +01:00
065d82a5f5 Merge pull request #45 from somoso/long-label-fix
Stop label from being long for joined subreddits
2021-01-12 08:46:35 -08:00
1895bbc025 Merge pull request #46 from somoso/ios-thumbnail-fix
Fix the thumbnail issue on iOS
2021-01-12 08:46:17 -08:00
65f1a2afb2 Stop label from being long for joined subreddits
Browsing with a long joined subreddit list will cause the label to look a bit weird.

Example: https://libredd.it/r/Android+AnimalsBeingBros+AnimalsBeingDerps+AnimalsBeingJerks+AppleWatch+CatSlaps+CatSmiles+CatsBeingAdorable+FreeTube+Games+Ijustwatched+IllegallySmol+IllegallySmolCats+IpodClassic+LearnRubyonRails+Megadrive+MovieDetails+Music+NetflixBestOf+NintendoSwitch+Possums+Teefers+UKPersonalFinance+airplaneears+apple+aww+brushybrushy+cats+catswhotrill+curledfeetsies+cyberpunkgame+dataisbeautiful+dechonkers+digital_ocean+dogs+dogsareliquid+dogswithjobs+emulation+greebles+happycowgifs+hardware+iOSDevelopment+iOSProgramming+iosdev+kittykankles+learnruby+likeus+mac+mashups+microsoft+movies+netflix+netsec+pihole+playstation+programming+rarepuppers+raspberry_pi+redditsync+rubyonrails+satelliteears+shittymoviedetails+spookyteefies+technology+teefies+vampirecats+velvethippos

That will cause the label to be excessively long
2021-01-12 15:47:39 +00:00
6eb9e6f0c0 Fix the thumbnail issue on iOS 2021-01-12 15:43:27 +00:00
eb735a42fe Handle comment parsing errors 2021-01-11 18:05:13 -08:00
541c741bde Parse GIFs correctly 2021-01-11 17:47:14 -08:00
7a33ed3434 Card thumbnails for users 2021-01-11 17:38:35 -08:00
48d2943f72 Fix subreddits not showing sidebars 2021-01-11 16:44:31 -08:00
6bbc90bc0d Clean Subreddit struct 2021-01-11 16:35:50 -08:00
4d18dc0bb8 Merge pull request #44 from robrobinbin/master
Make thumbnail clickable and bring behavior closer to reddit.
2021-01-11 16:35:14 -08:00
6dbd002acd Add direct link to thumbnail 2021-01-11 23:08:12 +01:00
bf6245a505 Fix multireddit sidebars 2021-01-11 10:39:36 -08:00
91746908a1 Switch to ureq 2021-01-11 10:33:48 -08:00
bb8273bab4 Fix #41 2021-01-11 10:33:42 -08:00
62bcc31305 Fix Wide UI on Mobile 2021-01-10 18:48:08 -08:00
08683fa5a6 Light theme 2021-01-10 18:15:34 -08:00
c58b077330 Update Dependencies 2021-01-10 13:20:47 -08:00
f445c42f55 Wide UI Mode 2021-01-10 13:08:36 -08:00
a0866b251e Update issue templates 2021-01-10 11:23:53 -08:00
aa819544f6 Update Matrix Address 2021-01-09 13:08:08 -08:00
fac56d7f87 Markdown spoilers and post footers on videos 2021-01-08 21:57:36 -08:00
ef1ad17234 Unknown path error handling 2021-01-08 21:11:20 -08:00
b8cdc605a2 Front page config and settings note 2021-01-08 20:55:40 -08:00
ef2f9ad12b Unify preferences under one struct 2021-01-08 17:50:03 -08:00
b13874d0db Add "hide nsfw" option 2021-01-08 17:35:04 -08:00
3d142afd03 Merge pull request #38 from somoso/patch-3
Break word to stop it disappearing on mobile
2021-01-08 17:07:50 -08:00
7fcb7fcfed Break word to stop it disappearing on mobile
This kept happening to me but I couldn't reproduce it in the iPad Simulator. Finally got it nailed down and sorted.

Tested this on Safari (mobile and desktop), Firefox, and Edge browser.
2021-01-08 23:08:58 +00:00
747d5a7c67 Merge pull request #37 from somoso/patch-2
Fix theming for all browsers
2021-01-08 12:56:00 -08:00
770c4d3630 Fix themeing for all browsers
Really noticable on iOS, but ensuring all browsers get the love.

The buttons and input aren't as flat as they usually are on my desktop Firefox. This patch should sort that out.
2021-01-08 20:26:29 +00:00
e7b448a282 Add shadow to navbar 2021-01-07 10:49:10 -08:00
c7c787dff1 Fix comment padding 2021-01-07 10:49:00 -08:00
59a34a0e85 Fixed navbar 2021-01-07 10:46:00 -08:00
6e8cf69227 Fix Default Comment Sorting 2021-01-07 10:32:55 -08:00
3444989f9a Default Comment Sort Setting 2021-01-07 08:38:05 -08:00
7e96bb3d80 Optimize use of Result<> 2021-01-06 21:27:24 -08:00
0adbb1556e Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-06 16:45:57 -08:00
710eecdb9d Add issue templates 2021-01-06 16:01:13 -08:00
8a57fa8a1d Remove "Safe" description from README 2021-01-06 14:19:55 -08:00
b33d79ed9b Cache robots.txt 2021-01-06 14:19:10 -08:00
0f506fc41b Cache proxied media 2021-01-06 11:11:04 -08:00
c9cd825d55 Create CSS variables for shadow and text color 2021-01-06 10:51:13 -08:00
e63384e6a6 Update cookie description 2021-01-06 09:54:32 -08:00
3260a4d596 Disable "secure" flag for cookies 2021-01-06 09:52:23 -08:00
da5c4603d9 Switch from chrono to time-rs 2021-01-05 20:01:21 -08:00
b50fa6f3ae Settings Button 2021-01-05 18:16:32 -08:00
aa7b4b2af7 Settings with Layouts 2021-01-05 18:04:49 -08:00
2b0193f5ea Fix proxying of NSFW images 2021-01-05 08:15:34 -08:00
2185d895c0 Prevent user datetimes from floating 2021-01-04 21:32:22 -08:00
9c1a932214 Clean Up Post Headers 2021-01-04 21:17:19 -08:00
8c0269af1c Fix post tags on mobile 2021-01-04 19:43:35 -08:00
df89c5076e Compact Libreddit Posts on Mobile 2021-01-04 19:26:41 -08:00
f819ad2bc6 Remove CSP "Upgrade Insecure Requests" Header 2021-01-04 10:11:07 -08:00
f5884a5270 Update Screenshot 2021-01-03 21:32:45 -08:00
c046d00060 Handle Unwrapping Errors 2021-01-03 21:31:21 -08:00
5934e34ea0 Merge pull request #30 from moosingin3space/master
Add controversial sort order
2021-01-03 21:15:27 -08:00
463b44ac52 Fix timeframe when sorting by controversial 2021-01-04 05:05:21 +00:00
b40d21e559 Add controversial sort order 2021-01-03 21:00:36 -08:00
a422a74747 Make Design More Compact 2021-01-03 19:44:44 -08:00
4124fa87d3 Correct Readme 2021-01-03 18:24:30 -08:00
1dd0c4ee20 Fix User Icon Proxy 2021-01-03 18:23:57 -08:00
0dd114c166 Post upvote ratio, permalink and reddit link 2021-01-03 13:06:49 -08:00
67090e9b08 Fix Proxied Icons 2021-01-03 10:22:41 -08:00
d97fb49fde Fix post::item IDs 2021-01-02 22:46:02 -08:00
9263b0657f Fix navbar padding 2021-01-02 22:40:22 -08:00
a3384cbaa6 Fix search pages 2021-01-02 22:37:54 -08:00
5d26b5c764 Upgrade Insecure Requests 2021-01-02 20:59:14 -08:00
516403ee47 Fix Readme 2021-01-02 20:59:04 -08:00
5ea504e6e8 Restrict Proxy to Reddit Domains 2021-01-02 20:50:23 -08:00
f49bff9853 Optimize Sequencing 2021-01-02 11:09:26 -08:00
4ec529cdb8 Rewrite Reddit Links to Libreddit 2021-01-02 10:58:21 -08:00
779de6f8af Fix Wiki Titles 2021-01-01 22:34:25 -08:00
0925a9b334 Add Wiki Pages 2021-01-01 22:21:43 -08:00
2f2ed6169d Optimize use of .unwrap() 2021-01-01 15:28:13 -08:00
59ef30c76d Remove .clone() in favor of borrowing 2021-01-01 12:55:09 -08:00
d43b49e7e4 Optimize Rust code with Clippy 2021-01-01 12:33:57 -08:00
64a92195dd Merge pull request #19 from somoso/patch-1
Fix posts overflowing on Safari on iOS
2021-01-01 11:52:21 -08:00
a7925ed62d Fix posts overflowing on Safari on iOS
In Safari, the value `anywhere` is not supported for property `overflow-wrap`. Once changed to `break-word`, it behaves like it does in Chrome and Firefox.
2021-01-01 15:46:36 +00:00
39ba50dada Error Page 2020-12-31 21:03:44 -08:00
bc1b29246d Update Screenshot 2020-12-31 20:23:19 -08:00
2d77a91150 Refactor Page Titles and Add Subreddit/User Titles 2020-12-31 20:21:56 -08:00
93c1db502d Fix Title and Navbar 2020-12-31 16:45:10 -08:00
a6dc7ee043 Rewrite + Searching 2020-12-31 15:54:13 -08:00
c7282520cd Add Focus Indicator 2020-12-30 10:53:27 -08:00
a866c1d068 Update Screenshot for v0.2.3 2020-12-29 19:40:49 -08:00
aa9aad6743 Stickied Posts 2020-12-29 19:01:02 -08:00
f65ee2eb6a Sort Top by Timeframe 2020-12-29 17:11:47 -08:00
44c4341e67 Update README.md 2020-12-29 13:29:24 -08:00
1c886f8003 Merge pull request #10 from StuffNoOneCaresAbout/add-instance
instances: add libreddit.kavin.rocks
2020-12-29 09:53:23 -08:00
b481d26be2 instances: add libreddit.kavin.rocks
And it's onion counterpart.
2020-12-29 12:53:49 +05:30
f00ef59404 Fix proxy-less deployment 2020-12-28 20:49:15 -08:00
3115ff3436 Update README.md 2020-12-28 18:45:46 -08:00
443b198c12 Markdown and Subreddit Sidebars 2020-12-28 18:42:46 -08:00
ac84d8d2db List Instances as a Table 2020-12-27 21:46:39 -08:00
e27cf94fbf Include Comment Lines in User History 2020-12-27 15:37:01 -08:00
68495fb280 Add Pages to User Profiles 2020-12-27 12:36:10 -08:00
bec5c78709 Persist Sort on Subreddit Pages 2020-12-26 12:43:51 -08:00
abfcfdf09e Merge pull request #8 from zachjmurphy/master
Add InsanityWtf Libreddit Instance
2020-12-25 18:55:41 -08:00
dad01749e6 Replace Responsive Feature with Secure 2020-12-25 18:06:33 -08:00
2efb73cee3 Update README.md 2020-12-25 17:09:11 -05:00
ace21b21d5 Redesign User/Subreddit About Boxes 2020-12-23 22:16:04 -08:00
280e16bd7f Fix Subreddit Icons 2020-12-23 20:36:49 -08:00
44d44a529c Add DotHQ Libreddit Instance 2020-12-23 13:34:05 -08:00
0957f2e339 NSFW Support 2020-12-22 18:29:43 -08:00
3516404a5f Update v0.2.2 2020-12-22 09:15:55 -08:00
d96daa335f Add Repl.it as Hosting Method 2020-12-22 09:11:03 -08:00
285d9da26d Further Document Libreddit Privacy 2020-12-22 08:45:21 -08:00
9ab7a72bce Fix Comparison Heading Position 2020-12-21 21:53:54 -08:00
46dd905509 Fix Grammar Mistakes 2020-12-21 21:52:50 -08:00
63d595c67d Revise README 2020-12-21 21:51:18 -08:00
dc0b5f42e6 Update README with Reddit Comparison 2020-12-21 21:40:06 -08:00
9ecbd25488 Reorganize CSS 2020-12-21 21:39:55 -08:00
83816fbcc6 Allow Indexing 2020-12-21 21:39:10 -08:00
11cfbdc3ed More Replies Button 2020-12-21 17:17:40 -08:00
4b7cbb3de2 Fix User Icons 2020-12-21 14:12:53 -08:00
b1a572072c Highlight Post Authors in Comments 2020-12-21 08:38:24 -08:00
b1071e9579 Switch Sorting System to Dropdown 2020-12-20 21:49:31 -08:00
da971f8680 Optimized Nested Comments for Mobile, Added IDs 2020-12-20 20:52:15 -08:00
b596f86cc2 Update Screenshot 2020-12-20 19:05:44 -08:00
3bcf0832a1 Correct README Regarding Multireddits 2020-12-20 17:45:52 -08:00
565f4f23b3 Multireddit Support & Referrer Policy 2020-12-20 17:45:26 -08:00
107 changed files with 12681 additions and 2590 deletions

View File

@ -0,0 +1,14 @@
{
"name": "Rust",
"image": "mcr.microsoft.com/devcontainers/rust:0-1-bullseye",
"features": {
"ghcr.io/devcontainers/features/docker-in-docker:2": {}
},
"portsAttributes": {
"8080": {
"label": "redlib",
"onAutoForward": "notify"
}
},
"postCreateCommand": "cargo build"
}

1
.dockerignore Normal file
View File

@ -0,0 +1 @@
target

56
.env.example Normal file
View File

@ -0,0 +1,56 @@
# Redlib configuration
# See the Configuration section of the README for a more detailed explanation of these settings.
# Instance-specific settings
# Enable SFW-only mode for the instance
REDLIB_SFW_ONLY=off
# Set a banner message for the instance
REDLIB_BANNER=
# Disable search engine indexing
REDLIB_ROBOTS_DISABLE_INDEXING=off
# Set the Pushshift frontend for "removed" links
REDLIB_PUSHSHIFT_FRONTEND=undelete.pullpush.io
# Default user settings
# Set the default theme (options: system, light, dark, black, dracula, nord, laserwave, violet, gold, rosebox, gruvboxdark, gruvboxlight)
REDLIB_DEFAULT_THEME=system
# Set the default mascot
REDLIB_DEFAULT_MASCOT=none
# Set the default front page (options: default, popular, all)
REDLIB_DEFAULT_FRONT_PAGE=default
# Set the default layout (options: card, clean, compact)
REDLIB_DEFAULT_LAYOUT=card
# Enable wide mode by default
REDLIB_DEFAULT_WIDE=off
# Set the default post sort method (options: hot, new, top, rising, controversial)
REDLIB_DEFAULT_POST_SORT=hot
# Set the default comment sort method (options: confidence, top, new, controversial, old)
REDLIB_DEFAULT_COMMENT_SORT=confidence
# Enable blurring Spoiler content by default
REDLIB_DEFAULT_BLUR_SPOILER=off
# Enable showing NSFW content by default
REDLIB_DEFAULT_SHOW_NSFW=off
# Enable blurring NSFW content by default
REDLIB_DEFAULT_BLUR_NSFW=off
# Enable HLS video format by default
REDLIB_DEFAULT_USE_HLS=off
# Enable audio+video downloads with ffmpeg.wasm
REDLIB_DEFAULT_FFMPEG_VIDEO_DOWNLOADS=off
# Hide HLS notification by default
REDLIB_DEFAULT_HIDE_HLS_NOTIFICATION=off
# Disable autoplay videos by default
REDLIB_DEFAULT_AUTOPLAY_VIDEOS=off
# Define a default list of subreddit subscriptions (format: sub1+sub2+sub3)
REDLIB_DEFAULT_SUBSCRIPTIONS=
# Define a default list of subreddit filters (format: sub1+sub2+sub3)
REDLIB_DEFAULT_FILTERS=
# Hide awards by default
REDLIB_DEFAULT_HIDE_AWARDS=off
# Hide sidebar and summary
REDLIB_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY=off
# Disable the confirmation before visiting Reddit
REDLIB_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION=off
# Hide score by default
REDLIB_DEFAULT_HIDE_SCORE=off
# Enable fixed navbar by default
REDLIB_DEFAULT_FIXED_NAVBAR=on

1
.envrc Normal file
View File

@ -0,0 +1 @@
use flake

1
.gitattributes vendored Normal file
View File

@ -0,0 +1 @@
Dockerfile.* linguist-language=Dockerfile

3
.github/FUNDING.yml vendored Normal file
View File

@ -0,0 +1,3 @@
liberapay: sigaloid
buy_me_a_coffee: sigaloid
github: sigaloid

41
.github/ISSUE_TEMPLATE/bug_report.md vendored Normal file
View File

@ -0,0 +1,41 @@
---
name: 🐛 Bug report
about: Create a report to help us improve
title: '🐛 Bug Report: '
labels: bug
assignees: ''
---
<!--
BEFORE FILING A BUG REPORT: Ensure that you are running the latest git commit. Visit /info on your instance, and ensure the git commit listed is the same commit listed on the home page.
-->
## Describe the bug
<!--
A clear and concise description of what the bug is.
-->
## Steps to reproduce the bug
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
-->
## What's the expected behavior?
<!--
A clear and concise description of what you expected to happen.
-->
## Additional context / screenshot
<!--
Add any other context about the problem here.
-->
<!-- Mandatory -->
- [ ] I checked that the instance that this was reported on is running the latest git commit, or I can reproduce it locally on the latest git commit

View File

@ -0,0 +1,28 @@
---
name: ✨ Feature parity
about: Suggest implementing a feature into Redlib that is found in Reddit.com
title: '✨ Feature parity: '
labels: feature parity
assignees: ''
---
## How does this feature work on Reddit?
<!--
A clear and concise description of what the feature is.
-->
## Describe how this could be implemented into Redlib
<!--
A clear and concise description of what you want to happen.
-->
## Describe alternatives you've considered
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
## Additional context / screenshot
<!--
Add any other context or screenshots about the feature parity request here.
-->

View File

@ -0,0 +1,28 @@
---
name: 💡 Feature request
about: Suggest a feature for Redlib that is not found in Reddit
title: '💡 Feature request: '
labels: enhancement
assignees: ''
---
## Is your feature request related to a problem? Please describe.
<!--
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
-->
## Describe the feature you would like to be implemented
<!--
A clear and concise description of what you want to happen.
-->
## Describe alternatives you've considered
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
## Additional context / screenshot
<!--
Add any other context or screenshots about the feature request here.
-->

76
.github/workflows/build-artifacts.yaml vendored Normal file
View File

@ -0,0 +1,76 @@
name: Release Build
on:
push:
paths-ignore:
- "*.md"
- "compose.*"
branches:
- "main"
release:
types: [published]
env:
CARGO_TERM_COLOR: always
CARGO_TARGET_AARCH64_UNKNOWN_LINUX_MUSL_LINKER: aarch64-linux-gnu-gcc
CC_aarch64_unknown_linux_musl: aarch64-linux-gnu-gcc
CARGO_TARGET_ARMV7_UNKNOWN_LINUX_MUSLEABIHF_LINKER: arm-linux-gnueabihf-gcc
CC_armv7_unknown_linux_musleabihf: arm-linux-gnueabihf-gcc
jobs:
build:
name: Rust project - latest
runs-on: ubuntu-latest
strategy:
matrix:
target:
- x86_64-unknown-linux-musl
- aarch64-unknown-linux-musl
- armv7-unknown-linux-musleabihf
steps:
- uses: actions/checkout@v4
- uses: actions-rust-lang/setup-rust-toolchain@v1
with:
target: ${{ matrix.target }}
- if: matrix.target == 'x86_64-unknown-linux-musl'
run: |
sudo apt-get update
sudo apt-get install -y --no-install-recommends musl-tools
- if: matrix.target == 'armv7-unknown-linux-musleabihf'
run: |
sudo apt update
sudo apt install -y gcc-arm-linux-gnueabihf musl-tools
- if: matrix.target == 'aarch64-unknown-linux-musl'
run: |
sudo apt update
sudo apt install -y gcc-aarch64-linux-gnu musl-tools
- name: Versions
id: version
run: echo "VERSION=$(cargo metadata --format-version 1 --no-deps | jq .packages[0].version -r | sed 's/^/v/')" >> "$GITHUB_OUTPUT"
- name: Build
run: cargo build --release --target ${{ matrix.target }}
- name: Package release
run: tar czf redlib-${{ matrix.target }}.tar.gz -C target/${{ matrix.target }}/release/ redlib
- name: Upload release
uses: softprops/action-gh-release@v1
with:
tag_name: ${{ steps.version.outputs.VERSION }}
name: ${{ steps.version.outputs.VERSION }} - ${{ github.event.head_commit.message }}
draft: true
files: |
redlib-${{ matrix.target }}.tar.gz
body: |
- ${{ github.event.head_commit.message }} ${{ github.sha }}
generate_release_notes: true

109
.github/workflows/main-docker.yml vendored Normal file
View File

@ -0,0 +1,109 @@
name: Container build
on:
workflow_run:
workflows: ["Release Build"]
types:
- completed
env:
REGISTRY_IMAGE: quay.io/redlib/redlib
jobs:
build:
runs-on: ubuntu-latest
strategy:
fail-fast: false
matrix:
include:
- { platform: linux/amd64, target: x86_64-unknown-linux-musl }
- { platform: linux/arm64, target: aarch64-unknown-linux-musl }
- { platform: linux/arm/v7, target: armv7-unknown-linux-musleabihf }
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Docker meta
id: meta
uses: docker/metadata-action@v5
with:
images: ${{ env.REGISTRY_IMAGE }}
tags: |
type=sha
type=raw,value=latest,enable={{is_default_branch}}
- name: Set up QEMU
uses: docker/setup-qemu-action@v3
- name: Set up Docker Buildx
uses: docker/setup-buildx-action@v3
- name: Login to Quay.io Container Registry
uses: docker/login-action@v3
with:
registry: quay.io
username: ${{ secrets.QUAY_USERNAME }}
password: ${{ secrets.QUAY_ROBOT_TOKEN }}
- name: Build and push
id: build
uses: docker/build-push-action@v5
with:
context: .
platforms: ${{ matrix.platform }}
labels: ${{ steps.meta.outputs.labels }}
outputs: type=image,name=${{ env.REGISTRY_IMAGE }},push-by-digest=true,name-canonical=true,push=true
file: Dockerfile
build-args: TARGET=${{ matrix.target }}
- name: Export digest
run: |
mkdir -p /tmp/digests
digest="${{ steps.build.outputs.digest }}"
touch "/tmp/digests/${digest#sha256:}"
- name: Upload digest
uses: actions/upload-artifact@v4
with:
name: digests-${{ matrix.target }}
path: /tmp/digests/*
if-no-files-found: error
retention-days: 1
merge:
runs-on: ubuntu-latest
needs:
- build
steps:
- name: Download digests
uses: actions/download-artifact@v4.1.7
with:
path: /tmp/digests
pattern: digests-*
merge-multiple: true
- name: Set up Docker Buildx
uses: docker/setup-buildx-action@v3
- name: Docker meta
id: meta
uses: docker/metadata-action@v5
with:
images: ${{ env.REGISTRY_IMAGE }}
tags: |
type=sha
type=raw,value=latest,enable={{is_default_branch}}
- name: Login to Quay.io Container Registry
uses: docker/login-action@v3
with:
registry: quay.io
username: ${{ secrets.QUAY_USERNAME }}
password: ${{ secrets.QUAY_ROBOT_TOKEN }}
- name: Create manifest list and push
working-directory: /tmp/digests
run: |
docker buildx imagetools create $(jq -cr '.tags | map("-t " + .) | join(" ")' <<< "$DOCKER_METADATA_OUTPUT_JSON") \
$(printf '${{ env.REGISTRY_IMAGE }}@sha256:%s ' *)
# - name: Push README to Quay.io
# uses: christian-korneck/update-container-description-action@v1
# env:
# DOCKER_APIKEY: ${{ secrets.APIKEY__QUAY_IO }}
# with:
# destination_container_repo: quay.io/redlib/redlib
# provider: quay
# readme_file: 'README.md'
- name: Inspect image
run: |
docker buildx imagetools inspect ${{ env.REGISTRY_IMAGE }}:${{ steps.meta.outputs.version }}

84
.github/workflows/main-rust.yml vendored Normal file
View File

@ -0,0 +1,84 @@
name: Rust Build & Publish
on:
push:
paths-ignore:
- "**.md"
branches:
- 'main'
release:
types: [published]
env:
CARGO_TERM_COLOR: always
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v3
- name: Cache Packages
uses: Swatinem/rust-cache@v2
- name: Install stable toolchain
uses: dtolnay/rust-toolchain@stable
with:
toolchain: stable
- name: Install musl-gcc
run: sudo apt-get install musl-tools
- name: Install cargo musl target
run: rustup target add x86_64-unknown-linux-musl
# Building actions
- name: Build
run: RUSTFLAGS='-C target-feature=+crt-static' cargo build --release --target x86_64-unknown-linux-musl
- name: Versions
id: version
run: echo "VERSION=$(cargo metadata --format-version 1 --no-deps | jq .packages[0].version -r | sed 's/^/v/')" >> "$GITHUB_OUTPUT"
# Publishing actions
- name: Publish to crates.io
if: github.event_name == 'release'
run: cargo publish --no-verify --token ${{ secrets.CARGO_REGISTRY_TOKEN }}
- name: Calculate SHA512 checksum
run: sha512sum target/x86_64-unknown-linux-musl/release/redlib > redlib.sha512
- name: Calculate SHA256 checksum
run: sha256sum target/x86_64-unknown-linux-musl/release/redlib > redlib.sha256
- uses: actions/upload-artifact@v4
name: Upload a Build Artifact
with:
name: redlib
path: |
target/x86_64-unknown-linux-musl/release/redlib
redlib.sha512
redlib.sha256
- name: Release
uses: softprops/action-gh-release@v1
if: github.base_ref != 'main' && github.event_name == 'release'
with:
tag_name: ${{ steps.version.outputs.VERSION }}
name: ${{ steps.version.outputs.VERSION }} - ${{ github.event.head_commit.message }}
draft: true
files: |
target/x86_64-unknown-linux-musl/release/redlib
redlib.sha512
redlib.sha256
body: |
- ${{ github.event.head_commit.message }} ${{ github.sha }}
generate_release_notes: true
env:
GITHUB_TOKEN: ${{ secrets.RELEASE_TOKEN }}

67
.github/workflows/pull-request.yml vendored Normal file
View File

@ -0,0 +1,67 @@
name: Pull Request
env:
CARGO_TERM_COLOR: always
NEXTEST_RETRIES: 10
on:
push:
branches:
- 'main'
pull_request:
branches:
- 'main'
jobs:
test:
name: cargo test
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v3
- name: Install stable toolchain
uses: dtolnay/rust-toolchain@stable
with:
toolchain: stable
- name: Install cargo-nextest
uses: taiki-e/install-action@nextest
- name: Run cargo nextest
run: cargo nextest run
format:
name: cargo fmt --all -- --check
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v3
- name: Install stable toolchain with rustfmt component
uses: dtolnay/rust-toolchain@stable
with:
toolchain: stable
components: rustfmt
- name: Run cargo fmt
run: cargo fmt --all -- --check
clippy:
name: cargo clippy -- -D warnings
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v3
- name: Install stable toolchain with clippy component
uses: dtolnay/rust-toolchain@stable
with:
toolchain: stable
components: clippy
- name: Run cargo clippy
run: cargo clippy -- -D warnings

View File

@ -1,29 +0,0 @@
name: Rust
on:
push:
branches: [master]
pull_request:
branches: [master]
env:
CARGO_TERM_COLOR: always
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: Cache Packages
uses: Swatinem/rust-cache@v1.0.1
- name: Build
run: cargo build --release
- uses: actions/upload-artifact@v2.2.1
name: Upload a Build Artifact
with:
name: libreddit
path: target/release/libreddit

9
.gitignore vendored
View File

@ -1 +1,10 @@
/target
.env
redlib.toml
# Idea Files
.idea/
# nix files
.direnv/
result

133
CREDITS Normal file
View File

@ -0,0 +1,133 @@
5trongthany <65565784+5trongthany@users.noreply.github.com>
674Y3r <87250374+674Y3r@users.noreply.github.com>
accountForIssues <52367365+accountForIssues@users.noreply.github.com>
Adrian Lebioda <adrianlebioda@gmail.com>
Akanksh Chitimalla <55909985+Akanksh12@users.noreply.github.com>
alefvanoon <53198048+alefvanoon@users.noreply.github.com>
Ales Lerch <13370338+axeII@users.noreply.github.com>
Alexandre Iooss <erdnaxe@crans.org>
alyaeanyx <alexandra.hollmeier@mailbox.org>
AndreVuillemot160 <84594011+AndreVuillemot160@users.noreply.github.com>
Andrew Kaufman <57281817+andrew-kaufman@users.noreply.github.com>
Artemis <51862164+artemislena@users.noreply.github.com>
arthomnix <35371030+arthomnix@users.noreply.github.com>
Arya K <73596856+gi-yt@users.noreply.github.com>
Austin Huang <im@austinhuang.me>
Ayaka <ayaka@kitty.community>
backfire-monism-net <development.0extl@simplelogin.com>
Basti <pred2k@users.noreply.github.com>
Ben Sherman <bennettmsherman@gmail.com>
Ben Smith <37027883+smithbm2316@users.noreply.github.com>
beucismis <beucismis@tutamail.com>
BobIsMyManager <ahoumatt@yahoo.com>
Butter Cat <butteredcats@protonmail.com>
Butter Cat <ButteredCats@protonmail.com>
Carbrex <95964955+Carbrex@users.noreply.github.com>
ccuser44 <68124053+ccuser44@users.noreply.github.com>
Connor Holloway <c.holloway314@outlook.com>
curlpipe <11898833+curlpipe@users.noreply.github.com>
dacousb <53299044+dacousb@users.noreply.github.com>
Daniel Nathan Gray <dng@disroot.org>
Daniel Valentine <Daniel-Valentine@users.noreply.github.com>
Daniel Valentine <daniel@vielle.ws>
dbrennand <52419383+dbrennand@users.noreply.github.com>
Dean Sallinen <deza604@gmail.com>
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Diego Magdaleno <38844659+DiegoMagdaleno@users.noreply.github.com>
domve <domve@posteo.net>
Dyras <jevwmguf@duck.com>
Edward <101938856+EdwardLangdon@users.noreply.github.com>
Éli Marshal <835958+EMarshal@users.noreply.github.com>
elliot <75391956+ellieeet123@users.noreply.github.com>
erdnaxe <erdnaxe@users.noreply.github.com>
Esmail EL BoB <github.defilable@simplelogin.co>
fawn <fawn@envs.net>
FireMasterK <20838718+FireMasterK@users.noreply.github.com>
George Roubos <cowkingdom@hotmail.com>
git-bruh <e817509a-8ee9-4332-b0ad-3a6bdf9ab63f@aleeas.com>
gmnsii <95436780+gmnsii@users.noreply.github.com>
gmnsii <github.gmnsii@pm.me>
gmnsii <gmnsii@void.noreply>
Gonçalo Valério <dethos@users.noreply.github.com>
guaddy <67671414+guaddy@users.noreply.github.com>
Harsh Mishra <erbeusgriffincasper@gmail.com>
hinto.janai <hinto.janai@protonmail.com>
igna <igna@intent.cool>
imabritishcow <bcow@protonmail.com>
invakid404 <invakid404@riseup.net>
İsmail Karslı <ismail@karsli.net>
Johannes Schleifenbaum <johannes@js-webcoding.de>
Jonathan Dahan <git@jonathan.is>
Josiah <70736638+fres7h@users.noreply.github.com>
JPyke3 <pyke.jacob1@gmail.com>
Kavin <20838718+FireMasterK@users.noreply.github.com>
Kazi <kzshantonu@users.noreply.github.com>
Kieran <42723993+EnderDev@users.noreply.github.com>
Kieran <kieran@dothq.co>
Kirk1984 <christoph-m@posteo.de>
kuanhulio <66286575+kuanhulio@users.noreply.github.com>
Kyle Roth <kylrth@gmail.com>
laazyCmd <laazy.pr00gramming@protonmail.com>
Laurențiu Nicola <lnicola@users.noreply.github.com>
Lena <102762572+MarshDeer@users.noreply.github.com>
Leopardus <leopardus3@pm.me>
Macic <46872282+Macic-Dev@users.noreply.github.com>
Mario A <10923513+Midblyte@users.noreply.github.com>
Márton <marton2@gmail.com>
Mathew Davies <ThePixelDeveloper@users.noreply.github.com>
Matthew Crossman <matt@crossman.page>
Matthew E <matt@matthew.science>
Matthew Esposito <matt@matthew.science>
Mennaruuk <52135169+Mennaruuk@users.noreply.github.com>
Midou36O <midou@midou.dev>
mikupls <93015331+mikupls@users.noreply.github.com>
Myzel394 <50424412+Myzel394@users.noreply.github.com>
Nainar <nainar.mb@gmail.com>
Nathan Moos <moosingin3space@gmail.com>
Nazar <63452145+Tokarak@users.noreply.github.com>
Nicholas Christopher <nchristopher@tuta.io>
Nick Lowery <ClockVapor@users.noreply.github.com>
Nico <github@dr460nf1r3.org>
NKIPSC <15067635+NKIPSC@users.noreply.github.com>
nohoster <136514837+nohoster@users.noreply.github.com>
o69mar <119129086+o69mar@users.noreply.github.com>
obeho <71698631+obeho@users.noreply.github.com>
obscurity <z@x4.pm>
Om G <34579088+OxyMagnesium@users.noreply.github.com>
Ondřej Pešek <iTzBoboCz@users.noreply.github.com>
perennial <mail@perennialte.ch>
Peter Sawyer <petersawyer314@gmail.com>
pin <90570748+0323pin@users.noreply.github.com>
potatoesAreGod <118043038+potatoesAreGod@users.noreply.github.com>
RiversideRocks <59586759+RiversideRocks@users.noreply.github.com>
robin <8597693+robrobinbin@users.noreply.github.com>
Robin <8597693+robrobinbin@users.noreply.github.com>
robrobinbin <>
robrobinbin <8597693+robrobinbin@users.noreply.github.com>
robrobinbin <robindepril@gmail.com>
Ruben Elshof <15641671+rubenelshof@users.noreply.github.com>
Rupert Angermeier <rangermeier@users.noreply.github.com>
Scoder12 <34356756+Scoder12@users.noreply.github.com>
Slayer <51095261+GhostSlayer@users.noreply.github.com>
Soheb <somoso@users.noreply.github.com>
somini <somini@users.noreply.github.com>
somoso <github@soheb.anonaddy.com>
Spenser Black <spenserblack01@gmail.com>
Spike <19519553+spikecodes@users.noreply.github.com>
spikecodes <19519553+spikecodes@users.noreply.github.com>
sybenx <syb@duck.com>
TheCultLeader666 <65368815+TheCultLeader666@users.noreply.github.com>
TheFrenchGhosty <47571719+TheFrenchGhosty@users.noreply.github.com>
The TwilightBlood <hwengerstickel@protonmail.com>
tirz <36501933+tirz@users.noreply.github.com>
tmak2002 <torben@tmak2002.dev>
Tokarak <63452145+Tokarak@users.noreply.github.com>
Tsvetomir Bonev <invakid404@riseup.net>
Vivek <vivek@revankar.net>
Vladislav Nepogodin <nepogodin.vlad@gmail.com>
Walkx <walkxnl@gmail.com>
Wichai <1482605+Chengings@users.noreply.github.com>
wsy2220 <wsy@dogben.com>
xatier <xatierlike@gmail.com>
Yaroslav Chvanov <yaroslav.chvanov@gmail.com>
Zach <72994911+zachjmurphy@users.noreply.github.com>

2737
Cargo.lock generated

File diff suppressed because it is too large Load Diff

View File

@ -1,23 +1,58 @@
[package]
name = "libreddit"
name = "redsunlib"
description = " Alternative private front-end to Reddit"
license = "AGPL-3.0"
repository = "https://github.com/spikecodes/libreddit"
version = "0.2.1"
authors = ["spikecodes <19519553+spikecodes@users.noreply.github.com>"]
edition = "2018"
[features]
default = ["proxy"]
proxy = ["actix-web/rustls", "base64"]
license = "AGPL-3.0-only"
repository = "https://git.stardust.wtf/iridium/redsunlib"
version = "0.35.3"
authors = [
"Matthew Esposito <matt+cargo@matthew.science>",
"spikecodes <19519553+spikecodes@users.noreply.github.com>",
]
edition = "2021"
[dependencies]
base64 = { version = "0.13.0", optional = true }
actix-web = "3.2.0"
reqwest = { version = "0.10", default_features = false, features = ["rustls-tls"] }
askama = "0.8.0"
serde = "1.0.117"
serde_json = "1.0"
pulldown-cmark = "0.8.0"
chrono = "0.4.19"
async-recursion = "0.3.1"
rinja = { version = "0.3.4", default-features = false }
cached = { version = "0.51.3", features = ["async"] }
clap = { version = "4.4.11", default-features = false, features = [
"std",
"env",
] }
regex = "1.10.2"
serde = { version = "1.0.193", features = ["derive"] }
cookie = "0.18.0"
futures-lite = "2.2.0"
hyper = { version = "0.14.28", features = ["full"] }
hyper-rustls = "0.24.2"
percent-encoding = "2.3.1"
route-recognizer = "0.3.1"
serde_json = "1.0.108"
tokio = { version = "1.35.1", features = ["full"] }
time = { version = "0.3.31", features = ["local-offset"] }
url = "2.5.0"
rust-embed = { version = "8.1.0", features = ["include-exclude"] }
libflate = "2.0.0"
brotli = { version = "7.0.0", features = ["std"] }
toml = "0.8.8"
once_cell = "1.19.0"
serde_yaml = "0.9.29"
build_html = "2.4.0"
uuid = { version = "1.6.1", features = ["v4"] }
base64 = "0.22.1"
fastrand = "2.0.1"
log = "0.4.20"
pretty_env_logger = "0.5.0"
dotenvy = "0.15.7"
rss = "2.0.7"
arc-swap = "1.7.1"
serde_json_path = "0.6.7"
async-recursion = "1.1.1"
[dev-dependencies]
lipsum = "0.9.0"
sealed_test = "1.0.0"
[profile.release]
codegen-units = 1
lto = true
strip = "symbols"

View File

@ -1,9 +1,34 @@
FROM rust:alpine as builder
WORKDIR /usr/src/libreddit
## Builder
FROM rust:alpine AS builder
RUN apk add --no-cache musl-dev git
WORKDIR /redsunlib
COPY . .
RUN apk add --no-cache g++ openssl-dev
RUN cargo install --path .
RUN cargo build --target x86_64-unknown-linux-musl --release
## Final image
FROM alpine:latest
COPY --from=builder /usr/local/cargo/bin/libreddit /usr/local/bin/libreddit
CMD ["libreddit"]
# Import ca-certificates from builder
COPY --from=builder /usr/share/ca-certificates /usr/share/ca-certificates
COPY --from=builder /etc/ssl/certs /etc/ssl/certs
# Copy our build
COPY --from=builder /redsunlib/target/x86_64-unknown-linux-musl/release/redsunlib /usr/local/bin/redsunlib
# Use an unprivileged user.
RUN adduser --home /nonexistent --no-create-home --disabled-password redsunlib
USER redsunlib
# Tell Docker to expose port 8080
EXPOSE 8080
# Run a healthcheck every minute to make sure redsunlib is functional
HEALTHCHECK --interval=1m --timeout=3s CMD wget --spider -q http://localhost:8080/settings || exit 1
CMD ["redsunlib"]

367
README.md
View File

@ -1,113 +1,342 @@
# Libreddit
<img align="left" width="128" height="128" src="https://git.stardust.wtf/attachments/842086e3-b718-4379-b718-c3a542842152" alt="logo">
> An alternative private front-end to Reddit
# Redsunlib
> An alternative private front-end to Reddit, a fork of [Redlib](https://github.com/redlib-org/redlib) with some <sup><sub>(minor)</sub></sup> function and cosmetic changes.
Libre + Reddit = Libreddit
<br>
- 🚀 Fast: written in Rust for blazing fast speeds and safety
- ☁️ Light: no javascript, no ads, no tracking
- 🕵 Private: all requests are proxied through the server, including media
- 🔒 Safe: does not rely on Reddit OAuth or require a Reddit API Key
- 📱 Responsive: works great on mobile!
![screenshot](https://git.stardust.wtf/attachments/7667e4e2-a32c-4269-9b5f-1d29cb3baf20)
Think Invidious but for Reddit. Watch your cat videos without being watched.
### Disclaimer
## Screenshot
There are rapid changes/features in this fork that can<sup>(will)</sup> change without notice. If you want to host this version, be aware that it's likely to break at some point. I still wouldn't recommend it in a production environment unless you know what you're doing. Or like living on the edge.......
![](https://i.ibb.co/Tgjb3w7/image.png)
> I would also like to thank the maintainers and contributors of both [Redlib](https://github.com/redlib-org/redlib) and [Libreddit](https://github.com/libreddit/libreddit) for all the work they did while I just added some low quality tacky features. ❤️
## About
---
### Elsewhere
Find Libreddit on...
- 💬 Matrix: [#libreddit:matrix.org](https://matrix.to/#/#libreddit:matrix.org)
- 🐋 Docker: [spikecodes/libreddit](https://hub.docker.com/r/spikecodes/libreddit)
- :octocat: GitHub: [spikecodes/libreddit](https://github.com/spikecodes/libreddit)
- 🦊 GitLab: [spikecodes/libreddit](https://gitlab.com/spikecodes/libreddit)
## Table of Contents
1. [Redsunlib](#redsunlib)
- [Disclaimer](#disclaimer)
2. [Table of Contents](#table-of-contents)
3. [Instances](#instances)
4. [About](#about)
- [The Name](#the-name)
- [Built with](#built-with)
- [How is it different from other Reddit front ends?](#how-is-it-different-from-other-reddit-front-ends)
- [Teddit](#teddit)
- [Libreddit](#libreddit)
5. [Comparison](#comparison)
- [Speed](#speed)
- [Privacy](#privacy)
- [Reddit](#reddit)
- [Redlib](#redlib-1)
- [Server](#server)
6. [Deployment](#deployment)
- [Docker](#docker)
- [Docker Compose](#docker-compose)
- [Docker CLI](#docker-cli)
- [Binary](#binary)
- [Running as a systemd service](#running-as-a-systemd-service)
- [Building from source](#building-from-source)
7. [Configuration](#configuration)
- [Instance settings](#instance-settings)
- [Default user settings](#default-user-settings)
### Info
Libreddit hopes to provide an easier way to browse Reddit, without the ads, trackers and bloat. Libreddit was inspired by other alternative front-ends to popular services such as [Invidious](https://github.com/iv-org/invidious) for YouTube, [Nitter](https://github.com/zedeus/nitter) for Twitter, and [Bibliogram](https://sr.ht/~cadence/bibliogram/) for Instagram.
---
Libreddit currently implements most of Reddit's functionalities but still lacks a few features that are being worked on below.
# Instances
### In Progress
- Searching
- Multireddits
> [!WARNING]
> 🔗 **Currently public Redsunlib instance are not available, consider using a [redlib](https://github.com/redlib-org/redlib-instances/blob/main/instances.md) instance if you are not comfortable running your own**
### How does it compare to Teddit?
You are more than welcome to host an instance and submit an issue if you want it added. That is, if you've read the [Disclaimer](#disclaimer) and it's within your "personal risk tolerance." ;)
Teddit is another awesome open source project designed to provide an alternative frontend to Reddit. There is no connection between the two and you're welcome to use whichever one you favor. Competition fosters innovation and Teddit's release has motivated me to build Libreddit into an even more polished product.
---
# About
Redlib hopes to provide an easier way to browse Reddit, without the ads, trackers, and bloat. Redlib was inspired by other alternative front-ends to popular services such as [Invidious](https://github.com/iv-org/invidious) for YouTube, [Nitter](https://github.com/zedeus/nitter) for Twitter, and [Bibliogram](https://sr.ht/~cadence/bibliogram/) for Instagram.
Redlib currently implements most of Reddit's (signed-out) functionalities but still lacks [a few features](https://github.com/redlib-org/redlib/issues).
## The Name
**Red sun** in the sky + Red**lib** = Redsunlib
And at the time, I was reading an excerpt from Mao Zedong, so the name seemed appropriate. But paradoxically named since Reddit is basically the sinophobia capital of the internet :/
## Built with
- [Rust](https://www.rust-lang.org/) - Programming language
- [Hyper](https://github.com/hyperium/hyper) - HTTP server and client
- [Rinja](https://github.com/rinja-rs/rinja) - Templating engine
- [Rustls](https://github.com/rustls/rustls) - TLS library
## How is it different from other Reddit front ends?
### Teddit
Teddit is another awesome open source project designed to provide an alternative frontend to Reddit. There is no connection between the two, and you're welcome to use whichever one you favor. Competition fosters innovation and Teddit's release has motivated me to build Redlib into an even more polished product.
If you are looking to compare, the biggest differences I have noticed are:
- Libreddit is themed around Reddit's redesign whereas Teddit appears to stick much closer to Reddit's old design. This may suit some users better as design is always subjective.
- Libreddit is written in Rust for speed and memory safety. It uses Actix Web, which was [benchmarked as the fastest web server for single queries](https://www.techempower.com/benchmarks/#hw=ph&test=db).
- Unlike Teddit (at the time of writing this), Libreddit does not require a Reddit API key to host.
## Instances
- Redlib is themed around Reddit's redesign whereas Teddit appears to stick much closer to Reddit's old design. This may suit some users better as design is always subjective.
- Redlib is written in [Rust](https://www.rust-lang.org) for speed and memory safety. It uses [Hyper](https://hyper.rs), a speedy and lightweight HTTP server/client implementation.
Feel free to [open an issue](https://github.com/spikecodes/libreddit/issues/new) to have your selfhosted instance listed here!
### Libreddit
- [libredd.it](https://libredd.it) 🇺🇸 (Thank you to [YeapGuy](https://github.com/YeapGuy)!)
- [libreddit.spike.codes](https://libreddit.spike.codes) 🇺🇸
While originating as a fork of Libreddit, the name "Redlib" was adopted to avoid legal issues, as Reddit only allows the use of their name if structured as "XYZ For Reddit".
## Installation
Several technical improvements have also been made, including:
### A) Cargo
- **OAuth token spoofing**: To circumvent rate limits imposed by Reddit, OAuth token spoofing is used to mimick the most common iOS and Android clients. While spoofing both iOS and Android clients was explored, only the Android client was chosen due to content restrictions when using an anonymous iOS client.
- **Token refreshing**: The authentication token is refreshed every 24 hours, emulating the behavior of the official Android app.
- **HTTP header mimicking**: Efforts are made to send along as many of the official app's headers as possible to reduce the likelihood of Reddit's crackdown on Redlib's requests.
Make sure Rust stable is installed along with `cargo`, Rust's package manager.
---
```
cargo install libreddit
# Comparison
This section outlines how Redlib compares to Reddit in terms of speed and privacy.
## Speed
Last tested on January 12, 2024.
Results from Google PageSpeed Insights ([Redlib Report](https://pagespeed.web.dev/report?url=https%3A%2F%2Fredlib.matthew.science%2F), [Reddit Report](https://pagespeed.web.dev/report?url=https://www.reddit.com)).
| Performance metric | Redlib | Reddit |
| ------------------- | -------- | --------- |
| Speed Index | 0.6s | 1.9s |
| Performance Score | 100% | 64% |
| Time to Interactive | **2.8s** | **12.4s** |
## Privacy
### Reddit
**Logging:** According to Reddit's [privacy policy](https://www.redditinc.com/policies/privacy-policy), they "may [automatically] log information" including:
- IP address
- User-agent string
- Browser type
- Operating system
- Referral URLs
- Device information (e.g., device IDs)
- Device settings
- Pages visited
- Links clicked
- The requested URL
- Search terms
**Location:** The same privacy policy goes on to describe that location data may be collected through the use of:
- GPS (consensual)
- Bluetooth (consensual)
- Content associated with a location (consensual)
- Your IP Address
**Cookies:** Reddit's [cookie notice](https://www.redditinc.com/policies/cookies) documents the array of cookies used by Reddit including/regarding:
- Authentication
- Functionality
- Analytics and Performance
- Advertising
- Third-Party Cookies
- Third-Party Site
### Redlib
For transparency, I hope to describe all the ways Redlib handles user privacy.
#### Server
- **Logging:** In production (when running the binary, hosting with docker, or using the official instances), Redlib logs nothing. When debugging (running from source without `--release`), Redlib logs post IDs fetched to aid with troubleshooting.
- **Cookies:** Redlib uses optional cookies to store any configured settings in [the settings menu](https://redlib.matthew.science/settings). These are not cross-site cookies and the cookies hold no personal data.
---
# Deployment
This section covers multiple ways of deploying Redlib. Using [Docker](#docker) is recommended for production.
For configuration options, see the [Configuration section](#Configuration).
## Docker
[Docker](https://www.docker.com) lets you run containerized applications. Containers are loosely isolated environments that are lightweight and contain everything needed to run the application, so there's no need to rely on what's installed on the host.
Docker images for Redsunlib are available at our [Gitea container registry](https://git.stardust.wtf/iridium/-/packages/container/redsunlib/latest), currently only with support for `amd64`, if you need `arm64`, or `armv7` platforms you can either build Redsunlib yourself or open an [issue](https://git.stardust.wtf/iridium/redsunlib/issues) :)
### Docker Compose
> [!IMPORTANT]
> These instructions assume the [Compose plugin](https://docs.docker.com/compose/migrate/#what-are-the-differences-between-compose-v1-and-compose-v2) has already been installed. If not, follow these [instructions on the Docker Docs](https://docs.docker.com/compose/install) for how to do so.
Copy `compose.yaml` and modify any relevant values (for example, the ports Redlib should listen on).
Start Redlib in detached mode (running in the background):
```bash
docker compose up -d
```
### B) Docker
Stream logs from the Redlib container:
Deploy the Docker image of Libreddit:
```
docker run -d --name libreddit -p 8080:8080 spikecodes/libreddit
```bash
docker logs -f redlib
```
Deploy using a different port (in this case, port 80):
```
docker run -d --name libreddit -p 80:8080 spikecodes/libreddit
### Docker CLI
Deploy Redlib:
```bash
docker pull git.stardust.wtf/iridium/redsunlib:latest
docker run -d --name redlib -p 8080:8080 git.stardust.wtf/iridium/redsunlib:latest
```
### C) AUR
Deploy using a different port on the host (in this case, port 80):
For ArchLinux users, Libreddit is available from the AUR as [`libreddit-git`](https://aur.archlinux.org/packages/libreddit-git).
Install:
```
yay -S libreddit-git
```bash
docker pull git.stardust.wtf/iridium/redsunlib:latest
docker run -d --name redlib -p 80:8080 git.stardust.wtf/iridium/redsunlib:latest
```
### D) GitHub Releases
If you're using a reverse proxy in front of Redlib, prefix the port numbers with `127.0.0.1` so that Redlib only listens on the host port **locally**. For example, if the host port for Redlib is `8080`, specify `127.0.0.1:8080:8080`.
If you're on Linux and none of these methods work for you, you can grab a Linux binary from [the newest release](https://github.com/spikecodes/libreddit/releases/latest).
Currently Libreddit does not have Windows or MacOS binaries but those will be available soon.
Stream logs from the Redlib container:
## Deploy an Instance
Once installed, deploy Libreddit (unless you're using Docker) by running:
```
libreddit
```bash
docker logs -f redlib
```
Specify a custom address for the server by passing the `-a` or `--address` argument:
```
libreddit --address=0.0.0.0:8111
## Binary
Currently binaries are not supplied at this moment but will be at some point in the future but can be [built from source](#building-from-source)
Copy the binary to `/usr/bin`:
```bash
sudo cp ./redlib /usr/bin/redlib
```
To disable the media proxy built into Libreddit, run:
```
libreddit --no-default-features
Deploy Redlib to `0.0.0.0:8080`:
```bash
redlib
```
## Building from Source
> [!IMPORTANT]
> If you're proxying Redlib through NGINX (see [issue #122](https://github.com/libreddit/libreddit/issues/122#issuecomment-782226853)), add
>
> ```nginx
> proxy_http_version 1.1;
> ```
>
> to your NGINX configuration file above your `proxy_pass` line.
### Running as a systemd service
You can use the systemd service available in `contrib/redlib.service`
(install it on `/etc/systemd/system/redlib.service`).
That service can be optionally configured in terms of environment variables by
creating a file in `/etc/redlib.conf`. Use the `contrib/redlib.conf` as a
template. You can also add the `REDLIB_DEFAULT__{X}` settings explained
above.
When "Proxying using NGINX" where the proxy is on the same machine, you should
guarantee nginx waits for this service to start. Edit
`/etc/systemd/system/redlib.service.d/reverse-proxy.conf`:
```conf
[Unit]
Before=nginx.service
```
git clone https://github.com/spikecodes/libreddit
cd libreddit
## Building from source
To deploy Redsunlib with changes not yet included in the latest release, you can build the application from source.
```bash
git clone https://git.stardust.wtf/iridium/redsunlib && cd redsunlib
cargo run
```
---
# Configuration
You can configure Redlib further using environment variables. For example:
```bash
REDLIB_DEFAULT_SHOW_NSFW=on redlib
```
```bash
REDLIB_DEFAULT_WIDE=on REDLIB_DEFAULT_THEME=dark redlib -r
```
You can also configure Redlib with a configuration file named `redlib.toml`. For example:
```toml
REDLIB_DEFAULT_WIDE = "on"
REDLIB_DEFAULT_USE_HLS = "on"
```
> [!NOTE]
> If you're deploying Redlib using the **Docker CLI or Docker Compose**, environment variables can be defined in a [`.env` file](https://docs.docker.com/compose/environment-variables/set-environment-variables/), allowing you to centralize and manage configuration in one place.
>
> To configure Redlib using a `.env` file, copy the `.env.example` file to `.env` and edit it accordingly.
>
> If using the Docker CLI, add ` --env-file .env` to the command that runs Redlib. For example:
>
> ```bash
> docker run -d --name redlib -p 8080:8080 --env-file .env git.stardust.wtf/iridium/redsunlib:latest
> ```
>
> If using Docker Compose, no changes are needed as the `.env` file is already referenced in `compose.yaml` via the `env_file: .env` line.
## Instance settings
Assign a default value for each instance-specific setting by passing environment variables to Redlib in the format `REDLIB_{X}`. Replace `{X}` with the setting name (see list below) in capital letters.
| Name | Possible values | Default value | Description |
| ------------------------- | --------------- | ---------------- | --------------------------------------------------------------------------------------------------------- |
| `SFW_ONLY` | `["on", "off"]` | `off` | Enables SFW-only mode for the instance, i.e. all NSFW content is filtered. |
| `BANNER` | String | (empty) | Allows the server to set a banner to be displayed. Currently this is displayed on the instance info page. |
| `ROBOTS_DISABLE_INDEXING` | `["on", "off"]` | `off` | Disables indexing of the instance by search engines. |
| `PUSHSHIFT_FRONTEND` | String | `undelete.pullpush.io` | Allows the server to set the Pushshift frontend to be used with "removed" links. |
| `PORT` | Integer 0-65535 | `8080` | The **internal** port Redlib listens on. |
| `ENABLE_RSS` | `["on", "off"]` | `off` | Enables RSS feed generation. |
| `FULL_URL` | String | (empty) | Allows for proper URLs (for now, only needed by RSS)
## Default user settings
Assign a default value for each user-modifiable setting by passing environment variables to Redlib in the format `REDLIB_DEFAULT_{Y}`. Replace `{Y}` with the setting name (see list below) in capital letters.
| Name | Possible values | Default value |
| ----------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------- | ------------- |
| `THEME` | `["system", "light", "dark", "black", "dracula", "nord", "laserwave", "violet", "gold", "rosebox", "gruvboxdark", "gruvboxlight", "tokyoNight", "catppuccin", "icebergDark", "doomone", "libredditBlack", "libredditDark", "libredditLight"]` | `system` |
| `MASCOT` | `["BoymoderBlahaj", "redsunlib" ... Add more at ./static/mascots] ` | _(none)_ |
| `FRONT_PAGE` | `["default", "popular", "all"]` | `default` |
| `LAYOUT` | `["card", "clean", "compact", "old", "waterfall"]` | `card` |
| `WIDE` | `["on", "off"]` | `off` |
| `POST_SORT` | `["hot", "new", "top", "rising", "controversial"]` | `hot` |
| `COMMENT_SORT` | `["confidence", "top", "new", "controversial", "old"]` | `confidence` |
| `BLUR_SPOILER` | `["on", "off"]` | `off` |
| `SHOW_NSFW` | `["on", "off"]` | `off` |
| `BLUR_NSFW` | `["on", "off"]` | `off` |
| `USE_HLS` | `["on", "off"]` | `off` |
| `FFMPEG_VIDEO_DOWNLOADS` | `["on", "off"]` | `off` |
| `HIDE_HLS_NOTIFICATION` | `["on", "off"]` | `off` |
| `AUTOPLAY_VIDEOS` | `["on", "off"]` | `off` |
| `SUBSCRIPTIONS` | `+`-delimited list of subreddits (`sub1+sub2+sub3+...`) | _(none)_ |
| `HIDE_AWARDS` | `["on", "off"]` | `off` |
| `DISABLE_VISIT_REDDIT_CONFIRMATION` | `["on", "off"]` | `off` |
| `HIDE_SCORE` | `["on", "off"]` | `off` |
| `HIDE_SIDEBAR_AND_SUMMARY` | `["on", "off"]` | `off` |
| `FIXED_NAVBAR` | `["on", "off"]` | `on` |

87
app.json Normal file
View File

@ -0,0 +1,87 @@
{
"name": "Redsunlib",
"description": "Private front-end for Reddit",
"buildpacks": [
{
"url": "https://github.com/emk/heroku-buildpack-rust"
},
{
"url": "emk/rust"
}
],
"stack": "container",
"env": {
"REDLIB_DEFAULT_THEME": {
"required": false
},
"REDLIB_DEFAULT_MASCOT": {
"required": false
},
"REDLIB_DEFAULT_FRONT_PAGE": {
"required": false
},
"REDLIB_DEFAULT_LAYOUT": {
"required": false
},
"REDLIB_DEFAULT_WIDE": {
"required": false
},
"REDLIB_DEFAULT_COMMENT_SORT": {
"required": false
},
"REDLIB_DEFAULT_POST_SORT": {
"required": false
},
"REDLIB_DEFAULT_BLUR_SPOILER": {
"required": false
},
"REDLIB_DEFAULT_SHOW_NSFW": {
"required": false
},
"REDLIB_DEFAULT_BLUR_NSFW": {
"required": false
},
"REDLIB_DEFAULT_USE_HLS": {
"required": false
},
"REDLIB_DEFAULT_FFMPEG_VIDEO_DOWNLOADS": {
"required": false
},
"REDLIB_HIDE_HLS_NOTIFICATION": {
"required": false
},
"REDLIB_SFW_ONLY": {
"required": false
},
"REDLIB_DEFAULT_HIDE_AWARDS": {
"required": false
},
"REDLIB_DEFAULT_HIDE_SCORE": {
"required": false
},
"REDLIB_BANNER": {
"required": false
},
"REDLIB_ROBOTS_DISABLE_INDEXING": {
"required": false
},
"REDLIB_DEFAULT_SUBSCRIPTIONS": {
"required": false
},
"REDLIB_DEFAULT_FILTERS": {
"required": false
},
"REDLIB_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION": {
"required": false
},
"REDLIB_PUSHSHIFT_FRONTEND": {
"required": false
},
"REDLIB_ENABLE_RSS": {
"required": false
},
"REDLIB_FULL_URL": {
"required": false
}
}
}

25
build.rs Normal file
View File

@ -0,0 +1,25 @@
use std::process::{Command, ExitStatus, Output};
#[cfg(not(target_os = "windows"))]
use std::os::unix::process::ExitStatusExt;
#[cfg(target_os = "windows")]
use std::os::windows::process::ExitStatusExt;
fn main() {
println!("cargo:rerun-if-changed=src/");
let output = String::from_utf8(
Command::new("git")
.args(["rev-parse", "HEAD"])
.output()
.unwrap_or(Output {
stdout: vec![],
stderr: vec![],
status: ExitStatus::from_raw(0),
})
.stdout,
)
.unwrap_or_default();
let git_hash = if output == String::default() { "dev".into() } else { output };
println!("cargo:rustc-env=GIT_HASH={git_hash}");
}

26
compose.dev.yaml Normal file
View File

@ -0,0 +1,26 @@
# docker-compose -f docker-compose.dev.yml up -d
version: "3.8"
services:
redsunlib:
build: .
restart: always
container_name: "redsunlib"
ports:
- 8080:8080 # Specify `127.0.0.1:8080:8080` instead if using a reverse proxy
user: nobody
read_only: true
security_opt:
- no-new-privileges:true
# - seccomp=seccomp-redsunlib.json
cap_drop:
- ALL
networks:
- redsunlib
healthcheck:
test: ["CMD", "wget", "--spider", "-q", "--tries=1", "http://localhost:8080/settings"]
interval: 5m
timeout: 3s
networks:
redsunlib:

24
compose.yaml Normal file
View File

@ -0,0 +1,24 @@
services:
redsunlib:
image: git.stardust.wtf/iridium/redsunlib:latest
restart: always
container_name: "redsunlib"
ports:
- 8080:8080 # Specify `127.0.0.1:8080:8080` instead if using a reverse proxy
user: nobody
read_only: true
security_opt:
- no-new-privileges:true
# - seccomp=seccomp-redsunlib.json
cap_drop:
- ALL
env_file: .env
networks:
- redsunlib
healthcheck:
test: ["CMD", "wget", "--spider", "-q", "--tries=1", "http://localhost:8080/settings"]
interval: 5m
timeout: 3s
networks:
redsunlib:

19
contrib/redlib.conf Normal file
View File

@ -0,0 +1,19 @@
ADDRESS=0.0.0.0
PORT=12345
#REDLIB_DEFAULT_THEME=default
#REDLIB_DEFAULT_MASCOT=none
#REDLIB_DEFAULT_FRONT_PAGE=default
#REDLIB_DEFAULT_LAYOUT=card
#REDLIB_DEFAULT_WIDE=off
#REDLIB_DEFAULT_POST_SORT=hot
#REDLIB_DEFAULT_COMMENT_SORT=confidence
#REDLIB_DEFAULT_BLUR_SPOILER=off
#REDLIB_DEFAULT_SHOW_NSFW=off
#REDLIB_DEFAULT_BLUR_NSFW=off
#REDLIB_DEFAULT_USE_HLS=off
#REDLIB_DEFAULT_FFMPEG_VIDEO_DOWNLOADS=off
#REDLIB_DEFAULT_HIDE_HLS_NOTIFICATION=off
#REDLIB_DEFAULT_AUTOPLAY_VIDEOS=off
#REDLIB_DEFAULT_SUBSCRIPTIONS=(sub1+sub2+sub3)
#REDLIB_DEFAULT_HIDE_AWARDS=off
#REDLIB_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION=off

19
contrib/redlib.plist Normal file
View File

@ -0,0 +1,19 @@
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key>
<string>redlib</string>
<key>Program</key>
<string>redlib</string>
<key>KeepAlive</key>
<true/>
<key>RunAtLoad</key>
<true/>
</dict>
</plist>

38
contrib/redlib.service Normal file
View File

@ -0,0 +1,38 @@
[Unit]
Description=redlib daemon
After=network.service
[Service]
DynamicUser=yes
# Default Values
#Environment=ADDRESS=0.0.0.0
#Environment=PORT=8080
# Optional Override
EnvironmentFile=-/etc/redlib.conf
ExecStart=/usr/bin/redlib -a ${ADDRESS} -p ${PORT}
# Hardening
DeviceAllow=
LockPersonality=yes
MemoryDenyWriteExecute=yes
PrivateDevices=yes
ProcSubset=pid
ProtectClock=yes
ProtectControlGroups=yes
ProtectHome=yes
ProtectHostname=yes
ProtectKernelLogs=yes
ProtectKernelModules=yes
ProtectKernelTunables=yes
ProtectProc=invisible
RestrictAddressFamilies=AF_INET AF_INET6
RestrictNamespaces=yes
RestrictRealtime=yes
RestrictSUIDSGID=yes
SystemCallArchitectures=native
SystemCallFilter=@system-service
SystemCallFilter=~@privileged @resources
UMask=0077
[Install]
WantedBy=default.target

106
flake.lock generated Normal file
View File

@ -0,0 +1,106 @@
{
"nodes": {
"crane": {
"inputs": {
"nixpkgs": [
"nixpkgs"
]
},
"locked": {
"lastModified": 1717025063,
"narHash": "sha256-dIubLa56W9sNNz0e8jGxrX3CAkPXsq7snuFA/Ie6dn8=",
"owner": "ipetkov",
"repo": "crane",
"rev": "480dff0be03dac0e51a8dfc26e882b0d123a450e",
"type": "github"
},
"original": {
"owner": "ipetkov",
"repo": "crane",
"type": "github"
}
},
"flake-utils": {
"inputs": {
"systems": "systems"
},
"locked": {
"lastModified": 1710146030,
"narHash": "sha256-SZ5L6eA7HJ/nmkzGG7/ISclqe6oZdOZTNoesiInkXPQ=",
"owner": "numtide",
"repo": "flake-utils",
"rev": "b1d9ab70662946ef0850d488da1c9019f3a9752a",
"type": "github"
},
"original": {
"owner": "numtide",
"repo": "flake-utils",
"type": "github"
}
},
"nixpkgs": {
"locked": {
"lastModified": 1717112898,
"narHash": "sha256-7R2ZvOnvd9h8fDd65p0JnB7wXfUvreox3xFdYWd1BnY=",
"owner": "NixOS",
"repo": "nixpkgs",
"rev": "6132b0f6e344ce2fe34fc051b72fb46e34f668e0",
"type": "github"
},
"original": {
"owner": "NixOS",
"ref": "nixpkgs-unstable",
"repo": "nixpkgs",
"type": "github"
}
},
"root": {
"inputs": {
"crane": "crane",
"flake-utils": "flake-utils",
"nixpkgs": "nixpkgs",
"rust-overlay": "rust-overlay"
}
},
"rust-overlay": {
"inputs": {
"flake-utils": [
"flake-utils"
],
"nixpkgs": [
"nixpkgs"
]
},
"locked": {
"lastModified": 1717121863,
"narHash": "sha256-/3sxIe7MZqF/jw1RTQCSmgTjwVod43mmrk84m50MJQ4=",
"owner": "oxalica",
"repo": "rust-overlay",
"rev": "2a7b53172ed08f856b8382d7dcfd36a4e0cbd866",
"type": "github"
},
"original": {
"owner": "oxalica",
"repo": "rust-overlay",
"type": "github"
}
},
"systems": {
"locked": {
"lastModified": 1681028828,
"narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=",
"owner": "nix-systems",
"repo": "default",
"rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e",
"type": "github"
},
"original": {
"owner": "nix-systems",
"repo": "default",
"type": "github"
}
}
},
"root": "root",
"version": 7
}

71
flake.nix Normal file
View File

@ -0,0 +1,71 @@
{
description = "Redlib: Private front-end for Reddit";
inputs = {
nixpkgs.url = "github:NixOS/nixpkgs/nixpkgs-unstable";
crane = {
url = "github:ipetkov/crane";
inputs.nixpkgs.follows = "nixpkgs";
};
flake-utils.url = "github:numtide/flake-utils";
rust-overlay = {
url = "github:oxalica/rust-overlay";
inputs = {
nixpkgs.follows = "nixpkgs";
flake-utils.follows = "flake-utils";
};
};
};
outputs = { nixpkgs, crane, flake-utils, rust-overlay, ... }:
flake-utils.lib.eachSystem [ "x86_64-linux" ] (system:
let
pkgs = import nixpkgs {
inherit system;
overlays = [ (import rust-overlay) ];
};
inherit (pkgs) lib;
rustToolchain = pkgs.rust-bin.stable.latest.default.override {
targets = [ "x86_64-unknown-linux-musl" ];
};
craneLib = (crane.mkLib pkgs).overrideToolchain rustToolchain;
src = lib.cleanSourceWith {
src = craneLib.path ./.;
filter = path: type:
(lib.hasInfix "/templates/" path) ||
(lib.hasInfix "/static/" path) ||
(craneLib.filterCargoSources path type);
};
redlib = craneLib.buildPackage {
inherit src;
strictDeps = true;
doCheck = false;
CARGO_BUILD_TARGET = "x86_64-unknown-linux-musl";
CARGO_BUILD_RUSTFLAGS = "-C target-feature=+crt-static";
};
in
{
checks = {
my-crate = redlib;
};
packages.default = redlib;
packages.docker = pkgs.dockerTools.buildImage {
name = "quay.io/redlib/redlib";
tag = "latest";
created = "now";
copyToRoot = with pkgs.dockerTools; [ caCertificates fakeNss ];
config.Cmd = "${redlib}/bin/redlib";
};
});
}

3
heroku.yml Normal file
View File

@ -0,0 +1,3 @@
build:
docker:
web: Dockerfile

15
scripts/gen-credits.sh Executable file
View File

@ -0,0 +1,15 @@
#!/usr/bin/env bash
# This scripts generates the CREDITS file in the repository root, which
# contains a list of all contributors ot the Redlib project.
#
# We use git-log to surface the names and emails of all authors and committers,
# and grep will filter any automated commits due to GitHub.
set -o pipefail
cd "$(dirname "${BASH_SOURCE[0]}")/../" || exit 1
git --no-pager log --pretty='%an <%ae>%n%cn <%ce>' main \
| sort -t'<' -u -k1,1 -k2,2 \
| grep -Fv -- 'GitHub <noreply@github.com>' \
> CREDITS

31
scripts/load_test.py Normal file
View File

@ -0,0 +1,31 @@
import requests
from bs4 import BeautifulSoup
from concurrent.futures import ThreadPoolExecutor
base_url = "http://localhost:8080"
full_path = f"{base_url}/r/politics"
ctr = 0
def fetch_url(url):
global ctr
response = requests.get(url)
ctr += 1
print(f"Request count: {ctr}")
return response
while full_path:
response = requests.get(full_path)
ctr += 1
print(f"Request count: {ctr}")
soup = BeautifulSoup(response.text, 'html.parser')
comment_links = soup.find_all('a', class_='post_comments')
comment_urls = [base_url + link['href'] for link in comment_links]
with ThreadPoolExecutor(max_workers=10) as executor:
executor.map(fetch_url, comment_urls)
next_link = soup.find('a', accesskey='N')
if next_link:
full_path = base_url + next_link['href']
else:
break

18
scripts/update_hls_js.sh Executable file
View File

@ -0,0 +1,18 @@
#!/bin/bash
cd "$(dirname "$0")"
LATEST_TAG=$(curl -s https://api.github.com/repos/video-dev/hls.js/releases/latest | jq -r '.tag_name')
if [[ -z "$LATEST_TAG" || "$LATEST_TAG" == "null" ]]; then
echo "Failed to fetch the latest release tag from GitHub."
exit 1
fi
LICENSE="// @license http://www.apache.org/licenses/LICENSE-2.0 Apache-2.0
// @source https://github.com/video-dev/hls.js/tree/$LATEST_TAG"
echo "$LICENSE" > ../static/hls.min.js
curl -s https://cdn.jsdelivr.net/npm/hls.js@${LATEST_TAG}/dist/hls.min.js >> ../static/hls.min.js
echo "Update complete. The latest hls.js (${LATEST_TAG}) has been saved to static/hls.min.js."

112
scripts/update_oauth_resources.sh Executable file
View File

@ -0,0 +1,112 @@
#!/bin/bash
# Requirements
# - curl
# - rg
# - jq
# Fetch iOS app versions
ios_version_list=$(curl -s "https://ipaarchive.com/app/usa/1064216828" | rg "(20\d{2}\.\d+.\d+) / (\d+)" --only-matching -r "Version \$1/Build \$2" | sort | uniq)
# Count the number of lines in the version list
ios_app_count=$(echo "$ios_version_list" | wc -l)
echo -e "Fetching \e[34m$ios_app_count iOS app versions...\e[0m"
# Specify the filename as a variable
filename="src/oauth_resources.rs"
# Add comment that it is user generated
echo "// This file was generated by scripts/update_oauth_resources.sh" > "$filename"
echo "// Rerun scripts/update_oauth_resources.sh to update this file" >> "$filename"
echo "// Please do not edit manually" >> "$filename"
echo "// Filled in with real app versions" >> "$filename"
# Open the array in the source file
echo "pub static _IOS_APP_VERSION_LIST: &[&str; $ios_app_count] = &[" >> "$filename"
num=0
# Append the version list to the source file
echo "$ios_version_list" | while IFS= read -r line; do
num=$((num+1))
echo " \"$line\"," >> "$filename"
echo -e "[$num/$ios_app_count] Fetched \e[34m$line\e[0m."
done
# Close the array in the source file
echo "];" >> "$filename"
# Fetch Android app versions
page_1=$(curl -s "https://apkcombo.com/reddit/com.reddit.frontpage/old-versions/" | rg "<a class=\"ver-item\" href=\"(/reddit/com\.reddit\.frontpage/download/phone-20\d{2}\.\d+\.\d+-apk)\" rel=\"nofollow\">" -r "https://apkcombo.com\$1" | sort | uniq)
# Append with pages
page_2=$(curl -s "https://apkcombo.com/reddit/com.reddit.frontpage/old-versions?page=2" | rg "<a class=\"ver-item\" href=\"(/reddit/com\.reddit\.frontpage/download/phone-20\d{2}\.\d+\.\d+-apk)\" rel=\"nofollow\">" -r "https://apkcombo.com\$1" | sort | uniq)
page_3=$(curl -s "https://apkcombo.com/reddit/com.reddit.frontpage/old-versions?page=3" | rg "<a class=\"ver-item\" href=\"(/reddit/com\.reddit\.frontpage/download/phone-20\d{2}\.\d+\.\d+-apk)\" rel=\"nofollow\">" -r "https://apkcombo.com\$1" | sort | uniq)
page_4=$(curl -s "https://apkcombo.com/reddit/com.reddit.frontpage/old-versions?page=4" | rg "<a class=\"ver-item\" href=\"(/reddit/com\.reddit\.frontpage/download/phone-20\d{2}\.\d+\.\d+-apk)\" rel=\"nofollow\">" -r "https://apkcombo.com\$1" | sort | uniq)
page_5=$(curl -s "https://apkcombo.com/reddit/com.reddit.frontpage/old-versions?page=5" | rg "<a class=\"ver-item\" href=\"(/reddit/com\.reddit\.frontpage/download/phone-20\d{2}\.\d+\.\d+-apk)\" rel=\"nofollow\">" -r "https://apkcombo.com\$1" | sort | uniq)
# Concatenate all pages
versions="${page_1}"
versions+=$'\n'
versions+="${page_2}"
versions+=$'\n'
versions+="${page_3}"
versions+=$'\n'
versions+="${page_4}"
versions+=$'\n'
versions+="${page_5}"
# Count the number of lines in the version list
android_count=$(echo "$versions" | wc -l)
echo -e "Fetching \e[32m$android_count Android app versions...\e[0m"
# Append to the source file
echo "pub static ANDROID_APP_VERSION_LIST: &[&str; $android_count] = &[" >> "$filename"
num=0
# For each in versions, curl the page and extract the build number
echo "$versions" | while IFS= read -r line; do
num=$((num+1))
fetch_page=$(curl -s "$line")
build=$(echo "$fetch_page" | rg "<span class=\"vercode\">\((\d+)\)</span>" --only-matching -r "\$1" | head -n1)
version=$(echo "$fetch_page" | rg "<span class=\"vername\">Reddit (20\d{2}\.\d+\.\d+)</span>" --only-matching -r "\$1" | head -n1)
echo " \"Version $version/Build $build\"," >> "$filename"
echo -e "[$num/$android_count] Fetched \e[32mVersion $version/Build $build\e[0m."
done
# Close the array in the source file
echo "];" >> "$filename"
# Retrieve iOS versions
table=$(curl -s "https://en.wikipedia.org/w/api.php?action=parse&page=IOS_17&prop=wikitext&section=31&format=json" | jq ".parse.wikitext.\"*\"" | rg "(17\.[\d\.]*)\\\n\|(\w*)\\\n\|" --only-matching -r "Version \$1 (Build \$2)")
# Count the number of lines in the version list
ios_count=$(echo "$table" | wc -l)
echo -e "Fetching \e[34m$ios_count iOS versions...\e[0m"
# Append to the source file
echo "pub static _IOS_OS_VERSION_LIST: &[&str; $ios_count] = &[" >> "$filename"
num=0
# For each in versions, curl the page and extract the build number
echo "$table" | while IFS= read -r line; do
num=$((num+1))
echo " \"$line\"," >> "$filename"
echo -e "\e[34m[$num/$ios_count] Fetched $line\e[0m."
done
# Close the array in the source file
echo "];" >> "$filename"
echo -e "\e[34mRetrieved $ios_app_count iOS app versions.\e[0m"
echo -e "\e[32mRetrieved $android_count Android app versions.\e[0m"
echo -e "\e[34mRetrieved $ios_count iOS versions.\e[0m"
echo -e "\e[34mTotal: $((ios_app_count + android_count + ios_count))\e[0m"
echo -e "\e[32mSuccess!\e[0m"

125
seccomp-redsunlib.json Normal file
View File

@ -0,0 +1,125 @@
{
"defaultAction": "SCMP_ACT_ERRNO",
"archMap": [
{
"architecture": "SCMP_ARCH_X86_64",
"subArchitectures": [
"SCMP_ARCH_X86",
"SCMP_ARCH_X32"
]
},
{
"architecture": "SCMP_ARCH_AARCH64",
"subArchitectures": [
"SCMP_ARCH_ARM"
]
},
{
"architecture": "SCMP_ARCH_MIPS64",
"subArchitectures": [
"SCMP_ARCH_MIPS",
"SCMP_ARCH_MIPS64N32"
]
},
{
"architecture": "SCMP_ARCH_MIPS64N32",
"subArchitectures": [
"SCMP_ARCH_MIPS",
"SCMP_ARCH_MIPS64"
]
},
{
"architecture": "SCMP_ARCH_MIPSEL64",
"subArchitectures": [
"SCMP_ARCH_MIPSEL",
"SCMP_ARCH_MIPSEL64N32"
]
},
{
"architecture": "SCMP_ARCH_MIPSEL64N32",
"subArchitectures": [
"SCMP_ARCH_MIPSEL",
"SCMP_ARCH_MIPSEL64"
]
},
{
"architecture": "SCMP_ARCH_S390X",
"subArchitectures": [
"SCMP_ARCH_S390"
]
}
],
"syscalls": [
{
"names": [
"accept4",
"arch_prctl",
"bind",
"brk",
"clock_gettime",
"clone",
"close",
"connect",
"epoll_create1",
"epoll_ctl",
"epoll_pwait",
"eventfd2",
"execve",
"exit",
"exit_group",
"fcntl",
"flock",
"fork",
"fstat",
"futex",
"getcwd",
"getpeername",
"getpid",
"getrandom",
"getsockname",
"getsockopt",
"getgid",
"getppid",
"gettid",
"getuid",
"ioctl",
"listen",
"lseek",
"madvise",
"mmap",
"mprotect",
"mremap",
"munmap",
"newfstatat",
"open",
"openat",
"prctl",
"poll",
"read",
"recvfrom",
"rt_sigaction",
"rt_sigprocmask",
"rt_sigreturn",
"sched_getaffinity",
"sched_yield",
"sendto",
"setitimer",
"setsockopt",
"set_tid_address",
"shutdown",
"sigaltstack",
"socket",
"socketpair",
"stat",
"wait4",
"write",
"writev"
],
"action": "SCMP_ACT_ALLOW",
"args": [],
"comment": "",
"includes": {},
"excludes": {}
}
]
}

520
src/client.rs Normal file
View File

@ -0,0 +1,520 @@
use arc_swap::ArcSwap;
use cached::proc_macro::cached;
use futures_lite::future::block_on;
use futures_lite::{future::Boxed, FutureExt};
use hyper::client::HttpConnector;
use hyper::header::HeaderValue;
use hyper::{body, body::Buf, client, header, Body, Client, Method, Request, Response, Uri};
use hyper_rustls::HttpsConnector;
use libflate::gzip;
use log::{error, trace, warn};
use once_cell::sync::Lazy;
use percent_encoding::{percent_encode, CONTROLS};
use serde_json::Value;
use std::sync::atomic::Ordering;
use std::sync::atomic::{AtomicBool, AtomicU16};
use std::{io, result::Result};
use crate::dbg_msg;
use crate::oauth::{force_refresh_token, token_daemon, Oauth};
use crate::server::RequestExt;
use crate::utils::format_url;
const REDDIT_URL_BASE: &str = "https://oauth.reddit.com";
const REDDIT_URL_BASE_HOST: &str = "oauth.reddit.com";
const REDDIT_SHORT_URL_BASE: &str = "https://redd.it";
const REDDIT_SHORT_URL_BASE_HOST: &str = "redd.it";
const ALTERNATIVE_REDDIT_URL_BASE: &str = "https://www.reddit.com";
const ALTERNATIVE_REDDIT_URL_BASE_HOST: &str = "www.reddit.com";
pub static CLIENT: Lazy<Client<HttpsConnector<HttpConnector>>> = Lazy::new(|| {
let https = hyper_rustls::HttpsConnectorBuilder::new().with_native_roots().https_only().enable_http1().build();
client::Client::builder().build(https)
});
pub static OAUTH_CLIENT: Lazy<ArcSwap<Oauth>> = Lazy::new(|| {
let client = block_on(Oauth::new());
tokio::spawn(token_daemon());
ArcSwap::new(client.into())
});
pub static OAUTH_RATELIMIT_REMAINING: AtomicU16 = AtomicU16::new(99);
pub static OAUTH_IS_ROLLING_OVER: AtomicBool = AtomicBool::new(false);
static URL_PAIRS: [(&str, &str); 2] = [
(ALTERNATIVE_REDDIT_URL_BASE, ALTERNATIVE_REDDIT_URL_BASE_HOST),
(REDDIT_SHORT_URL_BASE, REDDIT_SHORT_URL_BASE_HOST),
];
/// Gets the canonical path for a resource on Reddit. This is accomplished by
/// making a `HEAD` request to Reddit at the path given in `path`.
///
/// This function returns `Ok(Some(path))`, where `path`'s value is identical
/// to that of the value of the argument `path`, if Reddit responds to our
/// `HEAD` request with a 2xx-family HTTP code. It will also return an
/// `Ok(Some(String))` if Reddit responds to our `HEAD` request with a
/// `Location` header in the response, and the HTTP code is in the 3xx-family;
/// the `String` will contain the path as reported in `Location`. The return
/// value is `Ok(None)` if Reddit responded with a 3xx, but did not provide a
/// `Location` header. An `Err(String)` is returned if Reddit responds with a
/// 429, or if we were unable to decode the value in the `Location` header.
#[cached(size = 1024, time = 600, result = true)]
#[async_recursion::async_recursion]
pub async fn canonical_path(path: String, tries: i8) -> Result<Option<String>, String> {
if tries == 0 {
return Ok(None);
}
// for each URL pair, try the HEAD request
let res = {
// for url base and host in URL_PAIRS, try reddit_short_head(path.clone(), true, url_base, url_base_host) and if it succeeds, set res. else, res = None
let mut res = None;
for (url_base, url_base_host) in URL_PAIRS {
res = reddit_short_head(path.clone(), true, url_base, url_base_host).await.ok();
if let Some(res) = &res {
if !res.status().is_client_error() {
break;
}
}
}
res
};
let res = res.ok_or_else(|| "Unable to make HEAD request to Reddit.".to_string())?;
let status = res.status().as_u16();
let policy_error = res.headers().get(header::RETRY_AFTER).is_some();
match status {
// If Reddit responds with a 2xx, then the path is already canonical.
200..=299 => Ok(Some(path)),
// If Reddit responds with a 301, then the path is redirected.
301 => match res.headers().get(header::LOCATION) {
Some(val) => {
let Ok(original) = val.to_str() else {
return Err("Unable to decode Location header.".to_string());
};
// We need to strip the .json suffix from the original path.
// In addition, we want to remove share parameters.
// Cut it off here instead of letting it propagate all the way
// to main.rs
let stripped_uri = original.strip_suffix(".json").unwrap_or(original).split('?').next().unwrap_or_default();
// The reason why we now have to format_url, is because the new OAuth
// endpoints seem to return full paths, instead of relative paths.
// So we need to strip the .json suffix from the original path, and
// also remove all Reddit domain parts with format_url.
// Otherwise, it will literally redirect to Reddit.com.
let uri = format_url(stripped_uri);
// Decrement tries and try again
canonical_path(uri, tries - 1).await
}
None => Ok(None),
},
// If Reddit responds with anything other than 3xx (except for the 2xx and 301
// as above), return a None.
300..=399 => Ok(None),
// Rate limiting
429 => Err("Too many requests.".to_string()),
// Special condition rate limiting - https://github.com/redlib-org/redlib/issues/229
403 if policy_error => Err("Too many requests.".to_string()),
_ => Ok(
res
.headers()
.get(header::LOCATION)
.map(|val| percent_encode(val.as_bytes(), CONTROLS).to_string().trim_start_matches(REDDIT_URL_BASE).to_string()),
),
}
}
pub async fn proxy(req: Request<Body>, format: &str) -> Result<Response<Body>, String> {
let mut url = format!("{format}?{}", req.uri().query().unwrap_or_default());
// For each parameter in request
for (name, value) in &req.params() {
// Fill the parameter value in the url
url = url.replace(&format!("{{{name}}}"), value);
}
stream(&url, &req).await
}
async fn stream(url: &str, req: &Request<Body>) -> Result<Response<Body>, String> {
// First parameter is target URL (mandatory).
let parsed_uri = url.parse::<Uri>().map_err(|_| "Couldn't parse URL".to_string())?;
// Build the hyper client from the HTTPS connector.
let client: Client<_, Body> = CLIENT.clone();
let mut builder = Request::get(parsed_uri);
// Copy useful headers from original request
for &key in &["Range", "If-Modified-Since", "Cache-Control"] {
if let Some(value) = req.headers().get(key) {
builder = builder.header(key, value);
}
}
let stream_request = builder.body(Body::empty()).map_err(|_| "Couldn't build empty body in stream".to_string())?;
client
.request(stream_request)
.await
.map(|mut res| {
let mut rm = |key: &str| res.headers_mut().remove(key);
rm("access-control-expose-headers");
rm("server");
rm("vary");
rm("etag");
rm("x-cdn");
rm("x-cdn-client-region");
rm("x-cdn-name");
rm("x-cdn-server-region");
rm("x-reddit-cdn");
rm("x-reddit-video-features");
rm("Nel");
rm("Report-To");
res
})
.map_err(|e| e.to_string())
}
/// Makes a GET request to Reddit at `path`. By default, this will honor HTTP
/// 3xx codes Reddit returns and will automatically redirect.
fn reddit_get(path: String, quarantine: bool) -> Boxed<Result<Response<Body>, String>> {
request(&Method::GET, path, true, quarantine, REDDIT_URL_BASE, REDDIT_URL_BASE_HOST)
}
/// Makes a HEAD request to Reddit at `path, using the short URL base. This will not follow redirects.
fn reddit_short_head(path: String, quarantine: bool, base_path: &'static str, host: &'static str) -> Boxed<Result<Response<Body>, String>> {
request(&Method::HEAD, path, false, quarantine, base_path, host)
}
// /// Makes a HEAD request to Reddit at `path`. This will not follow redirects.
// fn reddit_head(path: String, quarantine: bool) -> Boxed<Result<Response<Body>, String>> {
// request(&Method::HEAD, path, false, quarantine, false)
// }
// Unused - reddit_head is only ever called in the context of a short URL
/// Makes a request to Reddit. If `redirect` is `true`, `request_with_redirect`
/// will recurse on the URL that Reddit provides in the Location HTTP header
/// in its response.
fn request(method: &'static Method, path: String, redirect: bool, quarantine: bool, base_path: &'static str, host: &'static str) -> Boxed<Result<Response<Body>, String>> {
// Build Reddit URL from path.
let url = format!("{base_path}{path}");
// Construct the hyper client from the HTTPS connector.
let client: Client<_, Body> = CLIENT.clone();
let (token, vendor_id, device_id, user_agent, loid) = {
let client = OAUTH_CLIENT.load_full();
(
client.token.clone(),
client.headers_map.get("Client-Vendor-Id").cloned().unwrap_or_default(),
client.headers_map.get("X-Reddit-Device-Id").cloned().unwrap_or_default(),
client.headers_map.get("User-Agent").cloned().unwrap_or_default(),
client.headers_map.get("x-reddit-loid").cloned().unwrap_or_default(),
)
};
// Build request to Reddit. When making a GET, request gzip compression.
// (Reddit doesn't do brotli yet.)
let builder = Request::builder()
.method(method)
.uri(&url)
.header("User-Agent", user_agent)
.header("Client-Vendor-Id", vendor_id)
.header("X-Reddit-Device-Id", device_id)
.header("x-reddit-loid", loid)
.header("Host", host)
.header("Authorization", &format!("Bearer {token}"))
.header("Accept-Encoding", if method == Method::GET { "gzip" } else { "identity" })
.header("Accept-Language", "en-US,en;q=0.5")
.header("Connection", "keep-alive")
.header(
"Cookie",
if quarantine {
"_options=%7B%22pref_quarantine_optin%22%3A%20true%2C%20%22pref_gated_sr_optin%22%3A%20true%7D"
} else {
""
},
)
.body(Body::empty());
async move {
match builder {
Ok(req) => match client.request(req).await {
Ok(mut response) => {
// Reddit may respond with a 3xx. Decide whether or not to
// redirect based on caller params.
if response.status().is_redirection() {
if !redirect {
return Ok(response);
};
let location_header = response.headers().get(header::LOCATION);
if location_header == Some(&HeaderValue::from_static("https://www.reddit.com/")) {
return Err("Reddit response was invalid".to_string());
}
return request(
method,
location_header
.map(|val| {
// We need to make adjustments to the URI
// we get back from Reddit. Namely, we
// must:
//
// 1. Remove the authority (e.g.
// https://www.reddit.com) that may be
// present, so that we recurse on the
// path (and query parameters) as
// required.
//
// 2. Percent-encode the path.
let new_path = percent_encode(val.as_bytes(), CONTROLS)
.to_string()
.trim_start_matches(REDDIT_URL_BASE)
.trim_start_matches(ALTERNATIVE_REDDIT_URL_BASE)
.to_string();
format!("{new_path}{}raw_json=1", if new_path.contains('?') { "&" } else { "?" })
})
.unwrap_or_default()
.to_string(),
true,
quarantine,
base_path,
host,
)
.await;
};
match response.headers().get(header::CONTENT_ENCODING) {
// Content not compressed.
None => Ok(response),
// Content encoded (hopefully with gzip).
Some(hdr) => {
match hdr.to_str() {
Ok(val) => match val {
"gzip" => {}
"identity" => return Ok(response),
_ => return Err("Reddit response was encoded with an unsupported compressor".to_string()),
},
Err(_) => return Err("Reddit response was invalid".to_string()),
}
// We get here if the body is gzip-compressed.
// The body must be something that implements
// std::io::Read, hence the conversion to
// bytes::buf::Buf and then transformation into a
// Reader.
let mut decompressed: Vec<u8>;
{
let mut aggregated_body = match body::aggregate(response.body_mut()).await {
Ok(b) => b.reader(),
Err(e) => return Err(e.to_string()),
};
let mut decoder = match gzip::Decoder::new(&mut aggregated_body) {
Ok(decoder) => decoder,
Err(e) => return Err(e.to_string()),
};
decompressed = Vec::<u8>::new();
if let Err(e) = io::copy(&mut decoder, &mut decompressed) {
return Err(e.to_string());
};
}
response.headers_mut().remove(header::CONTENT_ENCODING);
response.headers_mut().insert(header::CONTENT_LENGTH, decompressed.len().into());
*(response.body_mut()) = Body::from(decompressed);
Ok(response)
}
}
}
Err(e) => {
dbg_msg!("{method} {REDDIT_URL_BASE}{path}: {}", e);
Err(e.to_string())
}
},
Err(_) => Err("Post url contains non-ASCII characters".to_string()),
}
}
.boxed()
}
// Make a request to a Reddit API and parse the JSON response
#[cached(size = 100, time = 30, result = true)]
pub async fn json(path: String, quarantine: bool) -> Result<Value, String> {
// Closure to quickly build errors
let err = |msg: &str, e: String, path: String| -> Result<Value, String> {
// eprintln!("{} - {}: {}", url, msg, e);
Err(format!("{msg}: {e} | {path}"))
};
// First, handle rolling over the OAUTH_CLIENT if need be.
let current_rate_limit = OAUTH_RATELIMIT_REMAINING.load(Ordering::SeqCst);
let is_rolling_over = OAUTH_IS_ROLLING_OVER.load(Ordering::SeqCst);
if current_rate_limit < 10 && !is_rolling_over {
warn!("Rate limit {current_rate_limit} is low. Spawning force_refresh_token()");
tokio::spawn(force_refresh_token());
}
OAUTH_RATELIMIT_REMAINING.fetch_sub(1, Ordering::SeqCst);
// Fetch the url...
match reddit_get(path.clone(), quarantine).await {
Ok(response) => {
let status = response.status();
let reset: Option<String> = if let (Some(remaining), Some(reset), Some(used)) = (
response.headers().get("x-ratelimit-remaining").and_then(|val| val.to_str().ok().map(|s| s.to_string())),
response.headers().get("x-ratelimit-reset").and_then(|val| val.to_str().ok().map(|s| s.to_string())),
response.headers().get("x-ratelimit-used").and_then(|val| val.to_str().ok().map(|s| s.to_string())),
) {
trace!(
"Ratelimit remaining: Header says {remaining}, we have {current_rate_limit}. Resets in {reset}. Rollover: {}. Ratelimit used: {used}",
if is_rolling_over { "yes" } else { "no" },
);
Some(reset)
} else {
None
};
// asynchronously aggregate the chunks of the body
match hyper::body::aggregate(response).await {
Ok(body) => {
let has_remaining = body.has_remaining();
if !has_remaining {
// Rate limited, so spawn a force_refresh_token()
tokio::spawn(force_refresh_token());
return match reset {
Some(val) => Err(format!(
"Reddit rate limit exceeded. Try refreshing in a few seconds.\
Rate limit will reset in: {val}"
)),
None => Err("Reddit rate limit exceeded".to_string()),
};
}
// Parse the response from Reddit as JSON
match serde_json::from_reader(body.reader()) {
Ok(value) => {
let json: Value = value;
// If user is suspended
if let Some(data) = json.get("data") {
if let Some(is_suspended) = data.get("is_suspended").and_then(Value::as_bool) {
if is_suspended {
return Err("suspended".into());
}
}
}
// If Reddit returned an error
if json["error"].is_i64() {
// OAuth token has expired; http status 401
if json["message"] == "Unauthorized" {
error!("Forcing a token refresh");
let () = force_refresh_token().await;
return Err("OAuth token has expired. Please refresh the page!".to_string());
}
// Handle quarantined
if json["reason"] == "quarantined" {
return Err("quarantined".into());
}
// Handle gated
if json["reason"] == "gated" {
return Err("gated".into());
}
// Handle private subs
if json["reason"] == "private" {
return Err("private".into());
}
// Handle banned subs
if json["reason"] == "banned" {
return Err("banned".into());
}
Err(format!("Reddit error {} \"{}\": {} | {path}", json["error"], json["reason"], json["message"]))
} else {
Ok(json)
}
}
Err(e) => {
error!("Got an invalid response from reddit {e}. Status code: {status}");
if status.is_server_error() {
Err("Reddit is having issues, check if there's an outage".to_string())
} else {
err("Failed to parse page JSON data", e.to_string(), path)
}
}
}
}
Err(e) => err("Failed receiving body from Reddit", e.to_string(), path),
}
}
Err(e) => err("Couldn't send request to Reddit", e, path),
}
}
#[cfg(test)]
static POPULAR_URL: &str = "/r/popular/hot.json?&raw_json=1&geo_filter=GLOBAL";
#[tokio::test(flavor = "multi_thread")]
async fn test_localization_popular() {
let val = json(POPULAR_URL.to_string(), false).await.unwrap();
assert_eq!("GLOBAL", val["data"]["geo_filter"].as_str().unwrap());
}
#[tokio::test(flavor = "multi_thread")]
async fn test_obfuscated_share_link() {
let share_link = "/r/rust/s/kPgq8WNHRK".into();
// Correct link without share parameters
let canonical_link = "/r/rust/comments/18t5968/why_use_tuple_struct_over_standard_struct/kfbqlbc/".into();
assert_eq!(canonical_path(share_link, 3).await, Ok(Some(canonical_link)));
}
#[tokio::test(flavor = "multi_thread")]
async fn test_share_link_strip_json() {
let link = "/17krzvz".into();
let canonical_link = "/comments/17krzvz".into();
assert_eq!(canonical_path(link, 3).await, Ok(Some(canonical_link)));
}
#[tokio::test(flavor = "multi_thread")]
async fn test_private_sub() {
let link = json("/r/suicide/about.json?raw_json=1".into(), true).await;
assert!(link.is_err());
assert_eq!(link, Err("private".into()));
}
#[tokio::test(flavor = "multi_thread")]
async fn test_banned_sub() {
let link = json("/r/aaa/about.json?raw_json=1".into(), true).await;
assert!(link.is_err());
assert_eq!(link, Err("banned".into()));
}
#[tokio::test(flavor = "multi_thread")]
async fn test_gated_sub() {
// quarantine to false to specifically catch when we _don't_ catch it
let link = json("/r/drugs/about.json?raw_json=1".into(), false).await;
assert!(link.is_err());
assert_eq!(link, Err("gated".into()));
}

281
src/config.rs Normal file
View File

@ -0,0 +1,281 @@
use once_cell::sync::Lazy;
use serde::{Deserialize, Serialize};
use std::{env::var, fs::read_to_string};
// Waiting for https://github.com/rust-lang/rust/issues/74465 to land, so we
// can reduce reliance on once_cell.
//
// This is the local static that is initialized at runtime (technically at
// first request) and contains the instance settings.
pub static CONFIG: Lazy<Config> = Lazy::new(Config::load);
// This serves as the frontend for an archival API - on removed comments, this URL
// will be the base of a link, to display removed content (on another site).
pub const DEFAULT_PUSHSHIFT_FRONTEND: &str = "undelete.pullpush.io";
/// Stores the configuration parsed from the environment variables and the
/// config file. `Config::Default()` contains None for each setting.
/// When adding more config settings, add it to `Config::load`,
/// `get_setting_from_config`, both below, as well as
/// `instance_info::InstanceInfo.to_string`(), README.md and app.json.
#[derive(Default, Serialize, Deserialize, Clone, Debug)]
pub struct Config {
#[serde(rename = "REDLIB_SFW_ONLY")]
#[serde(alias = "LIBREDDIT_SFW_ONLY")]
pub(crate) sfw_only: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_THEME")]
#[serde(alias = "LIBREDDIT_DEFAULT_THEME")]
pub(crate) default_theme: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_MASCOT")]
#[serde(alias = "LIBREDDIT_DEFAULT_MASCOT")]
pub(crate) default_mascot: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_FRONT_PAGE")]
#[serde(alias = "LIBREDDIT_DEFAULT_FRONT_PAGE")]
pub(crate) default_front_page: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_LAYOUT")]
#[serde(alias = "LIBREDDIT_DEFAULT_LAYOUT")]
pub(crate) default_layout: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_WIDE")]
#[serde(alias = "LIBREDDIT_DEFAULT_WIDE")]
pub(crate) default_wide: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_COMMENT_SORT")]
#[serde(alias = "LIBREDDIT_DEFAULT_COMMENT_SORT")]
pub(crate) default_comment_sort: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_POST_SORT")]
#[serde(alias = "LIBREDDIT_DEFAULT_POST_SORT")]
pub(crate) default_post_sort: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_BLUR_SPOILER")]
#[serde(alias = "LIBREDDIT_DEFAULT_BLUR_SPOILER")]
pub(crate) default_blur_spoiler: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_SHOW_NSFW")]
#[serde(alias = "LIBREDDIT_DEFAULT_SHOW_NSFW")]
pub(crate) default_show_nsfw: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_BLUR_NSFW")]
#[serde(alias = "LIBREDDIT_DEFAULT_BLUR_NSFW")]
pub(crate) default_blur_nsfw: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_USE_HLS")]
#[serde(alias = "LIBREDDIT_DEFAULT_USE_HLS")]
pub(crate) default_use_hls: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_FFMPEG_VIDEO_DOWNLOADS")]
#[serde(alias = "LIBREDDIT_DEFAULT_FFMPEG_VIDEO_DOWNLOADS")]
pub(crate) default_ffmpeg_video_downloads: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_HIDE_HLS_NOTIFICATION")]
#[serde(alias = "LIBREDDIT_DEFAULT_HIDE_HLS_NOTIFICATION")]
pub(crate) default_hide_hls_notification: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_HIDE_AWARDS")]
#[serde(alias = "LIBREDDIT_DEFAULT_HIDE_AWARDS")]
pub(crate) default_hide_awards: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY")]
#[serde(alias = "LIBREDDIT_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY")]
pub(crate) default_hide_sidebar_and_summary: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_HIDE_SCORE")]
#[serde(alias = "LIBREDDIT_DEFAULT_HIDE_SCORE")]
pub(crate) default_hide_score: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_SUBSCRIPTIONS")]
#[serde(alias = "LIBREDDIT_DEFAULT_SUBSCRIPTIONS")]
pub(crate) default_subscriptions: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_FILTERS")]
#[serde(alias = "LIBREDDIT_DEFAULT_FILTERS")]
pub(crate) default_filters: Option<String>,
#[serde(rename = "REDLIB_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION")]
#[serde(alias = "LIBREDDIT_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION")]
pub(crate) default_disable_visit_reddit_confirmation: Option<String>,
#[serde(rename = "REDLIB_BANNER")]
#[serde(alias = "LIBREDDIT_BANNER")]
pub(crate) banner: Option<String>,
#[serde(rename = "REDLIB_ROBOTS_DISABLE_INDEXING")]
#[serde(alias = "LIBREDDIT_ROBOTS_DISABLE_INDEXING")]
pub(crate) robots_disable_indexing: Option<String>,
#[serde(rename = "REDLIB_PUSHSHIFT_FRONTEND")]
#[serde(alias = "LIBREDDIT_PUSHSHIFT_FRONTEND")]
pub(crate) pushshift: Option<String>,
#[serde(rename = "REDLIB_ENABLE_RSS")]
pub(crate) enable_rss: Option<String>,
#[serde(rename = "REDLIB_FULL_URL")]
pub(crate) full_url: Option<String>,
}
impl Config {
/// Load the configuration from the environment variables and the config file.
/// In the case that there are no environment variables set and there is no
/// config file, this function returns a Config that contains all None values.
pub fn load() -> Self {
let load_config = |name: &str| {
let new_file = read_to_string(name);
new_file.ok().and_then(|new_file| toml::from_str::<Self>(&new_file).ok())
};
let config = load_config("redlib.toml").or_else(|| load_config("libreddit.toml")).unwrap_or_default();
// This function defines the order of preference - first check for
// environment variables with "REDLIB", then check the legacy LIBREDDIT
// option, then check the config, then if all are `None`, return a `None`
let parse = |key: &str| -> Option<String> {
// Return the first non-`None` value
// If all are `None`, return `None`
let legacy_key = key.replace("REDLIB_", "LIBREDDIT_");
var(key).ok().or_else(|| var(legacy_key).ok()).or_else(|| get_setting_from_config(key, &config))
};
Self {
sfw_only: parse("REDLIB_SFW_ONLY"),
default_theme: parse("REDLIB_DEFAULT_THEME"),
default_mascot: parse("REDLIB_DEFAULT_MASCOT"),
default_front_page: parse("REDLIB_DEFAULT_FRONT_PAGE"),
default_layout: parse("REDLIB_DEFAULT_LAYOUT"),
default_post_sort: parse("REDLIB_DEFAULT_POST_SORT"),
default_wide: parse("REDLIB_DEFAULT_WIDE"),
default_comment_sort: parse("REDLIB_DEFAULT_COMMENT_SORT"),
default_blur_spoiler: parse("REDLIB_DEFAULT_BLUR_SPOILER"),
default_show_nsfw: parse("REDLIB_DEFAULT_SHOW_NSFW"),
default_blur_nsfw: parse("REDLIB_DEFAULT_BLUR_NSFW"),
default_use_hls: parse("REDLIB_DEFAULT_USE_HLS"),
default_ffmpeg_video_downloads: parse("REDLIB_DEFAULT_FFMPEG_VIDEO_DOWNLOADS"),
default_hide_hls_notification: parse("REDLIB_DEFAULT_HIDE_HLS_NOTIFICATION"),
default_hide_awards: parse("REDLIB_DEFAULT_HIDE_AWARDS"),
default_hide_sidebar_and_summary: parse("REDLIB_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY"),
default_hide_score: parse("REDLIB_DEFAULT_HIDE_SCORE"),
default_subscriptions: parse("REDLIB_DEFAULT_SUBSCRIPTIONS"),
default_filters: parse("REDLIB_DEFAULT_FILTERS"),
default_disable_visit_reddit_confirmation: parse("REDLIB_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION"),
banner: parse("REDLIB_BANNER"),
robots_disable_indexing: parse("REDLIB_ROBOTS_DISABLE_INDEXING"),
pushshift: parse("REDLIB_PUSHSHIFT_FRONTEND"),
enable_rss: parse("REDLIB_ENABLE_RSS"),
full_url: parse("REDLIB_FULL_URL"),
}
}
}
fn get_setting_from_config(name: &str, config: &Config) -> Option<String> {
match name {
"REDLIB_SFW_ONLY" => config.sfw_only.clone(),
"REDLIB_DEFAULT_THEME" => config.default_theme.clone(),
"REDLIB_DEFAULT_MASCOT" => config.default_mascot.clone(),
"REDLIB_DEFAULT_FRONT_PAGE" => config.default_front_page.clone(),
"REDLIB_DEFAULT_LAYOUT" => config.default_layout.clone(),
"REDLIB_DEFAULT_COMMENT_SORT" => config.default_comment_sort.clone(),
"REDLIB_DEFAULT_POST_SORT" => config.default_post_sort.clone(),
"REDLIB_DEFAULT_BLUR_SPOILER" => config.default_blur_spoiler.clone(),
"REDLIB_DEFAULT_SHOW_NSFW" => config.default_show_nsfw.clone(),
"REDLIB_DEFAULT_BLUR_NSFW" => config.default_blur_nsfw.clone(),
"REDLIB_DEFAULT_USE_HLS" => config.default_use_hls.clone(),
"REDLIB_DEFAULT_FFMPEG_VIDEO_DOWNLOADS" => config.default_ffmpeg_video_downloads.clone(),
"REDLIB_DEFAULT_HIDE_HLS_NOTIFICATION" => config.default_hide_hls_notification.clone(),
"REDLIB_DEFAULT_WIDE" => config.default_wide.clone(),
"REDLIB_DEFAULT_HIDE_AWARDS" => config.default_hide_awards.clone(),
"REDLIB_DEFAULT_HIDE_SIDEBAR_AND_SUMMARY" => config.default_hide_sidebar_and_summary.clone(),
"REDLIB_DEFAULT_HIDE_SCORE" => config.default_hide_score.clone(),
"REDLIB_DEFAULT_SUBSCRIPTIONS" => config.default_subscriptions.clone(),
"REDLIB_DEFAULT_FILTERS" => config.default_filters.clone(),
"REDLIB_DEFAULT_DISABLE_VISIT_REDDIT_CONFIRMATION" => config.default_disable_visit_reddit_confirmation.clone(),
"REDLIB_BANNER" => config.banner.clone(),
"REDLIB_ROBOTS_DISABLE_INDEXING" => config.robots_disable_indexing.clone(),
"REDLIB_PUSHSHIFT_FRONTEND" => config.pushshift.clone(),
"REDLIB_ENABLE_RSS" => config.enable_rss.clone(),
"REDLIB_FULL_URL" => config.full_url.clone(),
_ => None,
}
}
/// Retrieves setting from environment variable or config file.
pub fn get_setting(name: &str) -> Option<String> {
get_setting_from_config(name, &CONFIG)
}
#[cfg(test)]
use {sealed_test::prelude::*, std::fs::write};
#[test]
fn test_deserialize() {
// Must handle empty input
let result = toml::from_str::<Config>("");
assert!(result.is_ok(), "Error: {}", result.unwrap_err());
}
#[test]
#[sealed_test(env = [("REDLIB_SFW_ONLY", "on")])]
fn test_env_var() {
assert!(crate::utils::sfw_only())
}
#[test]
#[sealed_test]
fn test_config() {
let config_to_write = r#"REDLIB_DEFAULT_COMMENT_SORT = "best""#;
write("redlib.toml", config_to_write).unwrap();
assert_eq!(get_setting("REDLIB_DEFAULT_COMMENT_SORT"), Some("best".into()));
}
#[test]
#[sealed_test]
fn test_config_legacy() {
let config_to_write = r#"LIBREDDIT_DEFAULT_COMMENT_SORT = "best""#;
write("libreddit.toml", config_to_write).unwrap();
assert_eq!(get_setting("REDLIB_DEFAULT_COMMENT_SORT"), Some("best".into()));
}
#[test]
#[sealed_test(env = [("LIBREDDIT_SFW_ONLY", "on")])]
fn test_env_var_legacy() {
assert!(crate::utils::sfw_only())
}
#[test]
#[sealed_test(env = [("REDLIB_DEFAULT_COMMENT_SORT", "top")])]
fn test_env_config_precedence() {
let config_to_write = r#"REDLIB_DEFAULT_COMMENT_SORT = "best""#;
write("redlib.toml", config_to_write).unwrap();
assert_eq!(get_setting("REDLIB_DEFAULT_COMMENT_SORT"), Some("top".into()))
}
#[test]
#[sealed_test(env = [("REDLIB_DEFAULT_COMMENT_SORT", "top")])]
fn test_alt_env_config_precedence() {
let config_to_write = r#"REDLIB_DEFAULT_COMMENT_SORT = "best""#;
write("redlib.toml", config_to_write).unwrap();
assert_eq!(get_setting("REDLIB_DEFAULT_COMMENT_SORT"), Some("top".into()))
}
#[test]
#[sealed_test(env = [("REDLIB_DEFAULT_SUBSCRIPTIONS", "news+bestof")])]
fn test_default_subscriptions() {
assert_eq!(get_setting("REDLIB_DEFAULT_SUBSCRIPTIONS"), Some("news+bestof".into()));
}
#[test]
#[sealed_test(env = [("REDLIB_DEFAULT_FILTERS", "news+bestof")])]
fn test_default_filters() {
assert_eq!(get_setting("REDLIB_DEFAULT_FILTERS"), Some("news+bestof".into()));
}
#[test]
#[sealed_test]
fn test_pushshift() {
let config_to_write = r#"REDLIB_PUSHSHIFT_FRONTEND = "https://api.pushshift.io""#;
write("redlib.toml", config_to_write).unwrap();
assert!(get_setting("REDLIB_PUSHSHIFT_FRONTEND").is_some());
assert_eq!(get_setting("REDLIB_PUSHSHIFT_FRONTEND"), Some("https://api.pushshift.io".into()));
}

236
src/duplicates.rs Normal file
View File

@ -0,0 +1,236 @@
// Handler for post duplicates.
use crate::client::json;
use crate::server::RequestExt;
use crate::subreddit::{can_access_quarantine, quarantine};
use crate::utils::{error, filter_posts, get_filters, nsfw_landing, parse_post, template, Post, Preferences};
use hyper::{Body, Request, Response};
use rinja::Template;
use serde_json::Value;
use std::borrow::ToOwned;
use std::collections::HashSet;
use std::vec::Vec;
/// `DuplicatesParams` contains the parameters in the URL.
struct DuplicatesParams {
before: String,
after: String,
sort: String,
}
/// `DuplicatesTemplate` defines an Askama template for rendering duplicate
/// posts.
#[derive(Template)]
#[template(path = "duplicates.html")]
struct DuplicatesTemplate {
/// params contains the relevant request parameters.
params: DuplicatesParams,
/// post is the post whose ID is specified in the reqeust URL. Note that
/// this is not necessarily the "original" post.
post: Post,
/// duplicates is the list of posts that, per Reddit, are duplicates of
/// Post above.
duplicates: Vec<Post>,
/// prefs are the user preferences.
prefs: Preferences,
/// url is the request URL.
url: String,
/// num_posts_filtered counts how many posts were filtered from the
/// duplicates list.
num_posts_filtered: u64,
/// all_posts_filtered is true if every duplicate was filtered. This is an
/// edge case but can still happen.
all_posts_filtered: bool,
}
/// Make the GET request to Reddit. It assumes `req` is the appropriate Reddit
/// REST endpoint for enumerating post duplicates.
pub async fn item(req: Request<Body>) -> Result<Response<Body>, String> {
let path: String = format!("{}.json?{}&raw_json=1", req.uri().path(), req.uri().query().unwrap_or_default());
let sub = req.param("sub").unwrap_or_default();
let quarantined = can_access_quarantine(&req, &sub);
// Log the request in debugging mode
#[cfg(debug_assertions)]
req.param("id").unwrap_or_default();
// Send the GET, and await JSON.
match json(path, quarantined).await {
// Process response JSON.
Ok(response) => {
let post = parse_post(&response[0]["data"]["children"][0]).await;
let req_url = req.uri().to_string();
// Return landing page if this post if this Reddit deems this post
// NSFW, but we have also disabled the display of NSFW content
// or if the instance is SFW-only
if post.nsfw && crate::utils::should_be_nsfw_gated(&req, &req_url) {
return Ok(nsfw_landing(req, req_url).await.unwrap_or_default());
}
let filters = get_filters(&req);
let (duplicates, num_posts_filtered, all_posts_filtered) = parse_duplicates(&response[1], &filters).await;
// These are the values for the "before=", "after=", and "sort="
// query params, respectively.
let mut before: String = String::new();
let mut after: String = String::new();
let mut sort: String = String::new();
// FIXME: We have to perform a kludge to work around a Reddit API
// bug.
//
// The JSON object in "data" will never contain a "before" value so
// it is impossible to use it to determine our position in a
// listing. We'll make do by getting the ID of the first post in
// the listing, setting that as our "before" value, and ask Reddit
// to give us a batch of duplicate posts up to that post.
//
// Likewise, if we provide a "before" request in the GET, the
// result won't have an "after" in the JSON, in addition to missing
// the "before." So we will have to use the final post in the list
// of duplicates.
//
// That being said, we'll also need to capture the value of the
// "sort=" parameter as well, so we will need to inspect the
// query key-value pairs anyway.
let l = duplicates.len();
if l > 0 {
// This gets set to true if "before=" is one of the GET params.
let mut have_before: bool = false;
// This gets set to true if "after=" is one of the GET params.
let mut have_after: bool = false;
// Inspect the query key-value pairs. We will need to record
// the value of "sort=", along with checking to see if either
// one of "before=" or "after=" are given.
//
// If we're in the middle of the batch (evidenced by the
// presence of a "before=" or "after=" parameter in the GET),
// then use the first post as the "before" reference.
//
// We'll do this iteratively. Better than with .map_or()
// since a closure will continue to operate on remaining
// elements even after we've determined one of "before=" or
// "after=" (or both) are in the GET request.
//
// In practice, here should only ever be one of "before=" or
// "after=" and never both.
let query_str = req.uri().query().unwrap_or_default().to_string();
if !query_str.is_empty() {
for param in query_str.split('&') {
let kv: Vec<&str> = param.split('=').collect();
if kv.len() < 2 {
// Reject invalid query parameter.
continue;
}
let key: &str = kv[0];
match key {
"before" => have_before = true,
"after" => have_after = true,
"sort" => {
let val: &str = kv[1];
match val {
"new" | "num_comments" => sort = val.to_string(),
_ => {}
}
}
_ => {}
}
}
}
if have_after {
"t3_".clone_into(&mut before);
before.push_str(&duplicates[0].id);
}
// Address potentially missing "after". If "before=" is in the
// GET, then "after" will be null in the JSON (see FIXME
// above).
if have_before {
// The next batch will need to start from one after the
// last post in the current batch.
"t3_".clone_into(&mut after);
after.push_str(&duplicates[l - 1].id);
// Here is where things get terrible. Notice that we
// haven't set `before`. In order to do so, we will
// need to know if there is a batch that exists before
// this one, and doing so requires actually fetching the
// previous batch. In other words, we have to do yet one
// more GET to Reddit. There is no other way to determine
// whether or not to define `before`.
//
// We'll mitigate that by requesting at most one duplicate.
let new_path: String = format!(
"{}.json?before=t3_{}&sort={}&limit=1&raw_json=1",
req.uri().path(),
&duplicates[0].id,
if sort.is_empty() { "num_comments".to_string() } else { sort.clone() }
);
match json(new_path, true).await {
Ok(response) => {
if !response[1]["data"]["children"].as_array().unwrap_or(&Vec::new()).is_empty() {
"t3_".clone_into(&mut before);
before.push_str(&duplicates[0].id);
}
}
Err(msg) => {
// Abort entirely if we couldn't get the previous
// batch.
return error(req, &msg).await;
}
}
} else {
after = response[1]["data"]["after"].as_str().unwrap_or_default().to_string();
}
}
Ok(template(&DuplicatesTemplate {
params: DuplicatesParams { before, after, sort },
post,
duplicates,
prefs: Preferences::new(&req),
url: req_url,
num_posts_filtered,
all_posts_filtered,
}))
}
// Process error.
Err(msg) => {
if msg == "quarantined" || msg == "gated" {
let sub = req.param("sub").unwrap_or_default();
Ok(quarantine(&req, sub, &msg))
} else {
error(req, &msg).await
}
}
}
}
// DUPLICATES
async fn parse_duplicates(json: &Value, filters: &HashSet<String>) -> (Vec<Post>, u64, bool) {
let post_duplicates: &Vec<Value> = &json["data"]["children"].as_array().map_or(Vec::new(), ToOwned::to_owned);
let mut duplicates: Vec<Post> = Vec::new();
// Process each post and place them in the Vec<Post>.
for val in post_duplicates {
let post: Post = parse_post(val).await;
duplicates.push(post);
}
let (num_posts_filtered, all_posts_filtered) = filter_posts(&mut duplicates, filters);
(duplicates, num_posts_filtered, all_posts_filtered)
}

238
src/instance_info.rs Normal file
View File

@ -0,0 +1,238 @@
use crate::{
config::{Config, CONFIG},
server::RequestExt,
utils::{ErrorTemplate, Preferences},
};
use build_html::{Container, Html, HtmlContainer, Table};
use hyper::{http::Error, Body, Request, Response};
use once_cell::sync::Lazy;
use rinja::Template;
use serde::{Deserialize, Serialize};
use time::OffsetDateTime;
// This is the local static that is intialized at runtime (technically at
// the first request to the info endpoint) and contains the data
// retrieved from the info endpoint.
pub static INSTANCE_INFO: Lazy<InstanceInfo> = Lazy::new(InstanceInfo::new);
/// Handles instance info endpoint
pub async fn instance_info(req: Request<Body>) -> Result<Response<Body>, String> {
// This will retrieve the extension given, or create a new string - which will
// simply become the last option, an HTML page.
let extension = req.param("extension").unwrap_or_default();
let response = match extension.as_str() {
"yaml" | "yml" => info_yaml(),
"txt" => info_txt(),
"json" => info_json(),
"html" | "" => info_html(&req),
_ => {
let error = ErrorTemplate {
msg: "Error: Invalid info extension".into(),
prefs: Preferences::new(&req),
url: req.uri().to_string(),
}
.render()
.unwrap();
Response::builder().status(404).header("content-type", "text/html; charset=utf-8").body(error.into())
}
};
response.map_err(|err| format!("{err}"))
}
fn info_json() -> Result<Response<Body>, Error> {
if let Ok(body) = serde_json::to_string(&*INSTANCE_INFO) {
Response::builder().status(200).header("content-type", "application/json").body(body.into())
} else {
Response::builder()
.status(500)
.header("content-type", "text/plain")
.body(Body::from("Error serializing JSON"))
}
}
fn info_yaml() -> Result<Response<Body>, Error> {
if let Ok(body) = serde_yaml::to_string(&*INSTANCE_INFO) {
// We can use `application/yaml` as media type, though there is no guarantee
// that browsers will honor it. But we'll do it anyway. See:
// https://github.com/ietf-wg-httpapi/mediatypes/blob/main/draft-ietf-httpapi-yaml-mediatypes.md#media-type-applicationyaml-application-yaml
Response::builder().status(200).header("content-type", "application/yaml").body(body.into())
} else {
Response::builder()
.status(500)
.header("content-type", "text/plain")
.body(Body::from("Error serializing YAML."))
}
}
fn info_txt() -> Result<Response<Body>, Error> {
Response::builder()
.status(200)
.header("content-type", "text/plain")
.body(Body::from(INSTANCE_INFO.to_string(&StringType::Raw)))
}
fn info_html(req: &Request<Body>) -> Result<Response<Body>, Error> {
let message = MessageTemplate {
title: String::from("Instance information"),
body: INSTANCE_INFO.to_string(&StringType::Html),
prefs: Preferences::new(req),
url: req.uri().to_string(),
}
.render()
.unwrap();
Response::builder().status(200).header("content-type", "text/html; charset=utf8").body(Body::from(message))
}
#[derive(Serialize, Deserialize, Default)]
pub struct InstanceInfo {
package_name: String,
crate_version: String,
git_commit: String,
deploy_date: String,
compile_mode: String,
deploy_unix_ts: i64,
config: Config,
}
impl InstanceInfo {
pub fn new() -> Self {
Self {
package_name: env!("CARGO_PKG_NAME").to_string(),
crate_version: env!("CARGO_PKG_VERSION").to_string(),
git_commit: env!("GIT_HASH").to_string(),
deploy_date: OffsetDateTime::now_local().unwrap_or_else(|_| OffsetDateTime::now_utc()).to_string(),
#[cfg(debug_assertions)]
compile_mode: "Debug".into(),
#[cfg(not(debug_assertions))]
compile_mode: "Release".into(),
deploy_unix_ts: OffsetDateTime::now_local().unwrap_or_else(|_| OffsetDateTime::now_utc()).unix_timestamp(),
config: CONFIG.clone(),
}
}
fn to_table(&self) -> String {
let mut container = Container::default();
let convert = |o: &Option<String>| -> String { o.clone().unwrap_or_else(|| "<span class=\"unset\"><i>Unset</i></span>".to_owned()) };
if let Some(banner) = &self.config.banner {
container.add_header(3, "Instance banner");
container.add_raw("<br />");
container.add_paragraph(banner);
container.add_raw("<br />");
}
container.add_table(
Table::from([
["Package name", &self.package_name],
["Crate version", &self.crate_version],
["Git commit", &self.git_commit],
["Deploy date", &self.deploy_date],
["Deploy timestamp", &self.deploy_unix_ts.to_string()],
["Compile mode", &self.compile_mode],
["SFW only", &convert(&self.config.sfw_only)],
["Pushshift frontend", &convert(&self.config.pushshift)],
["RSS enabled", &convert(&self.config.enable_rss)],
["Full URL", &convert(&self.config.full_url)],
//TODO: fallback to crate::config::DEFAULT_PUSHSHIFT_FRONTEND
])
.with_header_row(["Settings"]),
);
container.add_raw("<br />");
container.add_table(
Table::from([
["Hide awards", &convert(&self.config.default_hide_awards)],
["Hide score", &convert(&self.config.default_hide_score)],
["Theme", &convert(&self.config.default_theme)],
["Mascot", &convert(&self.config.default_mascot)],
["Front page", &convert(&self.config.default_front_page)],
["Layout", &convert(&self.config.default_layout)],
["Wide", &convert(&self.config.default_wide)],
["Comment sort", &convert(&self.config.default_comment_sort)],
["Post sort", &convert(&self.config.default_post_sort)],
["Blur Spoiler", &convert(&self.config.default_blur_spoiler)],
["Show NSFW", &convert(&self.config.default_show_nsfw)],
["Blur NSFW", &convert(&self.config.default_blur_nsfw)],
["Use HLS", &convert(&self.config.default_use_hls)],
["Use FFmpeg", &convert(&self.config.default_ffmpeg_video_downloads)],
["Hide HLS notification", &convert(&self.config.default_hide_hls_notification)],
["Subscriptions", &convert(&self.config.default_subscriptions)],
["Filters", &convert(&self.config.default_filters)],
])
.with_header_row(["Default preferences"]),
);
container.to_html_string().replace("<th>", "<th colspan=\"2\">")
}
fn to_string(&self, string_type: &StringType) -> String {
match string_type {
StringType::Raw => {
format!(
"Package name: {}\n
Crate version: {}\n
Git commit: {}\n
Deploy date: {}\n
Deploy timestamp: {}\n
Compile mode: {}\n
SFW only: {:?}\n
Pushshift frontend: {:?}\n
RSS enabled: {:?}\n
Full URL: {:?}\n
Config:\n
Banner: {:?}\n
Hide awards: {:?}\n
Hide score: {:?}\n
Default theme: {:?}\n
Default mascot: {:?}\n
Default front page: {:?}\n
Default layout: {:?}\n
Default wide: {:?}\n
Default comment sort: {:?}\n
Default post sort: {:?}\n
Default blur Spoiler: {:?}\n
Default show NSFW: {:?}\n
Default blur NSFW: {:?}\n
Default use HLS: {:?}\n
Default use FFmpeg: {:?}\n
Default hide HLS notification: {:?}\n
Default subscriptions: {:?}\n
Default filters: {:?}\n",
self.package_name,
self.crate_version,
self.git_commit,
self.deploy_date,
self.deploy_unix_ts,
self.compile_mode,
self.config.sfw_only,
self.config.enable_rss,
self.config.full_url,
self.config.pushshift,
self.config.banner,
self.config.default_hide_awards,
self.config.default_hide_score,
self.config.default_theme,
self.config.default_mascot,
self.config.default_front_page,
self.config.default_layout,
self.config.default_wide,
self.config.default_comment_sort,
self.config.default_post_sort,
self.config.default_blur_spoiler,
self.config.default_show_nsfw,
self.config.default_blur_nsfw,
self.config.default_use_hls,
self.config.default_ffmpeg_video_downloads,
self.config.default_hide_hls_notification,
self.config.default_subscriptions,
self.config.default_filters,
)
}
StringType::Html => self.to_table(),
}
}
}
enum StringType {
Raw,
Html,
}
#[derive(Template)]
#[template(path = "message.html")]
struct MessageTemplate {
title: String,
body: String,
prefs: Preferences,
url: String,
}

View File

@ -1,68 +1,439 @@
// Import Crates
use actix_web::{get, middleware::NormalizePath, web, App, HttpResponse, HttpServer};
// Global specifiers
#![forbid(unsafe_code)]
#![allow(clippy::cmp_owned)]
// Reference local files
mod popular;
mod config;
mod duplicates;
mod instance_info;
mod oauth;
mod oauth_resources;
mod post;
mod proxy;
mod search;
mod settings;
mod subreddit;
mod user;
mod utils;
// Import Crates
use clap::{Arg, ArgAction, Command};
use futures_lite::FutureExt;
use hyper::{header::HeaderValue, Body, Request, Response};
mod client;
use client::{canonical_path, proxy};
use log::info;
use once_cell::sync::Lazy;
use server::RequestExt;
use utils::{error, redirect, ThemeAssets, MascotAssets};
use crate::client::OAUTH_CLIENT;
mod server;
// Create Services
async fn style() -> HttpResponse {
HttpResponse::Ok().content_type("text/css").body(include_str!("../static/style.css"))
// Required for the manifest to be valid
async fn pwa_logo() -> Result<Response<Body>, String> {
Ok(
Response::builder()
.status(200)
.header("content-type", "image/png")
.body(include_bytes!("../static/logo.png").as_ref().into())
.unwrap_or_default(),
)
}
async fn robots() -> HttpResponse {
HttpResponse::Ok().body(include_str!("../static/robots.txt"))
// Required for iOS App Icons
async fn iphone_logo() -> Result<Response<Body>, String> {
Ok(
Response::builder()
.status(200)
.header("content-type", "image/png")
.body(include_bytes!("../static/apple-touch-icon.png").as_ref().into())
.unwrap_or_default(),
)
}
#[get("/favicon.ico")]
async fn favicon() -> HttpResponse {
HttpResponse::Ok().body("")
async fn favicon() -> Result<Response<Body>, String> {
Ok(
Response::builder()
.status(200)
.header("content-type", "image/vnd.microsoft.icon")
.header("Cache-Control", "public, max-age=1209600, s-maxage=86400")
.body(include_bytes!("../static/favicon.ico").as_ref().into())
.unwrap_or_default(),
)
}
#[actix_web::main]
async fn main() -> std::io::Result<()> {
let args: Vec<String> = std::env::args().collect();
let mut address = "0.0.0.0:8080".to_string();
async fn font() -> Result<Response<Body>, String> {
Ok(
Response::builder()
.status(200)
.header("content-type", "font/woff2")
.header("Cache-Control", "public, max-age=1209600, s-maxage=86400")
.body(include_bytes!("../static/Inter.var.woff2").as_ref().into())
.unwrap_or_default(),
)
}
if args.len() > 1 {
for arg in args {
if arg.starts_with("--address=") || arg.starts_with("-a=") {
let split: Vec<&str> = arg.split("=").collect();
address = split[1].to_string();
}
async fn ffmpeg() -> Result<Response<Body>, String> {
Ok(
Response::builder()
.status(200)
.header("content-type", "application/wasm")
.header("Cache-Control", "public, max-age=1209600, s-maxage=86400")
.body(include_bytes!("../static/ffmpeg/ffmpeg-core.wasm").as_ref().into())
.unwrap_or_default(),
)
}
async fn resource(body: &str, content_type: &str, cache: bool) -> Result<Response<Body>, String> {
let mut res = Response::builder()
.status(200)
.header("content-type", content_type)
.body(body.to_string().into())
.unwrap_or_default();
if cache {
if let Ok(val) = HeaderValue::from_str("public, max-age=1209600, s-maxage=86400") {
res.headers_mut().insert("Cache-Control", val);
}
}
// start http server
println!("Running Libreddit v{} on {}!", env!("CARGO_PKG_VERSION"), address.clone());
HttpServer::new(|| {
App::new()
// TRAILING SLASH MIDDLEWARE
.wrap(NormalizePath::default())
// GENERAL SERVICES
.route("/style.css/", web::get().to(style))
.route("/favicon.ico/", web::get().to(|| HttpResponse::Ok()))
.route("/robots.txt/", web::get().to(robots))
// PROXY SERVICE
.route("/proxy/{url:.*}/", web::get().to(proxy::handler))
// USER SERVICES
.route("/u/{username}/", web::get().to(user::page))
.route("/user/{username}/", web::get().to(user::page))
// SUBREDDIT SERVICES
.route("/r/{sub}/", web::get().to(subreddit::page))
// POPULAR SERVICES
.route("/", web::get().to(popular::page))
// POST SERVICES
.route("/{id:.{5,6}}/", web::get().to(post::short))
.route("/r/{sub}/comments/{id}/{title}/", web::get().to(post::page))
})
.bind(address.clone())
.expect(format!("Cannot bind to the address: {}", address).as_str())
.run()
.await
Ok(res)
}
async fn style() -> Result<Response<Body>, String> {
let mut res = include_str!("../static/style.css").to_string();
for file in ThemeAssets::iter() {
res.push('\n');
let theme = ThemeAssets::get(file.as_ref()).unwrap();
res.push_str(std::str::from_utf8(theme.data.as_ref()).unwrap());
}
Ok(
Response::builder()
.status(200)
.header("content-type", "text/css")
.header("Cache-Control", "public, max-age=1209600, s-maxage=86400")
.body(res.to_string().into())
.unwrap_or_default(),
)
}
/// Serve mascot
async fn mascot_image(req: Request<Body>) -> Result<Response<Body>, String> {
let res = MascotAssets::get(&req.param("name").unwrap())
.unwrap_or(MascotAssets::get("redsunlib.png").unwrap());
Ok(
Response::builder()
.status(200)
.header("content-type", "image/png")
.header("Cache-Control", "public, max-age=1209600, s-maxage=86400")
.body(res.data.into())
.unwrap_or_default(),
)
}
#[tokio::main]
async fn main() {
// Load environment variables
_ = dotenvy::dotenv();
// Initialize logger
pretty_env_logger::init();
let matches = Command::new("Redlib")
.version(env!("CARGO_PKG_VERSION"))
.about("Private front-end for Reddit written in Rust ")
.arg(
Arg::new("redirect-https")
.short('r')
.long("redirect-https")
.help("Redirect all HTTP requests to HTTPS (no longer functional)")
.num_args(0),
)
.arg(
Arg::new("address")
.short('a')
.long("address")
.value_name("ADDRESS")
.help("Sets address to listen on")
.default_value("[::]")
.num_args(1),
)
.arg(
Arg::new("port")
.short('p')
.long("port")
.value_name("PORT")
.env("PORT")
.help("Port to listen on")
.default_value("8080")
.action(ArgAction::Set)
.num_args(1),
)
.arg(
Arg::new("hsts")
.short('H')
.long("hsts")
.value_name("EXPIRE_TIME")
.help("HSTS header to tell browsers that this site should only be accessed over HTTPS")
.default_value("604800")
.num_args(1),
)
.get_matches();
let address = matches.get_one::<String>("address").unwrap();
let port = matches.get_one::<String>("port").unwrap();
let hsts = matches.get_one("hsts").map(|m: &String| m.as_str());
let listener = [address, ":", port].concat();
println!("Starting Redsunlib...");
// Begin constructing a server
let mut app = server::Server::new();
// Force evaluation of statics. In instance_info case, we need to evaluate
// the timestamp so deploy date is accurate - in config case, we need to
// evaluate the configuration to avoid paying penalty at first request -
// in OAUTH case, we need to retrieve the token to avoid paying penalty
// at first request
info!("Evaluating config.");
Lazy::force(&config::CONFIG);
info!("Evaluating instance info.");
Lazy::force(&instance_info::INSTANCE_INFO);
info!("Creating OAUTH client.");
Lazy::force(&OAUTH_CLIENT);
// Define default headers (added to all responses)
app.default_headers = headers! {
"Referrer-Policy" => "no-referrer",
"X-Content-Type-Options" => "nosniff",
"X-Frame-Options" => "DENY",
"Content-Security-Policy" => "default-src 'none'; font-src 'self'; script-src 'self' 'wasm-unsafe-eval' blob:; manifest-src 'self'; media-src 'self' data: blob: about:; style-src 'self' 'unsafe-inline'; base-uri 'none'; img-src 'self' data:; form-action 'self'; frame-ancestors 'none'; connect-src 'self'; worker-src 'self' blob:;"
};
if let Some(expire_time) = hsts {
if let Ok(val) = HeaderValue::from_str(&format!("max-age={expire_time}")) {
app.default_headers.insert("Strict-Transport-Security", val);
}
}
// Read static files
app.at("/style.css").get(|_| style().boxed());
app
.at("/manifest.json")
.get(|_| resource(include_str!("../static/manifest.json"), "application/json", false).boxed());
app.at("/robots.txt").get(|_| {
resource(
if match config::get_setting("REDLIB_ROBOTS_DISABLE_INDEXING") {
Some(val) => val == "on",
None => false,
} {
"User-agent: *\nDisallow: /"
} else {
"User-agent: *\nDisallow: /u/\nDisallow: /user/"
},
"text/plain",
true,
)
.boxed()
});
app.at("/favicon.ico").get(|_| favicon().boxed());
app.at("/logo.png").get(|_| pwa_logo().boxed());
app.at("/Inter.var.woff2").get(|_| font().boxed());
app.at("/touch-icon-iphone.png").get(|_| iphone_logo().boxed());
app.at("/apple-touch-icon.png").get(|_| iphone_logo().boxed());
app
.at("/videoUtils.js")
.get(|_| resource(include_str!("../static/videoUtils.js"), "text/javascript", false).boxed());
app
.at("/hls.min.js")
.get(|_| resource(include_str!("../static/hls.min.js"), "text/javascript", false).boxed());
app
.at("/highlighted.js")
.get(|_| resource(include_str!("../static/highlighted.js"), "text/javascript", false).boxed());
// FFmpeg
app
.at("/ffmpeg/814.ffmpeg.js")
.get(|_| resource(include_str!("../static/ffmpeg/814.ffmpeg.js"), "text/javascript", false).boxed());
app
.at("/ffmpeg/814.ffmpeg.js.map")
.get(|_| resource(include_str!("../static/ffmpeg/814.ffmpeg.js.map"), "text/javascript", false).boxed());
app
.at("/ffmpeg/ffmpeg-core.js")
.get(|_| resource(include_str!("../static/ffmpeg/ffmpeg-core.js"), "text/javascript", false).boxed());
app.at("/ffmpeg/ffmpeg-core.wasm").get(|_| ffmpeg().boxed());
app
.at("/ffmpeg/ffmpeg-util.js")
.get(|_| resource(include_str!("../static/ffmpeg/ffmpeg-util.js"), "text/javascript", false).boxed());
app
.at("/ffmpeg/ffmpeg.js")
.get(|_| resource(include_str!("../static/ffmpeg/ffmpeg.js"), "text/javascript", false).boxed());
app
.at("/ffmpeg/ffmpeg.js.map")
.get(|_| resource(include_str!("../static/ffmpeg/ffmpeg.js.map"), "text/javascript", false).boxed());
// Proxy media through Redlib
app.at("/vid/:id/:size").get(|r| proxy(r, "https://v.redd.it/{id}/DASH_{size}").boxed());
app.at("/hls/:id/*path").get(|r| proxy(r, "https://v.redd.it/{id}/{path}").boxed());
app.at("/img/*path").get(|r| proxy(r, "https://i.redd.it/{path}").boxed());
app.at("/thumb/:point/:id").get(|r| proxy(r, "https://{point}.thumbs.redditmedia.com/{id}").boxed());
app.at("/emoji/:id/:name").get(|r| proxy(r, "https://emoji.redditmedia.com/{id}/{name}").boxed());
app
.at("/emote/:subreddit_id/:filename")
.get(|r| proxy(r, "https://reddit-econ-prod-assets-permanent.s3.amazonaws.com/asset-manager/{subreddit_id}/{filename}").boxed());
app
.at("/preview/:loc/award_images/:fullname/:id")
.get(|r| proxy(r, "https://{loc}view.redd.it/award_images/{fullname}/{id}").boxed());
app.at("/preview/:loc/:id").get(|r| proxy(r, "https://{loc}view.redd.it/{id}").boxed());
app.at("/style/*path").get(|r| proxy(r, "https://styles.redditmedia.com/{path}").boxed());
app.at("/static/*path").get(|r| proxy(r, "https://www.redditstatic.com/{path}").boxed());
// Browse user profile
app
.at("/u/:name")
.get(|r| async move { Ok(redirect(&format!("/user/{}", r.param("name").unwrap_or_default()))) }.boxed());
app.at("/u/:name/comments/:id/:title").get(|r| post::item(r).boxed());
app.at("/u/:name/comments/:id/:title/:comment_id").get(|r| post::item(r).boxed());
app.at("/user/[deleted]").get(|req| error(req, "User has deleted their account").boxed());
app.at("/user/:name.rss").get(|r| user::rss(r).boxed());
app.at("/user/:name").get(|r| user::profile(r).boxed());
app.at("/user/:name/:listing").get(|r| user::profile(r).boxed());
app.at("/user/:name/comments/:id").get(|r| post::item(r).boxed());
app.at("/user/:name/comments/:id/:title").get(|r| post::item(r).boxed());
app.at("/user/:name/comments/:id/:title/:comment_id").get(|r| post::item(r).boxed());
// Configure settings
app.at("/settings").get(|r| settings::get(r).boxed()).post(|r| settings::set(r).boxed());
app.at("/settings/restore").get(|r| settings::restore(r).boxed());
app.at("/settings/update").get(|r| settings::update(r).boxed());
// Mascots
app.at("/mascot/:name").get(|r| mascot_image(r).boxed());
// RSS Subscriptions
app.at("/r/:sub.rss").get(|r| subreddit::rss(r).boxed());
// Subreddit services
app
.at("/r/:sub")
.get(|r| subreddit::community(r).boxed())
.post(|r| subreddit::add_quarantine_exception(r).boxed());
app
.at("/r/u_:name")
.get(|r| async move { Ok(redirect(&format!("/user/{}", r.param("name").unwrap_or_default()))) }.boxed());
app.at("/r/:sub/subscribe").post(|r| subreddit::subscriptions_filters(r).boxed());
app.at("/r/:sub/unsubscribe").post(|r| subreddit::subscriptions_filters(r).boxed());
app.at("/r/:sub/filter").post(|r| subreddit::subscriptions_filters(r).boxed());
app.at("/r/:sub/unfilter").post(|r| subreddit::subscriptions_filters(r).boxed());
app.at("/r/:sub/comments/:id").get(|r| post::item(r).boxed());
app.at("/r/:sub/comments/:id/:title").get(|r| post::item(r).boxed());
app.at("/r/:sub/comments/:id/:title/:comment_id").get(|r| post::item(r).boxed());
app.at("/comments/:id").get(|r| post::item(r).boxed());
app.at("/comments/:id/comments").get(|r| post::item(r).boxed());
app.at("/comments/:id/comments/:comment_id").get(|r| post::item(r).boxed());
app.at("/comments/:id/:title").get(|r| post::item(r).boxed());
app.at("/comments/:id/:title/:comment_id").get(|r| post::item(r).boxed());
app.at("/r/:sub/duplicates/:id").get(|r| duplicates::item(r).boxed());
app.at("/r/:sub/duplicates/:id/:title").get(|r| duplicates::item(r).boxed());
app.at("/duplicates/:id").get(|r| duplicates::item(r).boxed());
app.at("/duplicates/:id/:title").get(|r| duplicates::item(r).boxed());
app.at("/r/:sub/search").get(|r| search::find(r).boxed());
app
.at("/r/:sub/w")
.get(|r| async move { Ok(redirect(&format!("/r/{}/wiki", r.param("sub").unwrap_or_default()))) }.boxed());
app
.at("/r/:sub/w/*page")
.get(|r| async move { Ok(redirect(&format!("/r/{}/wiki/{}", r.param("sub").unwrap_or_default(), r.param("wiki").unwrap_or_default()))) }.boxed());
app.at("/r/:sub/wiki").get(|r| subreddit::wiki(r).boxed());
app.at("/r/:sub/wiki/*page").get(|r| subreddit::wiki(r).boxed());
app.at("/r/:sub/about/sidebar").get(|r| subreddit::sidebar(r).boxed());
app.at("/r/:sub/:sort").get(|r| subreddit::community(r).boxed());
// Front page
app.at("/").get(|r| subreddit::community(r).boxed());
// View Reddit wiki
app.at("/w").get(|_| async { Ok(redirect("/wiki")) }.boxed());
app
.at("/w/*page")
.get(|r| async move { Ok(redirect(&format!("/wiki/{}", r.param("page").unwrap_or_default()))) }.boxed());
app.at("/wiki").get(|r| subreddit::wiki(r).boxed());
app.at("/wiki/*page").get(|r| subreddit::wiki(r).boxed());
// Search all of Reddit
app.at("/search").get(|r| search::find(r).boxed());
// Handle about pages
app.at("/about").get(|req| error(req, "About pages aren't added yet").boxed());
// Instance info page
app.at("/info").get(|r| instance_info::instance_info(r).boxed());
app.at("/info.:extension").get(|r| instance_info::instance_info(r).boxed());
// Handle obfuscated share links.
// Note that this still forces the server to follow the share link to get to the post, so maybe this wants to be updated with a warning before it follow it
app.at("/r/:sub/s/:id").get(|req: Request<Body>| {
Box::pin(async move {
let sub = req.param("sub").unwrap_or_default();
match req.param("id").as_deref() {
// Share link
Some(id) if (8..12).contains(&id.len()) => match canonical_path(format!("/r/{sub}/s/{id}"), 3).await {
Ok(Some(path)) => Ok(redirect(&path)),
Ok(None) => error(req, "Post ID is invalid. It may point to a post on a community that has been banned.").await,
Err(e) => error(req, &e).await,
},
// Error message for unknown pages
_ => error(req, "Nothing here").await,
}
})
});
app.at("/:id").get(|req: Request<Body>| {
Box::pin(async move {
match req.param("id").as_deref() {
// Sort front page
Some("best" | "hot" | "new" | "top" | "rising" | "controversial") => subreddit::community(req).await,
// Short link for post
Some(id) if (5..8).contains(&id.len()) => match canonical_path(format!("/{id}"), 3).await {
Ok(path_opt) => match path_opt {
Some(path) => Ok(redirect(&path)),
None => error(req, "Post ID is invalid. It may point to a post on a community that has been banned.").await,
},
Err(e) => error(req, &e).await,
},
// Error message for unknown pages
_ => error(req, "Nothing here").await,
}
})
});
// Default service in case no routes match
app.at("/*").get(|req| error(req, "Nothing here").boxed());
println!("Running Redsunlib v{} on {listener}!", env!("CARGO_PKG_VERSION"));
let server = app.listen(&listener);
// Run this server for... forever!
if let Err(e) = server.await {
eprintln!("Server error: {e}");
}
}

235
src/oauth.rs Normal file
View File

@ -0,0 +1,235 @@
use std::{collections::HashMap, sync::atomic::Ordering, time::Duration};
use crate::{
client::{CLIENT, OAUTH_CLIENT, OAUTH_IS_ROLLING_OVER, OAUTH_RATELIMIT_REMAINING},
oauth_resources::ANDROID_APP_VERSION_LIST,
};
use base64::{engine::general_purpose, Engine as _};
use hyper::{client, Body, Method, Request};
use log::{error, info, trace};
use serde_json::json;
use tokio::time::{error::Elapsed, timeout};
static REDDIT_ANDROID_OAUTH_CLIENT_ID: &str = "ohXpoqrZYub1kg";
static AUTH_ENDPOINT: &str = "https://www.reddit.com";
// Spoofed client for Android devices
#[derive(Debug, Clone, Default)]
pub struct Oauth {
pub(crate) initial_headers: HashMap<String, String>,
pub(crate) headers_map: HashMap<String, String>,
pub(crate) token: String,
expires_in: u64,
device: Device,
}
impl Oauth {
/// Create a new OAuth client
pub(crate) async fn new() -> Self {
// Call new_internal until it succeeds
loop {
let attempt = Self::new_with_timeout().await;
match attempt {
Ok(Some(oauth)) => {
info!("[✅] Successfully created OAuth client");
return oauth;
}
Ok(None) => {
error!("Failed to create OAuth client. Retrying in 5 seconds...");
continue;
}
Err(duration) => {
error!("Failed to create OAuth client in {duration:?}. Retrying in 5 seconds...");
}
}
}
}
async fn new_with_timeout() -> Result<Option<Self>, Elapsed> {
let mut oauth = Self::default();
timeout(Duration::from_secs(5), oauth.login()).await.map(|result| result.map(|_| oauth))
}
pub(crate) fn default() -> Self {
// Generate a device to spoof
let device = Device::new();
let headers_map = device.headers.clone();
let initial_headers = device.initial_headers.clone();
// For now, just insert headers - no token request
Self {
headers_map,
initial_headers,
token: String::new(),
expires_in: 0,
device,
}
}
async fn login(&mut self) -> Option<()> {
// Construct URL for OAuth token
let url = format!("{AUTH_ENDPOINT}/auth/v2/oauth/access-token/loid");
let mut builder = Request::builder().method(Method::POST).uri(&url);
// Add headers from spoofed client
for (key, value) in &self.initial_headers {
builder = builder.header(key, value);
}
// Set up HTTP Basic Auth - basically just the const OAuth ID's with no password,
// Base64-encoded. https://en.wikipedia.org/wiki/Basic_access_authentication
// This could be constant, but I don't think it's worth it. OAuth ID's can change
// over time and we want to be flexible.
let auth = general_purpose::STANDARD.encode(format!("{}:", self.device.oauth_id));
builder = builder.header("Authorization", format!("Basic {auth}"));
// Set JSON body. I couldn't tell you what this means. But that's what the client sends
let json = json!({
"scopes": ["*","email"]
});
let body = Body::from(json.to_string());
// Build request
let request = builder.body(body).unwrap();
trace!("Sending token request...");
// Send request
let client: client::Client<_, Body> = CLIENT.clone();
let resp = client.request(request).await.ok()?;
trace!("Received response with status {} and length {:?}", resp.status(), resp.headers().get("content-length"));
// Parse headers - loid header _should_ be saved sent on subsequent token refreshes.
// Technically it's not needed, but it's easy for Reddit API to check for this.
// It's some kind of header that uniquely identifies the device.
// Not worried about the privacy implications, since this is randomly changed
// and really only as privacy-concerning as the OAuth token itself.
if let Some(header) = resp.headers().get("x-reddit-loid") {
self.headers_map.insert("x-reddit-loid".to_owned(), header.to_str().ok()?.to_string());
}
// Same with x-reddit-session
if let Some(header) = resp.headers().get("x-reddit-session") {
self.headers_map.insert("x-reddit-session".to_owned(), header.to_str().ok()?.to_string());
}
trace!("Serializing response...");
// Serialize response
let body_bytes = hyper::body::to_bytes(resp.into_body()).await.ok()?;
let json: serde_json::Value = serde_json::from_slice(&body_bytes).ok()?;
trace!("Accessing relevant fields...");
// Save token and expiry
self.token = json.get("access_token")?.as_str()?.to_string();
self.expires_in = json.get("expires_in")?.as_u64()?;
self.headers_map.insert("Authorization".to_owned(), format!("Bearer {}", self.token));
info!("[✅] Success - Retrieved token \"{}...\", expires in {}", &self.token[..32], self.expires_in);
Some(())
}
}
pub async fn token_daemon() {
// Monitor for refreshing token
loop {
// Get expiry time - be sure to not hold the read lock
let expires_in = { OAUTH_CLIENT.load_full().expires_in };
// sleep for the expiry time minus 2 minutes
let duration = Duration::from_secs(expires_in - 120);
info!("[⏳] Waiting for {duration:?} seconds before refreshing OAuth token...");
tokio::time::sleep(duration).await;
info!("[⌛] {duration:?} Elapsed! Refreshing OAuth token...");
// Refresh token - in its own scope
{
force_refresh_token().await;
}
}
}
pub async fn force_refresh_token() {
if OAUTH_IS_ROLLING_OVER.compare_exchange(false, true, Ordering::SeqCst, Ordering::SeqCst).is_err() {
trace!("Skipping refresh token roll over, already in progress");
return;
}
trace!("Rolling over refresh token. Current rate limit: {}", OAUTH_RATELIMIT_REMAINING.load(Ordering::SeqCst));
let new_client = Oauth::new().await;
OAUTH_CLIENT.swap(new_client.into());
OAUTH_RATELIMIT_REMAINING.store(99, Ordering::SeqCst);
OAUTH_IS_ROLLING_OVER.store(false, Ordering::SeqCst);
}
#[derive(Debug, Clone, Default)]
struct Device {
oauth_id: String,
initial_headers: HashMap<String, String>,
headers: HashMap<String, String>,
}
impl Device {
fn android() -> Self {
// Generate uuid
let uuid = uuid::Uuid::new_v4().to_string();
// Generate random user-agent
let android_app_version = choose(ANDROID_APP_VERSION_LIST).to_string();
let android_version = fastrand::u8(9..=14);
let android_user_agent = format!("Reddit/{android_app_version}/Android {android_version}");
// Android device headers
let headers = HashMap::from([
("Client-Vendor-Id".into(), uuid.clone()),
("X-Reddit-Device-Id".into(), uuid.clone()),
("User-Agent".into(), android_user_agent),
]);
info!("[🔄] Spoofing Android client with headers: {headers:?}, uuid: \"{uuid}\", and OAuth ID \"{REDDIT_ANDROID_OAUTH_CLIENT_ID}\"");
Self {
oauth_id: REDDIT_ANDROID_OAUTH_CLIENT_ID.to_string(),
headers: headers.clone(),
initial_headers: headers,
}
}
fn new() -> Self {
// See https://github.com/redlib-org/redlib/issues/8
Self::android()
}
}
fn choose<T: Copy>(list: &[T]) -> T {
*fastrand::choose_multiple(list.iter(), 1)[0]
}
#[tokio::test(flavor = "multi_thread")]
async fn test_oauth_client() {
assert!(!OAUTH_CLIENT.load_full().token.is_empty());
}
#[tokio::test(flavor = "multi_thread")]
async fn test_oauth_client_refresh() {
force_refresh_token().await;
}
#[tokio::test(flavor = "multi_thread")]
async fn test_oauth_token_exists() {
assert!(!OAUTH_CLIENT.load_full().token.is_empty());
}
#[tokio::test(flavor = "multi_thread")]
async fn test_oauth_headers_len() {
assert!(OAUTH_CLIENT.load_full().headers_map.len() >= 3);
}
#[test]
fn test_creating_device() {
Device::new();
}

158
src/oauth_resources.rs Normal file
View File

@ -0,0 +1,158 @@
// This file was generated by scripts/update_oauth_resources.sh
// Rerun scripts/update_oauth_resources.sh to update this file
// Please do not edit manually
// Filled in with real app versions
pub static _IOS_APP_VERSION_LIST: &[&str; 1] = &[""];
pub static ANDROID_APP_VERSION_LIST: &[&str; 150] = &[
"Version 2023.48.0/Build 1319123",
"Version 2023.49.0/Build 1321715",
"Version 2023.49.1/Build 1322281",
"Version 2023.50.0/Build 1332338",
"Version 2023.50.1/Build 1345844",
"Version 2024.02.0/Build 1368985",
"Version 2024.03.0/Build 1379408",
"Version 2024.04.0/Build 1391236",
"Version 2024.05.0/Build 1403584",
"Version 2024.06.0/Build 1418489",
"Version 2024.07.0/Build 1429651",
"Version 2024.08.0/Build 1439531",
"Version 2024.10.0/Build 1470045",
"Version 2024.10.1/Build 1478645",
"Version 2024.11.0/Build 1480707",
"Version 2024.12.0/Build 1494694",
"Version 2024.13.0/Build 1505187",
"Version 2024.14.0/Build 1520556",
"Version 2024.15.0/Build 1536823",
"Version 2024.16.0/Build 1551366",
"Version 2024.17.0/Build 1568106",
"Version 2024.18.0/Build 1577901",
"Version 2024.18.1/Build 1585304",
"Version 2024.19.0/Build 1593346",
"Version 2024.20.0/Build 1612800",
"Version 2024.20.1/Build 1615586",
"Version 2024.20.2/Build 1624969",
"Version 2024.21.0/Build 1631686",
"Version 2024.22.0/Build 1645257",
"Version 2024.22.1/Build 1652272",
"Version 2023.21.0/Build 956283",
"Version 2023.22.0/Build 968223",
"Version 2023.23.0/Build 983896",
"Version 2023.24.0/Build 998541",
"Version 2023.25.0/Build 1014750",
"Version 2023.25.1/Build 1018737",
"Version 2023.26.0/Build 1019073",
"Version 2023.27.0/Build 1031923",
"Version 2023.28.0/Build 1046887",
"Version 2023.29.0/Build 1059855",
"Version 2023.30.0/Build 1078734",
"Version 2023.31.0/Build 1091027",
"Version 2023.32.0/Build 1109919",
"Version 2023.32.1/Build 1114141",
"Version 2023.33.1/Build 1129741",
"Version 2023.34.0/Build 1144243",
"Version 2023.35.0/Build 1157967",
"Version 2023.36.0/Build 1168982",
"Version 2023.37.0/Build 1182743",
"Version 2023.38.0/Build 1198522",
"Version 2023.39.0/Build 1211607",
"Version 2023.39.1/Build 1221505",
"Version 2023.40.0/Build 1221521",
"Version 2023.41.0/Build 1233125",
"Version 2023.41.1/Build 1239615",
"Version 2023.42.0/Build 1245088",
"Version 2023.43.0/Build 1257426",
"Version 2023.44.0/Build 1268622",
"Version 2023.45.0/Build 1281371",
"Version 2023.47.0/Build 1303604",
"Version 2022.42.0/Build 638508",
"Version 2022.43.0/Build 648277",
"Version 2022.44.0/Build 664348",
"Version 2022.45.0/Build 677985",
"Version 2023.01.0/Build 709875",
"Version 2023.02.0/Build 717912",
"Version 2023.03.0/Build 729220",
"Version 2023.04.0/Build 744681",
"Version 2023.05.0/Build 755453",
"Version 2023.06.0/Build 775017",
"Version 2023.07.0/Build 788827",
"Version 2023.07.1/Build 790267",
"Version 2023.08.0/Build 798718",
"Version 2023.09.0/Build 812015",
"Version 2023.09.1/Build 816833",
"Version 2023.10.0/Build 821148",
"Version 2023.11.0/Build 830610",
"Version 2023.12.0/Build 841150",
"Version 2023.13.0/Build 852246",
"Version 2023.14.0/Build 861593",
"Version 2023.14.1/Build 864826",
"Version 2023.15.0/Build 870628",
"Version 2023.16.0/Build 883294",
"Version 2023.16.1/Build 886269",
"Version 2023.17.0/Build 896030",
"Version 2023.17.1/Build 900542",
"Version 2023.18.0/Build 911877",
"Version 2023.19.0/Build 927681",
"Version 2023.20.0/Build 943980",
"Version 2023.20.1/Build 946732",
"Version 2022.20.0/Build 487703",
"Version 2022.21.0/Build 492436",
"Version 2022.22.0/Build 498700",
"Version 2022.23.0/Build 502374",
"Version 2022.23.1/Build 506606",
"Version 2022.24.0/Build 510950",
"Version 2022.24.1/Build 513462",
"Version 2022.25.0/Build 515072",
"Version 2022.25.1/Build 516394",
"Version 2022.25.2/Build 519915",
"Version 2022.26.0/Build 521193",
"Version 2022.27.0/Build 527406",
"Version 2022.27.1/Build 529687",
"Version 2022.28.0/Build 533235",
"Version 2022.30.0/Build 548620",
"Version 2022.31.0/Build 556666",
"Version 2022.31.1/Build 562612",
"Version 2022.32.0/Build 567875",
"Version 2022.33.0/Build 572600",
"Version 2022.34.0/Build 579352",
"Version 2022.35.0/Build 588016",
"Version 2022.35.1/Build 589034",
"Version 2022.36.0/Build 593102",
"Version 2022.37.0/Build 601691",
"Version 2022.38.0/Build 607460",
"Version 2022.39.0/Build 615385",
"Version 2022.39.1/Build 619019",
"Version 2022.40.0/Build 624782",
"Version 2022.41.0/Build 630468",
"Version 2022.41.1/Build 634168",
"Version 2021.39.1/Build 372418",
"Version 2021.41.0/Build 376052",
"Version 2021.42.0/Build 378193",
"Version 2021.43.0/Build 382019",
"Version 2021.44.0/Build 385129",
"Version 2021.45.0/Build 387663",
"Version 2021.46.0/Build 392043",
"Version 2021.47.0/Build 394342",
"Version 2022.10.0/Build 429896",
"Version 2022.1.0/Build 402829",
"Version 2022.11.0/Build 433004",
"Version 2022.12.0/Build 436848",
"Version 2022.13.0/Build 442084",
"Version 2022.13.1/Build 444621",
"Version 2022.14.1/Build 452742",
"Version 2022.15.0/Build 455453",
"Version 2022.16.0/Build 462377",
"Version 2022.17.0/Build 468480",
"Version 2022.18.0/Build 473740",
"Version 2022.19.1/Build 482464",
"Version 2022.2.0/Build 405543",
"Version 2022.3.0/Build 408637",
"Version 2022.4.0/Build 411368",
"Version 2022.5.0/Build 414731",
"Version 2022.6.0/Build 418391",
"Version 2022.6.1/Build 419585",
"Version 2022.6.2/Build 420562",
"Version 2022.7.0/Build 420849",
"Version 2022.8.0/Build 423906",
"Version 2022.9.0/Build 426592",
];
pub static _IOS_OS_VERSION_LIST: &[&str; 1] = &[""];

View File

@ -1,55 +0,0 @@
// CRATES
use crate::utils::{fetch_posts, ErrorTemplate, Params, Post};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use askama::Template;
// STRUCTS
#[derive(Template)]
#[template(path = "popular.html", escape = "none")]
struct PopularTemplate {
posts: Vec<Post>,
sort: String,
ends: (String, String),
}
// RENDER
async fn render(sub_name: String, sort: Option<String>, ends: (Option<String>, Option<String>)) -> Result<HttpResponse> {
let sorting = sort.unwrap_or("hot".to_string());
let before = ends.1.clone().unwrap_or(String::new()); // If there is an after, there must be a before
// Build the Reddit JSON API url
let url = match ends.0 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?before={}&count=25", sub_name, sorting, val),
None => match ends.1 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?after={}&count=25", sub_name, sorting, val),
None => format!("https://www.reddit.com/r/{}/{}.json", sub_name, sorting),
},
};
let items_result = fetch_posts(url, String::new()).await;
if items_result.is_err() {
let s = ErrorTemplate {
message: items_result.err().unwrap().to_string(),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s))
} else {
let items = items_result.unwrap();
let s = PopularTemplate {
posts: items.0,
sort: sorting,
ends: (before, items.1),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
}
}
// SERVICES
pub async fn page(params: web::Query<Params>) -> Result<HttpResponse> {
render("popular".to_string(), params.sort.clone(), (params.before.clone(), params.after.clone())).await
}

View File

@ -1,181 +1,255 @@
// CRATES
use crate::utils::{format_num, format_url, request, val, Comment, ErrorTemplate, Flair, Params, Post};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use crate::client::json;
use crate::config::get_setting;
use crate::server::RequestExt;
use crate::subreddit::{can_access_quarantine, quarantine};
use crate::utils::{
error, format_num, get_filters, nsfw_landing, param, parse_post, rewrite_emotes, setting, template, time, val, Author, Awards, Comment, Flair, FlairPart, Post, Preferences,
};
use hyper::{Body, Request, Response};
use async_recursion::async_recursion;
use askama::Template;
use chrono::{TimeZone, Utc};
use pulldown_cmark::{html, Options, Parser};
use once_cell::sync::Lazy;
use regex::Regex;
use rinja::Template;
use std::collections::{HashMap, HashSet};
// STRUCTS
#[derive(Template)]
#[template(path = "post.html", escape = "none")]
#[template(path = "post.html")]
struct PostTemplate {
comments: Vec<Comment>,
post: Post,
sort: String,
prefs: Preferences,
single_thread: bool,
url: String,
url_without_query: String,
comment_query: String,
}
async fn render(id: String, sort: Option<String>) -> Result<HttpResponse> {
// Log the post ID being fetched
dbg!(&id);
static COMMENT_SEARCH_CAPTURE: Lazy<Regex> = Lazy::new(|| Regex::new(r"\?q=(.*)&type=comment").unwrap());
// Handling sort paramater
let sorting: String = sort.unwrap_or("confidence".to_string());
pub async fn item(req: Request<Body>) -> Result<Response<Body>, String> {
// Build Reddit API path
let mut path: String = format!("{}.json?{}&raw_json=1", req.uri().path(), req.uri().query().unwrap_or_default());
let sub = req.param("sub").unwrap_or_default();
let quarantined = can_access_quarantine(&req, &sub);
let url = req.uri().to_string();
// Build the Reddit JSON API url
let url: String = format!("https://reddit.com/{}.json?sort={}", id, sorting);
// Set sort to sort query parameter
let sort = param(&path, "sort").unwrap_or_else(|| {
// Grab default comment sort method from Cookies
let default_sort = setting(&req, "comment_sort");
// If there's no sort query but there's a default sort, set sort to default_sort
if default_sort.is_empty() {
String::new()
} else {
path = format!("{}.json?{}&sort={}&raw_json=1", req.uri().path(), req.uri().query().unwrap_or_default(), default_sort);
default_sort
}
});
// Log the post ID being fetched in debug mode
#[cfg(debug_assertions)]
req.param("id").unwrap_or_default();
let single_thread = req.param("comment_id").is_some();
let highlighted_comment = &req.param("comment_id").unwrap_or_default();
// Send a request to the url, receive JSON in response
let req = request(url).await;
match json(path, quarantined).await {
// Otherwise, grab the JSON output from the request
Ok(response) => {
// Parse the JSON into Post and Comment structs
let post = parse_post(&response[0]["data"]["children"][0]).await;
// If the Reddit API returns an error, exit and send error page to user
if req.is_err() {
let s = ErrorTemplate {
message: req.err().unwrap().to_string(),
let req_url = req.uri().to_string();
// Return landing page if this post if this Reddit deems this post
// NSFW, but we have also disabled the display of NSFW content
// or if the instance is SFW-only.
if post.nsfw && crate::utils::should_be_nsfw_gated(&req, &req_url) {
return Ok(nsfw_landing(req, req_url).await.unwrap_or_default());
}
let query_body = match COMMENT_SEARCH_CAPTURE.captures(&url) {
Some(captures) => captures.get(1).unwrap().as_str().replace("%20", " ").replace('+', " "),
None => String::new(),
};
let query_string = format!("q={query_body}&type=comment");
let form = url::form_urlencoded::parse(query_string.as_bytes()).collect::<HashMap<_, _>>();
let query = form.get("q").unwrap().clone().to_string();
let comments = match query.as_str() {
"" => parse_comments(&response[1], &post.permalink, &post.author.name, highlighted_comment, &get_filters(&req), &req),
_ => query_comments(&response[1], &post.permalink, &post.author.name, highlighted_comment, &get_filters(&req), &query, &req),
};
// Use the Post and Comment structs to generate a website to show users
Ok(template(&PostTemplate {
comments,
post,
url_without_query: url.clone().trim_end_matches(&format!("?q={query}&type=comment")).to_string(),
sort,
prefs: Preferences::new(&req),
single_thread,
url: req_url,
comment_query: query,
}))
}
.render()
.unwrap();
return Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s));
}
// Otherwise, grab the JSON output from the request
let res = req.unwrap();
// Parse the JSON into Post and Comment structs
let post = parse_post(res[0].clone()).await;
let comments = parse_comments(res[1].clone()).await;
// Use the Post and Comment structs to generate a website to show users
let s = PostTemplate {
comments: comments.unwrap(),
post: post.unwrap(),
sort: sorting,
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
}
// SERVICES
pub async fn short(web::Path(id): web::Path<String>, params: web::Query<Params>) -> Result<HttpResponse> {
render(id, params.sort.clone()).await
}
pub async fn page(web::Path((_sub, id)): web::Path<(String, String)>, params: web::Query<Params>) -> Result<HttpResponse> {
render(id, params.sort.clone()).await
}
// UTILITIES
async fn media(data: &serde_json::Value) -> (String, String) {
let post_type: &str;
let url = if !data["preview"]["reddit_video_preview"]["fallback_url"].is_null() {
post_type = "video";
format_url(data["preview"]["reddit_video_preview"]["fallback_url"].as_str().unwrap()).await
} else if !data["secure_media"]["reddit_video"]["fallback_url"].is_null() {
post_type = "video";
format_url(data["secure_media"]["reddit_video"]["fallback_url"].as_str().unwrap()).await
} else if data["post_hint"].as_str().unwrap_or("") == "image" {
post_type = "image";
format_url(data["preview"]["images"][0]["source"]["url"].as_str().unwrap()).await
} else {
post_type = "link";
data["url"].as_str().unwrap().to_string()
};
(post_type.to_string(), url)
}
async fn markdown_to_html(md: &str) -> String {
let mut options = Options::empty();
options.insert(Options::ENABLE_TABLES);
options.insert(Options::ENABLE_FOOTNOTES);
options.insert(Options::ENABLE_STRIKETHROUGH);
options.insert(Options::ENABLE_TASKLISTS);
let parser = Parser::new_ext(md, options);
// Write to String buffer.
let mut html_output = String::new();
html::push_html(&mut html_output, parser);
html_output
}
// POSTS
async fn parse_post(json: serde_json::Value) -> Result<Post, &'static str> {
let post_data: &serde_json::Value = &json["data"]["children"][0];
let unix_time: i64 = post_data["data"]["created_utc"].as_f64().unwrap().round() as i64;
let score = post_data["data"]["score"].as_i64().unwrap();
let media = media(&post_data["data"]).await;
let post = Post {
title: val(post_data, "title").await,
community: val(post_data, "subreddit").await,
body: markdown_to_html(post_data["data"]["selftext"].as_str().unwrap()).await,
author: val(post_data, "author").await,
author_flair: Flair(
val(post_data, "author_flair_text").await,
val(post_data, "author_flair_background_color").await,
val(post_data, "author_flair_text_color").await,
),
url: val(post_data, "permalink").await,
score: format_num(score),
post_type: media.0,
media: media.1,
time: Utc.timestamp(unix_time, 0).format("%b %e %Y %H:%M UTC").to_string(),
flair: Flair(
val(post_data, "link_flair_text").await,
val(post_data, "link_flair_background_color").await,
if val(post_data, "link_flair_text_color").await == "dark" {
"black".to_string()
// If the Reddit API returns an error, exit and send error page to user
Err(msg) => {
if msg == "quarantined" || msg == "gated" {
let sub = req.param("sub").unwrap_or_default();
Ok(quarantine(&req, sub, &msg))
} else {
"white".to_string()
},
)
};
Ok(post)
error(req, &msg).await
}
}
}
}
// COMMENTS
#[async_recursion]
async fn parse_comments(json: serde_json::Value) -> Result<Vec<Comment>, &'static str> {
// Separate the comment JSON into a Vector of comments
let comment_data = json["data"]["children"].as_array().unwrap();
let mut comments: Vec<Comment> = Vec::new();
fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, highlighted_comment: &str, filters: &HashSet<String>, req: &Request<Body>) -> Vec<Comment> {
// Parse the comment JSON into a Vector of Comments
let comments = json["data"]["children"].as_array().map_or(Vec::new(), std::borrow::ToOwned::to_owned);
// For each comment, retrieve the values to build a Comment object
for comment in comment_data.iter() {
let unix_time: i64 = comment["data"]["created_utc"].as_f64().unwrap_or(0.0).round() as i64;
if unix_time == 0 {
continue;
comments
.into_iter()
.map(|comment| {
let data = &comment["data"];
let replies: Vec<Comment> = if data["replies"].is_object() {
parse_comments(&data["replies"], post_link, post_author, highlighted_comment, filters, req)
} else {
Vec::new()
};
build_comment(&comment, data, replies, post_link, post_author, highlighted_comment, filters, req)
})
.collect()
}
fn query_comments(
json: &serde_json::Value,
post_link: &str,
post_author: &str,
highlighted_comment: &str,
filters: &HashSet<String>,
query: &str,
req: &Request<Body>,
) -> Vec<Comment> {
let comments = json["data"]["children"].as_array().map_or(Vec::new(), std::borrow::ToOwned::to_owned);
let mut results = Vec::new();
for comment in comments {
let data = &comment["data"];
// If this comment contains replies, handle those too
if data["replies"].is_object() {
results.append(&mut query_comments(&data["replies"], post_link, post_author, highlighted_comment, filters, query, req));
}
let score = comment["data"]["score"].as_i64().unwrap_or(0);
let body = markdown_to_html(comment["data"]["body"].as_str().unwrap_or("")).await;
let replies: Vec<Comment> = if comment["data"]["replies"].is_object() {
parse_comments(comment["data"]["replies"].clone()).await.unwrap_or(Vec::new())
} else {
Vec::new()
};
comments.push(Comment {
body: body,
author: val(comment, "author").await,
score: format_num(score),
time: Utc.timestamp(unix_time, 0).format("%b %e %Y %H:%M UTC").to_string(),
replies: replies,
flair: Flair(
val(comment, "author_flair_text").await,
val(comment, "author_flair_background_color").await,
val(comment, "author_flair_text_color").await,
),
});
let c = build_comment(&comment, data, Vec::new(), post_link, post_author, highlighted_comment, filters, req);
if c.body.to_lowercase().contains(&query.to_lowercase()) {
results.push(c);
}
}
Ok(comments)
results
}
#[allow(clippy::too_many_arguments)]
fn build_comment(
comment: &serde_json::Value,
data: &serde_json::Value,
replies: Vec<Comment>,
post_link: &str,
post_author: &str,
highlighted_comment: &str,
filters: &HashSet<String>,
req: &Request<Body>,
) -> Comment {
let id = val(comment, "id");
let body = if (val(comment, "author") == "[deleted]" && val(comment, "body") == "[removed]") || val(comment, "body") == "[ Removed by Reddit ]" {
format!(
"<div class=\"md\"><p>[removed] — <a href=\"https://{}{post_link}{id}\">view removed comment</a></p></div>",
get_setting("REDLIB_PUSHSHIFT_FRONTEND").unwrap_or_else(|| String::from(crate::config::DEFAULT_PUSHSHIFT_FRONTEND)),
)
} else {
rewrite_emotes(&data["media_metadata"], val(comment, "body_html"))
};
let kind = comment["kind"].as_str().unwrap_or_default().to_string();
let unix_time = data["created_utc"].as_f64().unwrap_or_default();
let (rel_time, created) = time(unix_time);
let edited = data["edited"].as_f64().map_or((String::new(), String::new()), time);
let score = data["score"].as_i64().unwrap_or(0);
// The JSON API only provides comments up to some threshold.
// Further comments have to be loaded by subsequent requests.
// The "kind" value will be "more" and the "count"
// shows how many more (sub-)comments exist in the respective nesting level.
// Note that in certain (seemingly random) cases, the count is simply wrong.
let more_count = data["count"].as_i64().unwrap_or_default();
let awards: Awards = Awards::parse(&data["all_awardings"]);
let parent_kind_and_id = val(comment, "parent_id");
let parent_info = parent_kind_and_id.split('_').collect::<Vec<&str>>();
let highlighted = id == highlighted_comment;
let author = Author {
name: val(comment, "author"),
flair: Flair {
flair_parts: FlairPart::parse(
data["author_flair_type"].as_str().unwrap_or_default(),
data["author_flair_richtext"].as_array(),
data["author_flair_text"].as_str(),
),
text: val(comment, "link_flair_text"),
background_color: val(comment, "author_flair_background_color"),
foreground_color: val(comment, "author_flair_text_color"),
},
distinguished: val(comment, "distinguished"),
};
let is_filtered = filters.contains(&["u_", author.name.as_str()].concat());
// Many subreddits have a default comment posted about the sub's rules etc.
// Many Redlib users do not wish to see this kind of comment by default.
// Reddit does not tell us which users are "bots", so a good heuristic is to
// collapse stickied moderator comments.
let is_moderator_comment = data["distinguished"].as_str().unwrap_or_default() == "moderator";
let is_stickied = data["stickied"].as_bool().unwrap_or_default();
let collapsed = (is_moderator_comment && is_stickied) || is_filtered;
Comment {
id,
kind,
parent_id: parent_info[1].to_string(),
parent_kind: parent_info[0].to_string(),
post_link: post_link.to_string(),
post_author: post_author.to_string(),
body,
author,
score: if data["score_hidden"].as_bool().unwrap_or_default() {
("\u{2022}".to_string(), "Hidden".to_string())
} else {
format_num(score)
},
rel_time,
created,
edited,
replies,
highlighted,
awards,
collapsed,
is_filtered,
more_count,
prefs: Preferences::new(req),
}
}

View File

@ -1,29 +0,0 @@
use actix_web::{client::Client, web, Error, HttpResponse, Result};
#[cfg(feature = "proxy")]
use base64::decode;
pub async fn handler(web::Path(url): web::Path<String>) -> Result<HttpResponse> {
if cfg!(feature = "proxy") {
let media: String;
#[cfg(not(feature = "proxy"))]
let media = url;
#[cfg(feature = "proxy")]
match decode(url) {
Ok(bytes) => media = String::from_utf8(bytes).unwrap(),
Err(_e) => return Ok(HttpResponse::Ok().body("")),
};
let client = Client::default();
client
.get(media.replace("&amp;", "&"))
.send()
.await
.map_err(Error::from)
.and_then(|res| Ok(HttpResponse::build(res.status()).streaming(res)))
} else {
Ok(HttpResponse::Ok().body(""))
}
}

188
src/search.rs Normal file
View File

@ -0,0 +1,188 @@
// CRATES
use crate::utils::{self, catch_random, error, filter_posts, format_num, format_url, get_filters, param, redirect, setting, template, val, Post, Preferences};
use crate::{
client::json,
subreddit::{can_access_quarantine, quarantine},
RequestExt,
};
use hyper::{Body, Request, Response};
use once_cell::sync::Lazy;
use regex::Regex;
use rinja::Template;
// STRUCTS
struct SearchParams {
q: String,
sort: String,
t: String,
before: String,
after: String,
restrict_sr: String,
typed: String,
}
// STRUCTS
struct Subreddit {
name: String,
url: String,
icon: String,
description: String,
subscribers: (String, String),
}
#[derive(Template)]
#[template(path = "search.html")]
struct SearchTemplate {
posts: Vec<Post>,
subreddits: Vec<Subreddit>,
sub: String,
params: SearchParams,
prefs: Preferences,
url: String,
/// Whether the subreddit itself is filtered.
is_filtered: bool,
/// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place,
/// and all fetched posts being filtered).
all_posts_filtered: bool,
/// Whether all posts were hidden because they are NSFW (and user has disabled show NSFW)
all_posts_hidden_nsfw: bool,
no_posts: bool,
}
// Regex matched against search queries to determine if they are reddit urls.
static REDDIT_URL_MATCH: Lazy<Regex> = Lazy::new(|| Regex::new(r"^https?://([^\./]+\.)*reddit.com/").unwrap());
// SERVICES
pub async fn find(req: Request<Body>) -> Result<Response<Body>, String> {
// This ensures that during a search, no NSFW posts are fetched at all
let nsfw_results = if setting(&req, "show_nsfw") == "on" && !utils::sfw_only() {
"&include_over_18=on"
} else {
""
};
let uri_path = req.uri().path().replace("+", "%2B");
let path = format!("{}.json?{}{}&raw_json=1", uri_path, req.uri().query().unwrap_or_default(), nsfw_results);
let mut query = param(&path, "q").unwrap_or_default();
query = REDDIT_URL_MATCH.replace(&query, "").to_string();
if query.is_empty() {
return Ok(redirect("/"));
}
if query.starts_with("r/") || query.starts_with("user/") {
return Ok(redirect(&format!("/{query}")));
}
if query.starts_with("u/") {
return Ok(redirect(&format!("/user{}", &query[1..])));
}
let sub = req.param("sub").unwrap_or_default();
let quarantined = can_access_quarantine(&req, &sub);
// Handle random subreddits
if let Ok(random) = catch_random(&sub, "/find").await {
return Ok(random);
}
let typed = param(&path, "type").unwrap_or_default();
let sort = param(&path, "sort").unwrap_or_else(|| "relevance".to_string());
let filters = get_filters(&req);
// If search is not restricted to this subreddit, show other subreddits in search results
let subreddits = if param(&path, "restrict_sr").is_none() {
let mut subreddits = search_subreddits(&query, &typed).await;
subreddits.retain(|s| !filters.contains(s.name.as_str()));
subreddits
} else {
Vec::new()
};
let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str()));
// If all requested subs are filtered, we don't need to fetch posts.
if sub.split('+').all(|s| filters.contains(s)) {
Ok(template(&SearchTemplate {
posts: Vec::new(),
subreddits,
sub,
params: SearchParams {
q: query.replace('"', "&quot;"),
sort,
t: param(&path, "t").unwrap_or_default(),
before: param(&path, "after").unwrap_or_default(),
after: String::new(),
restrict_sr: param(&path, "restrict_sr").unwrap_or_default(),
typed,
},
prefs: Preferences::new(&req),
url,
is_filtered: true,
all_posts_filtered: false,
all_posts_hidden_nsfw: false,
no_posts: false,
}))
} else {
match Post::fetch(&path, quarantined).await {
Ok((mut posts, after)) => {
let (_, all_posts_filtered) = filter_posts(&mut posts, &filters);
let no_posts = posts.is_empty();
let all_posts_hidden_nsfw = !no_posts && (posts.iter().all(|p| p.flags.nsfw) && setting(&req, "show_nsfw") != "on");
Ok(template(&SearchTemplate {
posts,
subreddits,
sub,
params: SearchParams {
q: query.replace('"', "&quot;"),
sort,
t: param(&path, "t").unwrap_or_default(),
before: param(&path, "after").unwrap_or_default(),
after,
restrict_sr: param(&path, "restrict_sr").unwrap_or_default(),
typed,
},
prefs: Preferences::new(&req),
url,
is_filtered: false,
all_posts_filtered,
all_posts_hidden_nsfw,
no_posts,
}))
}
Err(msg) => {
if msg == "quarantined" || msg == "gated" {
let sub = req.param("sub").unwrap_or_default();
Ok(quarantine(&req, sub, &msg))
} else {
error(req, &msg).await
}
}
}
}
}
async fn search_subreddits(q: &str, typed: &str) -> Vec<Subreddit> {
let limit = if typed == "sr_user" { "50" } else { "3" };
let subreddit_search_path = format!("/subreddits/search.json?q={}&limit={limit}", q.replace(' ', "+"));
// Send a request to the url
json(subreddit_search_path, false).await.unwrap_or_default()["data"]["children"]
.as_array()
.map(ToOwned::to_owned)
.unwrap_or_default()
.iter()
.map(|subreddit| {
// For each subreddit from subreddit list
// Fetch subreddit icon either from the community_icon or icon_img value
let icon = subreddit["data"]["community_icon"].as_str().map_or_else(|| val(subreddit, "icon_img"), ToString::to_string);
Subreddit {
name: val(subreddit, "display_name"),
url: val(subreddit, "url"),
icon: format_url(&icon),
description: val(subreddit, "public_description"),
subscribers: format_num(subreddit["data"]["subscribers"].as_f64().unwrap_or_default() as i64),
}
})
.collect::<Vec<Subreddit>>()
}

737
src/server.rs Normal file
View File

@ -0,0 +1,737 @@
#![allow(dead_code)]
use brotli::enc::{BrotliCompress, BrotliEncoderParams};
use cached::proc_macro::cached;
use cookie::Cookie;
use core::f64;
use futures_lite::{future::Boxed, Future, FutureExt};
use hyper::{
body,
body::HttpBody,
header,
service::{make_service_fn, service_fn},
HeaderMap,
};
use hyper::{Body, Method, Request, Response, Server as HyperServer};
use libflate::gzip;
use route_recognizer::{Params, Router};
use std::{
cmp::Ordering,
fmt::Display,
io,
pin::Pin,
result::Result,
str::{from_utf8, Split},
string::ToString,
};
use time::Duration;
use crate::dbg_msg;
type BoxResponse = Pin<Box<dyn Future<Output = Result<Response<Body>, String>> + Send>>;
/// Compressors for the response Body, in ascending order of preference.
#[derive(Copy, Clone, Debug, Eq, Hash, Ord, PartialEq, PartialOrd)]
enum CompressionType {
Passthrough,
Gzip,
Brotli,
}
/// All browsers support gzip, so if we are given `Accept-Encoding: *`, deliver
/// gzipped-content.
///
/// Brotli would be nice universally, but Safari (iOS, iPhone, macOS) reportedly
/// doesn't support it yet.
const DEFAULT_COMPRESSOR: CompressionType = CompressionType::Gzip;
impl CompressionType {
/// Returns a `CompressionType` given a content coding
/// in [RFC 7231](https://datatracker.ietf.org/doc/html/rfc7231#section-5.3.4)
/// format.
fn parse(s: &str) -> Option<Self> {
let c = match s {
// Compressors we support.
"gzip" => Self::Gzip,
"br" => Self::Brotli,
// The wildcard means that we can choose whatever
// compression we prefer. In this case, use the
// default.
"*" => DEFAULT_COMPRESSOR,
// Compressor not supported.
_ => return None,
};
Some(c)
}
}
impl Display for CompressionType {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
match self {
Self::Gzip => write!(f, "gzip"),
Self::Brotli => write!(f, "br"),
Self::Passthrough => Ok(()),
}
}
}
pub struct Route<'a> {
router: &'a mut Router<fn(Request<Body>) -> BoxResponse>,
path: String,
}
pub struct Server {
pub default_headers: HeaderMap,
router: Router<fn(Request<Body>) -> BoxResponse>,
}
#[macro_export]
macro_rules! headers(
{ $($key:expr => $value:expr),+ } => {
{
let mut m = hyper::HeaderMap::new();
$(
if let Ok(val) = hyper::header::HeaderValue::from_str($value) {
m.insert($key, val);
}
)+
m
}
};
);
pub trait RequestExt {
fn params(&self) -> Params;
fn param(&self, name: &str) -> Option<String>;
fn set_params(&mut self, params: Params) -> Option<Params>;
fn cookies(&self) -> Vec<Cookie<'_>>;
fn cookie(&self, name: &str) -> Option<Cookie<'_>>;
}
pub trait ResponseExt {
fn cookies(&self) -> Vec<Cookie<'_>>;
fn insert_cookie(&mut self, cookie: Cookie<'_>);
fn remove_cookie(&mut self, name: String);
}
impl RequestExt for Request<Body> {
fn params(&self) -> Params {
self.extensions().get::<Params>().unwrap_or(&Params::new()).clone()
// self.extensions()
// .get::<RequestMeta>()
// .and_then(|meta| meta.route_params())
// .expect("Routerify: No RouteParams added while processing request")
}
fn param(&self, name: &str) -> Option<String> {
self.params().find(name).map(std::borrow::ToOwned::to_owned)
}
fn set_params(&mut self, params: Params) -> Option<Params> {
self.extensions_mut().insert(params)
}
fn cookies(&self) -> Vec<Cookie<'_>> {
self.headers().get("Cookie").map_or(Vec::new(), |header| {
header
.to_str()
.unwrap_or_default()
.split("; ")
.map(|cookie| Cookie::parse(cookie).unwrap_or_else(|_| Cookie::from("")))
.collect()
})
}
fn cookie(&self, name: &str) -> Option<Cookie<'_>> {
self.cookies().into_iter().find(|c| c.name() == name)
}
}
impl ResponseExt for Response<Body> {
fn cookies(&self) -> Vec<Cookie<'_>> {
self.headers().get("Cookie").map_or(Vec::new(), |header| {
header
.to_str()
.unwrap_or_default()
.split("; ")
.map(|cookie| Cookie::parse(cookie).unwrap_or_else(|_| Cookie::from("")))
.collect()
})
}
fn insert_cookie(&mut self, cookie: Cookie<'_>) {
if let Ok(val) = header::HeaderValue::from_str(&cookie.to_string()) {
self.headers_mut().append("Set-Cookie", val);
}
}
fn remove_cookie(&mut self, name: String) {
let mut cookie = Cookie::from(name);
cookie.set_path("/");
cookie.set_max_age(Duration::seconds(1));
if let Ok(val) = header::HeaderValue::from_str(&cookie.to_string()) {
self.headers_mut().append("Set-Cookie", val);
}
}
}
impl Route<'_> {
fn method(&mut self, method: &Method, dest: fn(Request<Body>) -> BoxResponse) -> &mut Self {
self.router.add(&format!("/{}{}", method.as_str(), self.path), dest);
self
}
/// Add an endpoint for `GET` requests
pub fn get(&mut self, dest: fn(Request<Body>) -> BoxResponse) -> &mut Self {
self.method(&Method::GET, dest)
}
/// Add an endpoint for `POST` requests
pub fn post(&mut self, dest: fn(Request<Body>) -> BoxResponse) -> &mut Self {
self.method(&Method::POST, dest)
}
}
impl Server {
pub fn new() -> Self {
Self {
default_headers: HeaderMap::new(),
router: Router::new(),
}
}
pub fn at(&mut self, path: &str) -> Route<'_> {
Route {
path: path.to_owned(),
router: &mut self.router,
}
}
pub fn listen(self, addr: &str) -> Boxed<Result<(), hyper::Error>> {
let make_svc = make_service_fn(move |_conn| {
// For correct borrowing, these values need to be borrowed
let router = self.router.clone();
let default_headers = self.default_headers.clone();
// This is the `Service` that will handle the connection.
// `service_fn` is a helper to convert a function that
// returns a Response into a `Service`.
// let shared_router = router.clone();
async move {
Ok::<_, String>(service_fn(move |req: Request<Body>| {
let req_headers = req.headers().clone();
let def_headers = default_headers.clone();
// Remove double slashes and decode encoded slashes
let mut path = req.uri().path().replace("//", "/").replace("%2F", "/");
// Remove trailing slashes
if path != "/" && path.ends_with('/') {
path.pop();
}
// Match the visited path with an added route
match router.recognize(&format!("/{}{}", req.method().as_str(), path)) {
// If a route was configured for this path
Ok(found) => {
let mut parammed = req;
parammed.set_params(found.params().clone());
// Run the route's function
let func = (found.handler().to_owned().to_owned())(parammed);
async move {
match func.await {
Ok(mut res) => {
res.headers_mut().extend(def_headers);
let _ = compress_response(&req_headers, &mut res).await;
Ok(res)
}
Err(msg) => new_boilerplate(def_headers, req_headers, 500, Body::from(msg)).await,
}
}
.boxed()
}
// If there was a routing error
Err(e) => new_boilerplate(def_headers, req_headers, 404, e.into()).boxed(),
}
}))
}
});
// Build SocketAddr from provided address
let address = &addr.parse().unwrap_or_else(|_| panic!("Cannot parse {addr} as address (example format: 0.0.0.0:8080)"));
// Bind server to address specified above. Gracefully shut down if CTRL+C is pressed
let server = HyperServer::bind(address).serve(make_svc).with_graceful_shutdown(async {
// Wait for the CTRL+C signal
tokio::signal::ctrl_c().await.expect("Failed to install CTRL+C signal handler");
});
server.boxed()
}
}
/// Create a boilerplate Response for error conditions. This response will be
/// compressed if requested by client.
async fn new_boilerplate(
default_headers: HeaderMap<header::HeaderValue>,
req_headers: HeaderMap<header::HeaderValue>,
status: u16,
body: Body,
) -> Result<Response<Body>, String> {
match Response::builder().status(status).body(body) {
Ok(mut res) => {
let _ = compress_response(&req_headers, &mut res).await;
res.headers_mut().extend(default_headers.clone());
Ok(res)
}
Err(msg) => Err(msg.to_string()),
}
}
/// Determines the desired compressor based on the Accept-Encoding header.
///
/// This function will honor the [q-value](https://developer.mozilla.org/en-US/docs/Glossary/Quality_values)
/// for each compressor. The q-value is an optional parameter, a decimal value
/// on \[0..1\], to order the compressors by preference. An Accept-Encoding value
/// with no q-values is also accepted.
///
/// Here are [examples](https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Accept-Encoding#examples)
/// of valid Accept-Encoding headers.
///
/// ```http
/// Accept-Encoding: gzip
/// Accept-Encoding: gzip, compress, br
/// Accept-Encoding: br;q=1.0, gzip;q=0.8, *;q=0.1
/// ```
#[cached]
fn determine_compressor(accept_encoding: String) -> Option<CompressionType> {
if accept_encoding.is_empty() {
return None;
};
// Keep track of the compressor candidate based on both the client's
// preference and our own. Concrete examples:
//
// 1. "Accept-Encoding: gzip, br" => assuming we like brotli more than
// gzip, and the browser supports brotli, we choose brotli
//
// 2. "Accept-Encoding: gzip;q=0.8, br;q=0.3" => the client has stated a
// preference for gzip over brotli, so we choose gzip
//
// To do this, we need to define a struct which contains the requested
// requested compressor (abstracted as a CompressionType enum) and the
// q-value. If no q-value is defined for the compressor, we assume one of
// 1.0. We first compare compressor candidates by comparing q-values, and
// then CompressionTypes. We keep track of whatever is the greatest per our
// ordering.
struct CompressorCandidate {
alg: CompressionType,
q: f64,
}
impl Ord for CompressorCandidate {
fn cmp(&self, other: &Self) -> Ordering {
// Compare q-values. Break ties with the
// CompressionType values.
match self.q.total_cmp(&other.q) {
Ordering::Equal => self.alg.cmp(&other.alg),
ord => ord,
}
}
}
impl PartialOrd for CompressorCandidate {
fn partial_cmp(&self, other: &Self) -> Option<Ordering> {
Some(self.cmp(other))
}
}
impl PartialEq for CompressorCandidate {
fn eq(&self, other: &Self) -> bool {
(self.q == other.q) && (self.alg == other.alg)
}
}
impl Eq for CompressorCandidate {}
// This is the current candidate.
//
// Assmume no candidate so far. We do this by assigning the sentinel value
// of negative infinity to the q-value. If this value is negative infinity,
// that means there was no viable compressor candidate.
let mut cur_candidate = CompressorCandidate {
alg: CompressionType::Passthrough,
q: f64::NEG_INFINITY,
};
// This loop reads the requested compressors and keeps track of whichever
// one has the highest priority per our heuristic.
for val in accept_encoding.split(',') {
let mut q: f64 = 1.0;
// The compressor and q-value (if the latter is defined)
// will be delimited by semicolons.
let mut spl: Split<'_, char> = val.split(';');
// Get the compressor. For example, in
// gzip;q=0.8
// this grabs "gzip" in the string. It
// will further validate the compressor against the
// list of those we support. If it is not supported,
// we move onto the next one.
let compressor: CompressionType = match spl.next() {
// CompressionType::parse will return the appropriate enum given
// a string. For example, it will return CompressionType::Gzip
// when given "gzip".
Some(s) => match CompressionType::parse(s.trim()) {
Some(candidate) => candidate,
// We don't support the requested compression algorithm.
None => continue,
},
// We should never get here, but I'm paranoid.
None => continue,
};
// Get the q-value. This might not be defined, in which case assume
// 1.0.
if let Some(s) = spl.next() {
if !(s.len() > 2 && s.starts_with("q=")) {
// If the q-value is malformed, the header is malformed, so
// abort.
return None;
}
match s[2..].parse::<f64>() {
Ok(val) => {
if (0.0..=1.0).contains(&val) {
q = val;
} else {
// If the value is outside [0..1], header is malformed.
// Abort.
return None;
};
}
Err(_) => {
// If this isn't a f64, then assume a malformed header
// value and abort.
return None;
}
}
};
// If new_candidate > cur_candidate, make new_candidate the new
// cur_candidate. But do this safely! It is very possible that
// someone gave us the string "NAN", which (&str).parse::<f64>
// will happily translate to f64::NAN.
let new_candidate = CompressorCandidate { alg: compressor, q };
if let Some(ord) = new_candidate.partial_cmp(&cur_candidate) {
if ord == Ordering::Greater {
cur_candidate = new_candidate;
}
};
}
if cur_candidate.q == f64::NEG_INFINITY {
None
} else {
Some(cur_candidate.alg)
}
}
/// Compress the response body, if possible or desirable. The Body will be
/// compressed in place, and a new header Content-Encoding will be set
/// indicating the compression algorithm.
///
/// This function deems Body eligible compression if and only if the following
/// conditions are met:
///
/// 1. the HTTP client requests a compression encoding in the Content-Encoding
/// header (hence the need for the `req_headers`);
///
/// 2. the content encoding corresponds to a compression algorithm we support;
///
/// 3. the Media type in the Content-Type response header is text with any
/// subtype (e.g. text/plain) or application/json.
///
/// `compress_response` returns Ok on successful compression, or if not all three
/// conditions above are met. It returns Err if there was a problem decoding
/// any header in either `req_headers` or res, but res will remain intact.
///
/// This function logs errors to stderr, but only in debug mode. No information
/// is logged in release builds.
async fn compress_response(req_headers: &HeaderMap<header::HeaderValue>, res: &mut Response<Body>) -> Result<(), String> {
// Check if the data is eligible for compression.
if let Some(hdr) = res.headers().get(header::CONTENT_TYPE) {
match from_utf8(hdr.as_bytes()) {
Ok(val) => {
let s = val.to_string();
// TODO: better determination of what is eligible for compression
if !(s.starts_with("text/") || s.starts_with("application/json")) {
return Ok(());
};
}
Err(e) => {
dbg_msg!(e);
return Err(e.to_string());
}
};
} else {
// Response declares no Content-Type. Assume for simplicity that it
// cannot be compressed.
return Ok(());
};
// Don't bother if the size of the size of the response body will fit
// within an IP frame (less the bytes that make up the TCP/IP and HTTP
// headers).
if res.body().size_hint().lower() < 1452 {
return Ok(());
};
// Check to see which compressor is requested, and if we can use it.
let accept_encoding: String = match req_headers.get(header::ACCEPT_ENCODING) {
None => return Ok(()), // Client requested no compression.
Some(hdr) => match String::from_utf8(hdr.as_bytes().into()) {
Ok(val) => val,
#[cfg(debug_assertions)]
Err(e) => {
dbg_msg!(e);
return Ok(());
}
#[cfg(not(debug_assertions))]
Err(_) => return Ok(()),
},
};
let compressor: CompressionType = match determine_compressor(accept_encoding) {
Some(c) => c,
None => return Ok(()),
};
// Get the body from the response.
let body_bytes: Vec<u8> = match body::to_bytes(res.body_mut()).await {
Ok(b) => b.to_vec(),
Err(e) => {
dbg_msg!(e);
return Err(e.to_string());
}
};
// Compress!
match compress_body(compressor, body_bytes) {
Ok(compressed) => {
// We get here iff the compression was successful. Replace the body
// with the compressed payload, and add the appropriate
// Content-Encoding header in the response.
res.headers_mut().insert(header::CONTENT_ENCODING, compressor.to_string().parse().unwrap());
*(res.body_mut()) = Body::from(compressed);
}
Err(e) => return Err(e),
}
Ok(())
}
/// Compresses a `Vec<u8>` given a [`CompressionType`].
///
/// This is a helper function for [`compress_response`] and should not be
/// called directly.
// I've chosen a TTL of 600 (== 10 minutes) since compression is
// computationally expensive and we don't want to be doing it often. This is
// larger than client::json's TTL, but that's okay, because if client::json
// returns a new serde_json::Value, body_bytes changes, so this function will
// execute again.
#[cached(size = 100, time = 600, result = true)]
fn compress_body(compressor: CompressionType, body_bytes: Vec<u8>) -> Result<Vec<u8>, String> {
// io::Cursor implements io::Read, required for our encoders.
let mut reader = io::Cursor::new(body_bytes);
let compressed: Vec<u8> = match compressor {
CompressionType::Gzip => {
let mut gz: gzip::Encoder<Vec<u8>> = match gzip::Encoder::new(Vec::new()) {
Ok(gz) => gz,
Err(e) => {
dbg_msg!(e);
return Err(e.to_string());
}
};
match io::copy(&mut reader, &mut gz) {
Ok(_) => match gz.finish().into_result() {
Ok(compressed) => compressed,
Err(e) => {
dbg_msg!(e);
return Err(e.to_string());
}
},
Err(e) => {
dbg_msg!(e);
return Err(e.to_string());
}
}
}
CompressionType::Brotli => {
// We may want to make the compression parameters configurable
// in the future. For now, the defaults are sufficient.
let brotli_params = BrotliEncoderParams::default();
let mut compressed = Vec::<u8>::new();
match BrotliCompress(&mut reader, &mut compressed, &brotli_params) {
Ok(_) => compressed,
Err(e) => {
dbg_msg!(e);
return Err(e.to_string());
}
}
}
// This arm is for any requested compressor for which we don't yet
// have an implementation.
CompressionType::Passthrough => {
let msg = "unsupported compressor".to_string();
return Err(msg);
}
};
Ok(compressed)
}
#[cfg(test)]
mod tests {
use super::*;
use brotli::Decompressor as BrotliDecompressor;
use futures_lite::future::block_on;
use lipsum::lipsum;
use std::{boxed::Box, io};
#[test]
fn test_determine_compressor() {
// Single compressor given.
assert_eq!(determine_compressor("unsupported".to_string()), None);
assert_eq!(determine_compressor("gzip".to_string()), Some(CompressionType::Gzip));
assert_eq!(determine_compressor("*".to_string()), Some(DEFAULT_COMPRESSOR));
// Multiple compressors.
assert_eq!(determine_compressor("gzip, br".to_string()), Some(CompressionType::Brotli));
assert_eq!(determine_compressor("gzip;q=0.8, br;q=0.3".to_string()), Some(CompressionType::Gzip));
assert_eq!(determine_compressor("br, gzip".to_string()), Some(CompressionType::Brotli));
assert_eq!(determine_compressor("br;q=0.3, gzip;q=0.4".to_string()), Some(CompressionType::Gzip));
// Invalid q-values.
assert_eq!(determine_compressor("gzip;q=NAN".to_string()), None);
}
#[test]
fn test_compress_response() {
// This macro generates an Accept-Encoding header value given any number of
// compressors.
macro_rules! ae_gen {
($x:expr) => {
$x.to_string().as_str()
};
($x:expr, $($y:expr),+) => {
format!("{}, {}", $x.to_string(), ae_gen!($($y),+)).as_str()
};
}
for accept_encoding in [
"*",
ae_gen!(CompressionType::Gzip),
ae_gen!(CompressionType::Brotli, CompressionType::Gzip),
ae_gen!(CompressionType::Brotli),
] {
// Determine what the expected encoding should be based on both the
// specific encodings we accept.
let expected_encoding: CompressionType = match determine_compressor(accept_encoding.to_string()) {
Some(s) => s,
None => panic!("determine_compressor(accept_encoding.to_string()) => None"),
};
// Build headers with our Accept-Encoding.
let mut req_headers = HeaderMap::new();
req_headers.insert(header::ACCEPT_ENCODING, header::HeaderValue::from_str(accept_encoding).unwrap());
// Build test response.
let lorem_ipsum: String = lipsum(10000);
let expected_lorem_ipsum = Vec::<u8>::from(lorem_ipsum.as_str());
let mut res = Response::builder()
.status(200)
.header(header::CONTENT_TYPE, "text/plain")
.body(Body::from(lorem_ipsum))
.unwrap();
// Perform the compression.
if let Err(e) = block_on(compress_response(&req_headers, &mut res)) {
panic!("compress_response(&req_headers, &mut res) => Err(\"{e}\")");
};
// If the content was compressed, we expect the Content-Encoding
// header to be modified.
assert_eq!(
res
.headers()
.get(header::CONTENT_ENCODING)
.unwrap_or_else(|| panic!("missing content-encoding header"))
.to_str()
.unwrap_or_else(|_| panic!("failed to convert Content-Encoding header::HeaderValue to String")),
expected_encoding.to_string()
);
// Decompress body and make sure it's equal to what we started
// with.
//
// In the case of no compression, just make sure the "new" body in
// the Response is the same as what with which we start.
let body_vec = match block_on(body::to_bytes(res.body_mut())) {
Ok(b) => b.to_vec(),
Err(e) => panic!("{e}"),
};
if expected_encoding == CompressionType::Passthrough {
assert!(body_vec.eq(&expected_lorem_ipsum));
continue;
}
// This provides an io::Read for the underlying body.
let mut body_cursor: io::Cursor<Vec<u8>> = io::Cursor::new(body_vec);
// Match the appropriate decompresor for the given
// expected_encoding.
let mut decoder: Box<dyn io::Read> = match expected_encoding {
CompressionType::Gzip => match gzip::Decoder::new(&mut body_cursor) {
Ok(dgz) => Box::new(dgz),
Err(e) => panic!("{e}"),
},
CompressionType::Brotli => Box::new(BrotliDecompressor::new(body_cursor, expected_lorem_ipsum.len())),
_ => panic!("no decompressor for {}", expected_encoding.to_string()),
};
let mut decompressed = Vec::<u8>::new();
if let Err(e) = io::copy(&mut decoder, &mut decompressed) {
panic!("{e}");
};
assert!(decompressed.eq(&expected_lorem_ipsum));
}
}
}

148
src/settings.rs Normal file
View File

@ -0,0 +1,148 @@
use std::collections::HashMap;
// CRATES
use crate::server::ResponseExt;
use crate::utils::{redirect, template, Preferences};
use cookie::Cookie;
use futures_lite::StreamExt;
use hyper::{Body, Request, Response};
use rinja::Template;
use time::{Duration, OffsetDateTime};
// STRUCTS
#[derive(Template)]
#[template(path = "settings.html")]
struct SettingsTemplate {
prefs: Preferences,
url: String,
}
// CONSTANTS
const PREFS: [&str; 19] = [
"theme",
"mascot",
"front_page",
"layout",
"wide",
"comment_sort",
"post_sort",
"blur_spoiler",
"show_nsfw",
"blur_nsfw",
"use_hls",
"ffmpeg_video_downloads",
"hide_hls_notification",
"autoplay_videos",
"hide_sidebar_and_summary",
"fixed_navbar",
"hide_awards",
"hide_score",
"disable_visit_reddit_confirmation",
];
// FUNCTIONS
// Retrieve cookies from request "Cookie" header
pub async fn get(req: Request<Body>) -> Result<Response<Body>, String> {
let url = req.uri().to_string();
Ok(template(&SettingsTemplate {
prefs: Preferences::new(&req),
url,
}))
}
// Set cookies using response "Set-Cookie" header
pub async fn set(req: Request<Body>) -> Result<Response<Body>, String> {
// Split the body into parts
let (parts, mut body) = req.into_parts();
// Grab existing cookies
let _cookies: Vec<Cookie<'_>> = parts
.headers
.get_all("Cookie")
.iter()
.filter_map(|header| Cookie::parse(header.to_str().unwrap_or_default()).ok())
.collect();
// Aggregate the body...
// let whole_body = hyper::body::aggregate(req).await.map_err(|e| e.to_string())?;
let body_bytes = body
.try_fold(Vec::new(), |mut data, chunk| {
data.extend_from_slice(&chunk);
Ok(data)
})
.await
.map_err(|e| e.to_string())?;
let form = url::form_urlencoded::parse(&body_bytes).collect::<HashMap<_, _>>();
let mut response = redirect("/settings");
for &name in &PREFS {
match form.get(name) {
Some(value) => response.insert_cookie(
Cookie::build((name.to_owned(), value.clone()))
.path("/")
.http_only(name != "ffmpeg_video_downloads")
.expires(OffsetDateTime::now_utc() + Duration::weeks(52))
.into(),
),
None => response.remove_cookie(name.to_string()),
};
}
Ok(response)
}
fn set_cookies_method(req: Request<Body>, remove_cookies: bool) -> Response<Body> {
// Split the body into parts
let (parts, _) = req.into_parts();
// Grab existing cookies
let _cookies: Vec<Cookie<'_>> = parts
.headers
.get_all("Cookie")
.iter()
.filter_map(|header| Cookie::parse(header.to_str().unwrap_or_default()).ok())
.collect();
let query = parts.uri.query().unwrap_or_default().as_bytes();
let form = url::form_urlencoded::parse(query).collect::<HashMap<_, _>>();
let path = match form.get("redirect") {
Some(value) => format!("/{}", value.replace("%26", "&").replace("%23", "#")),
None => "/".to_string(),
};
let mut response = redirect(&path);
for name in [PREFS.to_vec(), vec!["subscriptions", "filters"]].concat() {
match form.get(name) {
Some(value) => response.insert_cookie(
Cookie::build((name.to_owned(), value.clone()))
.path("/")
.http_only(name != "ffmpeg_video_downloads")
.expires(OffsetDateTime::now_utc() + Duration::weeks(52))
.into(),
),
None => {
if remove_cookies {
response.remove_cookie(name.to_string());
}
}
};
}
response
}
// Set cookies using response "Set-Cookie" header
pub async fn restore(req: Request<Body>) -> Result<Response<Body>, String> {
Ok(set_cookies_method(req, true))
}
pub async fn update(req: Request<Body>) -> Result<Response<Body>, String> {
Ok(set_cookies_method(req, false))
}

View File

@ -1,97 +1,530 @@
use crate::{config, utils};
// CRATES
use crate::utils::{fetch_posts, format_num, format_url, request, val, ErrorTemplate, Params, Post, Subreddit};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use askama::Template;
use std::convert::TryInto;
use crate::utils::{
catch_random, error, filter_posts, format_num, format_url, get_filters, nsfw_landing, param, redirect, rewrite_urls, setting, template, val, Post, Preferences, Subreddit,
};
use crate::{client::json, server::ResponseExt, RequestExt};
use cookie::Cookie;
use hyper::{Body, Request, Response};
use rinja::Template;
use once_cell::sync::Lazy;
use regex::Regex;
use time::{Duration, OffsetDateTime};
// STRUCTS
#[derive(Template)]
#[template(path = "subreddit.html", escape = "none")]
#[template(path = "subreddit.html")]
struct SubredditTemplate {
sub: Subreddit,
posts: Vec<Post>,
sort: String,
sort: (String, String),
ends: (String, String),
prefs: Preferences,
url: String,
redirect_url: String,
/// Whether the subreddit itself is filtered.
is_filtered: bool,
/// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place,
/// and all fetched posts being filtered).
all_posts_filtered: bool,
/// Whether all posts were hidden because they are NSFW (and user has disabled show NSFW)
all_posts_hidden_nsfw: bool,
no_posts: bool,
}
#[derive(Template)]
#[template(path = "wiki.html")]
struct WikiTemplate {
sub: String,
wiki: String,
page: String,
prefs: Preferences,
url: String,
}
#[derive(Template)]
#[template(path = "wall.html")]
struct WallTemplate {
title: String,
sub: String,
msg: String,
prefs: Preferences,
url: String,
}
static GEO_FILTER_MATCH: Lazy<Regex> = Lazy::new(|| Regex::new(r"geo_filter=(?<region>\w+)").unwrap());
// SERVICES
#[allow(dead_code)]
pub async fn page(web::Path(sub): web::Path<String>, params: web::Query<Params>) -> Result<HttpResponse> {
render(sub, params.sort.clone(), (params.before.clone(), params.after.clone())).await
}
pub async fn community(req: Request<Body>) -> Result<Response<Body>, String> {
// Build Reddit API path
let root = req.uri().path() == "/";
let query = req.uri().query().unwrap_or_default().to_string();
let subscribed = setting(&req, "subscriptions");
let front_page = setting(&req, "front_page");
let post_sort = req.cookie("post_sort").map_or_else(|| "hot".to_string(), |c| c.value().to_string());
let sort = req.param("sort").unwrap_or_else(|| req.param("id").unwrap_or(post_sort));
pub async fn render(sub_name: String, sort: Option<String>, ends: (Option<String>, Option<String>)) -> Result<HttpResponse> {
let sorting = sort.unwrap_or("hot".to_string());
let before = ends.1.clone().unwrap_or(String::new()); // If there is an after, there must be a before
let sub_name = req.param("sub").unwrap_or(if front_page == "default" || front_page.is_empty() {
if subscribed.is_empty() {
"popular".to_string()
} else {
subscribed.clone()
}
} else {
front_page.clone()
});
let quarantined = can_access_quarantine(&req, &sub_name) || root;
// Build the Reddit JSON API url
let url = match ends.0 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?before={}&count=25", sub_name, sorting, val),
None => match ends.1 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?after={}&count=25", sub_name, sorting, val),
None => format!("https://www.reddit.com/r/{}/{}.json", sub_name, sorting),
},
// Handle random subreddits
if let Ok(random) = catch_random(&sub_name, "").await {
return Ok(random);
}
if req.param("sub").is_some() && sub_name.starts_with("u_") {
return Ok(redirect(&["/user/", &sub_name[2..]].concat()));
}
// Request subreddit metadata
let sub = if !sub_name.contains('+') && sub_name != subscribed && sub_name != "popular" && sub_name != "all" {
// Regular subreddit
subreddit(&sub_name, quarantined).await.unwrap_or_default()
} else if sub_name == subscribed {
// Subscription feed
if req.uri().path().starts_with("/r/") {
subreddit(&sub_name, quarantined).await.unwrap_or_default()
} else {
Subreddit::default()
}
} else {
// Multireddit, all, popular
Subreddit {
name: sub_name.clone(),
..Subreddit::default()
}
};
let sub_result = subreddit(&sub_name).await;
let items_result = fetch_posts(url, String::new()).await;
let req_url = req.uri().to_string();
// Return landing page if this post if this is NSFW community but the user
// has disabled the display of NSFW content or if the instance is SFW-only.
if sub.nsfw && crate::utils::should_be_nsfw_gated(&req, &req_url) {
return Ok(nsfw_landing(req, req_url).await.unwrap_or_default());
}
if sub_result.is_err() || items_result.is_err() {
let s = ErrorTemplate {
message: sub_result.err().unwrap().to_string(),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s))
let mut params = String::from("&raw_json=1");
if sub_name == "popular" {
let geo_filter = match GEO_FILTER_MATCH.captures(&query) {
Some(geo_filter) => geo_filter["region"].to_string(),
None => "GLOBAL".to_owned(),
};
params.push_str(&format!("&geo_filter={geo_filter}"));
}
let path = format!("/r/{}/{sort}.json?{}{params}", sub_name.replace('+', "%2B"), req.uri().query().unwrap_or_default());
let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str()));
let redirect_url = url[1..].replace('?', "%3F").replace('&', "%26").replace('+', "%2B");
let filters = get_filters(&req);
// If all requested subs are filtered, we don't need to fetch posts.
if sub_name.split('+').all(|s| filters.contains(s)) {
Ok(template(&SubredditTemplate {
sub,
posts: Vec::new(),
sort: (sort, param(&path, "t").unwrap_or_default()),
ends: (param(&path, "after").unwrap_or_default(), String::new()),
prefs: Preferences::new(&req),
url,
redirect_url,
is_filtered: true,
all_posts_filtered: false,
all_posts_hidden_nsfw: false,
no_posts: false,
}))
} else {
let mut sub = sub_result.unwrap();
let items = items_result.unwrap();
match Post::fetch(&path, quarantined).await {
Ok((mut posts, after)) => {
let (_, all_posts_filtered) = filter_posts(&mut posts, &filters);
let no_posts = posts.is_empty();
let all_posts_hidden_nsfw = !no_posts && (posts.iter().all(|p| p.flags.nsfw) && setting(&req, "show_nsfw") != "on");
if sort == "new" {
posts.sort_by(|a, b| b.created_ts.cmp(&a.created_ts));
posts.sort_by(|a, b| b.flags.stickied.cmp(&a.flags.stickied));
}
Ok(template(&SubredditTemplate {
sub,
posts,
sort: (sort, param(&path, "t").unwrap_or_default()),
ends: (param(&path, "after").unwrap_or_default(), after),
prefs: Preferences::new(&req),
url,
redirect_url,
is_filtered: false,
all_posts_filtered,
all_posts_hidden_nsfw,
no_posts,
}))
}
Err(msg) => match msg.as_str() {
"quarantined" | "gated" => Ok(quarantine(&req, sub_name, &msg)),
"private" => error(req, &format!("r/{sub_name} is a private community")).await,
"banned" => error(req, &format!("r/{sub_name} has been banned from Reddit")).await,
_ => error(req, &msg).await,
},
}
}
}
sub.icon = if sub.icon != "" {
format!(r#"<img class="subreddit_icon" src="{}">"#, sub.icon)
pub fn quarantine(req: &Request<Body>, sub: String, restriction: &str) -> Response<Body> {
let wall = WallTemplate {
title: format!("r/{sub} is {restriction}"),
msg: "Please click the button below to continue to this subreddit.".to_string(),
url: req.uri().to_string(),
sub,
prefs: Preferences::new(req),
};
Response::builder()
.status(403)
.header("content-type", "text/html")
.body(wall.render().unwrap_or_default().into())
.unwrap_or_default()
}
pub async fn add_quarantine_exception(req: Request<Body>) -> Result<Response<Body>, String> {
let subreddit = req.param("sub").ok_or("Invalid URL")?;
let redir = param(&format!("?{}", req.uri().query().unwrap_or_default()), "redir").ok_or("Invalid URL")?;
let mut response = redirect(&redir);
response.insert_cookie(
Cookie::build((&format!("allow_quaran_{}", subreddit.to_lowercase()), "true"))
.path("/")
.http_only(true)
.expires(cookie::Expiration::Session)
.into(),
);
Ok(response)
}
pub fn can_access_quarantine(req: &Request<Body>, sub: &str) -> bool {
// Determine if the subreddit can be accessed
setting(req, &format!("allow_quaran_{}", sub.to_lowercase())).parse().unwrap_or_default()
}
// Sub, filter, unfilter, or unsub by setting subscription cookie using response "Set-Cookie" header
pub async fn subscriptions_filters(req: Request<Body>) -> Result<Response<Body>, String> {
let sub = req.param("sub").unwrap_or_default();
let action: Vec<String> = req.uri().path().split('/').map(String::from).collect();
// Handle random subreddits
if sub == "random" || sub == "randnsfw" {
if action.contains(&"filter".to_string()) || action.contains(&"unfilter".to_string()) {
return Err("Can't filter random subreddit!".to_string());
}
return Err("Can't subscribe to random subreddit!".to_string());
}
let query = req.uri().query().unwrap_or_default().to_string();
let preferences = Preferences::new(&req);
let mut sub_list = preferences.subscriptions;
let mut filters = preferences.filters;
// Retrieve list of posts for these subreddits to extract display names
let posts = json(format!("/r/{sub}/hot.json?raw_json=1"), true).await;
let display_lookup: Vec<(String, &str)> = match &posts {
Ok(posts) => posts["data"]["children"]
.as_array()
.map(|list| {
list
.iter()
.map(|post| {
let display_name = post["data"]["subreddit"].as_str().unwrap_or_default();
(display_name.to_lowercase(), display_name)
})
.collect::<Vec<_>>()
})
.unwrap_or_default(),
Err(_) => vec![],
};
// Find each subreddit name (separated by '+') in sub parameter
for part in sub.split('+').filter(|x| x != &"") {
// Retrieve display name for the subreddit
let display;
let part = if part.starts_with("u_") {
part
} else if let Some(&(_, display)) = display_lookup.iter().find(|x| x.0 == part.to_lowercase()) {
// This is already known, doesn't require separate request
display
} else {
String::new()
// This subreddit display name isn't known, retrieve it
let path: String = format!("/r/{part}/about.json?raw_json=1");
display = json(path, true).await;
match &display {
Ok(display) => display["data"]["display_name"].as_str(),
Err(_) => None,
}
.unwrap_or(part)
};
let s = SubredditTemplate {
sub: sub,
posts: items.0,
sort: sorting,
ends: (before, items.1),
// Modify sub list based on action
if action.contains(&"subscribe".to_string()) && !sub_list.contains(&part.to_owned()) {
// Add each sub name to the subscribed list
sub_list.push(part.to_owned());
filters.retain(|s| s.to_lowercase() != part.to_lowercase());
// Reorder sub names alphabetically
sub_list.sort_by_key(|a| a.to_lowercase());
filters.sort_by_key(|a| a.to_lowercase());
} else if action.contains(&"unsubscribe".to_string()) {
// Remove sub name from subscribed list
sub_list.retain(|s| s.to_lowercase() != part.to_lowercase());
} else if action.contains(&"filter".to_string()) && !filters.contains(&part.to_owned()) {
// Add each sub name to the filtered list
filters.push(part.to_owned());
sub_list.retain(|s| s.to_lowercase() != part.to_lowercase());
// Reorder sub names alphabetically
filters.sort_by_key(|a| a.to_lowercase());
sub_list.sort_by_key(|a| a.to_lowercase());
} else if action.contains(&"unfilter".to_string()) {
// Remove sub name from filtered list
filters.retain(|s| s.to_lowercase() != part.to_lowercase());
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
}
}
// SUBREDDIT
async fn subreddit(sub: &String) -> Result<Subreddit, &'static str> {
// Build the Reddit JSON API url
let url: String = format!("https://www.reddit.com/r/{}/about.json", sub);
// Send a request to the url, receive JSON in response
let req = request(url).await;
// If the Reddit API returns an error, exit this function
if req.is_err() {
return Err(req.err().unwrap());
}
// Otherwise, grab the JSON output from the request
let res = req.unwrap();
let members = res["data"]["subscribers"].as_u64().unwrap_or(0);
let active = res["data"]["accounts_active"].as_u64().unwrap_or(0);
let sub = Subreddit {
name: val(&res, "display_name").await,
title: val(&res, "title").await,
description: val(&res, "public_description").await,
icon: format_url(val(&res, "icon_img").await.as_str()).await,
members: format_num(members.try_into().unwrap()),
active: format_num(active.try_into().unwrap()),
// Redirect back to subreddit
// check for redirect parameter if unsubscribing/unfiltering from outside sidebar
let path = if let Some(redirect_path) = param(&format!("?{query}"), "redirect") {
format!("/{redirect_path}")
} else {
format!("/r/{sub}")
};
Ok(sub)
let mut response = redirect(&path);
// Delete cookie if empty, else set
if sub_list.is_empty() {
response.remove_cookie("subscriptions".to_string());
} else {
response.insert_cookie(
Cookie::build(("subscriptions", sub_list.join("+")))
.path("/")
.http_only(true)
.expires(OffsetDateTime::now_utc() + Duration::weeks(52))
.into(),
);
}
if filters.is_empty() {
response.remove_cookie("filters".to_string());
} else {
response.insert_cookie(
Cookie::build(("filters", filters.join("+")))
.path("/")
.http_only(true)
.expires(OffsetDateTime::now_utc() + Duration::weeks(52))
.into(),
);
}
Ok(response)
}
pub async fn wiki(req: Request<Body>) -> Result<Response<Body>, String> {
let sub = req.param("sub").unwrap_or_else(|| "reddit.com".to_string());
let quarantined = can_access_quarantine(&req, &sub);
// Handle random subreddits
if let Ok(random) = catch_random(&sub, "/wiki").await {
return Ok(random);
}
let page = req.param("page").unwrap_or_else(|| "index".to_string());
let path: String = format!("/r/{sub}/wiki/{page}.json?raw_json=1");
let url = req.uri().to_string();
match json(path, quarantined).await {
Ok(response) => Ok(template(&WikiTemplate {
sub,
wiki: rewrite_urls(response["data"]["content_html"].as_str().unwrap_or("<h3>Wiki not found</h3>")),
page,
prefs: Preferences::new(&req),
url,
})),
Err(msg) => {
if msg == "quarantined" || msg == "gated" {
Ok(quarantine(&req, sub, &msg))
} else {
error(req, &msg).await
}
}
}
}
pub async fn sidebar(req: Request<Body>) -> Result<Response<Body>, String> {
let sub = req.param("sub").unwrap_or_else(|| "reddit.com".to_string());
let quarantined = can_access_quarantine(&req, &sub);
// Handle random subreddits
if let Ok(random) = catch_random(&sub, "/about/sidebar").await {
return Ok(random);
}
// Build the Reddit JSON API url
let path: String = format!("/r/{sub}/about.json?raw_json=1");
let url = req.uri().to_string();
// Send a request to the url
match json(path, quarantined).await {
// If success, receive JSON in response
Ok(response) => Ok(template(&WikiTemplate {
wiki: rewrite_urls(&val(&response, "description_html")),
// wiki: format!(
// "{}<hr><h1>Moderators</h1><br><ul>{}</ul>",
// rewrite_urls(&val(&response, "description_html"),
// moderators(&sub, quarantined).await.unwrap_or(vec!["Could not fetch moderators".to_string()]).join(""),
// ),
sub,
page: "Sidebar".to_string(),
prefs: Preferences::new(&req),
url,
})),
Err(msg) => {
if msg == "quarantined" || msg == "gated" {
Ok(quarantine(&req, sub, &msg))
} else {
error(req, &msg).await
}
}
}
}
// pub async fn moderators(sub: &str, quarantined: bool) -> Result<Vec<String>, String> {
// // Retrieve and format the html for the moderators list
// Ok(
// moderators_list(sub, quarantined)
// .await?
// .iter()
// .map(|m| format!("<li><a style=\"color: var(--accent)\" href=\"/u/{name}\">{name}</a></li>", name = m))
// .collect(),
// )
// }
// async fn moderators_list(sub: &str, quarantined: bool) -> Result<Vec<String>, String> {
// // Build the moderator list URL
// let path: String = format!("/r/{}/about/moderators.json?raw_json=1", sub);
// // Retrieve response
// json(path, quarantined).await.map(|response| {
// // Traverse json tree and format into list of strings
// response["data"]["children"]
// .as_array()
// .unwrap_or(&Vec::new())
// .iter()
// .filter_map(|moderator| {
// let name = moderator["name"].as_str().unwrap_or_default();
// if name.is_empty() {
// None
// } else {
// Some(name.to_string())
// }
// })
// .collect::<Vec<_>>()
// })
// }
// SUBREDDIT
async fn subreddit(sub: &str, quarantined: bool) -> Result<Subreddit, String> {
// Build the Reddit JSON API url
let path: String = format!("/r/{sub}/about.json?raw_json=1");
// Send a request to the url
let res = json(path, quarantined).await?;
// Metadata regarding the subreddit
let members: i64 = res["data"]["subscribers"].as_u64().unwrap_or_default() as i64;
let active: i64 = res["data"]["accounts_active"].as_u64().unwrap_or_default() as i64;
// Fetch subreddit icon either from the community_icon or icon_img value
let community_icon: &str = res["data"]["community_icon"].as_str().unwrap_or_default();
let icon = if community_icon.is_empty() { val(&res, "icon_img") } else { community_icon.to_string() };
Ok(Subreddit {
name: val(&res, "display_name"),
title: val(&res, "title"),
description: val(&res, "public_description"),
info: rewrite_urls(&val(&res, "description_html")),
// moderators: moderators_list(sub, quarantined).await.unwrap_or_default(),
icon: format_url(&icon),
members: format_num(members),
active: format_num(active),
wiki: res["data"]["wiki_enabled"].as_bool().unwrap_or_default(),
nsfw: res["data"]["over18"].as_bool().unwrap_or_default(),
})
}
pub async fn rss(req: Request<Body>) -> Result<Response<Body>, String> {
if config::get_setting("REDLIB_ENABLE_RSS").is_none() {
return Ok(error(req, "RSS is disabled on this instance.").await.unwrap_or_default());
}
use hyper::header::CONTENT_TYPE;
use rss::{ChannelBuilder, Item};
// Get subreddit
let sub = req.param("sub").unwrap_or_default();
let post_sort = req.cookie("post_sort").map_or_else(|| "hot".to_string(), |c| c.value().to_string());
let sort = req.param("sort").unwrap_or_else(|| req.param("id").unwrap_or(post_sort));
// Get path
let path = format!("/r/{sub}/{sort}.json?{}", req.uri().query().unwrap_or_default());
// Get subreddit data
let subreddit = subreddit(&sub, false).await?;
// Get posts
let (posts, _) = Post::fetch(&path, false).await?;
// Build the RSS feed
let channel = ChannelBuilder::default()
.title(&subreddit.title)
.description(&subreddit.description)
.items(
posts
.into_iter()
.map(|post| Item {
title: Some(post.title.to_string()),
link: Some(utils::get_post_url(&post)),
author: Some(post.author.name),
content: Some(rewrite_urls(&post.body)),
description: Some(format!(
"<a href='{}{}'>Comments</a>",
config::get_setting("REDLIB_FULL_URL").unwrap_or_default(),
post.permalink
)),
..Default::default()
})
.collect::<Vec<_>>(),
)
.build();
// Serialize the feed to RSS
let body = channel.to_string().into_bytes();
// Create the HTTP response
let mut res = Response::new(Body::from(body));
res.headers_mut().insert(CONTENT_TYPE, hyper::header::HeaderValue::from_static("application/rss+xml"));
Ok(res)
}
#[tokio::test(flavor = "multi_thread")]
async fn test_fetching_subreddit() {
let subreddit = subreddit("rust", false).await;
assert!(subreddit.is_ok());
}
#[tokio::test(flavor = "multi_thread")]
async fn test_gated_and_quarantined() {
let quarantined = subreddit("edgy", true).await;
assert!(quarantined.is_ok());
let gated = subreddit("drugs", true).await;
assert!(gated.is_ok());
}

View File

@ -1,73 +1,188 @@
// CRATES
use crate::utils::{fetch_posts, nested_val, request, ErrorTemplate, Params, Post, User};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use askama::Template;
use crate::client::json;
use crate::server::RequestExt;
use crate::utils::{error, filter_posts, format_url, get_filters, nsfw_landing, param, setting, template, Post, Preferences, User};
use crate::{config, utils};
use hyper::{Body, Request, Response};
use rinja::Template;
use time::{macros::format_description, OffsetDateTime};
// STRUCTS
#[derive(Template)]
#[template(path = "user.html", escape = "none")]
#[template(path = "user.html")]
struct UserTemplate {
user: User,
posts: Vec<Post>,
sort: String,
sort: (String, String),
ends: (String, String),
/// "overview", "comments", or "submitted"
listing: String,
prefs: Preferences,
url: String,
redirect_url: String,
/// Whether the user themself is filtered.
is_filtered: bool,
/// Whether all fetched posts are filtered (to differentiate between no posts fetched in the first place,
/// and all fetched posts being filtered).
all_posts_filtered: bool,
/// Whether all posts were hidden because they are NSFW (and user has disabled show NSFW)
all_posts_hidden_nsfw: bool,
no_posts: bool,
}
async fn render(username: String, sort: String) -> Result<HttpResponse> {
// Build the Reddit JSON API url
let url: String = format!("https://www.reddit.com/user/{}/.json?sort={}", username, sort);
// FUNCTIONS
pub async fn profile(req: Request<Body>) -> Result<Response<Body>, String> {
let listing = req.param("listing").unwrap_or_else(|| "overview".to_string());
let user = user(&username).await;
let posts = fetch_posts(url, "Comment".to_string()).await;
// Build the Reddit JSON API path
let path = format!(
"/user/{}/{listing}.json?{}&raw_json=1",
req.param("name").unwrap_or_else(|| "reddit".to_string()),
req.uri().query().unwrap_or_default(),
);
let url = String::from(req.uri().path_and_query().map_or("", |val| val.as_str()));
let redirect_url = url[1..].replace('?', "%3F").replace('&', "%26");
if user.is_err() || posts.is_err() {
let s = ErrorTemplate {
message: user.err().unwrap().to_string(),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s))
} else {
let s = UserTemplate {
user: user.unwrap(),
posts: posts.unwrap().0,
sort: sort,
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
// Retrieve other variables from Redlib request
let sort = param(&path, "sort").unwrap_or_default();
let username = req.param("name").unwrap_or_default();
// Retrieve info from user about page.
let user = user(&username).await.unwrap_or_default();
let req_url = req.uri().to_string();
// Return landing page if this post if this Reddit deems this user NSFW,
// but we have also disabled the display of NSFW content or if the instance
// is SFW-only.
if user.nsfw && crate::utils::should_be_nsfw_gated(&req, &req_url) {
return Ok(nsfw_landing(req, req_url).await.unwrap_or_default());
}
}
// SERVICES
pub async fn page(web::Path(username): web::Path<String>, params: web::Query<Params>) -> Result<HttpResponse> {
match &params.sort {
Some(sort) => render(username, sort.to_string()).await,
None => render(username, "hot".to_string()).await,
let filters = get_filters(&req);
if filters.contains(&["u_", &username].concat()) {
Ok(template(&UserTemplate {
user,
posts: Vec::new(),
sort: (sort, param(&path, "t").unwrap_or_default()),
ends: (param(&path, "after").unwrap_or_default(), String::new()),
listing,
prefs: Preferences::new(&req),
url,
redirect_url,
is_filtered: true,
all_posts_filtered: false,
all_posts_hidden_nsfw: false,
no_posts: false,
}))
} else {
// Request user posts/comments from Reddit
match Post::fetch(&path, false).await {
Ok((mut posts, after)) => {
let (_, all_posts_filtered) = filter_posts(&mut posts, &filters);
let no_posts = posts.is_empty();
let all_posts_hidden_nsfw = !no_posts && (posts.iter().all(|p| p.flags.nsfw) && setting(&req, "show_nsfw") != "on");
Ok(template(&UserTemplate {
user,
posts,
sort: (sort, param(&path, "t").unwrap_or_default()),
ends: (param(&path, "after").unwrap_or_default(), after),
listing,
prefs: Preferences::new(&req),
url,
redirect_url,
is_filtered: false,
all_posts_filtered,
all_posts_hidden_nsfw,
no_posts,
}))
}
// If there is an error show error page
Err(msg) => error(req, &msg).await,
}
}
}
// USER
async fn user(name: &String) -> Result<User, &'static str> {
// Build the Reddit JSON API url
let url: String = format!("https://www.reddit.com/user/{}/about.json", name);
async fn user(name: &str) -> Result<User, String> {
// Build the Reddit JSON API path
let path: String = format!("/user/{name}/about.json?raw_json=1");
// Send a request to the url, receive JSON in response
let req = request(url).await;
// Send a request to the url
json(path, false).await.map(|res| {
// Grab creation date as unix timestamp
let created_unix = res["data"]["created"].as_f64().unwrap_or(0.0).round() as i64;
let created = OffsetDateTime::from_unix_timestamp(created_unix).unwrap_or(OffsetDateTime::UNIX_EPOCH);
// If the Reddit API returns an error, exit this function
if req.is_err() {
return Err(req.err().unwrap());
}
// Closure used to parse JSON from Reddit APIs
let about = |item| res["data"]["subreddit"][item].as_str().unwrap_or_default().to_string();
// Otherwise, grab the JSON output from the request
let res = req.unwrap();
// Parse the JSON output into a User struct
Ok(User {
name: name.to_string(),
icon: nested_val(&res, "subreddit", "icon_img").await,
karma: res["data"]["total_karma"].as_i64().unwrap(),
banner: nested_val(&res, "subreddit", "banner_img").await,
description: nested_val(&res, "subreddit", "public_description").await,
// Parse the JSON output into a User struct
User {
name: res["data"]["name"].as_str().unwrap_or(name).to_owned(),
title: about("title"),
icon: format_url(&about("icon_img")),
karma: res["data"]["total_karma"].as_i64().unwrap_or(0),
created: created.format(format_description!("[month repr:short] [day] '[year repr:last_two]")).unwrap_or_default(),
banner: about("banner_img"),
description: about("public_description"),
nsfw: res["data"]["subreddit"]["over_18"].as_bool().unwrap_or_default(),
}
})
}
pub async fn rss(req: Request<Body>) -> Result<Response<Body>, String> {
if config::get_setting("REDLIB_ENABLE_RSS").is_none() {
return Ok(error(req, "RSS is disabled on this instance.").await.unwrap_or_default());
}
use crate::utils::rewrite_urls;
use hyper::header::CONTENT_TYPE;
use rss::{ChannelBuilder, Item};
// Get user
let user_str = req.param("name").unwrap_or_default();
let listing = req.param("listing").unwrap_or_else(|| "overview".to_string());
// Get path
let path = format!("/user/{user_str}/{listing}.json?{}&raw_json=1", req.uri().query().unwrap_or_default(),);
// Get user
let user_obj = user(&user_str).await.unwrap_or_default();
// Get posts
let (posts, _) = Post::fetch(&path, false).await?;
// Build the RSS feed
let channel = ChannelBuilder::default()
.title(user_str)
.description(user_obj.description)
.items(
posts
.into_iter()
.map(|post| Item {
title: Some(post.title.to_string()),
link: Some(utils::get_post_url(&post)),
author: Some(post.author.name),
content: Some(rewrite_urls(&post.body)),
..Default::default()
})
.collect::<Vec<_>>(),
)
.build();
// Serialize the feed to RSS
let body = channel.to_string().into_bytes();
// Create the HTTP response
let mut res = Response::new(Body::from(body));
res.headers_mut().insert(CONTENT_TYPE, hyper::header::HeaderValue::from_static("application/rss+xml"));
Ok(res)
}
#[tokio::test(flavor = "multi_thread")]
async fn test_fetching_user() {
let user = user("spez").await;
assert!(user.is_ok());
assert!(user.unwrap().karma > 100);
}

File diff suppressed because it is too large Load Diff

BIN
static/Inter.var.woff2 Normal file

Binary file not shown.

BIN
static/apple-touch-icon.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 60 KiB

BIN
static/favicon.ico Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 13 KiB

BIN
static/favicon.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 4.3 KiB

View File

@ -0,0 +1,2 @@
!function(e,t){"object"==typeof exports&&"object"==typeof module?module.exports=t():"function"==typeof define&&define.amd?define([],t):"object"==typeof exports?exports.FFmpegWASM=t():e.FFmpegWASM=t()}(self,(()=>(()=>{var e={454:e=>{function t(e){return Promise.resolve().then((()=>{var t=new Error("Cannot find module '"+e+"'");throw t.code="MODULE_NOT_FOUND",t}))}t.keys=()=>[],t.resolve=t,t.id=454,e.exports=t}},t={};function r(a){var o=t[a];if(void 0!==o)return o.exports;var s=t[a]={exports:{}};return e[a](s,s.exports,r),s.exports}return r.o=(e,t)=>Object.prototype.hasOwnProperty.call(e,t),(()=>{"use strict";var e;!function(e){e.LOAD="LOAD",e.EXEC="EXEC",e.WRITE_FILE="WRITE_FILE",e.READ_FILE="READ_FILE",e.DELETE_FILE="DELETE_FILE",e.RENAME="RENAME",e.CREATE_DIR="CREATE_DIR",e.LIST_DIR="LIST_DIR",e.DELETE_DIR="DELETE_DIR",e.ERROR="ERROR",e.DOWNLOAD="DOWNLOAD",e.PROGRESS="PROGRESS",e.LOG="LOG",e.MOUNT="MOUNT",e.UNMOUNT="UNMOUNT"}(e||(e={}));const t=new Error("unknown message type"),a=new Error("ffmpeg is not loaded, call `await ffmpeg.load()` first"),o=(new Error("called FFmpeg.terminate()"),new Error("failed to import ffmpeg-core.js"));let s;self.onmessage=async({data:{id:n,type:E,data:i}})=>{const c=[];let p;try{if(E!==e.LOAD&&!s)throw a;switch(E){case e.LOAD:p=await(async({coreURL:t="https://unpkg.com/@ffmpeg/core@0.12.1/dist/umd/ffmpeg-core.js",wasmURL:a,workerURL:n})=>{const E=!s,i=t,c=a||t.replace(/.js$/g,".wasm"),p=n||t.replace(/.js$/g,".worker.js");try{importScripts(i)}catch{if(self.createFFmpegCore=(await r(454)(i)).default,!self.createFFmpegCore)throw o}return s=await self.createFFmpegCore({mainScriptUrlOrBlob:`${i}#${btoa(JSON.stringify({wasmURL:c,workerURL:p}))}`}),s.setLogger((t=>self.postMessage({type:e.LOG,data:t}))),s.setProgress((t=>self.postMessage({type:e.PROGRESS,data:t}))),E})(i);break;case e.EXEC:p=(({args:e,timeout:t=-1})=>{s.setTimeout(t),s.exec(...e);const r=s.ret;return s.reset(),r})(i);break;case e.WRITE_FILE:p=(({path:e,data:t})=>(s.FS.writeFile(e,t),!0))(i);break;case e.READ_FILE:p=(({path:e,encoding:t})=>s.FS.readFile(e,{encoding:t}))(i);break;case e.DELETE_FILE:p=(({path:e})=>(s.FS.unlink(e),!0))(i);break;case e.RENAME:p=(({oldPath:e,newPath:t})=>(s.FS.rename(e,t),!0))(i);break;case e.CREATE_DIR:p=(({path:e})=>(s.FS.mkdir(e),!0))(i);break;case e.LIST_DIR:p=(({path:e})=>{const t=s.FS.readdir(e),r=[];for(const a of t){const t=s.FS.stat(`${e}/${a}`),o=s.FS.isDir(t.mode);r.push({name:a,isDir:o})}return r})(i);break;case e.DELETE_DIR:p=(({path:e})=>(s.FS.rmdir(e),!0))(i);break;case e.MOUNT:p=(({fsType:e,options:t,mountPoint:r})=>{let a=e,o=s.FS.filesystems[a];return!!o&&(s.FS.mount(o,t,r),!0)})(i);break;case e.UNMOUNT:p=(({mountPoint:e})=>(s.FS.unmount(e),!0))(i);break;default:throw t}}catch(t){return void self.postMessage({id:n,type:e.ERROR,data:t.toString()})}p instanceof Uint8Array&&c.push(p.buffer),self.postMessage({id:n,type:E,data:p},c)}})(),{}})()));
//# sourceMappingURL=814.ffmpeg.js.map

File diff suppressed because one or more lines are too long

File diff suppressed because one or more lines are too long

BIN
static/ffmpeg/ffmpeg-core.wasm Executable file

Binary file not shown.

View File

@ -0,0 +1 @@
!function(e,t){"object"==typeof exports&&"object"==typeof module?module.exports=t():"function"==typeof define&&define.amd?define([],t):"object"==typeof exports?exports.FFmpegUtil=t():e.FFmpegUtil=t()}(self,(()=>(()=>{"use strict";var e={591:(e,t)=>{Object.defineProperty(t,"__esModule",{value:!0}),t.HeaderContentLength=void 0,t.HeaderContentLength="Content-Length"},431:(e,t)=>{Object.defineProperty(t,"__esModule",{value:!0}),t.ERROR_INCOMPLETED_DOWNLOAD=t.ERROR_RESPONSE_BODY_READER=void 0,t.ERROR_RESPONSE_BODY_READER=new Error("failed to get response body reader"),t.ERROR_INCOMPLETED_DOWNLOAD=new Error("failed to complete download")},915:function(e,t,o){var r=this&&this.__awaiter||function(e,t,o,r){return new(o||(o=Promise))((function(n,i){function d(e){try{l(r.next(e))}catch(e){i(e)}}function a(e){try{l(r.throw(e))}catch(e){i(e)}}function l(e){var t;e.done?n(e.value):(t=e.value,t instanceof o?t:new o((function(e){e(t)}))).then(d,a)}l((r=r.apply(e,t||[])).next())}))};Object.defineProperty(t,"__esModule",{value:!0}),t.toBlobURL=t.downloadWithProgress=t.importScript=t.fetchFile=void 0;const n=o(431),i=o(591);t.fetchFile=e=>r(void 0,void 0,void 0,(function*(){let t;if("string"==typeof e)t=/data:_data\/([a-zA-Z]*);base64,([^"]*)/.test(e)?atob(e.split(",")[1]).split("").map((e=>e.charCodeAt(0))):yield(yield fetch(e)).arrayBuffer();else if(e instanceof URL)t=yield(yield fetch(e)).arrayBuffer();else{if(!(e instanceof File||e instanceof Blob))return new Uint8Array;t=yield(o=e,new Promise(((e,t)=>{const r=new FileReader;r.onload=()=>{const{result:t}=r;t instanceof ArrayBuffer?e(new Uint8Array(t)):e(new Uint8Array)},r.onerror=e=>{var o,r;t(Error(`File could not be read! Code=${(null===(r=null===(o=null==e?void 0:e.target)||void 0===o?void 0:o.error)||void 0===r?void 0:r.code)||-1}`))},r.readAsArrayBuffer(o)})))}var o;return new Uint8Array(t)})),t.importScript=e=>r(void 0,void 0,void 0,(function*(){return new Promise((t=>{const o=document.createElement("script"),r=()=>{o.removeEventListener("load",r),t()};o.src=e,o.type="text/javascript",o.addEventListener("load",r),document.getElementsByTagName("head")[0].appendChild(o)}))})),t.downloadWithProgress=(e,t)=>r(void 0,void 0,void 0,(function*(){var o;const r=yield fetch(e);let d;try{const a=parseInt(r.headers.get(i.HeaderContentLength)||"-1"),l=null===(o=r.body)||void 0===o?void 0:o.getReader();if(!l)throw n.ERROR_RESPONSE_BODY_READER;const c=[];let s=0;for(;;){const{done:o,value:r}=yield l.read(),i=r?r.length:0;if(o){if(-1!=a&&a!==s)throw n.ERROR_INCOMPLETED_DOWNLOAD;t&&t({url:e,total:a,received:s,delta:i,done:o});break}c.push(r),s+=i,t&&t({url:e,total:a,received:s,delta:i,done:o})}const f=new Uint8Array(s);let u=0;for(const e of c)f.set(e,u),u+=e.length;d=f.buffer}catch(o){console.log("failed to send download progress event: ",o),d=yield r.arrayBuffer(),t&&t({url:e,total:d.byteLength,received:d.byteLength,delta:0,done:!0})}return d})),t.toBlobURL=(e,o,n=!1,i)=>r(void 0,void 0,void 0,(function*(){const r=n?yield(0,t.downloadWithProgress)(e,i):yield(yield fetch(e)).arrayBuffer(),d=new Blob([r],{type:o});return URL.createObjectURL(d)}))}},t={};return function o(r){var n=t[r];if(void 0!==n)return n.exports;var i=t[r]={exports:{}};return e[r].call(i.exports,i,i.exports,o),i.exports}(915)})()));

2
static/ffmpeg/ffmpeg.js Normal file
View File

@ -0,0 +1,2 @@
!function(e,t){"object"==typeof exports&&"object"==typeof module?module.exports=t():"function"==typeof define&&define.amd?define([],t):"object"==typeof exports?exports.FFmpegWASM=t():e.FFmpegWASM=t()}(self,(()=>(()=>{"use strict";var e={m:{},d:(t,s)=>{for(var r in s)e.o(s,r)&&!e.o(t,r)&&Object.defineProperty(t,r,{enumerable:!0,get:s[r]})},u:e=>e+".ffmpeg.js"};e.g=function(){if("object"==typeof globalThis)return globalThis;try{return this||new Function("return this")()}catch(e){if("object"==typeof window)return window}}(),e.o=(e,t)=>Object.prototype.hasOwnProperty.call(e,t),e.r=e=>{"undefined"!=typeof Symbol&&Symbol.toStringTag&&Object.defineProperty(e,Symbol.toStringTag,{value:"Module"}),Object.defineProperty(e,"__esModule",{value:!0})},(()=>{var t;e.g.importScripts&&(t=e.g.location+"");var s=e.g.document;if(!t&&s&&(s.currentScript&&(t=s.currentScript.src),!t)){var r=s.getElementsByTagName("script");if(r.length)for(var a=r.length-1;a>-1&&!t;)t=r[a--].src}if(!t)throw new Error("Automatic publicPath is not supported in this browser");t=t.replace(/#.*$/,"").replace(/\?.*$/,"").replace(/\/[^\/]+$/,"/"),e.p=t})(),e.b=document.baseURI||self.location.href;var t,s={};e.r(s),e.d(s,{FFmpeg:()=>i}),function(e){e.LOAD="LOAD",e.EXEC="EXEC",e.WRITE_FILE="WRITE_FILE",e.READ_FILE="READ_FILE",e.DELETE_FILE="DELETE_FILE",e.RENAME="RENAME",e.CREATE_DIR="CREATE_DIR",e.LIST_DIR="LIST_DIR",e.DELETE_DIR="DELETE_DIR",e.ERROR="ERROR",e.DOWNLOAD="DOWNLOAD",e.PROGRESS="PROGRESS",e.LOG="LOG",e.MOUNT="MOUNT",e.UNMOUNT="UNMOUNT"}(t||(t={}));const r=(()=>{let e=0;return()=>e++})(),a=(new Error("unknown message type"),new Error("ffmpeg is not loaded, call `await ffmpeg.load()` first")),o=new Error("called FFmpeg.terminate()");new Error("failed to import ffmpeg-core.js");class i{#e=null;#t={};#s={};#r=[];#a=[];loaded=!1;#o=()=>{this.#e&&(this.#e.onmessage=({data:{id:e,type:s,data:r}})=>{switch(s){case t.LOAD:this.loaded=!0,this.#t[e](r);break;case t.MOUNT:case t.UNMOUNT:case t.EXEC:case t.WRITE_FILE:case t.READ_FILE:case t.DELETE_FILE:case t.RENAME:case t.CREATE_DIR:case t.LIST_DIR:case t.DELETE_DIR:this.#t[e](r);break;case t.LOG:this.#r.forEach((e=>e(r)));break;case t.PROGRESS:this.#a.forEach((e=>e(r)));break;case t.ERROR:this.#s[e](r)}delete this.#t[e],delete this.#s[e]})};#i=({type:e,data:t},s=[],o)=>this.#e?new Promise(((a,i)=>{const n=r();this.#e&&this.#e.postMessage({id:n,type:e,data:t},s),this.#t[n]=a,this.#s[n]=i,o?.addEventListener("abort",(()=>{i(new DOMException(`Message # ${n} was aborted`,"AbortError"))}),{once:!0})})):Promise.reject(a);on(e,t){"log"===e?this.#r.push(t):"progress"===e&&this.#a.push(t)}off(e,t){"log"===e?this.#r=this.#r.filter((e=>e!==t)):"progress"===e&&(this.#a=this.#a.filter((e=>e!==t)))}load=(s={},{signal:r}={})=>(this.#e||(this.#e=new Worker(new URL(e.p+e.u(814),e.b),{type:void 0}),this.#o()),this.#i({type:t.LOAD,data:s},void 0,r));exec=(e,s=-1,{signal:r}={})=>this.#i({type:t.EXEC,data:{args:e,timeout:s}},void 0,r);terminate=()=>{const e=Object.keys(this.#s);for(const t of e)this.#s[t](o),delete this.#s[t],delete this.#t[t];this.#e&&(this.#e.terminate(),this.#e=null,this.loaded=!1)};writeFile=(e,s,{signal:r}={})=>{const a=[];return s instanceof Uint8Array&&a.push(s.buffer),this.#i({type:t.WRITE_FILE,data:{path:e,data:s}},a,r)};mount=(e,s,r)=>this.#i({type:t.MOUNT,data:{fsType:e,options:s,mountPoint:r}},[]);unmount=e=>this.#i({type:t.UNMOUNT,data:{mountPoint:e}},[]);readFile=(e,s="binary",{signal:r}={})=>this.#i({type:t.READ_FILE,data:{path:e,encoding:s}},void 0,r);deleteFile=(e,{signal:s}={})=>this.#i({type:t.DELETE_FILE,data:{path:e}},void 0,s);rename=(e,s,{signal:r}={})=>this.#i({type:t.RENAME,data:{oldPath:e,newPath:s}},void 0,r);createDir=(e,{signal:s}={})=>this.#i({type:t.CREATE_DIR,data:{path:e}},void 0,s);listDir=(e,{signal:s}={})=>this.#i({type:t.LIST_DIR,data:{path:e}},void 0,s);deleteDir=(e,{signal:s}={})=>this.#i({type:t.DELETE_DIR,data:{path:e}},void 0,s)}return s})()));
//# sourceMappingURL=ffmpeg.js.map

File diff suppressed because one or more lines are too long

1
static/highlighted.js Normal file
View File

@ -0,0 +1 @@
document.querySelector('#commentQueryForms').scrollIntoView();

3
static/hls.min.js vendored Normal file

File diff suppressed because one or more lines are too long

BIN
static/logo.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 60 KiB

33
static/logo.svg Normal file
View File

@ -0,0 +1,33 @@
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<svg
version="1.1"
viewBox="0 0 512 512"
id="svg2"
width="512"
height="512"
xmlns="http://www.w3.org/2000/svg"
xmlns:svg="http://www.w3.org/2000/svg">
<defs id="defs2" />
<rect width="512" height="512" fill="#4c082a" />
<g
transform="matrix(0.75272,0,0,0.75272,-1.1596187,-0.37987125)"
id="g2">
<circle
fill="#1a1a1a"
id="circle1"
style="fill:#4c082a;fill-opacity:0"
r="340.10001"
cy="340.32001"
cx="341.10999" />
<path
d="m 320.64,126.73 v 300.8 h 92.264 V 219.61 h 75.803 v -92.83 h -75.803 v -0.0508 z"
fill="#f83240"
id="path1"
style="fill:#f83240;fill-opacity:1" />
<path
d="M 193.1,126.74 V 510.7 h 0.006 v 43.543 h 295.82 v -92.338 h -202.74 v -335.16 z"
fill="#f83240"
id="path2"
style="fill:#f83240;fill-opacity:1" />
</g>
</svg>

After

Width:  |  Height:  |  Size: 943 B

BIN
static/logo_full.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 219 KiB

24
static/manifest.json Normal file
View File

@ -0,0 +1,24 @@
{
"name": "Redlib",
"short_name": "Redlib",
"display": "standalone",
"background_color": "#1f1f1f",
"description": "An alternative private front-end to Reddit",
"theme_color": "#1f1f1f",
"start_url": "/",
"icons": [
{
"src": "logo.png",
"sizes": "512x512",
"type": "image/png"
},
{
"src": "apple-touch-icon.png",
"sizes": "180x180"
},
{
"src": "favicon.ico",
"sizes": "32x32"
}
]
}

Binary file not shown.

After

Width:  |  Height:  |  Size: 9.5 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 4.8 KiB

11
static/opensearch.xml Normal file
View File

@ -0,0 +1,11 @@
<OpenSearchDescription xmlns="http://a9.com/-/spec/opensearch/1.1/"
xmlns:moz="http://www.mozilla.org/2006/browser/search/">
<ShortName>Search Redlib</ShortName>
<Description>Search for whatever you want on Redlib, awesome Reddit frontend</Description>
<InputEncoding>UTF-8</InputEncoding>
<Image width="32" height="32" type="image/x-icon">/favicon.ico</Image>
<Url type="text/html" template="/search">
<Param name="q" value="{searchTerms}"/>
</Url>
<moz:SearchForm>/search</moz:SearchForm>
</OpenSearchDescription>

View File

@ -1,2 +0,0 @@
User-Agent: *
Disallow: /

File diff suppressed because it is too large Load Diff

14
static/themes/black.css Normal file
View File

@ -0,0 +1,14 @@
/* Black theme setting */
.black {
--accent: #bb2b3b;
--green: #00a229;
--text: white;
--foreground: #0f0f0f;
--background: black;
--outside: black;
--post: black;
--panel-border: 2px solid #0f0f0f;
--highlighted: #0f0f0f;
--visited: #aaa;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}

View File

@ -0,0 +1,17 @@
/* Catppuccin theme setting */
.catppuccin {
--accent: #b4befe; /* lavender */
--green: #a6e3a1; /* green */
--text: #cdd6f4; /* text */
--foreground: #181825; /* mantle */
--background: #1e1e2e; /* base */
--outside: #11111b; /* crust */
--post: #11111b; /* crust */
--panel-border: none;
--highlighted: #313244; /* surface0 */
--visited: #6c7086; /* overlay0 */
--shadow: 0 0 0 transparent;
--nsfw: #fab387; /* peach */
--admin: #eba0ac; /* maroon */
}

14
static/themes/dark.css Normal file
View File

@ -0,0 +1,14 @@
/* Dark theme setting */
.dark{
--accent: #d54455;
--green: #5cff85;
--text: white;
--foreground: #222;
--background: #0f0f0f;
--outside: #1f1f1f;
--post: #161616;
--panel-border: 1px solid #333;
--highlighted: #333;
--visited: #aaa;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.5);
}

13
static/themes/doomone.css Normal file
View File

@ -0,0 +1,13 @@
.doomone {
--accent: #51afef;
--green: #00a229;
--text: #bbc2cf;
--foreground: #3d4148;
--background: #282c34;
--outside: #52565c;
--post: #24272e;
--panel-border: 2px solid #52565c;
--highlighted: #686b70;
--visited: #969692;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}

14
static/themes/dracula.css Normal file
View File

@ -0,0 +1,14 @@
/* Dracula theme setting */
.dracula {
--accent: #bd93f9;
--green: #50fa7b;
--text: #f8f8f2;
--foreground: #3d4051;
--background: #282a36;
--outside: #393c4d;
--post: #333544;
--panel-border: 2px solid #44475a;
--highlighted: #4e5267;
--visited: #969692;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}

14
static/themes/gold.css Normal file
View File

@ -0,0 +1,14 @@
/* Gold theme setting */
.gold {
--accent: #f2aa4c;
--green: #5cff85;
--text: white;
--foreground: #234;
--background: #101820;
--outside: #1b2936;
--post: #1b2936;
--panel-border: 0px solid black;
--highlighted: #234;
--visited: #aaa;
--shadow: 0 2px 5px rgba(0, 0, 0, 0.5);
}

View File

@ -0,0 +1,13 @@
/* Gruvbox-Dark theme setting */
.gruvboxdark {
--accent: #8ec07c;
--green: #b8bb26;
--text: #ebdbb2;
--foreground: #3c3836;
--background: #282828;
--outside: #3c3836;
--post: #3c3836;
--panel-border: 1px solid #504945;
--highlighted: #282828;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.5);
}

View File

@ -0,0 +1,18 @@
/* Gruvbox-Light theme setting */
.gruvboxlight {
--accent: #427b58;
--green: #79740e;
--text: #3c3836;
--foreground: #ebdbb2;
--background: #fbf1c7;
--outside: #ebdbb2;
--post: #ebdbb2;
--panel-border: 1px solid #d5c4a1;
--highlighted: #fbf1c7;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.25);
}
html:has(> .gruvboxlight) {
/* Hint color theme to browser for scrollbar */
color-scheme: light;
}

View File

@ -0,0 +1,14 @@
/* icebergDark theme setting */
.icebergDark {
--accent: #85a0c7;
--green: #b5bf82;
--text: #c6c8d1;
--foreground: #454d73;
--background: #161821;
--outside: #1f2233;
--post: #1f2233;
--panel-border: 1px solid #454d73;
--highlighted: #0f1117;
--visited: #0f1117;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.5);
}

View File

@ -0,0 +1,14 @@
/* Laserwave theme setting */
.laserwave {
--accent: #eb64b9;
--green: #74dfc4;
--text: #e0dfe1;
--foreground: #302a36;
--background: #27212e;
--outside: #3e3647;
--post: #3e3647;
--panel-border: 2px solid #2f2738;
--highlighted: #302a36;
--visited: #91889b;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}

View File

@ -0,0 +1,14 @@
/* Libreddit black theme setting */
.libredditBlack {
--accent: #009a9a;
--green: #00a229;
--text: white;
--foreground: #0f0f0f;
--background: black;
--outside: black;
--post: black;
--panel-border: 2px solid #0f0f0f;
--highlighted: #0f0f0f;
--visited: #aaa;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}

View File

@ -0,0 +1,14 @@
/* Libreddit dark theme setting */
.libredditDark{
--accent: aqua;
--green: #5cff85;
--text: white;
--foreground: #222;
--background: #0f0f0f;
--outside: #1f1f1f;
--post: #161616;
--panel-border: 1px solid #333;
--highlighted: #333;
--visited: #aaa;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.5);
}

View File

@ -0,0 +1,19 @@
/* Libreddit light theme setting */
.libredditLight {
--accent: #009a9a;
--green: #00a229;
--text: black;
--foreground: #f5f5f5;
--background: #ddd;
--outside: #ececec;
--post: #eee;
--panel-border: 1px solid #ccc;
--highlighted: white;
--visited: #555;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}
html:has(> .libredditLight) {
/* Hint color theme to browser for scrollbar */
color-scheme: light;
}

19
static/themes/light.css Normal file
View File

@ -0,0 +1,19 @@
/* Light theme setting */
.light {
--accent: #bb2b3b;
--green: #00a229;
--text: black;
--foreground: #f5f5f5;
--background: #ddd;
--outside: #ececec;
--post: #eee;
--panel-border: 1px solid #ccc;
--highlighted: white;
--visited: #555;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}
html:has(> .light) {
/* Hint color theme to browser for scrollbar */
color-scheme: light;
}

14
static/themes/nord.css Normal file
View File

@ -0,0 +1,14 @@
/* Nord theme setting */
.nord {
--accent: #8fbcbb;
--green: #a3be8c;
--text: #eceff4;
--foreground: #3b4252;
--background: #2e3440;
--outside: #434c5e;
--post: #434c5e;
--panel-border: 2px solid #4c566a;
--highlighted: #3b4252;
--visited: #a3a5aa;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
}

13
static/themes/rosebox.css Normal file
View File

@ -0,0 +1,13 @@
/* Rosebox theme setting */
.rosebox {
--accent: #a57562;
--green: #a3be8c;
--text: white;
--foreground: #222;
--background: #262626;
--outside: #222;
--post: #222;
--panel-border: 1px solid #222;
--highlighted: #262626;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.5);
}

View File

@ -0,0 +1,14 @@
/* Tokyo Night theme setting */
.tokyoNight {
--accent: #565f89;
--green: #73daca;
--text: #a9b1d6;
--foreground: #24283b;
--background: #1a1b26;
--outside: #24283b;
--post: #1a1b26;
--panel-border: 1px solid #a9b1d6;
--highlighted: #414868;
--visited: #414868;
--shadow: 0 1px 3px rgba(0, 0, 0, 0.5);
}

14
static/themes/violet.css Normal file
View File

@ -0,0 +1,14 @@
/* Violet theme setting */
.violet {
--accent: #7c71dd;
--green: #5cff85;
--text: white;
--foreground: #1F2347;
--background: #12152b;
--outside: #181c3a;
--post: #181c3a;
--panel-border: 1px solid #1F2347;
--highlighted: #1F2347;
--visited: #aaa;
--shadow: 0 2px 5px rgba(0, 0, 0, 0.5);
}

228
static/videoUtils.js Normal file
View File

@ -0,0 +1,228 @@
// @license http://www.gnu.org/licenses/agpl-3.0.html AGPL-3.0
let ffmpeg = null;
(function () {
if (Hls.isSupported()) {
var downloadsEnabled = document.cookie.split("; ").find((row) => row.startsWith("ffmpeg_video_downloads="))?.split("=")[1] == "on";
var videoSources = document.querySelectorAll("video source[type='application/vnd.apple.mpegurl']");
videoSources.forEach(function (source) {
var playlist = source.src;
var oldVideo = source.parentNode;
var autoplay = oldVideo.classList.contains("hls_autoplay");
// If HLS is supported natively then don't use hls.js
if (oldVideo.canPlayType(source.type) === "probably") {
if (autoplay) {
oldVideo.play();
}
return;
}
// Replace video with copy that will have all "source" elements removed
var newVideo = oldVideo.cloneNode(true);
var allSources = newVideo.querySelectorAll("source");
allSources.forEach(function (source) {
source.remove();
});
// Empty source to enable play event
newVideo.src = "about:blank";
oldVideo.parentNode.replaceChild(newVideo, oldVideo);
function initializeHls() {
newVideo.removeEventListener('play', initializeHls);
var hls = new Hls({ autoStartLoad: false });
hls.loadSource(playlist);
hls.attachMedia(newVideo);
hls.on(Hls.Events.MANIFEST_PARSED, function () {
hls.loadLevel = hls.levels.length - 1;
var availableLevels = hls.levels.map(function(level) {
return {
height: level.height,
width: level.width,
bitrate: level.bitrate,
};
});
addQualitySelector(newVideo, hls, availableLevels);
if (downloadsEnabled){ addVideoDownload(newVideo, hls); }
hls.startLoad();
newVideo.play();
});
hls.on(Hls.Events.ERROR, function (event, data) {
var errorType = data.type;
var errorFatal = data.fatal;
if (errorFatal) {
switch (errorType) {
case Hls.ErrorType.NETWORK_ERROR:
hls.startLoad();
break;
case Hls.ErrorType.MEDIA_ERROR:
hls.recoverMediaError();
break;
default:
hls.destroy();
break;
}
}
console.error("HLS error", data);
});
}
if (downloadsEnabled){
const { fetchFile } = FFmpegUtil;
const { FFmpeg } = FFmpegWASM;
function addVideoDownload(videoElement, hlsInstance) {
var mediaStream = [];
var downloadButton = document.createElement('button');
downloadButton.classList.add('video-options','download');
downloadButton.innerText = "⏳"
const mergeStreams = async () => {
if (ffmpeg === null) {
ffmpeg = new FFmpeg();
await ffmpeg.load({
coreURL: "/ffmpeg/ffmpeg-core.js",
});
ffmpeg.on("log", ({ message }) => {
console.log(message); // This is quite noisy but i will include it
})
ffmpeg.on("progress", ({ progress, time }) => { // Progress TODO: show progress ring around button not just ⏳
// console.log("ffmpeg prog:",progress * 100)
});
}
// Combine Video Audio Streams
await ffmpeg.writeFile("video", await fetchFile(concatBlob(mediaStream['video'])));
await ffmpeg.writeFile("audio", await fetchFile(concatBlob(mediaStream['audio'])));
console.time('ffmpeg-exec');
await ffmpeg.exec(['-i', "video", '-i', "audio",'-c:v', "copy", '-c:a', "aac", 'output.mp4']);
console.timeEnd('ffmpeg-exec')
// Save
const toSlug = (str) => {
return str
.normalize('NFD')
.replace(/[\u0300-\u036f]/g, '')
.replace(/[\W_]+/g, '-')
.toLowerCase()
.replace(/^-+|-+$/g, '');
}
var filename = toSlug(videoElement.parentNode.parentNode.id || document.title)
const data = await ffmpeg.readFile('output.mp4');
saveAs(new Blob([data.buffer]),filename, {type: 'video/mp4'});
return
}
function saveAs(blob, filename) { // Yeah ok...
var url = URL.createObjectURL(blob);
var a = document.createElement("a");
document.body.appendChild(a);
a.style = "display: none";
a.href = url;
a.download = filename;
a.click();
window.URL.revokeObjectURL(url);
}
function concatBlob(inputArray) {
var totalLength = inputArray.reduce(function (prev, cur) {
return prev + cur.length
}, 0);
var result = new Uint8Array(totalLength);
var offset = 0;
inputArray.forEach(function (element) {
result.set(element, offset);
offset += element.length;
});
return new Blob([result], {
type: 'application/octet-stream'
});
}
function getStreams() {
var video = document.createElement('video');
video.autoplay = true;
var dataStreams = {
'video': [],
'audio': []
};
mediaStream = dataStreams; // Update stream
hlsInstance.on(Hls.Events.BUFFER_APPENDING, function (event, data) {
dataStreams[data.type].push(data.data);
});
var isDownloading = false
function startDownload() {
if (!isDownloading) { isDownloading = true } else { return }
downloadButton.innerText = "⏳"
mergeStreams()
.then(_ => {
isDownloading = false
downloadButton.innerText = "⭳"
});
}
function waitForLoad() {
const poll = resolve => {
if(hlsInstance._media.buffered.length === 1 &&
hlsInstance._media.buffered.start(0) === 0 &&
hlsInstance._media.buffered.end(0) === hlsInstance._media.duration)
resolve();
else setTimeout(_ => poll(resolve), 400);
}
return new Promise(poll);
}
waitForLoad(_ => flag === true)
.then(_ => {
downloadButton.innerText = "⭳"
downloadButton.addEventListener('click', startDownload);
});
}
videoElement.parentNode.appendChild(downloadButton);
getStreams()
}
}
function addQualitySelector(videoElement, hlsInstance, availableLevels) {
var qualitySelector = document.createElement('select');
qualitySelector.classList.add('video-options');
var last = availableLevels.length - 1;
availableLevels.forEach(function (level, index) {
var option = document.createElement('option');
option.value = index.toString();
var bitrate = (level.bitrate / 1_000).toFixed(0);
option.text = level.height + 'p (' + bitrate + ' kbps)';
if (index === last) {
option.selected = "selected";
}
qualitySelector.appendChild(option);
});
qualitySelector.selectedIndex = availableLevels.length - 1;
qualitySelector.addEventListener('change', function () {
var selectedIndex = qualitySelector.selectedIndex;
hlsInstance.nextLevel = selectedIndex;
hlsInstance.startLoad();
});
videoElement.parentNode.appendChild(qualitySelector);
}
newVideo.addEventListener('play', initializeHls);
if (autoplay) {
newVideo.play();
}
});
} else {
var videos = document.querySelectorAll("video.hls_autoplay");
videos.forEach(function (video) {
video.setAttribute("autoplay", "");
});
}
})();
// @license-end

View File

@ -1,36 +1,94 @@
{% import "utils.html" as utils %}
<!DOCTYPE html>
<html lang="en">
<html lang="en" class="{% if prefs.fixed_navbar == "on" %}fixed_navbar{% endif %}">
<head>
{% block head %}
<title>{% block title %}{% endblock %}</title>
<meta http-equiv="Content-Security-Policy" content="default-src 'self'; style-src 'self' 'unsafe-inline';">
<meta name="description" content="View on Libreddit, an alternative private front-end to Reddit.">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>{% block title %}Redlib{% endblock %}</title>
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
<link rel="stylesheet" href="/style.css">
{% block sortstyle %}
<style>
#sort > #sort_{{ sort }} {
background: aqua;
color: black;
}
</style>
{% endblock %}
{% endblock %}
</head>
<body>
{% block header %}
<header>
<a href="/"><span id="lib">lib</span>reddit. <span id="version">v{{ env!("CARGO_PKG_VERSION") }}</span></a>
<a id="github" href="https://github.com/spikecodes/libreddit">GITHUB</a>
</header>
<meta name="description" content="View on Redlib, an alternative private front-end to Reddit.">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
{% if crate::utils::disable_indexing() %}
<meta name="robots" content="noindex, nofollow">
{% endif %}
<!-- General PWA -->
<meta name="theme-color" content="#1F1F1F">
<!-- iOS Application -->
<meta name="apple-mobile-web-app-title" content="Redlib">
<meta name="apple-mobile-web-app-capable" content="yes">
<meta name="apple-mobile-web-app-status-bar-style" content="default">
<!-- Android -->
<meta name="mobile-web-app-capable" content="yes">
<!-- iOS Logo -->
<link href="/touch-icon-iphone.png" rel="apple-touch-icon">
<!-- OpenSearch description file -->
<link rel="search" type="application/opensearchdescription+xml" title="Search Redlib" href="/opensearch.xml">
<!-- PWA Manifest -->
<link rel="manifest" type="application/json" href="/manifest.json">
<link rel="shortcut icon" type="image/x-icon" href="/favicon.ico">
<link rel="stylesheet" type="text/css" href="/style.css?v={{ env!("CARGO_PKG_VERSION") }}">
{% endblock %}
</head>
<body class="
{% if prefs.layout != "" %}{{ prefs.layout }}{% endif %}
{% if prefs.wide == "on" || prefs.layout == "old" || prefs.layout == "waterfall" %} wide{% endif %}
{% if prefs.theme != "system" %} {{ prefs.theme }}{% endif %}
{% if prefs.fixed_navbar == "on" %} fixed_navbar{% endif %}">
<!-- NAVIGATION BAR -->
<nav class="
{% if prefs.fixed_navbar == "on" %} fixed_navbar{% endif %}">
<div id="logo">
<a id="redlib" href="/"><span id="lib">red</span><span id="reddit">sun</span><span id="lib">lib.</span></a>
{% block subscriptions %}{% endblock %}
</div>
{% block search %}{% endblock %}
<div id="links">
<a id="reddit_link" {% if prefs.disable_visit_reddit_confirmation != "on" %}href="#popup"{% else %}href="https://www.reddit.com{{ url }}" rel="nofollow"{% endif %}>
<span>reddit</span>
<svg xmlns="http://www.w3.org/2000/svg" width="24" height="24" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round">
<path d="M22 2L12 22"/>
<path d="M2 6.70587C3.33333 8.07884 3.33333 11.5971 3.33333 11.5971M3.33333 19.647V11.5971M3.33333 11.5971C3.33333 11.5971 5.125 7.47817 8 7.47817C10.875 7.47817 12 8.85114 12 8.85114"/>
</svg>
</a>
{% if prefs.disable_visit_reddit_confirmation != "on" %}
{% call utils::visit_reddit_confirmation(url) %}
{% endif %}
<a id="settings_link" href="/settings">
<span>settings</span>
<svg xmlns="http://www.w3.org/2000/svg" width="24" height="24" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round">
<title>settings</title>
<circle cx="12" cy="12" r="3"/><path d="M19.4 15a1.65 1.65 0 0 0 .33 1.82l.06.06a2 2 0 0 1 0 2.83 2 2 0 0 1-2.83 0l-.06-.06a1.65 1.65 0 0 0-1.82-.33 1.65 1.65 0 0 0-1 1.51V21a2 2 0 0 1-2 2 2 2 0 0 1-2-2v-.09A1.65 1.65 0 0 0 9 19.4a1.65 1.65 0 0 0-1.82.33l-.06.06a2 2 0 0 1-2.83 0 2 2 0 0 1 0-2.83l.06-.06a1.65 1.65 0 0 0 .33-1.82 1.65 1.65 0 0 0-1.51-1H3a2 2 0 0 1-2-2 2 2 0 0 1 2-2h.09A1.65 1.65 0 0 0 4.6 9a1.65 1.65 0 0 0-.33-1.82l-.06-.06a2 2 0 0 1 0-2.83 2 2 0 0 1 2.83 0l.06.06a1.65 1.65 0 0 0 1.82.33H9a1.65 1.65 0 0 0 1-1.51V3a2 2 0 0 1 2-2 2 2 0 0 1 2 2v.09a1.65 1.65 0 0 0 1 1.51 1.65 1.65 0 0 0 1.82-.33l.06-.06a2 2 0 0 1 2.83 0 2 2 0 0 1 0 2.83l-.06.06a1.65 1.65 0 0 0-.33 1.82V9a1.65 1.65 0 0 0 1.51 1H21a2 2 0 0 1 2 2 2 2 0 0 1-2 2h-.09a1.65 1.65 0 0 0-1.51 1z"/>
</svg>
</a>
</div>
</nav>
{% if prefs.mascot != "none" && prefs.mascot != "" %}
<!-- MASCOT -->
<div class="mascot">
<img src="/mascot/{{ prefs.mascot }}.png">
</div>
{% endif %}
<!-- MAIN CONTENT -->
{% block body %}
<main>
{% block content %}
{% endblock %}
</main>
{% endblock %}
<!-- FOOTER -->
{% block footer %}
<footer>
<p id="version">v{{ env!("CARGO_PKG_VERSION") }}</p>
<div class="footer-button">
<a href="/info" title="View instance information">ⓘ View instance info</a>
</div>
<div class="footer-button">
<a href="https://git.stardust.wtf/iridium/redsunlib" title="View code on git.stardust.wtf">&lt;&gt; Code</a>
</div>
</footer>
{% endblock %}
</body>
</html>
</html>

47
templates/comment.html Normal file
View File

@ -0,0 +1,47 @@
{% import "utils.html" as utils %}
{% if kind == "more" && parent_kind == "t1" %}
<a class="deeper_replies" href="{{ post_link }}{{ parent_id }}">&rarr; More replies ({{ more_count }})</a>
{% else if kind == "t1" %}
<div id="{{ id }}" class="comment">
<div class="comment_left">
<p class="comment_score" title="{{ score.1 }}">
{% if prefs.hide_score != "on" %}
{{ score.0 }}
{% else %}
&#x2022;
{% endif %}
</p>
<div class="line"></div>
</div>
<details class="comment_right" {% if !collapsed || highlighted %}open{% endif %}>
<summary class="comment_data">
{% if author.name != "[deleted]" %}
<a class="comment_author {{ author.distinguished }} {% if author.name == post_author %}op{% endif %}" href="/user/{{ author.name }}">u/{{ author.name }}</a>
{% else %}
<span class="comment_author {{ author.distinguished }}">u/[deleted]</span>
{% endif %}
{% if author.flair.flair_parts.len() > 0 %}
<small class="author_flair">{% call utils::render_flair(author.flair.flair_parts) %}</small>
{% endif %}
<a href="{{ post_link }}{{ id }}/?context=3#{{ id }}" class="created" title="{{ created }}">{{ rel_time }}</a>
{% if edited.0 != "".to_string() %}<span class="edited" title="{{ edited.1 }}">edited {{ edited.0 }}</span>{% endif %}
{% if !awards.is_empty() && prefs.hide_awards != "on" %}
<span class="dot">&bull;</span>
{% for award in awards.clone() %}
<span class="award" title="{{ award.name }}">
<img alt="{{ award.name }}" src="{{ award.icon_url }}" width="16" height="16"/>
</span>
{% endfor %}
{% endif %}
</summary>
{% if is_filtered %}
<div class="comment_body_filtered {% if highlighted %}highlighted{% endif %}">(Filtered content)</div>
{% else %}
<div class="comment_body {% if highlighted %}highlighted{% endif %}">{{ body|safe }}</div>
{% endif %}
<blockquote class="replies">{% for c in replies -%}{{ c.render().unwrap()|safe }}{%- endfor %}
</bockquote>
</details>
</div>
{% endif %}

113
templates/duplicates.html Normal file
View File

@ -0,0 +1,113 @@
{% extends "base.html" %}
{% import "utils.html" as utils %}
{% block title %}{{ post.title }} - r/{{ post.community }}{% endblock %}
{% block search %}
{% call utils::search(["/r/", post.community.as_str()].concat(), "") %}
{% endblock %}
{% block root %}/r/{{ post.community }}{% endblock %}{% block location %}r/{{ post.community }}{% endblock %}
{% block head %}
{% call super() %}
{% endblock %}
{% block subscriptions %}
{% call utils::sub_list(post.community.as_str()) %}
{% endblock %}
{% block content %}
<div id="column_one">
{% call utils::post(post) %}
<!-- DUPLICATES -->
{% if post.num_duplicates == 0 %}
<span class="listing_warn">(No duplicates found)</span>
{% else if post.flags.nsfw && prefs.show_nsfw != "on" %}
<span class="listing_warn">(Enable "Show NSFW posts" in <a href="/settings">settings</a> to show duplicates)</span>
{% else %}
<div id="duplicates_msg"><h3>Duplicates</h3></div>
{% if num_posts_filtered > 0 %}
<span class="listing_warn">
{% if all_posts_filtered %}
(All posts have been filtered)
{% else %}
(Some posts have been filtered)
{% endif %}
</span>
{% endif %}
<div id="sort">
<div id="sort_options">
<a {% if params.sort.is_empty() || params.sort.eq("num_comments") %}class="selected"{% endif %} href="?sort=num_comments">
Number of comments
</a>
<a {% if params.sort.eq("new") %}class="selected"{% endif %} href="?sort=new">
New
</a>
</div>
</div>
<div id="posts">
{% for post in duplicates -%}
{# TODO: utils::post should be reworked to permit a truncated display of a post as below #}
{% if !(post.flags.nsfw) || prefs.show_nsfw == "on" %}
<div class="post {% if post.flags.stickied %}stickied{% endif %}" id="{{ post.id }}">
<p class="post_header">
{% let community -%}
{% if post.community.starts_with("u_") -%}
{% let community = format!("u/{}", &post.community[2..]) -%}
{% else -%}
{% let community = format!("r/{}", post.community) -%}
{% endif -%}
<a class="post_subreddit" href="/r/{{ post.community }}">{{ post.community }}</a>
<span class="dot">&bull;</span>
<a class="post_author {{ post.author.distinguished }}" href="/u/{{ post.author.name }}">u/{{ post.author.name }}</a>
<span class="dot">&bull;</span>
<span class="created" title="{{ post.created }}">{{ post.rel_time }}</span>
{% if !post.awards.is_empty() && prefs.hide_awards != "on" %}
{% for award in post.awards.clone() %}
<span class="award" title="{{ award.name }}">
<img alt="{{ award.name }}" src="{{ award.icon_url }}" width="16" height="16"/>
</span>
{% endfor %}
{% endif %}
</p>
<h2 class="post_title">
{% if post.flair.flair_parts.len() > 0 %}
<a href="/r/{{ post.community }}/search?q=flair_name%3A%22{{ post.flair.text }}%22&restrict_sr=on"
class="post_flair"
style="color:{{ post.flair.foreground_color }}; background:{{ post.flair.background_color }};"
dir="ltr">{% call utils::render_flair(post.flair.flair_parts) %}</a>
{% endif %}
<a href="{{ post.permalink }}">{{ post.title }}</a>{% if post.flags.nsfw %} <small class="nsfw">NSFW</small>{% endif %}
</h2>
<div class="post_score" title="{{ post.score.1 }}">
{% if prefs.hide_score != "on" %}
{{ post.score.0 }}
{% else %}
&#x2022;
{% endif %}
<span class="label"> Upvotes</span></div>
<div class="post_footer">
<a href="{{ post.permalink }}" class="post_comments" title="{{ post.comments.1 }} comments">{{ post.comments.0 }} comments</a>
</div>
</div>
{% endif %}
{%- endfor %}
</div>
<footer>
{% if params.before != "" %}
<a href="?before={{ params.before }}{% if !params.sort.is_empty() %}&sort={{ params.sort }}{% endif %}" accesskey="P">PREV</a>
{% endif %}
{% if params.after != "" %}
<a href="?after={{ params.after }}{% if !params.sort.is_empty() %}&sort={{ params.sort }}{% endif %}" accesskey="N">NEXT</a>
{% endif %}
</footer>
{% endif %}
</div>
{% endblock %}

View File

@ -1,6 +1,15 @@
{% extends "base.html" %}
{% block title %}Error: {{ message }}{% endblock %}
{% block title %}Error: {{ msg }}{% endblock %}
{% block sortstyle %}{% endblock %}
{% block content %}
<h1 style="text-align: center; font-size: 50px;">{{ message }}</h1>
<div id="error">
<h1>{{ msg }}</h1>
<h3><a href="https://www.redditstatus.com/">Reddit Status</a></h3>
<br />
<h3>Expected something to work? <a
href="https://github.com/redlib-org/redlib/issues/new?assignees=&labels=bug&projects=&template=bug_report.md&title=%F0%9F%90%9B+Bug+Report%3A+{{ msg }}">Report
an issue</a></h3>
<br />
<h3>Head back <a href="/">home</a>?</h3>
</div>
{% endblock %}

10
templates/message.html Normal file
View File

@ -0,0 +1,10 @@
{% extends "base.html" %}
{% block title %}{{ title }}{% endblock %}
{% block sortstyle %}{% endblock %}
{% block content %}
<div id="message">
<h1>{{ title }}</h1>
<br>
{{ body|safe }}
</div>
{% endblock %}

View File

@ -0,0 +1,31 @@
{% extends "base.html" %}
{% block title %}NSFW content gated{% endblock %}
{% block sortstyle %}{% endblock %}
{% block content %}
<div id="nsfw_landing">
<h1>
&#128561;
{% if res_type == crate::utils::ResourceType::Subreddit %}
r/{{ res }} is a <b class="nsfw-tag">NSFW</b> community!
{% else if res_type == crate::utils::ResourceType::User %}
u/{{ res }}'s content is <b class="nsfw-tag">NSFW</b>!
{% else if res_type == crate::utils::ResourceType::Post %}
This post is <b class="nsfw-tag">NSFW</b>!
{% endif %}
</h1>
<br />
<p>
{% if crate::utils::sfw_only() %}
This instance of Redlib is SFW-only.</p>
{% else %}
Enable "Show NSFW posts" in <a href="/settings">settings</a> to view this {% if res_type == crate::utils::ResourceType::Subreddit %}subreddit{% else if res_type == crate::utils::ResourceType::User %}user's posts or comments{% else if res_type == crate::utils::ResourceType::Post %}post{% endif %}. <br>
<div>Alternatively <a href="/settings/update/?show_nsfw=on&redirect={{self.url[1..self.url.len()]}}">enable NSFW posts</a> now and view this {% if res_type == crate::utils::ResourceType::Subreddit %}subreddit{% else if res_type == crate::utils::ResourceType::User %}profile{% else if res_type == crate::utils::ResourceType::Post %}post{% endif %} immediately</div>
{% if res_type == crate::utils::ResourceType::Post %} You can also temporarily bypass this gate and view the post by clicking on this <a href="{{url}}&bypass_nsfw_landing">link</a>.{% endif %}
{% endif %}
</p>
</div>
{% endblock %}
{% block footer %}
{% endblock %}

View File

@ -1,44 +0,0 @@
{% extends "base.html" %}
{% block title %}Libreddit{% endblock %}
{% block content %}
<div id="sort">
<div id="sort_hot"><a href="?sort=hot">Hot</a></div>
<div id="sort_top"><a href="?sort=top">Top</a></div>
<div id="sort_new"><a href="?sort=new">New</a></div>
<div id="sort_rising"><a href="?sort=rising">Rising</a></div>
</div>
{% for post in posts %}
<div class="post">
<div class="post_left">
<h3 class="post_score">{{ post.score }}</h3>
</div>
<div class="post_right">
<h4>
<b><a class="post_subreddit" href="/r/{{ post.community }}">r/{{ post.community }}</a></b>
&bull; <a class="post_author" href="/u/{{ post.author }}">u/{{ post.author }}</a>
{% if post.author_flair.0 != "" %}
<small class="author_flair">{{ post.author_flair.0 }}</small>
{% endif %}
<span class="datetime" style="float: right;">{{ post.time }}</span>
</h4>
<h3 class="post_title">
{% if post.flair.0 != "" %}
<small class="post_flair" style="color:{{ post.flair.2 }}; background:{{ post.flair.1 }}">{{ post.flair.0 }}</small>
{% endif %}
<a href="{{ post.url }}">{{ post.title }}</a>
</h3>
</div>
<img class="post_thumbnail" src="{{ post.media }}">
</div><br>
{% endfor %}
<footer>
{% if ends.0 != "" %}
<a href="?before={{ ends.0 }}">PREV</a>
{% endif %}
{% if ends.1 != "" %}
<a href="?after={{ ends.1 }}">NEXT</a>
{% endif %}
</footer>
{% endblock %}

View File

@ -1,82 +1,91 @@
{% extends "base.html" %}
{% import "utils.html" as utils %}
{% block title %}{{ post.title }} - r/{{ post.community }}{% endblock %}
{% block head %}
{% call super() %}
<meta name="author" content="u/{{ post.author }}">
{% block search %}
{% call utils::search(["/r/", post.community.as_str()].concat(), "") %}
{% endblock %}
{% macro comment(item) -%}
{% block root %}/r/{{ post.community }}{% endblock %}{% block location %}r/{{ post.community }}{% endblock %}
{% block head %}
{% call super() %}
<!-- Meta Tags -->
<meta name="author" content="u/{{ post.author.name }}">
<meta name="title" content="{{ post.title }} - r/{{ post.community }}">
<meta property="og:title" content="{{ post.title }} - r/{{ post.community }}">
<meta property="og:description" content="View on Redlib, an alternative private front-end to Reddit.">
<meta property="og:url" content="{{ post.permalink }}">
<meta property="twitter:url" content="{{ post.permalink }}">
<meta property="twitter:title" content="{{ post.title }} - r/{{ post.community }}">
<meta property="twitter:description" content="View on Redlib, an alternative private front-end to Reddit.">
{% if post.post_type == "image" %}
<meta property="og:type" content="image">
<meta property="og:image" content="{{ post.thumbnail.url }}">
<meta property="twitter:card" content="summary_large_image">
<meta property="twitter:image" content="{{ post.thumbnail.url }}">
{% else if post.post_type == "video" || post.post_type == "gif" %}
<meta property="twitter:card" content="video">
<meta property="og:type" content="video">
<meta property="og:video" content="{{ post.media.url }}">
<meta property="og:video:type" content="video/mp4">
{% else %}
<meta property="og:type" content="website">
{% if single_thread %}
<script src="/highlighted.js" defer></script>
{% endif %}
{% endif %}
{% endblock %}
<div class="comment">
<div class="comment_left">
<h3 class="comment_score">{{ item.score }}</h3>
<div class="line"></div>
</div>
<details class="comment_right" open>
<summary class="comment_data">
<a class="comment_author" href="/u/{{ item.author }}">u/{{ item.author }}</a>
{% if item.flair.0 != "" %}
<small class="author_flair">{{ item.flair.0 }}</small>
{% endif %}
&bull; <span class="datetime">{{ item.time }}</span>
</summary>
<h4 class="comment_body">{{ item.body }}</h4>
{%- endmacro %}
{% block subscriptions %}
{% call utils::sub_list(post.community.as_str()) %}
{% endblock %}
{% block content %}
<div class="post highlighted">
<div class="post_left">
<h3 class="post_score">{{ post.score }}</h3>
</div>
<div class="post_right">
<h4>
<b><a class="post_subreddit" href="/r/{{ post.community }}">r/{{ post.community }}</a></b>
&bull;
<a class="post_author" href="/u/{{ post.author }}">u/{{ post.author }}</a>
{% if post.author_flair.0 != "" %}
<small class="author_flair">{{ post.author_flair.0 }}</small>
{% endif %}
<span class="datetime">{{ post.time }}</span>
</h4>
<a href="{{ post.url }}" class="post_title">
{{ post.title }}
{% if post.flair.0 != "" %}
<small class="post_flair" style="color:{{ post.flair.2 }}; background:{{ post.flair.1 }}">{{ post.flair.0 }}</small>
{% endif %}
</a>
{% if post.post_type == "image" %}
<img class="post_media" src="{{ post.media }}"/>
{% else if post.post_type == "video" %}
<video class="post_media" src="{{ post.media }}" controls autoplay loop>
{% else if post.post_type == "link" %}
<a id="post_url" href="{{ post.media }}">{{ post.media }}</a>
{% endif %}
<h4 class="post_body">{{ post.body }}</h4>
</div>
</div>
<div id="sort">
<div id="sort_confidence"><a href="?sort=confidence">Best</a></div>
<div id="sort_top"><a href="?sort=top">Top</a></div>
<div id="sort_new"><a href="?sort=new">New</a></div>
<div id="sort_controversial"><a href="?sort=controversial">Controversial</a></div>
<div id="sort_old"><a href="?sort=old">Old</a></div>
</div>
<div id="column_one">
{% call utils::post(post) %}
{% for c in comments -%}
<!-- SORT FORM -->
<div id="commentQueryForms">
<form id="sort">
<p id="comment_count">{{post.comments.0}} {% if post.comments.0 == "1" %}comment{% else %}comments{% endif %} <span id="sorted_by">sorted by </span></p>
<select name="sort" title="Sort comments by" id="commentSortSelect">
{% call utils::options(sort, ["confidence", "top", "new", "controversial", "old"], "confidence") %}
</select>
<button id="sort_submit" class="submit">
<svg width="15" viewBox="0 0 110 100" fill="none" stroke-width="10" stroke-linecap="round">
<path d="M20 50 H100" />
<path d="M75 15 L100 50 L75 85" />
&rarr;
</svg>
</button>
</form>
<!-- SEARCH FORM -->
<form id="sort">
<input id="search" class="commentQuery" type="search" name="q" value="{{ comment_query }}" placeholder="Search comments">
<input type="hidden" name="type" value="comment">
</form>
</div>
<div>
{% if comment_query != "" %}
Comments containing "{{ comment_query }}"&nbsp;|&nbsp;<a id="allCommentsLink" href="{{ url_without_query }}">All comments</a>
{% endif %}
</div>
<!-- COMMENTS -->
{% for c in comments -%}
<div class="thread">
{% call comment(c) %}
<div class="replies">
{% for reply in c.replies %}
{% call comment(reply) %}
<div class="replies">
{% for response in reply.replies %}
{% call comment(response) %}</details></div>
{% endfor %}
</div></details></div>
{% endfor %}
</div></details></div>
{% if single_thread %}
<p class="thread_nav"><a href="{{ post.permalink }}">View all comments</a></p>
{% if c.parent_kind == "t1" %}
<p class="thread_nav"><a href="?context=9999">Show parent comments</a></p>
{% endif %}
{% endif %}
{{ c.render().unwrap()|safe }}
</div>
{%- endfor %}
{%- endfor %}
{% endblock %}
</div>
{% endblock %}

Some files were not shown because too many files have changed in this diff Show More