Compare commits

...

409 Commits

Author SHA1 Message Date
f0b69f8a4a Update to v0.4 2021-03-10 20:51:08 -08:00
118ff9485c Document proxy.rs 2021-03-10 19:02:03 -08:00
4a51b7cfb0 Horizontally squish comments 2021-03-10 15:10:59 -08:00
f877face80 Update README.md 2021-03-10 20:56:33 +00:00
f0e8deb000 Add alt attribute to user icon 2021-03-10 11:29:36 -08:00
e70dfe2c0b Fix <video> size attributes 2021-03-10 10:49:18 -08:00
2e89a85858 Handle alternative status codes 2021-03-09 22:23:26 -08:00
e59b2b1346 Custom HTTP client with Rustls 2021-03-09 22:13:46 -08:00
1c36549134 Fix #146 2021-03-09 07:22:17 -08:00
5fb88d4744 Allow certain clippy lints 2021-03-08 19:22:10 -08:00
6c7188a1b9 Prevent pushing of Cargo.lock 2021-03-08 18:50:03 -08:00
84009fbb8e Remove Cargo.lock 2021-03-08 18:49:35 -08:00
bf783c2f3a Optimize type casting 2021-03-08 18:49:06 -08:00
213babb057 Update dependencies 2021-03-08 16:30:34 -08:00
7dbc02d930 Update himiko instances' location 2021-03-05 18:39:31 +00:00
10873dd0c6 Fix #144 2021-03-05 06:24:40 -08:00
c0d1519341 Update screenshot in README.md (#143)
* Update screenshot in README.md

New screenshot for v0.3.1. Also 35% lighter!

* Update screenshot

Co-authored-by: Spike <19519553+spikecodes@users.noreply.github.com>
2021-03-05 04:30:32 +00:00
8709c49f39 NGINX reverse proxy notice 2021-03-04 03:53:49 +00:00
56cfeba9e5 Update to v0.3.1 2021-03-03 09:35:40 -08:00
890d5ae625 Revert to curl-client 2021-03-03 09:24:31 -08:00
caa8f1d49e Test h1-client 2021-03-03 09:15:19 -08:00
dd51b23dc4 Switch Surf backend to h1-client-rustls 2021-03-02 19:44:51 -08:00
52d9698879 Add british cow instances. Closes #141 2021-03-02 08:56:01 -08:00
20f6945160 Fix #140 2021-02-27 13:34:02 -08:00
10c73fad7f Switch favicon to ico 2021-02-26 12:04:11 -08:00
2bddc952cb Link comment timestamps. Closes #137 2021-02-25 21:53:27 -08:00
1de01d7283 Log errors to io::stderr 2021-02-25 16:21:56 -08:00
9183ce1921 Handle links in Media::parse 2021-02-25 11:43:58 -08:00
a197df89ff Turn off media logging 2021-02-25 11:30:15 -08:00
be2a1d876b Fix url rewrites 2021-02-25 11:01:25 -08:00
686d61801f Fix #110 2021-02-25 10:24:37 -08:00
5d643277bc Geometric logo 2021-02-25 09:07:45 -08:00
a3ec44149c Categorize utilities 2021-02-24 21:29:23 -08:00
83ba0fb913 Reduce dependencies 2021-02-24 20:26:12 -08:00
55e9915bb0 Refactor post_body width 2021-02-24 11:28:26 -08:00
5cd5b553b0 Handle suspended users 2021-02-24 11:26:23 -08:00
2b2bd8421b Added multi-staged container build and a user (#134)
* Added multi-staged container build and a user

This reduces the image size from 2Gb to 92Mb by only put the necessary files into the container.
Container now runs without root priviliges.

* Add EXPOSE to Dockerfile
2021-02-24 19:17:36 +00:00
47d01a0dca Fix logging 2021-02-24 11:00:04 -08:00
0a69937238 Specify Ubuntu version of workflow 2021-02-24 18:44:50 +00:00
6d08f2dd24 Fix post body overflow on mobile 2021-02-24 09:31:58 -08:00
4a06882dc8 Simplify routes in main.rs 2021-02-24 09:26:01 -08:00
3e567d9acf Add new instance 2021-02-23 12:54:39 -08:00
8034594006 Better subreddit error messages. Closes #131 2021-02-22 16:43:32 -08:00
2f3315dcfc Fixes #130 2021-02-22 12:56:23 -08:00
df118764df Update README.md 2021-02-22 05:45:56 +00:00
d78f82649e List other images in manifest 2021-02-21 20:37:53 -08:00
80fb3a5c18 Fix #110 2021-02-21 20:28:04 -08:00
518d5753a7 Handle about pages 2021-02-21 10:13:20 -08:00
de38f7ef18 Fix post flairs 2021-02-21 10:11:17 -08:00
dd67b52199 Fix #126 2021-02-20 18:36:30 -08:00
9cfab348eb Filter by flair. Closes #124 2021-02-20 13:59:16 -08:00
e1f7b6d0c0 Remove defunct Libreddit instance 2021-02-20 20:19:56 +00:00
a606e48435 Handle 4 more unwraps 2021-02-20 12:14:32 -08:00
2091f26bda Fix media previews 2021-02-19 21:49:02 -08:00
b3341b49c0 Individually proxy subreddit and user icons 2021-02-19 21:46:44 -08:00
65e4ceff7b Individually proxy previews 2021-02-19 20:50:55 -08:00
bacb22f7f9 Fix post url indentation 2021-02-19 18:19:04 -08:00
902c9a6e42 Individually proxy custom emojis 2021-02-19 18:18:09 -08:00
c586de66ba Individually proxy images and thumbnails 2021-02-19 12:55:07 -08:00
e466be8946 Fix manifest and update dependencies 2021-02-19 11:10:48 -08:00
bed3465475 Remove "Options" and fix "Built With" 2021-02-19 09:20:09 -08:00
8560e8a37a Add "port" command line argument 2021-02-18 11:49:50 -08:00
3652342f46 Use clap for arg parsing 2021-02-18 11:40:10 -08:00
58127b17d8 Individually proxy videos 2021-02-18 10:04:59 -08:00
2f4deb221a Reduce dependencies 2021-02-16 11:39:56 -08:00
38230ed473 Add more rich meta tags (#121) 2021-02-16 19:16:32 +00:00
71501b064c Update dependencies 2021-02-15 14:09:45 -08:00
47a58ea05c Add CODEOWNERS file 2021-02-15 13:46:07 -08:00
14ecf3cf60 Edit indicator 2021-02-14 14:53:09 -08:00
aa7c8c85df Templatize redirects 2021-02-13 15:02:38 -08:00
0cb7031c36 Fix focus outline 2021-02-13 13:38:12 -08:00
93cfc713c6 Generate URL to restore settings, including subscriptions. Closes #89 (#116)
* Start recursive comments

* Update comment.html

* Fix move error

* Comment improvements

* Fix merge

* Remove extra endif from post.html

* Fix post.html

* Restore setting from link

* Tweak settings page

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-02-13 20:55:23 +00:00
ff8685ae4c Add tooltips for accessibility 2021-02-12 20:53:33 -08:00
f06320a4ae Subscribe to multireddit button. Closes #104 2021-02-12 20:47:54 -08:00
809be42e01 Add "View all comments" and "Show parent comments" buttons when viewing a single thread. Closes #65 (#115)
* Start recursive comments

* Update comment.html

* Fix move error

* Comment improvements

* Fix merge

* Remove extra endif from post.html

* Fix post.html

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-02-12 09:16:59 -08:00
58ca085521 Fix subscription list overflow 2021-02-11 09:18:32 -08:00
4a40e16277 Fix comment structuring (#113)
* Start recursive comments

* Update comment.html

* Fix move error

Co-authored-by: spikecodes <19519553+spikecodes@users.noreply.github.com>
2021-02-10 10:48:51 -08:00
fee2cb1b56 Split subscription names by + 2021-02-09 21:56:38 -08:00
8785bc95f5 Fix extra slashes in post bodies 2021-02-09 21:54:55 -08:00
16454213cf Fix Dockerfile 2021-02-10 01:31:28 +00:00
6feb347c27 Fix post ID parsing 2021-02-09 12:08:38 -08:00
e731cfbac4 Support post links without titles 2021-02-09 10:11:39 -08:00
008924fff8 Fix listen address 2021-02-09 09:54:13 -08:00
ebbdd7185f Move from Actix Web to Tide (#99)
* Initial commit

* Port posts

* Pinpoint Tide Bug

* Revert testing

* Add basic sub support

* Unwrap nested routes

* Front page & sync templates

* Port remaining functions

* Log request errors

* Clean main and settings

* Handle /w/ requests

* Create template() util

* Reduce caching time to 30s

* Fix subscription redirects

* Handle frontpage sorting
2021-02-09 17:38:52 +00:00
402b3149e1 Fix 'no entry found for key' error 2021-02-07 17:56:06 -08:00
ac5ef89dff Fix gallery unwrapping 2021-02-07 17:33:54 -08:00
7edca18f8d Inline videos/gifs for card view (#107)
* Basic gallery support

* Inline videos for card view
2021-02-08 00:22:14 +00:00
cf45d53fdd Basic gallery support (#103) 2021-02-06 20:05:11 +00:00
2a475d127a Added black theme (#101) 2021-02-06 20:04:29 +00:00
3fa523e67b Add a multi-architecture docker build. (#98)
* Add a multi-arch docker build.

* Remove caching
2021-02-04 22:17:04 +00:00
3fbb433e37 Update dependencies 2021-02-03 21:48:56 -08:00
5fbcfd850f Merge pull request #97 from Mennaruuk/master
Fix installation method links
2021-02-03 21:46:28 -08:00
c758db84ec Merge pull request #95 from robrobinbin/patch-5
Remove tiny "padding" below inline images
2021-02-03 21:45:11 -08:00
90d3063f93 Fix #96 2021-02-03 21:42:43 -08:00
82a601d534 Minor update to README.md 2021-02-04 00:39:46 -05:00
12a1b3f459 Update style.css 2021-02-03 22:01:46 +01:00
e23eaf0be0 Merge pull request #92 from robrobinbin/master
Fix layout on browsers with faulty "display: contents" support (Safari)
2021-02-03 11:55:04 -08:00
821709c8d2 Resolve merge conflict 2021-02-03 20:14:37 +01:00
653b0e7024 Don't use display contents and remove duplication 2021-02-03 20:11:04 +01:00
c7a2c43287 Merge pull request #93 from mcrossman/master
Make feed sorting case-insensitive
2021-02-02 19:13:35 -08:00
9824370771 Fix feed sorting to be case-insensitive. 2021-02-03 10:53:09 +11:00
d87b96d0ea Update style.css 2021-02-02 21:33:23 +01:00
6eae4bc47a Update style.css 2021-02-02 21:23:51 +01:00
1bcb070fbb Update user.html 2021-02-02 21:21:29 +01:00
24bc758090 Update subreddit.html 2021-02-02 21:21:11 +01:00
ffbb1cf7cd Update search.html 2021-02-02 21:20:38 +01:00
cbf1f540d6 Update post.html 2021-02-02 21:20:03 +01:00
f8e0d2d4b9 Merge pull request #7 from spikecodes/master
Merge upstream
2021-02-02 21:18:21 +01:00
8a27b2bac8 Support /w/ for Wikis 2021-02-02 08:59:50 -08:00
69941d9efd Implement #88 2021-02-01 17:50:00 -08:00
956de50419 Change Libreddit PWA Theme Color 2021-02-01 16:26:35 -08:00
d790264a62 Merge pull request #87 from robrobinbin/master
Add fallback to thumbnails when SVG is not supported
2021-02-01 12:53:03 -08:00
f4f2d8a377 Update style.css 2021-02-01 21:02:38 +01:00
dd908c9f68 Update user.html 2021-02-01 21:00:03 +01:00
9e1948733d Update subreddit.html 2021-02-01 20:59:31 +01:00
9df1dfae32 Update search.html 2021-02-01 20:58:59 +01:00
cfbee1bb81 Merge pull request #6 from spikecodes/master
Merge upstream
2021-02-01 20:57:06 +01:00
8430cbc6f3 Merge pull request #86 from robrobinbin/patch-4
Place noscript placeholder into grid
2021-02-01 11:45:34 -08:00
a9dd2e6f2c Place noscript placeholder into grid 2021-02-01 20:43:32 +01:00
36964982fb Merge pull request #85 from robrobinbin/master
Improve support for browsers without inline SVG support
2021-02-01 11:39:35 -08:00
0742a33304 Update base.html 2021-02-01 20:32:57 +01:00
7f320b3143 Update style.css 2021-02-01 20:27:56 +01:00
58f4fc4e77 Update user.html 2021-02-01 20:27:08 +01:00
7d8faefad0 Update search.html 2021-02-01 20:26:35 +01:00
ba9b5afd4e Update post.html 2021-02-01 20:25:57 +01:00
ae09f77bf6 Update subreddit.html 2021-02-01 20:25:06 +01:00
5030c418de Merge pull request #5 from spikecodes/master
Merge upstream
2021-02-01 20:23:35 +01:00
4ccd6b1751 Merge pull request #84 from JPyke3/tor-service
Added new Tor Hidden Service
2021-02-01 10:25:52 -08:00
7d17aa0627 Merge pull request #83 from JPyke3/master
iOS "Add to Homescreen" functionality
2021-02-01 10:25:28 -08:00
4b73e2d914 Added personal Hidden Service 2021-02-01 18:44:55 +01:00
0a140a6ffc Merge branch 'master' of github.com:JPyke3/libreddit into master 2021-02-01 11:13:36 +01:00
e837d84105 Add Support for iOS "Add to Homescreen"
* Adds basic Manifest.json
 * Adds Meta Tags for iOS
 * Adds Meta Tags for Android
 * Adds Logo for Manifest.json
 * Adds iOS Logo for homescreen
2021-02-01 11:10:53 +01:00
f6d791ccd9 Style focus outline 2021-01-31 20:56:13 -08:00
effaeb7508 Fix debug logging error 2021-01-31 19:08:50 -08:00
6257faf9dc Update screenshot 2021-01-31 19:05:06 -08:00
ee0da63862 Update version and screenshot 2021-01-31 18:50:31 -08:00
971f14bb55 Fix #66 2021-01-31 14:10:13 -08:00
9a1733ac99 Fix cog class 2021-01-31 13:49:55 -08:00
c32d62fbd5 Shrink links to icons on mobile 2021-01-31 13:22:11 -08:00
1a0d12d2ff Merge pull request #82 from robrobinbin/patch-2
Minor style tweak
2021-01-31 12:40:32 -08:00
2a27850914 Minor style tweak
on small screens, subscriptions are put below the logo, which looks a bit strange and makes it somehow harder to hit
2021-01-31 19:41:32 +01:00
bfcc4c985d Fix subreddit sorting and media 2021-01-30 21:46:35 -08:00
1653d4fb4c Hide NSFW content by default 2021-01-30 21:43:46 -08:00
79027c4c75 Merge pull request #78 from mcrossman/subscriptions
Subscribing to subreddits (favorites)
2021-01-30 21:28:51 -08:00
269bb0bfb6 Convert subscription requests to POST 2021-01-30 21:21:46 -08:00
7933d840b3 Squish navbar further 2021-01-30 20:42:12 -08:00
b875e9377e Merge branch 'master' into subscriptions 2021-01-30 20:25:48 -08:00
8c80946121 Fix multireddit sidebars 2021-01-30 20:24:09 -08:00
21d96e261f Set subscriptions as default front page 2021-01-30 20:18:57 -08:00
9c58d23b41 Unwrap #subscriptions 2021-01-30 18:16:42 -08:00
4ae2191392 Refactor subscriptions 2021-01-30 18:10:38 -08:00
d62a3ab86b Refactor redirect path setting 2021-01-30 17:50:26 -08:00
9b7cd1da5a Prevent adding duplicate subs. 2021-01-31 12:45:19 +11:00
a301f1ecb6 Simplify redirect path 2021-01-30 17:38:47 -08:00
f14639ee00 Simplify sub list editing. 2021-01-31 12:25:07 +11:00
b527735f6f Fix cookie deletion when removing last sub. 2021-01-31 12:17:43 +11:00
8cc01c58f3 Move the subs list into the navbar. 2021-01-31 09:52:18 +11:00
a1d800a0f0 Fix video loading 2021-01-30 11:00:55 -08:00
449899962a Change subscription to get. Add subs to settings. 2021-01-30 22:27:49 +11:00
dc2030e6f3 Vertical list subscriptions. 2021-01-30 21:21:54 +11:00
ef5a1cd66e Rename block sub_list > subscriptions 2021-01-30 19:47:30 +11:00
11e4ff42ed Clean-up and more consistent styles. 2021-01-30 19:33:38 +11:00
c71df35b22 Add basic unsubscribe. 2021-01-30 18:18:53 +11:00
345308a9ac Basic subscribe functionality. 2021-01-30 18:00:00 +11:00
75bbcefbec Display sub list from list in cookie.
Very basic sub list setup. Cookie must be manually added in devtools.
2021-01-30 16:00:55 +11:00
49a6168607 Improve CSP 2021-01-29 14:39:03 -08:00
f55ea5a353 Specify default headers for security 2021-01-28 21:53:10 -08:00
30c33d91e1 Specify video MIME in posts 2021-01-28 20:06:35 -08:00
00b135fb0f Fix #75 2021-01-28 15:50:18 -08:00
5fe9ce8d7b Refactor JSON parsing error logging 2021-01-27 21:48:32 -08:00
8c04365049 Improve error logging 2021-01-27 17:48:58 -08:00
d5b1c3a5bb Fix connection reset error 2021-01-26 11:00:07 -08:00
f038aa61f4 Log JSON parse errors and Reddit request failures 2021-01-26 10:54:57 -08:00
f72c9d39be Simplify proxy 2021-01-25 17:01:02 -08:00
e6c2d08425 Add libreddit.himiko.cloud instance 2021-01-24 23:33:18 +00:00
e901e99278 Merge pull request #71 from robrobinbin/patch-1
Lower number of cached items
2021-01-24 16:22:58 +00:00
acd2cff747 Lower number of cached items 2021-01-24 11:08:24 +01:00
8f913e696c Merge pull request #70 from robrobinbin/master
Experiment with caching requests to Reddit
2021-01-23 21:50:16 +00:00
226d39328c Make it 60 2021-01-23 22:45:41 +01:00
b2ad2f636c Resolve merge conflicts 2021-01-23 10:55:11 +01:00
18fe7ff8cf Merge pull request #4 from spikecodes/master
Merge upstream
2021-01-23 10:48:54 +01:00
077c222a4e Experiment with caching 2021-01-23 10:48:33 +01:00
2270b6cf95 Reduce post padding 2021-01-21 21:25:51 -08:00
758b627660 Merge pull request #68 from robrobinbin/master
Absolutly no jumping when images (fail to) load
2021-01-21 23:32:59 +00:00
baf7272cfd Absolutly no jumping when images (fail to) load 2021-01-21 22:04:06 +01:00
6641e242af Allow NSFW searching 2021-01-20 18:26:58 -08:00
610fcfbf87 Fix #67 2021-01-20 17:55:04 -08:00
dea7f33910 Add code optimizations 2021-01-20 17:38:34 -08:00
c299e128ab Inline images rework 2021-01-21 00:12:54 +01:00
53fa946c75 Merge pull request #64 from robrobinbin/master
Should fix the strange overflowing placeholder when an image fails to load
2021-01-20 21:46:58 +00:00
5d44a071f9 Why?? 2021-01-20 20:59:57 +01:00
e29e203188 Merge pull request #3 from spikecodes/master
Merge upstream
2021-01-20 20:18:08 +01:00
6ead6e08dc Update README.md 2021-01-20 05:55:59 +00:00
7360503234 Make posts single-color 2021-01-18 21:32:25 -08:00
140c1b1bfa Small fixes 2021-01-18 13:12:59 -08:00
040982f1fd Document command line options in Readme 2021-01-18 18:44:03 +00:00
4b0677d10e Merge pull request #2 from spikecodes/master
Merge upstream
2021-01-18 18:48:23 +01:00
616751e054 Embolden markdown blockquotes 2021-01-17 22:32:42 -08:00
5df957f193 Update README.md 2021-01-18 06:30:34 +00:00
7f9cb1b35a Fix post score and thumbnail backgrounds 2021-01-17 20:35:49 -08:00
c030771d36 Refine transitions 2021-01-17 19:16:15 -08:00
a562395c26 Refactor system theme 2021-01-17 18:11:36 -08:00
2bcdf68e40 Merge pull request #62 from mcrossman/master
Add "system" theme that matches browser/OS theme
2021-01-18 01:59:12 +00:00
72eaa685d0 Prevent "system" class from being added to body. 2021-01-18 11:25:39 +11:00
899a414cf6 Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-18 11:24:49 +11:00
524538eeb8 Update Google Lighthouse Report 2021-01-17 16:22:34 -08:00
a184559c21 Clean CSS 2021-01-17 16:20:17 -08:00
1c9fd46e98 Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-18 11:16:02 +11:00
738941d830 Fix arrow alignment 2021-01-17 16:03:14 -08:00
06ab7a4181 Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-18 10:59:26 +11:00
6981d94417 Clean up theme section of CSS. 2021-01-18 10:56:30 +11:00
dd60cb5b2b SVG arrow 2021-01-17 15:51:03 -08:00
1d57e29d56 Add "system" theme to settings. 2021-01-18 10:44:51 +11:00
2d973707f3 Add auto theme selection in style. 2021-01-18 10:42:55 +11:00
cbb937b494 Merge pull request #61 from robrobinbin/master
Improve accessibility rating
2021-01-17 23:00:25 +00:00
d45ee03122 Resolve merge conflicts 2021-01-17 23:54:48 +01:00
162e00b243 Improve accessibility rating 2021-01-17 23:49:36 +01:00
7a32ba087e Merge pull request #1 from spikecodes/master
Merge upstream
2021-01-17 23:47:09 +01:00
801216dfe9 Include code link on mobile 2021-01-17 13:24:44 -08:00
21763c51cd Make number formatting inclusive 2021-01-17 12:59:40 -08:00
138f8320e9 Create media struct 2021-01-17 12:58:12 -08:00
571ba3392c Merge pull request #60 from robrobinbin/master
Add comment counter and other post improvements
2021-01-17 20:05:01 +00:00
090ca1a140 Add comment counter and other post improvements 2021-01-17 20:39:57 +01:00
6127f2a90c Support forced HTTPS redirects #39 2021-01-16 22:04:03 -08:00
ef9bc791e1 Fix inline style tags 2021-01-16 19:21:47 -08:00
894323becf Update Screenshot 2021-01-16 17:47:45 -08:00
4c89d31948 Fix version number 2021-01-16 17:45:45 -08:00
471d181284 Disable production error logging 2021-01-16 15:17:08 -08:00
0e48c66b8c Fix user agent 2021-01-16 15:13:34 -08:00
a0bc1732cf Moderator and admin distinguishers 2021-01-16 15:02:24 -08:00
6d5fd1dbf6 Use main instance more in Readme 2021-01-16 11:56:13 -08:00
0f6e73dd87 Reformat code 2021-01-16 11:50:12 -08:00
151490faf0 Add space next to comment collapse marker 2021-01-16 11:49:49 -08:00
fdf60e7255 Separate datetime into relative and absolute 2021-01-16 11:40:32 -08:00
ab102ca32c Merge pull request #57 from robrobinbin/master
Improve support for text-only browsers
2021-01-16 10:49:41 -08:00
998b301229 Improve support for text-only browsers 2021-01-16 11:00:15 +01:00
d7839899e6 Merge pull request #3 from spikecodes/master
Merge upstream
2021-01-16 09:20:14 +01:00
2385fa33ec Use ureq until AWC IO error is fixed 2021-01-15 21:26:51 -08:00
1fd688eeed Improve awc error log 2021-01-15 20:57:51 -08:00
65543a43b2 Make User-Agent Reddit-compliant 2021-01-15 20:29:34 -08:00
0099021478 Refactor flair spacing 2021-01-15 15:55:10 -08:00
3a9b2dba32 Fix error log 2021-01-15 15:35:09 -08:00
59021b9331 Switch back to ureq temporarily 2021-01-15 15:28:51 -08:00
078d6fe25b Request about pages before posts 2021-01-15 15:05:55 -08:00
373ce55203 Recommend secondary instance 2021-01-15 11:27:06 -08:00
aef0442e9d Add rate-limit warning 2021-01-15 11:24:12 -08:00
21ff8d7b6f Fix #56 2021-01-15 11:21:59 -08:00
bca2a7e540 Error logging 2021-01-15 10:58:53 -08:00
0c014ad41b Comment utils.rs 2021-01-14 15:13:52 -08:00
32b8637c7e Handle failed redirects 2021-01-14 14:56:28 -08:00
5ed122d92c Merge pull request #55 from robrobinbin/master
Add placeholder image for posts without thumbnail
2021-01-14 14:26:40 -08:00
45660816ce Add cardview to search results too 2021-01-14 21:53:07 +01:00
d19e73f059 Add placeholder for posts without thumbnail 2021-01-14 21:45:43 +01:00
18684c934b Refactor subreddit searching 2021-01-14 11:45:04 -08:00
cf4c5e1fe8 Implement #53 2021-01-14 10:57:50 -08:00
7ef4a20aff Merge pull request #54 from robrobinbin/master
Add subreddits to search results, closes #18
2021-01-14 10:54:20 -08:00
292f8fbbb7 remove lines that aren't used yet 2021-01-14 19:33:17 +01:00
735f79d80b Merge pull request #2 from spikecodes/master
Merge upstream into code
2021-01-14 19:29:06 +01:00
a85a4278f6 Add subreddits to search results 2021-01-14 19:22:50 +01:00
dbe617d7eb Switch to awc 2021-01-14 09:53:54 -08:00
842d97e9fa Fix short post IDs 2021-01-13 21:02:48 -08:00
0bf5576427 Categorize routes and refactor error handlers 2021-01-13 19:53:52 -08:00
dd027bff4b Refactor flair parsing 2021-01-13 18:19:40 -08:00
f95ef51017 Add days to time() 2021-01-13 16:31:24 -08:00
740641cb4e Move nested_val() to user.rs 2021-01-13 15:55:10 -08:00
09c98c8da6 Refactor code 2021-01-13 12:52:00 -08:00
33c8bdffb9 Merge pull request #1 from spikecodes/master
Merge upstream
2021-01-13 20:22:43 +01:00
5ab88567de Merge pull request #50 from robrobinbin/rich-flairs
Add support for rich flairs with "Emoji"
2021-01-13 10:48:25 -08:00
c6627ceece Merge branch 'master' into rich-flairs 2021-01-13 08:27:39 +01:00
d9affcdefc Rich flairs 2021-01-13 08:23:48 +01:00
96607256fc Add Favicon 2021-01-12 20:18:20 -08:00
eb9a0dcb4a Fix GIFs 2021-01-12 19:52:02 -08:00
89fa0d5489 Merge pull request #47 from robrobinbin/scrollbar-for-overflowing-code
Add overflow "auto" to "pre"
2021-01-12 15:16:02 -08:00
22589c8296 Merge pull request #48 from robrobinbin/relative_timestamps
Relative timestamps for posts younger than 24h
2021-01-12 15:15:37 -08:00
b0540d2c57 Rich flairs 2021-01-13 00:10:06 +01:00
41c4661bbb Rich flairs 2021-01-12 23:55:35 +01:00
d2314580a9 Rich flairs 2021-01-12 23:50:50 +01:00
a4d77926b6 Rich flairs 2021-01-12 23:34:16 +01:00
bbe7024323 Start richtext flairs 2021-01-12 22:43:03 +01:00
32e1469e11 whitespace 2021-01-12 20:03:54 +01:00
2d4ca2379f whitespace 2021-01-12 20:02:35 +01:00
374f53eb32 Relative timestamps for recent posts 2021-01-12 19:59:32 +01:00
add7efea3c Update style.css 2021-01-12 18:53:10 +01:00
065d82a5f5 Merge pull request #45 from somoso/long-label-fix
Stop label from being long for joined subreddits
2021-01-12 08:46:35 -08:00
1895bbc025 Merge pull request #46 from somoso/ios-thumbnail-fix
Fix the thumbnail issue on iOS
2021-01-12 08:46:17 -08:00
65f1a2afb2 Stop label from being long for joined subreddits
Browsing with a long joined subreddit list will cause the label to look a bit weird.

Example: https://libredd.it/r/Android+AnimalsBeingBros+AnimalsBeingDerps+AnimalsBeingJerks+AppleWatch+CatSlaps+CatSmiles+CatsBeingAdorable+FreeTube+Games+Ijustwatched+IllegallySmol+IllegallySmolCats+IpodClassic+LearnRubyonRails+Megadrive+MovieDetails+Music+NetflixBestOf+NintendoSwitch+Possums+Teefers+UKPersonalFinance+airplaneears+apple+aww+brushybrushy+cats+catswhotrill+curledfeetsies+cyberpunkgame+dataisbeautiful+dechonkers+digital_ocean+dogs+dogsareliquid+dogswithjobs+emulation+greebles+happycowgifs+hardware+iOSDevelopment+iOSProgramming+iosdev+kittykankles+learnruby+likeus+mac+mashups+microsoft+movies+netflix+netsec+pihole+playstation+programming+rarepuppers+raspberry_pi+redditsync+rubyonrails+satelliteears+shittymoviedetails+spookyteefies+technology+teefies+vampirecats+velvethippos

That will cause the label to be excessively long
2021-01-12 15:47:39 +00:00
6eb9e6f0c0 Fix the thumbnail issue on iOS 2021-01-12 15:43:27 +00:00
eb735a42fe Handle comment parsing errors 2021-01-11 18:05:13 -08:00
541c741bde Parse GIFs correctly 2021-01-11 17:47:14 -08:00
7a33ed3434 Card thumbnails for users 2021-01-11 17:38:35 -08:00
48d2943f72 Fix subreddits not showing sidebars 2021-01-11 16:44:31 -08:00
6bbc90bc0d Clean Subreddit struct 2021-01-11 16:35:50 -08:00
4d18dc0bb8 Merge pull request #44 from robrobinbin/master
Make thumbnail clickable and bring behavior closer to reddit.
2021-01-11 16:35:14 -08:00
6dbd002acd Add direct link to thumbnail 2021-01-11 23:08:12 +01:00
bf6245a505 Fix multireddit sidebars 2021-01-11 10:39:36 -08:00
91746908a1 Switch to ureq 2021-01-11 10:33:48 -08:00
bb8273bab4 Fix #41 2021-01-11 10:33:42 -08:00
62bcc31305 Fix Wide UI on Mobile 2021-01-10 18:48:08 -08:00
08683fa5a6 Light theme 2021-01-10 18:15:34 -08:00
c58b077330 Update Dependencies 2021-01-10 13:20:47 -08:00
f445c42f55 Wide UI Mode 2021-01-10 13:08:36 -08:00
a0866b251e Update issue templates 2021-01-10 11:23:53 -08:00
aa819544f6 Update Matrix Address 2021-01-09 13:08:08 -08:00
fac56d7f87 Markdown spoilers and post footers on videos 2021-01-08 21:57:36 -08:00
ef1ad17234 Unknown path error handling 2021-01-08 21:11:20 -08:00
b8cdc605a2 Front page config and settings note 2021-01-08 20:55:40 -08:00
ef2f9ad12b Unify preferences under one struct 2021-01-08 17:50:03 -08:00
b13874d0db Add "hide nsfw" option 2021-01-08 17:35:04 -08:00
3d142afd03 Merge pull request #38 from somoso/patch-3
Break word to stop it disappearing on mobile
2021-01-08 17:07:50 -08:00
7fcb7fcfed Break word to stop it disappearing on mobile
This kept happening to me but I couldn't reproduce it in the iPad Simulator. Finally got it nailed down and sorted.

Tested this on Safari (mobile and desktop), Firefox, and Edge browser.
2021-01-08 23:08:58 +00:00
747d5a7c67 Merge pull request #37 from somoso/patch-2
Fix theming for all browsers
2021-01-08 12:56:00 -08:00
770c4d3630 Fix themeing for all browsers
Really noticable on iOS, but ensuring all browsers get the love.

The buttons and input aren't as flat as they usually are on my desktop Firefox. This patch should sort that out.
2021-01-08 20:26:29 +00:00
e7b448a282 Add shadow to navbar 2021-01-07 10:49:10 -08:00
c7c787dff1 Fix comment padding 2021-01-07 10:49:00 -08:00
59a34a0e85 Fixed navbar 2021-01-07 10:46:00 -08:00
6e8cf69227 Fix Default Comment Sorting 2021-01-07 10:32:55 -08:00
3444989f9a Default Comment Sort Setting 2021-01-07 08:38:05 -08:00
7e96bb3d80 Optimize use of Result<> 2021-01-06 21:27:24 -08:00
0adbb1556e Merge branch 'master' of https://github.com/spikecodes/libreddit 2021-01-06 16:45:57 -08:00
710eecdb9d Add issue templates 2021-01-06 16:01:13 -08:00
8a57fa8a1d Remove "Safe" description from README 2021-01-06 14:19:55 -08:00
b33d79ed9b Cache robots.txt 2021-01-06 14:19:10 -08:00
0f506fc41b Cache proxied media 2021-01-06 11:11:04 -08:00
c9cd825d55 Create CSS variables for shadow and text color 2021-01-06 10:51:13 -08:00
e63384e6a6 Update cookie description 2021-01-06 09:54:32 -08:00
3260a4d596 Disable "secure" flag for cookies 2021-01-06 09:52:23 -08:00
da5c4603d9 Switch from chrono to time-rs 2021-01-05 20:01:21 -08:00
b50fa6f3ae Settings Button 2021-01-05 18:16:32 -08:00
aa7b4b2af7 Settings with Layouts 2021-01-05 18:04:49 -08:00
2b0193f5ea Fix proxying of NSFW images 2021-01-05 08:15:34 -08:00
2185d895c0 Prevent user datetimes from floating 2021-01-04 21:32:22 -08:00
9c1a932214 Clean Up Post Headers 2021-01-04 21:17:19 -08:00
8c0269af1c Fix post tags on mobile 2021-01-04 19:43:35 -08:00
df89c5076e Compact Libreddit Posts on Mobile 2021-01-04 19:26:41 -08:00
f819ad2bc6 Remove CSP "Upgrade Insecure Requests" Header 2021-01-04 10:11:07 -08:00
f5884a5270 Update Screenshot 2021-01-03 21:32:45 -08:00
c046d00060 Handle Unwrapping Errors 2021-01-03 21:31:21 -08:00
5934e34ea0 Merge pull request #30 from moosingin3space/master
Add controversial sort order
2021-01-03 21:15:27 -08:00
463b44ac52 Fix timeframe when sorting by controversial 2021-01-04 05:05:21 +00:00
b40d21e559 Add controversial sort order 2021-01-03 21:00:36 -08:00
a422a74747 Make Design More Compact 2021-01-03 19:44:44 -08:00
4124fa87d3 Correct Readme 2021-01-03 18:24:30 -08:00
1dd0c4ee20 Fix User Icon Proxy 2021-01-03 18:23:57 -08:00
0dd114c166 Post upvote ratio, permalink and reddit link 2021-01-03 13:06:49 -08:00
67090e9b08 Fix Proxied Icons 2021-01-03 10:22:41 -08:00
d97fb49fde Fix post::item IDs 2021-01-02 22:46:02 -08:00
9263b0657f Fix navbar padding 2021-01-02 22:40:22 -08:00
a3384cbaa6 Fix search pages 2021-01-02 22:37:54 -08:00
5d26b5c764 Upgrade Insecure Requests 2021-01-02 20:59:14 -08:00
516403ee47 Fix Readme 2021-01-02 20:59:04 -08:00
5ea504e6e8 Restrict Proxy to Reddit Domains 2021-01-02 20:50:23 -08:00
f49bff9853 Optimize Sequencing 2021-01-02 11:09:26 -08:00
4ec529cdb8 Rewrite Reddit Links to Libreddit 2021-01-02 10:58:21 -08:00
779de6f8af Fix Wiki Titles 2021-01-01 22:34:25 -08:00
0925a9b334 Add Wiki Pages 2021-01-01 22:21:43 -08:00
2f2ed6169d Optimize use of .unwrap() 2021-01-01 15:28:13 -08:00
59ef30c76d Remove .clone() in favor of borrowing 2021-01-01 12:55:09 -08:00
d43b49e7e4 Optimize Rust code with Clippy 2021-01-01 12:33:57 -08:00
64a92195dd Merge pull request #19 from somoso/patch-1
Fix posts overflowing on Safari on iOS
2021-01-01 11:52:21 -08:00
a7925ed62d Fix posts overflowing on Safari on iOS
In Safari, the value `anywhere` is not supported for property `overflow-wrap`. Once changed to `break-word`, it behaves like it does in Chrome and Firefox.
2021-01-01 15:46:36 +00:00
39ba50dada Error Page 2020-12-31 21:03:44 -08:00
bc1b29246d Update Screenshot 2020-12-31 20:23:19 -08:00
2d77a91150 Refactor Page Titles and Add Subreddit/User Titles 2020-12-31 20:21:56 -08:00
93c1db502d Fix Title and Navbar 2020-12-31 16:45:10 -08:00
a6dc7ee043 Rewrite + Searching 2020-12-31 15:54:13 -08:00
c7282520cd Add Focus Indicator 2020-12-30 10:53:27 -08:00
a866c1d068 Update Screenshot for v0.2.3 2020-12-29 19:40:49 -08:00
aa9aad6743 Stickied Posts 2020-12-29 19:01:02 -08:00
f65ee2eb6a Sort Top by Timeframe 2020-12-29 17:11:47 -08:00
44c4341e67 Update README.md 2020-12-29 13:29:24 -08:00
1c886f8003 Merge pull request #10 from StuffNoOneCaresAbout/add-instance
instances: add libreddit.kavin.rocks
2020-12-29 09:53:23 -08:00
b481d26be2 instances: add libreddit.kavin.rocks
And it's onion counterpart.
2020-12-29 12:53:49 +05:30
f00ef59404 Fix proxy-less deployment 2020-12-28 20:49:15 -08:00
3115ff3436 Update README.md 2020-12-28 18:45:46 -08:00
443b198c12 Markdown and Subreddit Sidebars 2020-12-28 18:42:46 -08:00
ac84d8d2db List Instances as a Table 2020-12-27 21:46:39 -08:00
e27cf94fbf Include Comment Lines in User History 2020-12-27 15:37:01 -08:00
68495fb280 Add Pages to User Profiles 2020-12-27 12:36:10 -08:00
bec5c78709 Persist Sort on Subreddit Pages 2020-12-26 12:43:51 -08:00
abfcfdf09e Merge pull request #8 from zachjmurphy/master
Add InsanityWtf Libreddit Instance
2020-12-25 18:55:41 -08:00
dad01749e6 Replace Responsive Feature with Secure 2020-12-25 18:06:33 -08:00
2efb73cee3 Update README.md 2020-12-25 17:09:11 -05:00
ace21b21d5 Redesign User/Subreddit About Boxes 2020-12-23 22:16:04 -08:00
280e16bd7f Fix Subreddit Icons 2020-12-23 20:36:49 -08:00
44d44a529c Add DotHQ Libreddit Instance 2020-12-23 13:34:05 -08:00
0957f2e339 NSFW Support 2020-12-22 18:29:43 -08:00
3516404a5f Update v0.2.2 2020-12-22 09:15:55 -08:00
d96daa335f Add Repl.it as Hosting Method 2020-12-22 09:11:03 -08:00
285d9da26d Further Document Libreddit Privacy 2020-12-22 08:45:21 -08:00
9ab7a72bce Fix Comparison Heading Position 2020-12-21 21:53:54 -08:00
46dd905509 Fix Grammar Mistakes 2020-12-21 21:52:50 -08:00
63d595c67d Revise README 2020-12-21 21:51:18 -08:00
dc0b5f42e6 Update README with Reddit Comparison 2020-12-21 21:40:06 -08:00
9ecbd25488 Reorganize CSS 2020-12-21 21:39:55 -08:00
83816fbcc6 Allow Indexing 2020-12-21 21:39:10 -08:00
11cfbdc3ed More Replies Button 2020-12-21 17:17:40 -08:00
4b7cbb3de2 Fix User Icons 2020-12-21 14:12:53 -08:00
b1a572072c Highlight Post Authors in Comments 2020-12-21 08:38:24 -08:00
b1071e9579 Switch Sorting System to Dropdown 2020-12-20 21:49:31 -08:00
da971f8680 Optimized Nested Comments for Mobile, Added IDs 2020-12-20 20:52:15 -08:00
b596f86cc2 Update Screenshot 2020-12-20 19:05:44 -08:00
3bcf0832a1 Correct README Regarding Multireddits 2020-12-20 17:45:52 -08:00
565f4f23b3 Multireddit Support & Referrer Policy 2020-12-20 17:45:26 -08:00
ef3820a2e1 User Flairs 2020-12-20 11:29:23 -08:00
1678245750 Add Sorting to Short Links 2020-12-20 09:10:37 -08:00
3594b6d41f Fix CSS and CSP 2020-12-19 22:25:00 -08:00
a754d42b9e Enforce Content Security Policy 2020-12-19 21:49:10 -08:00
c7e0234d33 Fix comment hover color 2020-12-19 21:44:30 -08:00
11a9ff53e4 Update README 2020-12-19 21:44:07 -08:00
37 changed files with 3393 additions and 3294 deletions

24
.github/ISSUE_TEMPLATE/bug_report.md vendored Normal file
View File

@ -0,0 +1,24 @@
---
name: Bug report
about: Create a report to help us improve
title: Bug Report | [title]
labels: bug
assignees: ''
---
**Describe the bug**
A clear and concise description of what the bug is.
**To reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Additional context**
Add any other context about the problem here.

View File

@ -0,0 +1,20 @@
---
name: Feature request
about: Suggest an idea for this project
title: Feature Request | [title]
labels: enhancement
assignees: ''
---
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.

36
.github/workflows/docker-build.yml vendored Normal file
View File

@ -0,0 +1,36 @@
name: Docker Multi-Architecture Build
on:
push:
paths-ignore:
- "**.md"
branches:
- master
jobs:
build-docker:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: Set up QEMU
uses: docker/setup-qemu-action@v1
with:
platforms: all
- name: Set up Docker Buildx
id: buildx
uses: docker/setup-buildx-action@v1
with:
version: latest
- name: Login to DockerHub
uses: docker/login-action@v1
with:
username: ${{ secrets.DOCKER_USERNAME }}
password: ${{ secrets.DOCKER_PASSWORD }}
- name: Build and push
uses: docker/build-push-action@v2
with:
context: .
file: ./Dockerfile
platforms: linux/amd64,linux/arm64
push: true
tags: spikecodes/libreddit:latest

View File

@ -11,7 +11,7 @@ env:
jobs:
build:
runs-on: ubuntu-latest
runs-on: ubuntu-18.04
steps:
- uses: actions/checkout@v2

1
.gitignore vendored
View File

@ -1 +1,2 @@
/target
Cargo.lock

1
CODEOWNERS Normal file
View File

@ -0,0 +1 @@
* @spikecodes

2263
Cargo.lock generated

File diff suppressed because it is too large Load Diff

View File

@ -3,21 +3,19 @@ name = "libreddit"
description = " Alternative private front-end to Reddit"
license = "AGPL-3.0"
repository = "https://github.com/spikecodes/libreddit"
version = "0.2.0"
version = "0.4.0"
authors = ["spikecodes <19519553+spikecodes@users.noreply.github.com>"]
edition = "2018"
[features]
default = ["proxy"]
proxy = ["actix-web/rustls", "base64"]
[dependencies]
base64 = { version = "0.13.0", optional = true }
actix-web = "3.2.0"
reqwest = { version = "0.10", default_features = false, features = ["rustls-tls"] }
askama = "0.8.0"
serde = "1.0.117"
serde_json = "1.0"
pulldown-cmark = "0.8.0"
chrono = "0.4.19"
async-recursion = "0.3.1"
askama = { version = "0.10.5", default-features = false }
async-recursion = "0.3.2"
async-std = { version = "1.9.0", features = ["attributes"] }
async-tls = { version = "0.11.0", default-features = false, features = ["client"] }
cached = "0.23.0"
clap = { version = "2.33.3", default-features = false }
regex = "1.4.3"
serde = { version = "1.0.124", features = ["derive"] }
serde_json = "1.0.64"
tide = { version = "0.16.0", default-features = false, features = ["h1-server", "cookies"] }
time = "0.2.25"

View File

@ -1,9 +1,17 @@
FROM rust:alpine as builder
FROM rust:latest as builder
WORKDIR /usr/src/libreddit
COPY . .
RUN apk add --no-cache g++ openssl-dev
RUN cargo install --path .
FROM alpine:latest
FROM debian:buster-slim
RUN apt-get update && apt-get install -y libcurl4 && rm -rf /var/lib/apt/lists/*
COPY --from=builder /usr/local/cargo/bin/libreddit /usr/local/bin/libreddit
CMD ["libreddit"]
RUN useradd --system --user-group --home-dir /nonexistent --no-create-home --shell /usr/sbin/nologin libreddit
USER libreddit
EXPOSE 8080
CMD ["libreddit"]

194
README.md
View File

@ -2,58 +2,150 @@
> An alternative private front-end to Reddit
Libre + Reddit = Libreddit
![screenshot](https://i.ibb.co/74gZ4pd/libreddit-rust.png)
- 🚀 Fast: written in Rust for blazing fast speeds and safety
- ☁️ Light: no javascript, no ads, no tracking
---
**10 second pitch:** Libreddit is a portmanteau of "libre" (meaning freedom) and "Reddit". It is a private front-end like [Invidious](https://github.com/iv-org/invidious) but for Reddit. Browse the coldest takes of [r/unpopularopinion](https://libredd.it/r/unpopularopinion) without being [tracked](#reddit).
- 🚀 Fast: written in Rust for blazing fast speeds and memory safety
- ☁️ Light: no JavaScript, no ads, no tracking, no bloat
- 🕵 Private: all requests are proxied through the server, including media
- 🔒 Safe: does not rely on Reddit OAuth or require a Reddit API Key
- 📱 Responsive: works great on mobile!
- 🔒 Secure: strong [Content Security Policy](https://developer.mozilla.org/en-US/docs/Web/HTTP/CSP) prevents browser requests to Reddit
Think Invidious but for Reddit. Watch your cat videos without being watched.
---
## Screenshot
**BTC:** bc1qwyxjnafpu3gypcpgs025cw9wa7ryudtecmwa6y
![](https://i.ibb.co/Tgjb3w7/image.png)
**XMR:** 45FJrEuFPtG2o7QZz2Nps77TbHD4sPqxViwbdyV9A6ktfHiWs47UngG5zXPcLoDXAc8taeuBgeNjfeprwgeXYXhN3C9tVSR
## About
---
### Elsewhere
Find Libreddit on...
- 💬 Matrix: [#libreddit:matrix.org](https://matrix.to/#/#libreddit:matrix.org)
- 🐋 Docker: [spikecodes/libreddit](https://hub.docker.com/r/spikecodes/libreddit)
- :octocat: GitHub: [spikecodes/libreddit](https://github.com/spikecodes/libreddit)
- 🦊 GitLab: [spikecodes/libreddit](https://gitlab.com/spikecodes/libreddit)
## Jump to...
- [About](#about)
- [Teddit Comparison](#how-does-it-compare-to-teddit)
- [Comparison](#comparison)
- [Installation](#installation)
- [Cargo](#1-cargo)
- [Docker](#2-docker)
- [AUR](#3-aur)
- [GitHub Releases](#4-github-releases)
- [Repl.it](#5-replit)
- [Deployment](#deployment)
### Info
Libreddit hopes to provide an easier way to browse Reddit, without the ads, trackers and bloat. Libreddit was inspired by other alternative front-ends to popular services such as [Invidious](https://github.com/iv-org/invidious) for YouTube, [Nitter](https://github.com/zedeus/nitter) for Twitter, and [Bibliogram](https://sr.ht/~cadence/bibliogram/) for Instagram.
---
Libreddit currently implements most of Reddit's functionalities but still lacks a few features that are being worked on below.
# Instances
### In Progress
- Nested comments
- User flairs
- Searching
Feel free to [open an issue](https://github.com/spikecodes/libreddit/issues/new) to have your [selfhosted instance](#deployment) listed here!
### How does it compare to Teddit?
| Website | Country | Cloudflare |
|-|-|-|
| [libredd.it](https://libredd.it) (official) | 🇺🇸 US | |
| [libreddit.spike.codes](https://libreddit.spike.codes) (official) | 🇺🇸 US | |
| [libreddit.dothq.co](https://libreddit.dothq.co) | 🇺🇸 US | |
| [libreddit.kavin.rocks](https://libreddit.kavin.rocks) | 🇮🇳 IN | ✅ |
| [libreddit.himiko.cloud](https://libreddit.himiko.cloud) | 🇫🇮 FI | |
| [libreddit.bcow.xyz](https://libreddit.bcow.xyz) | 🇺🇸 US | |
| [spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion](http://spjmllawtheisznfs7uryhxumin26ssv2draj7oope3ok3wuhy43eoyd.onion) | 🇮🇳 IN | |
| [fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion](http://fwhhsbrbltmrct5hshrnqlqygqvcgmnek3cnka55zj4y7nuus5muwyyd.onion) | 🇩🇪 DE | |
| [libreddit.himiko7xl2skojc6odi7hykl626gt4qki3vxdbv33u2u3af76d6k32ad.onion](http://libreddit.himiko7xl2skojc6odi7hykl626gt4qki3vxdbv33u2u3af76d6k32ad.onion) | 🇫🇮 FI | |
| [dflv6yjt7il3n3tggf4qhcmkzbti2ppytqx3o7pjrzwgntutpewscyid.onion](http://dflv6yjt7il3n3tggf4qhcmkzbti2ppytqx3o7pjrzwgntutpewscyid.onion/) | 🇺🇸 US | |
A checkmark in the "Cloudflare" category here refers to the use of the reverse proxy, [Cloudflare](https://cloudflare). The checkmark will not be listed for a site which uses Cloudflare DNS but rather the proxying service which grants Cloudflare the ability to monitor traffic to the website.
---
# About
Find Libreddit on 💬 [Matrix](https://matrix.to/#/#libreddit:kde.org), 🐋 [Docker](https://hub.docker.com/r/spikecodes/libreddit), :octocat: [GitHub](https://github.com/spikecodes/libreddit), and 🦊 [GitLab](https://gitlab.com/spikecodes/libreddit).
## Built with
- [Rust](https://www.rust-lang.org/) - Programming language
- [Tide](https://github.com/http-rs/tide) - Web server
- [Askama](https://github.com/djc/askama) - Templating engine
- [Surf](https://github.com/http-rs/surf) - HTTP client
## Info
Libreddit hopes to provide an easier way to browse Reddit, without the ads, trackers, and bloat. Libreddit was inspired by other alternative front-ends to popular services such as [Invidious](https://github.com/iv-org/invidious) for YouTube, [Nitter](https://github.com/zedeus/nitter) for Twitter, and [Bibliogram](https://sr.ht/~cadence/bibliogram/) for Instagram.
Libreddit currently implements most of Reddit's (signed-out) functionalities but still lacks [a few features](https://github.com/spikecodes/libreddit/issues).
## How does it compare to Teddit?
Teddit is another awesome open source project designed to provide an alternative frontend to Reddit. There is no connection between the two and you're welcome to use whichever one you favor. Competition fosters innovation and Teddit's release has motivated me to build Libreddit into an even more polished product.
If you are looking to compare, the biggest differences I have noticed are:
- Libreddit is themed around Reddit's redesign whereas Teddit appears to stick much closer to Reddit's old design. This may suit some users better as design is always subjective.
- Libreddit is written in Rust for speed and memory safety. It uses Actix Web, which was [benchmarked as the fastest web server for single queries](https://www.techempower.com/benchmarks/#hw=ph&test=db).
- Unlike Teddit (at the time of writing this), Libreddit does not require a Reddit API key to host.
- Libreddit is written in [Rust](https://www.rust-lang.org) for speed and memory safety. It uses [Actix Web](https://actix.rs), which was [benchmarked as the fastest web server for single queries](https://www.techempower.com/benchmarks/#hw=ph&test=db).
## Instances
---
Feel free to [open an issue](https://github.com/spikecodes/libreddit/issues/new) to have your selfhosted instance listed here!
# Comparison
- [libredd.it](https://libredd.it) 🇺🇸 (Thank you to [YeapGuy](https://github.com/YeapGuy)!)
- [libreddit.spike.codes](https://libreddit.spike.codes) 🇺🇸
This section outlines how Libreddit compares to Reddit.
## Installation
## Speed
### A) Cargo
Lasted tested Jan 17, 2021.
Results from Google Lighthouse ([Libreddit Report](https://lighthouse-dot-webdotdevsite.appspot.com/lh/html?url=https%3A%2F%2Flibredd.it), [Reddit Report](https://lighthouse-dot-webdotdevsite.appspot.com/lh/html?url=https%3A%2F%2Fwww.reddit.com%2F)).
| | Libreddit | Reddit |
|------------------------|---------------|------------|
| Requests | 20 | 70 |
| Resource Size (card ui)| 1,224 KiB | 1,690 KiB |
| Time to Interactive | **1.5 s** | **11.2 s** |
## Privacy
### Reddit
**Logging:** According to Reddit's [privacy policy](https://www.redditinc.com/policies/privacy-policy), they "may [automatically] log information" including:
- IP address
- User-agent string
- Browser type
- Operating system
- Referral URLs
- Device information (e.g., device IDs)
- Device settings
- Pages visited
- Links clicked
- The requested URL
- Search terms
**Location:** The same privacy policy goes on to describe location data may be collected through the use of:
- GPS (consensual)
- Bluetooth (consensual)
- Content associated with a location (consensual)
- Your IP Address
**Cookies:** Reddit's [cookie notice](https://www.redditinc.com/policies/cookies) documents the array of cookies used by Reddit including/regarding:
- Authentication
- Functionality
- Analytics and Performance
- Advertising
- Third-Party Cookies
- Third-Party Site
### Libreddit
For transparency, I hope to describe all the ways Libreddit handles user privacy.
**Logging:** In production (when running the binary, hosting with docker, or using the official instances), Libreddit logs when Reddit is ratelimiting Libreddit and when Reddit's JSON responses can't be parsed. When debugging (running from source without `--release`), Libreddit logs post IDs and URL paths fetched to aid with troubleshooting.
**DNS:** Both official domains (`libredd.it` and `libreddit.spike.codes`) use Cloudflare as the DNS resolver. Though, the sites are not proxied through Cloudflare meaning Cloudflare doesn't have access to user traffic.
**Cookies:** Libreddit uses optional cookies to store any configured settings in [the settings menu](https://libredd.it/settings). This is not a cross-site cookie and the cookie holds no personal data, only a value of the possible layout.
**Hosting:** The official instances are hosted on [Repl.it](https://repl.it/) which monitors usage to prevent abuse. I can understand if this invalidates certain users' threat models and therefore, selfhosting and browsing through Tor are welcomed.
---
# Installation
## 1) Cargo
Make sure Rust stable is installed along with `cargo`, Rust's package manager.
@ -61,9 +153,9 @@ Make sure Rust stable is installed along with `cargo`, Rust's package manager.
cargo install libreddit
```
### B) Docker
## 2) Docker
Deploy the Docker image of Libreddit:
Deploy the [Docker image](https://hub.docker.com/r/spikecodes/libreddit) of Libreddit:
```
docker run -d --name libreddit -p 8080:8080 spikecodes/libreddit
```
@ -73,39 +165,47 @@ Deploy using a different port (in this case, port 80):
docker run -d --name libreddit -p 80:8080 spikecodes/libreddit
```
### C) AUR
## 3) AUR
For ArchLinux users, Libreddit is available from the AUR as [`libreddit-git`](https://aur.archlinux.org/packages/libreddit-git).
Install:
```
yay -S libreddit-git
```
### D) GitHub Releases
## 4) GitHub Releases
If you're on Linux and none of these methods work for you, you can grab a Linux binary from [the newest release](https://github.com/spikecodes/libreddit/releases/latest).
Currently Libreddit does not have Windows or MacOS binaries but those will be available soon.
## Deploy an Instance
## 5) Repl.it
Once installed, deploy Libreddit (unless you're using Docker) by running:
**Note:** Repl.it is a free option but they are *not* private and will monitor server usage to prevent abuse. If you need a free and easy setup, this method may work best for you.
1. Create a Repl.it account (see note above)
2. Visit [the official Repl](https://repl.it/@spikethecoder/libreddit) and fork it
3. Hit the run button to download the latest Libreddit version and start it
In the web preview (defaults to top right), you should see your instance hosted where you can assign a [custom domain](https://docs.repl.it/repls/web-hosting#custom-domains).
---
# Deployment
Once installed, deploy Libreddit to `0.0.0.0:8080` by running:
```
libreddit
```
Specify a custom address for the server by passing the `-a` or `--address` argument:
```
libreddit --address=0.0.0.0:8111
```
## Proxying using NGINX
To disable the media proxy built into Libreddit, run:
```
libreddit --no-default-features
**NOTE** If you're [proxying Libreddit through a NGINX Reverse Proxy](https://github.com/spikecodes/libreddit/issues/122#issuecomment-782226853), add
```nginx
proxy_http_version 1.1;
```
to your NGINX configuration file above your `proxy_pass` line.
## Building from Source
## Building
```
git clone https://github.com/spikecodes/libreddit

View File

@ -1,68 +1,277 @@
// Import Crates
use actix_web::{get, middleware::NormalizePath, web, App, HttpResponse, HttpServer};
// Global specifiers
#![forbid(unsafe_code)]
#![warn(clippy::pedantic, clippy::all)]
#![allow(
clippy::needless_pass_by_value,
clippy::match_wildcard_for_single_variants,
clippy::cast_possible_truncation,
clippy::similar_names,
clippy::cast_possible_wrap
)]
// Reference local files
mod popular;
mod post;
mod proxy;
mod search;
mod settings;
mod subreddit;
mod user;
mod utils;
// Create Services
async fn style() -> HttpResponse {
HttpResponse::Ok().content_type("text/css").body(include_str!("../static/style.css"))
}
// Import Crates
use clap::{App, Arg};
use proxy::handler;
use tide::{
utils::{async_trait, After},
Middleware, Next, Request, Response,
};
use utils::{error, redirect};
async fn robots() -> HttpResponse {
HttpResponse::Ok().body(include_str!("../static/robots.txt"))
}
// Build middleware
struct HttpsRedirect<HttpsOnly>(HttpsOnly);
struct NormalizePath;
#[get("/favicon.ico")]
async fn favicon() -> HttpResponse {
HttpResponse::Ok().body("")
}
#[async_trait]
impl<State, HttpsOnly> Middleware<State> for HttpsRedirect<HttpsOnly>
where
State: Clone + Send + Sync + 'static,
HttpsOnly: Into<bool> + Copy + Send + Sync + 'static,
{
async fn handle(&self, request: Request<State>, next: Next<'_, State>) -> tide::Result {
let secure = request.url().scheme() == "https";
#[actix_web::main]
async fn main() -> std::io::Result<()> {
let args: Vec<String> = std::env::args().collect();
let mut address = "0.0.0.0:8080".to_string();
if self.0.into() && !secure {
let mut secured = request.url().to_owned();
secured.set_scheme("https").unwrap_or_default();
if args.len() > 1 {
for arg in args {
if arg.starts_with("--address=") || arg.starts_with("-a=") {
let split: Vec<&str> = arg.split("=").collect();
address = split[1].to_string();
}
Ok(redirect(secured.to_string()))
} else {
Ok(next.run(request).await)
}
}
// start http server
println!("Running Libreddit v{} on {}!", env!("CARGO_PKG_VERSION"), address.clone());
HttpServer::new(|| {
App::new()
// TRAILING SLASH MIDDLEWARE
.wrap(NormalizePath::default())
// GENERAL SERVICES
.route("/style.css/", web::get().to(style))
.route("/favicon.ico/", web::get().to(|| HttpResponse::Ok()))
.route("/robots.txt/", web::get().to(robots))
// PROXY SERVICE
.route("/proxy/{url:.*}/", web::get().to(proxy::handler))
// USER SERVICES
.route("/u/{username}/", web::get().to(user::page))
.route("/user/{username}/", web::get().to(user::page))
// SUBREDDIT SERVICES
.route("/r/{sub}/", web::get().to(subreddit::page))
// POPULAR SERVICES
.route("/", web::get().to(popular::page))
// POST SERVICES
.route("/{id:.{5,6}}/", web::get().to(post::short))
.route("/r/{sub}/comments/{id}/{title}/", web::get().to(post::page))
})
.bind(address.clone())
.expect(format!("Cannot bind to the address: {}", address).as_str())
.run()
.await
}
#[async_trait]
impl<State: Clone + Send + Sync + 'static> Middleware<State> for NormalizePath {
async fn handle(&self, request: Request<State>, next: Next<'_, State>) -> tide::Result {
let path = request.url().path();
let query = request.url().query().unwrap_or_default();
if path.ends_with('/') {
Ok(next.run(request).await)
} else {
let normalized = if query.is_empty() {
format!("{}/", path.replace("//", "/"))
} else {
format!("{}/?{}", path.replace("//", "/"), query)
};
Ok(redirect(normalized))
}
}
}
// Create Services
// Required for the manifest to be valid
async fn pwa_logo(_req: Request<()>) -> tide::Result {
Ok(Response::builder(200).content_type("image/png").body(include_bytes!("../static/logo.png").as_ref()).build())
}
// Required for iOS App Icons
async fn iphone_logo(_req: Request<()>) -> tide::Result {
Ok(
Response::builder(200)
.content_type("image/png")
.body(include_bytes!("../static/apple-touch-icon.png").as_ref())
.build(),
)
}
async fn favicon(_req: Request<()>) -> tide::Result {
Ok(
Response::builder(200)
.content_type("image/vnd.microsoft.icon")
.header("Cache-Control", "public, max-age=1209600, s-maxage=86400")
.body(include_bytes!("../static/favicon.ico").as_ref())
.build(),
)
}
async fn resource(body: &str, content_type: &str, cache: bool) -> tide::Result {
let mut res = Response::new(200);
if cache {
res.insert_header("Cache-Control", "public, max-age=1209600, s-maxage=86400");
}
res.set_content_type(content_type);
res.set_body(body);
Ok(res)
}
#[async_std::main]
async fn main() -> tide::Result<()> {
let matches = App::new("Libreddit")
.version(env!("CARGO_PKG_VERSION"))
.about("Private front-end for Reddit written in Rust ")
.arg(
Arg::with_name("address")
.short("a")
.long("address")
.value_name("ADDRESS")
.help("Sets address to listen on")
.default_value("0.0.0.0")
.takes_value(true),
)
.arg(
Arg::with_name("port")
.short("p")
.long("port")
.value_name("PORT")
.help("Port to listen on")
.default_value("8080")
.takes_value(true),
)
.arg(
Arg::with_name("redirect-https")
.short("r")
.long("redirect-https")
.help("Redirect all HTTP requests to HTTPS")
.takes_value(false),
)
.get_matches();
let address = matches.value_of("address").unwrap_or("0.0.0.0");
let port = matches.value_of("port").unwrap_or("8080");
let force_https = matches.is_present("redirect-https");
let listener = format!("{}:{}", address, port);
println!("Starting Libreddit...");
// Start HTTP server
let mut app = tide::new();
// Redirect to HTTPS if "--redirect-https" enabled
app.with(HttpsRedirect(force_https));
// Append trailing slash and remove double slashes
app.with(NormalizePath);
// Apply default headers for security
app.with(After(|mut res: Response| async move {
res.insert_header("Referrer-Policy", "no-referrer");
res.insert_header("X-Content-Type-Options", "nosniff");
res.insert_header("X-Frame-Options", "DENY");
res.insert_header(
"Content-Security-Policy",
"default-src 'none'; manifest-src 'self'; media-src 'self'; style-src 'self' 'unsafe-inline'; base-uri 'none'; img-src 'self' data:; form-action 'self'; frame-ancestors 'none';",
);
Ok(res)
}));
// Read static files
app.at("/style.css/").get(|_| resource(include_str!("../static/style.css"), "text/css", false));
app
.at("/manifest.json/")
.get(|_| resource(include_str!("../static/manifest.json"), "application/json", false));
app.at("/robots.txt/").get(|_| resource("User-agent: *\nAllow: /", "text/plain", true));
app.at("/favicon.ico/").get(favicon);
app.at("/logo.png/").get(pwa_logo);
app.at("/touch-icon-iphone.png/").get(iphone_logo);
app.at("/apple-touch-icon.png/").get(iphone_logo);
// Proxy media through Libreddit
app
.at("/vid/:id/:size/") /* */
.get(|req| handler(req, "https://v.redd.it/{}/DASH_{}", vec!["id", "size"]));
app
.at("/img/:id/") /* */
.get(|req| handler(req, "https://i.redd.it/{}", vec!["id"]));
app
.at("/thumb/:point/:id/") /* */
.get(|req| handler(req, "https://{}.thumbs.redditmedia.com/{}", vec!["point", "id"]));
app
.at("/emoji/:id/:name/") /* */
.get(|req| handler(req, "https://emoji.redditmedia.com/{}/{}", vec!["id", "name"]));
app
.at("/preview/:loc/:id/:query/")
.get(|req| handler(req, "https://{}view.redd.it/{}?{}", vec!["loc", "id", "query"]));
app
.at("/style/*path/") /* */
.get(|req| handler(req, "https://styles.redditmedia.com/{}", vec!["path"]));
app
.at("/static/*path/") /* */
.get(|req| handler(req, "https://www.redditstatic.com/{}", vec!["path"]));
// Browse user profile
app.at("/u/:name/").get(user::profile);
app.at("/u/:name/comments/:id/:title/").get(post::item);
app.at("/u/:name/comments/:id/:title/:comment_id/").get(post::item);
app.at("/user/:name/").get(user::profile);
app.at("/user/:name/comments/:id/").get(post::item);
app.at("/user/:name/comments/:id/:title/").get(post::item);
app.at("/user/:name/comments/:id/:title/:comment_id/").get(post::item);
// Configure settings
app.at("/settings/").get(settings::get).post(settings::set);
app.at("/settings/restore/").get(settings::restore);
// Subreddit services
app.at("/r/:sub/").get(subreddit::page);
app.at("/r/:sub/subscribe/").post(subreddit::subscriptions);
app.at("/r/:sub/unsubscribe/").post(subreddit::subscriptions);
app.at("/r/:sub/comments/:id/").get(post::item);
app.at("/r/:sub/comments/:id/:title/").get(post::item);
app.at("/r/:sub/comments/:id/:title/:comment_id/").get(post::item);
app.at("/r/:sub/search/").get(search::find);
app.at("/r/:sub/wiki/").get(subreddit::wiki);
app.at("/r/:sub/wiki/:page/").get(subreddit::wiki);
app.at("/r/:sub/w/").get(subreddit::wiki);
app.at("/r/:sub/w/:page/").get(subreddit::wiki);
app.at("/r/:sub/:sort/").get(subreddit::page);
// Comments handler
app.at("/comments/:id/").get(post::item);
// Front page
app.at("/").get(subreddit::page);
// View Reddit wiki
app.at("/w/").get(subreddit::wiki);
app.at("/w/:page/").get(subreddit::wiki);
app.at("/wiki/").get(subreddit::wiki);
app.at("/wiki/:page/").get(subreddit::wiki);
// Search all of Reddit
app.at("/search/").get(search::find);
// Handle about pages
app.at("/about/").get(|req| error(req, "About pages aren't here yet".to_string()));
app.at("/:id/").get(|req: Request<()>| async {
match req.param("id") {
// Sort front page
Ok("best") | Ok("hot") | Ok("new") | Ok("top") | Ok("rising") | Ok("controversial") => subreddit::page(req).await,
// Short link for post
Ok(id) if id.len() > 4 && id.len() < 7 => post::item(req).await,
// Error message for unknown pages
_ => error(req, "Nothing here".to_string()).await,
}
});
// Default service in case no routes match
app.at("*").get(|req| error(req, "Nothing here".to_string()));
println!("Running Libreddit v{} on {}!", env!("CARGO_PKG_VERSION"), listener);
app.listen(&listener).await?;
Ok(())
}

View File

@ -1,55 +0,0 @@
// CRATES
use crate::utils::{fetch_posts, ErrorTemplate, Params, Post};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use askama::Template;
// STRUCTS
#[derive(Template)]
#[template(path = "popular.html", escape = "none")]
struct PopularTemplate {
posts: Vec<Post>,
sort: String,
ends: (String, String),
}
// RENDER
async fn render(sub_name: String, sort: Option<String>, ends: (Option<String>, Option<String>)) -> Result<HttpResponse> {
let sorting = sort.unwrap_or("hot".to_string());
let before = ends.1.clone().unwrap_or(String::new()); // If there is an after, there must be a before
// Build the Reddit JSON API url
let url = match ends.0 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?before={}&count=25", sub_name, sorting, val),
None => match ends.1 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?after={}&count=25", sub_name, sorting, val),
None => format!("https://www.reddit.com/r/{}/{}.json", sub_name, sorting),
},
};
let items_result = fetch_posts(url, String::new()).await;
if items_result.is_err() {
let s = ErrorTemplate {
message: items_result.err().unwrap().to_string(),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s))
} else {
let items = items_result.unwrap();
let s = PopularTemplate {
posts: items.0,
sort: sorting,
ends: (before, items.1),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
}
}
// SERVICES
pub async fn page(params: web::Query<Params>) -> Result<HttpResponse> {
render("popular".to_string(), params.sort.clone(), (params.before.clone(), params.after.clone())).await
}

View File

@ -1,12 +1,12 @@
// CRATES
use crate::utils::{format_num, format_url, request, val, Comment, ErrorTemplate, Flair, Params, Post};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use crate::utils::{
cookie, error, format_num, format_url, param, request, rewrite_urls, template, time, val, Author, Comment, Flags, Flair, FlairPart, Media, Post, Preferences,
};
use tide::Request;
use async_recursion::async_recursion;
use askama::Template;
use chrono::{TimeZone, Utc};
use pulldown_cmark::{html, Options, Parser};
// STRUCTS
#[derive(Template)]
@ -15,157 +15,200 @@ struct PostTemplate {
comments: Vec<Comment>,
post: Post,
sort: String,
prefs: Preferences,
single_thread: bool,
}
async fn render(id: String, sort: String) -> Result<HttpResponse> {
// Log the post ID being fetched
dbg!(&id);
pub async fn item(req: Request<()>) -> tide::Result {
// Build Reddit API path
let mut path: String = format!("{}.json?{}&raw_json=1", req.url().path(), req.url().query().unwrap_or_default());
// Build the Reddit JSON API url
let url: String = format!("https://reddit.com/{}.json?sort={}", id, sort);
// Set sort to sort query parameter
let mut sort: String = param(&path, "sort");
// Grab default comment sort method from Cookies
let default_sort = cookie(&req, "comment_sort");
// If there's no sort query but there's a default sort, set sort to default_sort
if sort.is_empty() && !default_sort.is_empty() {
sort = default_sort;
path = format!("{}.json?{}&sort={}&raw_json=1", req.url().path(), req.url().query().unwrap_or_default(), sort);
}
// Log the post ID being fetched in debug mode
#[cfg(debug_assertions)]
dbg!(req.param("id").unwrap_or(""));
let single_thread = &req.param("comment_id").is_ok();
let highlighted_comment = &req.param("comment_id").unwrap_or_default();
// Send a request to the url, receive JSON in response
let req = request(url).await;
match request(path).await {
// Otherwise, grab the JSON output from the request
Ok(res) => {
// Parse the JSON into Post and Comment structs
let post = parse_post(&res[0]).await;
let comments = parse_comments(&res[1], &post.permalink, &post.author.name, *highlighted_comment).await;
// If the Reddit API returns an error, exit and send error page to user
if req.is_err() {
let s = ErrorTemplate {
message: req.err().unwrap().to_string(),
// Use the Post and Comment structs to generate a website to show users
template(PostTemplate {
comments,
post,
sort,
prefs: Preferences::new(req),
single_thread: *single_thread,
})
}
.render()
.unwrap();
return Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s));
// If the Reddit API returns an error, exit and send error page to user
Err(msg) => error(req, msg).await,
}
// Otherwise, grab the JSON output from the request
let res = req.unwrap();
// Parse the JSON into Post and Comment structs
let post = parse_post(res[0].clone()).await;
let comments = parse_comments(res[1].clone()).await;
// Use the Post and Comment structs to generate a website to show users
let s = PostTemplate {
comments: comments.unwrap(),
post: post.unwrap(),
sort: sort,
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
}
// SERVICES
pub async fn short(web::Path(id): web::Path<String>) -> Result<HttpResponse> {
render(id.to_string(), "confidence".to_string()).await
}
pub async fn page(web::Path((_sub, id)): web::Path<(String, String)>, params: web::Query<Params>) -> Result<HttpResponse> {
match &params.sort {
Some(sort) => render(id, sort.to_string()).await,
None => render(id, "confidence".to_string()).await,
}
}
// UTILITIES
async fn media(data: &serde_json::Value) -> (String, String) {
let post_type: &str;
let url = if !data["preview"]["reddit_video_preview"]["fallback_url"].is_null() {
post_type = "video";
format_url(data["preview"]["reddit_video_preview"]["fallback_url"].as_str().unwrap()).await
} else if !data["secure_media"]["reddit_video"]["fallback_url"].is_null() {
post_type = "video";
format_url(data["secure_media"]["reddit_video"]["fallback_url"].as_str().unwrap()).await
} else if data["post_hint"].as_str().unwrap_or("") == "image" {
post_type = "image";
format_url(data["preview"]["images"][0]["source"]["url"].as_str().unwrap()).await
} else {
post_type = "link";
data["url"].as_str().unwrap().to_string()
};
(post_type.to_string(), url)
}
async fn markdown_to_html(md: &str) -> String {
let mut options = Options::empty();
options.insert(Options::ENABLE_TABLES);
options.insert(Options::ENABLE_FOOTNOTES);
options.insert(Options::ENABLE_STRIKETHROUGH);
options.insert(Options::ENABLE_TASKLISTS);
let parser = Parser::new_ext(md, options);
// Write to String buffer.
let mut html_output = String::new();
html::push_html(&mut html_output, parser);
html_output
}
// POSTS
async fn parse_post(json: serde_json::Value) -> Result<Post, &'static str> {
let post_data: &serde_json::Value = &json["data"]["children"][0];
async fn parse_post(json: &serde_json::Value) -> Post {
// Retrieve post (as opposed to comments) from JSON
let post: &serde_json::Value = &json["data"]["children"][0];
let unix_time: i64 = post_data["data"]["created_utc"].as_f64().unwrap().round() as i64;
let score = post_data["data"]["score"].as_i64().unwrap();
// Grab UTC time as unix timestamp
let (rel_time, created) = time(post["data"]["created_utc"].as_f64().unwrap_or_default());
// Parse post score and upvote ratio
let score = post["data"]["score"].as_i64().unwrap_or_default();
let ratio: f64 = post["data"]["upvote_ratio"].as_f64().unwrap_or(1.0) * 100.0;
let media = media(&post_data["data"]).await;
// Determine the type of media along with the media URL
let (post_type, media, gallery) = Media::parse(&post["data"]).await;
let post = Post {
title: val(post_data, "title").await,
community: val(post_data, "subreddit").await,
body: markdown_to_html(post_data["data"]["selftext"].as_str().unwrap()).await,
author: val(post_data, "author").await,
url: val(post_data, "permalink").await,
// Build a post using data parsed from Reddit post API
Post {
id: val(post, "id"),
title: val(post, "title"),
community: val(post, "subreddit"),
body: rewrite_urls(&val(post, "selftext_html")).replace("\\", ""),
author: Author {
name: val(post, "author"),
flair: Flair {
flair_parts: FlairPart::parse(
post["data"]["author_flair_type"].as_str().unwrap_or_default(),
post["data"]["author_flair_richtext"].as_array(),
post["data"]["author_flair_text"].as_str(),
),
text: val(post, "link_flair_text"),
background_color: val(post, "author_flair_background_color"),
foreground_color: val(post, "author_flair_text_color"),
},
distinguished: val(post, "distinguished"),
},
permalink: val(post, "permalink"),
score: format_num(score),
post_type: media.0,
media: media.1,
time: Utc.timestamp(unix_time, 0).format("%b %e %Y %H:%M UTC").to_string(),
flair: Flair(
val(post_data, "link_flair_text").await,
val(post_data, "link_flair_background_color").await,
if val(post_data, "link_flair_text_color").await == "dark" {
upvote_ratio: ratio as i64,
post_type,
media,
thumbnail: Media {
url: format_url(val(post, "thumbnail").as_str()),
width: post["data"]["thumbnail_width"].as_i64().unwrap_or_default(),
height: post["data"]["thumbnail_height"].as_i64().unwrap_or_default(),
poster: "".to_string(),
},
flair: Flair {
flair_parts: FlairPart::parse(
post["data"]["link_flair_type"].as_str().unwrap_or_default(),
post["data"]["link_flair_richtext"].as_array(),
post["data"]["link_flair_text"].as_str(),
),
text: val(post, "link_flair_text"),
background_color: val(post, "link_flair_background_color"),
foreground_color: if val(post, "link_flair_text_color") == "dark" {
"black".to_string()
} else {
"white".to_string()
},
),
};
Ok(post)
},
flags: Flags {
nsfw: post["data"]["over_18"].as_bool().unwrap_or(false),
stickied: post["data"]["stickied"].as_bool().unwrap_or(false),
},
domain: val(post, "domain"),
rel_time,
created,
comments: format_num(post["data"]["num_comments"].as_i64().unwrap_or_default()),
gallery,
}
}
// COMMENTS
#[async_recursion]
async fn parse_comments(json: serde_json::Value) -> Result<Vec<Comment>, &'static str> {
async fn parse_comments(json: &serde_json::Value, post_link: &str, post_author: &str, highlighted_comment: &str) -> Vec<Comment> {
// Separate the comment JSON into a Vector of comments
let comment_data = json["data"]["children"].as_array().unwrap();
let comment_data = match json["data"]["children"].as_array() {
Some(f) => f.to_owned(),
None => Vec::new(),
};
let mut comments: Vec<Comment> = Vec::new();
// For each comment, retrieve the values to build a Comment object
for comment in comment_data.iter() {
let unix_time: i64 = comment["data"]["created_utc"].as_f64().unwrap_or(0.0).round() as i64;
if unix_time == 0 {
continue;
}
for comment in comment_data {
let kind = comment["kind"].as_str().unwrap_or_default().to_string();
let data = &comment["data"];
let score = comment["data"]["score"].as_i64().unwrap_or(0);
let body = markdown_to_html(comment["data"]["body"].as_str().unwrap_or("")).await;
let unix_time = data["created_utc"].as_f64().unwrap_or_default();
let (rel_time, created) = time(unix_time);
let replies: Vec<Comment> = if comment["data"]["replies"].is_object() {
parse_comments(comment["data"]["replies"].clone()).await.unwrap_or(Vec::new())
let edited = match data["edited"].as_f64() {
Some(stamp) => time(stamp),
None => (String::new(), String::new()),
};
let score = data["score"].as_i64().unwrap_or(0);
let body = rewrite_urls(&val(&comment, "body_html"));
// If this comment contains replies, handle those too
let replies: Vec<Comment> = if data["replies"].is_object() {
parse_comments(&data["replies"], post_link, post_author, highlighted_comment).await
} else {
Vec::new()
};
let parent_kind_and_id = val(&comment, "parent_id");
let parent_info = parent_kind_and_id.split('_').collect::<Vec<&str>>();
let id = val(&comment, "id");
let highlighted = id == highlighted_comment;
comments.push(Comment {
body: body,
author: val(comment, "author").await,
score: format_num(score),
time: Utc.timestamp(unix_time, 0).format("%b %e %Y %H:%M UTC").to_string(),
replies: replies,
id,
kind,
parent_id: parent_info[1].to_string(),
parent_kind: parent_info[0].to_string(),
post_link: post_link.to_string(),
post_author: post_author.to_string(),
body,
author: Author {
name: val(&comment, "author"),
flair: Flair {
flair_parts: FlairPart::parse(
data["author_flair_type"].as_str().unwrap_or_default(),
data["author_flair_richtext"].as_array(),
data["author_flair_text"].as_str(),
),
text: val(&comment, "link_flair_text"),
background_color: val(&comment, "author_flair_background_color"),
foreground_color: val(&comment, "author_flair_text_color"),
},
distinguished: val(&comment, "distinguished"),
},
score: if data["score_hidden"].as_bool().unwrap_or_default() {
"\u{2022}".to_string()
} else {
format_num(score)
},
rel_time,
created,
edited,
replies,
highlighted,
});
}
Ok(comments)
comments
}

View File

@ -1,29 +1,88 @@
use actix_web::{client::Client, web, Error, HttpResponse, Result};
use async_std::{io, net::TcpStream, prelude::*};
use async_tls::TlsConnector;
use tide::{http::url::Url, Request, Response};
#[cfg(feature = "proxy")]
use base64::decode;
/// Handle tide routes to proxy by parsing `params` from `req`uest.
pub async fn handler(req: Request<()>, format: &str, params: Vec<&str>) -> tide::Result {
let mut url = format.to_string();
pub async fn handler(web::Path(url): web::Path<String>) -> Result<HttpResponse> {
if cfg!(feature = "proxy") {
let media: String;
for name in params {
let param = req.param(name).unwrap_or_default();
url = url.replacen("{}", param, 1);
}
#[cfg(not(feature = "proxy"))]
let media = url;
request(url).await
}
#[cfg(feature = "proxy")]
match decode(url) {
Ok(bytes) => media = String::from_utf8(bytes).unwrap(),
Err(_e) => return Ok(HttpResponse::Ok().body("")),
};
/// Sends a request to a Reddit media domain and proxy the response.
///
/// Relays the `Content-Length` and `Content-Type` header.
async fn request(url: String) -> tide::Result {
// Parse url into parts
let parts = Url::parse(&url).unwrap();
let host = parts.host().unwrap().to_string();
let domain = parts.domain().unwrap_or_default();
let path = format!("{}?{}", parts.path(), parts.query().unwrap_or_default());
// Build reddit-compliant user agent for Libreddit
let user_agent = format!("web:libreddit:{}", env!("CARGO_PKG_VERSION"));
let client = Client::default();
client
.get(media.replace("&amp;", "&"))
.send()
.await
.map_err(Error::from)
.and_then(|res| Ok(HttpResponse::build(res.status()).streaming(res)))
} else {
Ok(HttpResponse::Ok().body(""))
// Construct a request body
let req = format!(
"GET {} HTTP/1.1\r\nHost: {}\r\nAccept: */*\r\nConnection: close\r\nUser-Agent: {}\r\n\r\n",
path, host, user_agent
);
// Initialize TLS connector for requests
let connector = TlsConnector::default();
// Open a TCP connection
let tcp_stream = TcpStream::connect(format!("{}:443", domain)).await.unwrap();
// Use the connector to start the handshake process
let mut tls_stream = connector.connect(domain, tcp_stream).await.unwrap();
// Write the aforementioned HTTP request to the stream
tls_stream.write_all(req.as_bytes()).await.unwrap();
// And read the response
let mut writer = Vec::new();
io::copy(&mut tls_stream, &mut writer).await.unwrap();
// Find the delimiter which separates the body and headers
match (0..writer.len()).find(|i| writer[i.to_owned()] == 10_u8 && writer[i - 2] == 10_u8) {
Some(delim) => {
// Split the response into the body and headers
let split = writer.split_at(delim);
let headers_str = String::from_utf8_lossy(split.0);
let headers = headers_str.split("\r\n").collect::<Vec<&str>>();
let body = split.1[1..split.1.len()].to_vec();
// Parse the status code from the first header line
let status: u16 = headers[0].split(' ').collect::<Vec<&str>>()[1].parse().unwrap_or_default();
// Define a closure for easier header fetching
let header = |name: &str| {
headers
.iter()
.find(|x| x.starts_with(name))
.map(|f| f.split(": ").collect::<Vec<&str>>()[1])
.unwrap_or_default()
};
// Parse Content-Length and Content-Type from headers
let content_length = header("Content-Length");
let content_type = header("Content-Type");
// Build response
Ok(
Response::builder(status)
.body(tide::http::Body::from_bytes(body))
.header("Cache-Control", "public, max-age=1209600, s-maxage=86400")
.header("Content-Length", content_length)
.header("Content-Type", content_type)
.build(),
)
}
None => Ok(Response::builder(503).body("Couldn't parse media".to_string()).build()),
}
}

96
src/search.rs Normal file
View File

@ -0,0 +1,96 @@
// CRATES
use crate::utils::{cookie, error, param, request, template, val, Post, Preferences};
use askama::Template;
use tide::Request;
// STRUCTS
struct SearchParams {
q: String,
sort: String,
t: String,
before: String,
after: String,
restrict_sr: String,
}
// STRUCTS
struct Subreddit {
name: String,
url: String,
description: String,
subscribers: i64,
}
#[derive(Template)]
#[template(path = "search.html", escape = "none")]
struct SearchTemplate {
posts: Vec<Post>,
subreddits: Vec<Subreddit>,
sub: String,
params: SearchParams,
prefs: Preferences,
}
// SERVICES
pub async fn find(req: Request<()>) -> tide::Result {
let nsfw_results = if cookie(&req, "show_nsfw") == "on" { "&include_over_18=on" } else { "" };
let path = format!("{}.json?{}{}", req.url().path(), req.url().query().unwrap_or_default(), nsfw_results);
let sub = req.param("sub").unwrap_or("").to_string();
let query = param(&path, "q");
let sort = if param(&path, "sort").is_empty() {
"relevance".to_string()
} else {
param(&path, "sort")
};
let subreddits = if param(&path, "restrict_sr").is_empty() {
search_subreddits(&query).await
} else {
Vec::new()
};
match Post::fetch(&path, String::new()).await {
Ok((posts, after)) => template(SearchTemplate {
posts,
subreddits,
sub,
params: SearchParams {
q: query.replace('"', "&quot;"),
sort,
t: param(&path, "t"),
before: param(&path, "after"),
after,
restrict_sr: param(&path, "restrict_sr"),
},
prefs: Preferences::new(req),
}),
Err(msg) => error(req, msg).await,
}
}
async fn search_subreddits(q: &str) -> Vec<Subreddit> {
let subreddit_search_path = format!("/subreddits/search.json?q={}&limit=3", q.replace(' ', "+"));
// Send a request to the url
match request(subreddit_search_path).await {
// If success, receive JSON in response
Ok(response) => {
match response["data"]["children"].as_array() {
// For each subreddit from subreddit list
Some(list) => list
.iter()
.map(|subreddit| Subreddit {
name: val(subreddit, "display_name_prefixed"),
url: val(subreddit, "url"),
description: val(subreddit, "public_description"),
subscribers: subreddit["data"]["subscribers"].as_f64().unwrap_or_default() as i64,
})
.collect::<Vec<Subreddit>>(),
_ => Vec::new(),
}
}
// If the Reddit API returns an error, exit this function
_ => Vec::new(),
}
}

87
src/settings.rs Normal file
View File

@ -0,0 +1,87 @@
// CRATES
use crate::utils::{redirect, template, Preferences};
use askama::Template;
use tide::{http::Cookie, Request};
use time::{Duration, OffsetDateTime};
// STRUCTS
#[derive(Template)]
#[template(path = "settings.html")]
struct SettingsTemplate {
prefs: Preferences,
}
#[derive(serde::Deserialize, Default)]
#[serde(default)]
pub struct Form {
theme: Option<String>,
front_page: Option<String>,
layout: Option<String>,
wide: Option<String>,
comment_sort: Option<String>,
show_nsfw: Option<String>,
redirect: Option<String>,
subscriptions: Option<String>,
}
// FUNCTIONS
// Retrieve cookies from request "Cookie" header
pub async fn get(req: Request<()>) -> tide::Result {
template(SettingsTemplate { prefs: Preferences::new(req) })
}
// Set cookies using response "Set-Cookie" header
pub async fn set(mut req: Request<()>) -> tide::Result {
let form: Form = req.body_form().await.unwrap_or_default();
let mut res = redirect("/settings".to_string());
let names = vec!["theme", "front_page", "layout", "wide", "comment_sort", "show_nsfw"];
let values = vec![form.theme, form.front_page, form.layout, form.wide, form.comment_sort, form.show_nsfw];
for (i, name) in names.iter().enumerate() {
match values.get(i) {
Some(value) => res.insert_cookie(
Cookie::build(name.to_owned(), value.to_owned().unwrap_or_default())
.path("/")
.http_only(true)
.expires(OffsetDateTime::now_utc() + Duration::weeks(52))
.finish(),
),
None => res.remove_cookie(Cookie::named(name.to_owned())),
};
}
Ok(res)
}
// Set cookies using response "Set-Cookie" header
pub async fn restore(req: Request<()>) -> tide::Result {
let form: Form = req.query()?;
let path = match form.redirect {
Some(value) => format!("/{}/", value),
None => "/".to_string(),
};
let mut res = redirect(path);
let names = vec!["theme", "front_page", "layout", "wide", "comment_sort", "show_nsfw", "subscriptions"];
let values = vec![form.theme, form.front_page, form.layout, form.wide, form.comment_sort, form.show_nsfw, form.subscriptions];
for (i, name) in names.iter().enumerate() {
match values.get(i) {
Some(value) => res.insert_cookie(
Cookie::build(name.to_owned(), value.to_owned().unwrap_or_default())
.path("/")
.http_only(true)
.expires(OffsetDateTime::now_utc() + Duration::weeks(52))
.finish(),
),
None => res.remove_cookie(Cookie::named(name.to_owned())),
};
}
Ok(res)
}

View File

@ -1,8 +1,8 @@
// CRATES
use crate::utils::{fetch_posts, format_num, format_url, request, val, ErrorTemplate, Params, Post, Subreddit};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use crate::utils::{cookie, error, format_num, format_url, param, redirect, request, rewrite_urls, template, val, Post, Preferences, Subreddit};
use askama::Template;
use std::convert::TryInto;
use tide::{http::Cookie, Request};
use time::{Duration, OffsetDateTime};
// STRUCTS
#[derive(Template)]
@ -10,88 +10,176 @@ use std::convert::TryInto;
struct SubredditTemplate {
sub: Subreddit,
posts: Vec<Post>,
sort: String,
sort: (String, String),
ends: (String, String),
prefs: Preferences,
}
#[derive(Template)]
#[template(path = "wiki.html", escape = "none")]
struct WikiTemplate {
sub: String,
wiki: String,
page: String,
prefs: Preferences,
}
// SERVICES
#[allow(dead_code)]
pub async fn page(web::Path(sub): web::Path<String>, params: web::Query<Params>) -> Result<HttpResponse> {
render(sub, params.sort.clone(), (params.before.clone(), params.after.clone())).await
pub async fn page(req: Request<()>) -> tide::Result {
// Build Reddit API path
let subscribed = cookie(&req, "subscriptions");
let front_page = cookie(&req, "front_page");
let sort = req.param("sort").unwrap_or_else(|_| req.param("id").unwrap_or("hot")).to_string();
let sub = req.param("sub").map(String::from).unwrap_or(if front_page == "default" || front_page.is_empty() {
if subscribed.is_empty() {
"popular".to_string()
} else {
subscribed.to_owned()
}
} else {
front_page.to_owned()
});
let path = format!("/r/{}/{}.json?{}&raw_json=1", sub, sort, req.url().query().unwrap_or_default());
match Post::fetch(&path, String::new()).await {
Ok((posts, after)) => {
// If you can get subreddit posts, also request subreddit metadata
let sub = if !sub.contains('+') && sub != subscribed && sub != "popular" && sub != "all" {
// Regular subreddit
subreddit(&sub).await.unwrap_or_default()
} else if sub == subscribed {
// Subscription feed
if req.url().path().starts_with("/r/") {
subreddit(&sub).await.unwrap_or_default()
} else {
Subreddit::default()
}
} else if sub.contains('+') {
// Multireddit
Subreddit {
name: sub,
..Subreddit::default()
}
} else {
Subreddit::default()
};
template(SubredditTemplate {
sub,
posts,
sort: (sort, param(&path, "t")),
ends: (param(&path, "after"), after),
prefs: Preferences::new(req),
})
}
Err(msg) => match msg.as_str() {
"quarantined" => error(req, format!("r/{} has been quarantined by Reddit", sub)).await,
"private" => error(req, format!("r/{} is a private community", sub)).await,
"banned" => error(req, format!("r/{} has been banned from Reddit", sub)).await,
_ => error(req, msg).await,
},
}
}
pub async fn render(sub_name: String, sort: Option<String>, ends: (Option<String>, Option<String>)) -> Result<HttpResponse> {
let sorting = sort.unwrap_or("hot".to_string());
let before = ends.1.clone().unwrap_or(String::new()); // If there is an after, there must be a before
// Sub or unsub by setting subscription cookie using response "Set-Cookie" header
pub async fn subscriptions(req: Request<()>) -> tide::Result {
let sub = req.param("sub").unwrap_or_default().to_string();
let query = req.url().query().unwrap_or_default().to_string();
let action: Vec<String> = req.url().path().split('/').map(String::from).collect();
// Build the Reddit JSON API url
let url = match ends.0 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?before={}&count=25", sub_name, sorting, val),
None => match ends.1 {
Some(val) => format!("https://www.reddit.com/r/{}/{}.json?after={}&count=25", sub_name, sorting, val),
None => format!("https://www.reddit.com/r/{}/{}.json", sub_name, sorting),
},
let mut sub_list = Preferences::new(req).subscriptions;
// Find each subreddit name (separated by '+') in sub parameter
for part in sub.split('+') {
// Modify sub list based on action
if action.contains(&"subscribe".to_string()) && !sub_list.contains(&part.to_owned()) {
// Add each sub name to the subscribed list
sub_list.push(part.to_owned());
// Reorder sub names alphabettically
sub_list.sort_by_key(|a| a.to_lowercase())
} else if action.contains(&"unsubscribe".to_string()) {
// Remove sub name from subscribed list
sub_list.retain(|s| s != part);
}
}
// Redirect back to subreddit
// check for redirect parameter if unsubscribing from outside sidebar
let redirect_path = param(&format!("/?{}", query), "redirect");
let path = if redirect_path.is_empty() {
format!("/r/{}", sub)
} else {
format!("/{}/", redirect_path)
};
let sub_result = subreddit(&sub_name).await;
let items_result = fetch_posts(url, String::new()).await;
let mut res = redirect(path);
if sub_result.is_err() || items_result.is_err() {
let s = ErrorTemplate {
message: sub_result.err().unwrap().to_string(),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s))
// Delete cookie if empty, else set
if sub_list.is_empty() {
// res.del_cookie(&Cookie::build("subscriptions", "").path("/").finish());
res.remove_cookie(Cookie::build("subscriptions", "").path("/").finish());
} else {
let mut sub = sub_result.unwrap();
let items = items_result.unwrap();
res.insert_cookie(
Cookie::build("subscriptions", sub_list.join("+"))
.path("/")
.http_only(true)
.expires(OffsetDateTime::now_utc() + Duration::weeks(52))
.finish(),
);
}
sub.icon = if sub.icon != "" {
format!(r#"<img class="subreddit_icon" src="{}">"#, sub.icon)
} else {
String::new()
};
Ok(res)
}
let s = SubredditTemplate {
sub: sub,
posts: items.0,
sort: sorting,
ends: (before, items.1),
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
pub async fn wiki(req: Request<()>) -> tide::Result {
let sub = req.param("sub").unwrap_or("reddit.com").to_string();
let page = req.param("page").unwrap_or("index").to_string();
let path: String = format!("/r/{}/wiki/{}.json?raw_json=1", sub, page);
match request(path).await {
Ok(response) => template(WikiTemplate {
sub,
wiki: rewrite_urls(response["data"]["content_html"].as_str().unwrap_or_default()),
page,
prefs: Preferences::new(req),
}),
Err(msg) => error(req, msg).await,
}
}
// SUBREDDIT
async fn subreddit(sub: &String) -> Result<Subreddit, &'static str> {
async fn subreddit(sub: &str) -> Result<Subreddit, String> {
// Build the Reddit JSON API url
let url: String = format!("https://www.reddit.com/r/{}/about.json", sub);
let path: String = format!("/r/{}/about.json?raw_json=1", sub);
// Send a request to the url, receive JSON in response
let req = request(url).await;
// Send a request to the url
match request(path).await {
// If success, receive JSON in response
Ok(res) => {
// Metadata regarding the subreddit
let members: i64 = res["data"]["subscribers"].as_u64().unwrap_or_default() as i64;
let active: i64 = res["data"]["accounts_active"].as_u64().unwrap_or_default() as i64;
// If the Reddit API returns an error, exit this function
if req.is_err() {
return Err(req.err().unwrap());
// Fetch subreddit icon either from the community_icon or icon_img value
let community_icon: &str = res["data"]["community_icon"].as_str().map_or("", |s| s.split('?').collect::<Vec<&str>>()[0]);
let icon = if community_icon.is_empty() { val(&res, "icon_img") } else { community_icon.to_string() };
let sub = Subreddit {
name: val(&res, "display_name"),
title: val(&res, "title"),
description: val(&res, "public_description"),
info: rewrite_urls(&val(&res, "description_html").replace("\\", "")),
icon: format_url(&icon),
members: format_num(members),
active: format_num(active),
wiki: res["data"]["wiki_enabled"].as_bool().unwrap_or_default(),
};
Ok(sub)
}
// If the Reddit API returns an error, exit this function
Err(msg) => return Err(msg),
}
// Otherwise, grab the JSON output from the request
let res = req.unwrap();
let members = res["data"]["subscribers"].as_u64().unwrap_or(0);
let active = res["data"]["accounts_active"].as_u64().unwrap_or(0);
let sub = Subreddit {
name: val(&res, "display_name").await,
title: val(&res, "title").await,
description: val(&res, "public_description").await,
icon: format_url(val(&res, "icon_img").await.as_str()).await,
members: format_num(members.try_into().unwrap()),
active: format_num(active.try_into().unwrap()),
};
Ok(sub)
}

View File

@ -1,7 +1,8 @@
// CRATES
use crate::utils::{fetch_posts, nested_val, request, ErrorTemplate, Params, Post, User};
use actix_web::{http::StatusCode, web, HttpResponse, Result};
use crate::utils::{error, format_url, param, request, template, Post, Preferences, User};
use askama::Template;
use tide::Request;
use time::OffsetDateTime;
// STRUCTS
#[derive(Template)]
@ -9,65 +10,68 @@ use askama::Template;
struct UserTemplate {
user: User,
posts: Vec<Post>,
sort: String,
sort: (String, String),
ends: (String, String),
prefs: Preferences,
}
async fn render(username: String, sort: String) -> Result<HttpResponse> {
// Build the Reddit JSON API url
let url: String = format!("https://www.reddit.com/user/{}/.json?sort={}", username, sort);
// FUNCTIONS
pub async fn profile(req: Request<()>) -> tide::Result {
// Build the Reddit JSON API path
let path = format!("{}.json?{}&raw_json=1", req.url().path(), req.url().query().unwrap_or_default());
let user = user(&username).await;
let posts = fetch_posts(url, "Comment".to_string()).await;
// Retrieve other variables from Libreddit request
let sort = param(&path, "sort");
let username = req.param("name").unwrap_or("").to_string();
if user.is_err() || posts.is_err() {
let s = ErrorTemplate {
message: user.err().unwrap().to_string(),
// Request user posts/comments from Reddit
let posts = Post::fetch(&path, "Comment".to_string()).await;
match posts {
Ok((posts, after)) => {
// If you can get user posts, also request user data
let user = user(&username).await.unwrap_or_default();
template(UserTemplate {
user,
posts,
sort: (sort, param(&path, "t")),
ends: (param(&path, "after"), after),
prefs: Preferences::new(req),
})
}
.render()
.unwrap();
Ok(HttpResponse::Ok().status(StatusCode::NOT_FOUND).content_type("text/html").body(s))
} else {
let s = UserTemplate {
user: user.unwrap(),
posts: posts.unwrap().0,
sort: sort,
}
.render()
.unwrap();
Ok(HttpResponse::Ok().content_type("text/html").body(s))
}
}
// SERVICES
pub async fn page(web::Path(username): web::Path<String>, params: web::Query<Params>) -> Result<HttpResponse> {
match &params.sort {
Some(sort) => render(username, sort.to_string()).await,
None => render(username, "hot".to_string()).await,
// If there is an error show error page
Err(msg) => error(req, msg).await,
}
}
// USER
async fn user(name: &String) -> Result<User, &'static str> {
// Build the Reddit JSON API url
let url: String = format!("https://www.reddit.com/user/{}/about.json", name);
async fn user(name: &str) -> Result<User, String> {
// Build the Reddit JSON API path
let path: String = format!("/user/{}/about.json?raw_json=1", name);
// Send a request to the url, receive JSON in response
let req = request(url).await;
// Send a request to the url
match request(path).await {
// If success, receive JSON in response
Ok(res) => {
// Grab creation date as unix timestamp
let created: i64 = res["data"]["created"].as_f64().unwrap_or(0.0).round() as i64;
// If the Reddit API returns an error, exit this function
if req.is_err() {
return Err(req.err().unwrap());
// nested_val function used to parse JSON from Reddit APIs
let about = |item| res["data"]["subreddit"][item].as_str().unwrap_or_default().to_string();
// Parse the JSON output into a User struct
Ok(User {
name: name.to_string(),
title: about("title"),
icon: format_url(&about("icon_img")),
karma: res["data"]["total_karma"].as_i64().unwrap_or(0),
created: OffsetDateTime::from_unix_timestamp(created).format("%b %d '%y"),
banner: about("banner_img"),
description: about("public_description"),
})
}
// If the Reddit API returns an error, exit this function
Err(msg) => return Err(msg),
}
// Otherwise, grab the JSON output from the request
let res = req.unwrap();
// Parse the JSON output into a User struct
Ok(User {
name: name.to_string(),
icon: nested_val(&res, "subreddit", "icon_img").await,
karma: res["data"]["total_karma"].as_i64().unwrap(),
banner: nested_val(&res, "subreddit", "banner_img").await,
description: nested_val(&res, "subreddit", "public_description").await,
})
}

View File

@ -1,215 +1,602 @@
//
// CRATES
//
use chrono::{TimeZone, Utc};
use serde_json::{from_str, Value};
// use surf::{client, get, middleware::Redirect};
use askama::Template;
use async_recursion::async_recursion;
use async_std::{io, net::TcpStream, prelude::*};
use async_tls::TlsConnector;
use cached::proc_macro::cached;
use regex::Regex;
use serde_json::{from_str, Error, Value};
use std::collections::HashMap;
use tide::{http::url::Url, http::Cookie, Request, Response};
use time::{Duration, OffsetDateTime};
#[cfg(feature = "proxy")]
use base64::encode;
// Post flair with content, background color and foreground color
pub struct Flair {
pub flair_parts: Vec<FlairPart>,
pub text: String,
pub background_color: String,
pub foreground_color: String,
}
//
// STRUCTS
//
#[allow(dead_code)]
// Post flair with text, background color and foreground color
pub struct Flair(pub String, pub String, pub String);
// Part of flair, either emoji or text
pub struct FlairPart {
pub flair_part_type: String,
pub value: String,
}
impl FlairPart {
pub fn parse(flair_type: &str, rich_flair: Option<&Vec<Value>>, text_flair: Option<&str>) -> Vec<Self> {
// Parse type of flair
match flair_type {
// If flair contains emojis and text
"richtext" => match rich_flair {
Some(rich) => rich
.iter()
// For each part of the flair, extract text and emojis
.map(|part| {
let value = |name: &str| part[name].as_str().unwrap_or_default();
Self {
flair_part_type: value("e").to_string(),
value: match value("e") {
"text" => value("t").to_string(),
"emoji" => format_url(value("u")),
_ => String::new(),
},
}
})
.collect::<Vec<Self>>(),
None => Vec::new(),
},
// If flair contains only text
"text" => match text_flair {
Some(text) => vec![Self {
flair_part_type: "text".to_string(),
value: text.to_string(),
}],
None => Vec::new(),
},
_ => Vec::new(),
}
}
}
pub struct Author {
pub name: String,
pub flair: Flair,
pub distinguished: String,
}
// Post flags with nsfw and stickied
pub struct Flags {
pub nsfw: bool,
pub stickied: bool,
}
pub struct Media {
pub url: String,
pub width: i64,
pub height: i64,
pub poster: String,
}
impl Media {
pub async fn parse(data: &Value) -> (String, Self, Vec<GalleryMedia>) {
let mut gallery = Vec::new();
// If post is a video, return the video
let (post_type, url_val) = if data["preview"]["reddit_video_preview"]["fallback_url"].is_string() {
// Return reddit video
("video", &data["preview"]["reddit_video_preview"]["fallback_url"])
} else if data["secure_media"]["reddit_video"]["fallback_url"].is_string() {
// Return reddit video
("video", &data["secure_media"]["reddit_video"]["fallback_url"])
} else if data["post_hint"].as_str().unwrap_or("") == "image" {
// Handle images, whether GIFs or pics
let preview = &data["preview"]["images"][0];
let mp4 = &preview["variants"]["mp4"];
if mp4.is_object() {
// Return the mp4 if the media is a gif
("gif", &mp4["source"]["url"])
} else {
// Return the picture if the media is an image
if data["domain"] == "i.redd.it" {
("image", &data["url"])
} else {
("image", &preview["source"]["url"])
}
}
} else if data["is_self"].as_bool().unwrap_or_default() {
// If type is self, return permalink
("self", &data["permalink"])
} else if data["is_gallery"].as_bool().unwrap_or_default() {
// If this post contains a gallery of images
gallery = GalleryMedia::parse(&data["gallery_data"]["items"], &data["media_metadata"]);
("gallery", &data["url"])
} else {
// If type can't be determined, return url
("link", &data["url"])
};
let source = &data["preview"]["images"][0]["source"];
let url = if post_type == "self" || post_type == "link" {
url_val.as_str().unwrap_or_default().to_string()
} else {
format_url(url_val.as_str().unwrap_or_default())
};
(
post_type.to_string(),
Self {
url,
width: source["width"].as_i64().unwrap_or_default(),
height: source["height"].as_i64().unwrap_or_default(),
poster: format_url(source["url"].as_str().unwrap_or_default()),
},
gallery,
)
}
}
pub struct GalleryMedia {
pub url: String,
pub width: i64,
pub height: i64,
pub caption: String,
pub outbound_url: String,
}
impl GalleryMedia {
fn parse(items: &Value, metadata: &Value) -> Vec<Self> {
items
.as_array()
.unwrap_or(&Vec::new())
.iter()
.map(|item| {
// For each image in gallery
let media_id = item["media_id"].as_str().unwrap_or_default();
let image = &metadata[media_id]["s"];
// Construct gallery items
Self {
url: format_url(image["u"].as_str().unwrap_or_default()),
width: image["x"].as_i64().unwrap_or_default(),
height: image["y"].as_i64().unwrap_or_default(),
caption: item["caption"].as_str().unwrap_or_default().to_string(),
outbound_url: item["outbound_url"].as_str().unwrap_or_default().to_string(),
}
})
.collect::<Vec<Self>>()
}
}
#[allow(dead_code)]
// Post containing content, metadata and media
pub struct Post {
pub id: String,
pub title: String,
pub community: String,
pub body: String,
pub author: String,
pub url: String,
pub author: Author,
pub permalink: String,
pub score: String,
pub upvote_ratio: i64,
pub post_type: String,
pub media: String,
pub time: String,
pub flair: Flair,
pub flags: Flags,
pub thumbnail: Media,
pub media: Media,
pub domain: String,
pub rel_time: String,
pub created: String,
pub comments: String,
pub gallery: Vec<GalleryMedia>,
}
#[allow(dead_code)]
impl Post {
// Fetch posts of a user or subreddit and return a vector of posts and the "after" value
pub async fn fetch(path: &str, fallback_title: String) -> Result<(Vec<Self>, String), String> {
let res;
let post_list;
// Send a request to the url
match request(path.to_string()).await {
// If success, receive JSON in response
Ok(response) => {
res = response;
}
// If the Reddit API returns an error, exit this function
Err(msg) => return Err(msg),
}
// Fetch the list of posts from the JSON response
match res["data"]["children"].as_array() {
Some(list) => post_list = list,
None => return Err("No posts found".to_string()),
}
let mut posts: Vec<Self> = Vec::new();
// For each post from posts list
for post in post_list {
let data = &post["data"];
let (rel_time, created) = time(data["created_utc"].as_f64().unwrap_or_default());
let score = data["score"].as_i64().unwrap_or_default();
let ratio: f64 = data["upvote_ratio"].as_f64().unwrap_or(1.0) * 100.0;
let title = val(post, "title");
// Determine the type of media along with the media URL
let (post_type, media, gallery) = Media::parse(&data).await;
posts.push(Self {
id: val(post, "id"),
title: if title.is_empty() { fallback_title.to_owned() } else { title },
community: val(post, "subreddit"),
body: rewrite_urls(&val(post, "body_html")),
author: Author {
name: val(post, "author"),
flair: Flair {
flair_parts: FlairPart::parse(
data["author_flair_type"].as_str().unwrap_or_default(),
data["author_flair_richtext"].as_array(),
data["author_flair_text"].as_str(),
),
text: val(post, "link_flair_text"),
background_color: val(post, "author_flair_background_color"),
foreground_color: val(post, "author_flair_text_color"),
},
distinguished: val(post, "distinguished"),
},
score: if data["hide_score"].as_bool().unwrap_or_default() {
"\u{2022}".to_string()
} else {
format_num(score)
},
upvote_ratio: ratio as i64,
post_type,
thumbnail: Media {
url: format_url(val(post, "thumbnail").as_str()),
width: data["thumbnail_width"].as_i64().unwrap_or_default(),
height: data["thumbnail_height"].as_i64().unwrap_or_default(),
poster: "".to_string(),
},
media,
domain: val(post, "domain"),
flair: Flair {
flair_parts: FlairPart::parse(
data["link_flair_type"].as_str().unwrap_or_default(),
data["link_flair_richtext"].as_array(),
data["link_flair_text"].as_str(),
),
text: val(post, "link_flair_text"),
background_color: val(post, "link_flair_background_color"),
foreground_color: if val(post, "link_flair_text_color") == "dark" {
"black".to_string()
} else {
"white".to_string()
},
},
flags: Flags {
nsfw: data["over_18"].as_bool().unwrap_or_default(),
stickied: data["stickied"].as_bool().unwrap_or_default(),
},
permalink: val(post, "permalink"),
rel_time,
created,
comments: format_num(data["num_comments"].as_i64().unwrap_or_default()),
gallery,
});
}
Ok((posts, res["data"]["after"].as_str().unwrap_or_default().to_string()))
}
}
#[derive(Template)]
#[template(path = "comment.html", escape = "none")]
// Comment with content, post, score and data/time that it was posted
pub struct Comment {
pub id: String,
pub kind: String,
pub parent_id: String,
pub parent_kind: String,
pub post_link: String,
pub post_author: String,
pub body: String,
pub author: String,
pub author: Author,
pub score: String,
pub time: String,
pub rel_time: String,
pub created: String,
pub edited: (String, String),
pub replies: Vec<Comment>,
pub highlighted: bool,
}
#[allow(dead_code)]
#[derive(Template)]
#[template(path = "error.html", escape = "none")]
pub struct ErrorTemplate {
pub msg: String,
pub prefs: Preferences,
}
#[derive(Default)]
// User struct containing metadata about user
pub struct User {
pub name: String,
pub title: String,
pub icon: String,
pub karma: i64,
pub created: String,
pub banner: String,
pub description: String,
}
#[allow(dead_code)]
#[derive(Default)]
// Subreddit struct containing metadata about community
pub struct Subreddit {
pub name: String,
pub title: String,
pub description: String,
pub info: String,
pub icon: String,
pub members: String,
pub active: String,
pub wiki: bool,
}
// Parser for query params, used in sorting (eg. /r/rust/?sort=hot)
#[derive(serde::Deserialize)]
pub struct Params {
pub t: Option<String>,
pub q: Option<String>,
pub sort: Option<String>,
pub after: Option<String>,
pub before: Option<String>,
}
// Error template
#[derive(askama::Template)]
#[template(path = "error.html", escape = "none")]
pub struct ErrorTemplate {
pub message: String,
#[derive(Default)]
pub struct Preferences {
pub theme: String,
pub front_page: String,
pub layout: String,
pub wide: String,
pub show_nsfw: String,
pub comment_sort: String,
pub subscriptions: Vec<String>,
}
impl Preferences {
// Build preferences from cookies
pub fn new(req: Request<()>) -> Self {
Self {
theme: cookie(&req, "theme"),
front_page: cookie(&req, "front_page"),
layout: cookie(&req, "layout"),
wide: cookie(&req, "wide"),
show_nsfw: cookie(&req, "show_nsfw"),
comment_sort: cookie(&req, "comment_sort"),
subscriptions: cookie(&req, "subscriptions").split('+').map(String::from).filter(|s| !s.is_empty()).collect(),
}
}
}
//
// FORMATTING
//
pub async fn format_url(url: &str) -> String {
#[cfg(feature = "proxy")]
return "/proxy/".to_string() + encode(url).as_str();
#[cfg(not(feature = "proxy"))]
return url.to_string();
// Grab a query parameter from a url
pub fn param(path: &str, value: &str) -> String {
match Url::parse(format!("https://libredd.it/{}", path).as_str()) {
Ok(url) => url.query_pairs().into_owned().collect::<HashMap<_, _>>().get(value).unwrap_or(&String::new()).to_owned(),
_ => String::new(),
}
}
// Parse a cookie value from request
pub fn cookie(req: &Request<()>, name: &str) -> String {
let cookie = req.cookie(name).unwrap_or_else(|| Cookie::named(name));
cookie.value().to_string()
}
// Direct urls to proxy if proxy is enabled
pub fn format_url(url: &str) -> String {
if url.is_empty() || url == "self" || url == "default" || url == "nsfw" || url == "spoiler" {
String::new()
} else {
match Url::parse(url) {
Ok(parsed) => {
let domain = parsed.domain().unwrap_or_default();
let capture = |regex: &str, format: &str, segments: i16| {
Regex::new(regex)
.map(|re| match re.captures(url) {
Some(caps) => match segments {
1 => [format, &caps[1], "/"].join(""),
2 => [format, &caps[1], "/", &caps[2], "/"].join(""),
_ => String::new(),
},
None => String::new(),
})
.unwrap_or_default()
};
match domain {
"v.redd.it" => capture(r"https://v\.redd\.it/(.*)/DASH_([0-9]{2,4}(\.mp4|$))", "/vid/", 2),
"i.redd.it" => capture(r"https://i\.redd\.it/(.*)", "/img/", 1),
"a.thumbs.redditmedia.com" => capture(r"https://a\.thumbs\.redditmedia\.com/(.*)", "/thumb/a/", 1),
"b.thumbs.redditmedia.com" => capture(r"https://b\.thumbs\.redditmedia\.com/(.*)", "/thumb/b/", 1),
"emoji.redditmedia.com" => capture(r"https://emoji\.redditmedia\.com/(.*)/(.*)", "/emoji/", 2),
"preview.redd.it" => capture(r"https://preview\.redd\.it/(.*)\?(.*)", "/preview/pre/", 2),
"external-preview.redd.it" => capture(r"https://external\-preview\.redd\.it/(.*)\?(.*)", "/preview/external-pre/", 2),
"styles.redditmedia.com" => capture(r"https://styles\.redditmedia\.com/(.*)", "/style/", 1),
"www.redditstatic.com" => capture(r"https://www\.redditstatic\.com/(.*)", "/static/", 1),
_ => String::new(),
}
}
Err(_) => String::new(),
}
}
}
// Rewrite Reddit links to Libreddit in body of text
pub fn rewrite_urls(text: &str) -> String {
match Regex::new(r#"href="(https|http|)://(www.|old.|np.|)(reddit).(com)/"#) {
Ok(re) => re.replace_all(text, r#"href="/"#).to_string(),
Err(_) => String::new(),
}
}
// Append `m` and `k` for millions and thousands respectively
pub fn format_num(num: i64) -> String {
if num > 1000000 {
format!("{}m", num / 1000000)
} else if num > 1000 {
format!("{}k", num / 1000)
if num >= 1_000_000 {
format!("{}m", num / 1_000_000)
} else if num >= 1000 {
format!("{}k", num / 1_000)
} else {
num.to_string()
}
}
//
// JSON PARSING
//
// Parse a relative and absolute time from a UNIX timestamp
pub fn time(created: f64) -> (String, String) {
let time = OffsetDateTime::from_unix_timestamp(created.round() as i64);
let time_delta = OffsetDateTime::now_utc() - time;
// If the time difference is more than a month, show full date
let rel_time = if time_delta > Duration::days(30) {
time.format("%b %d '%y")
// Otherwise, show relative date/time
} else if time_delta.whole_days() > 0 {
format!("{}d ago", time_delta.whole_days())
} else if time_delta.whole_hours() > 0 {
format!("{}h ago", time_delta.whole_hours())
} else {
format!("{}m ago", time_delta.whole_minutes())
};
(rel_time, time.format("%b %d %Y, %H:%M:%S UTC"))
}
#[allow(dead_code)]
// val() function used to parse JSON from Reddit APIs
pub async fn val(j: &serde_json::Value, k: &str) -> String {
String::from(j["data"][k].as_str().unwrap_or(""))
}
#[allow(dead_code)]
// nested_val() function used to parse JSON from Reddit APIs
pub async fn nested_val(j: &serde_json::Value, n: &str, k: &str) -> String {
String::from(j["data"][n][k].as_str().unwrap())
}
#[allow(dead_code)]
pub async fn fetch_posts(url: String, fallback_title: String) -> Result<(Vec<Post>, String), &'static str> {
// Send a request to the url, receive JSON in response
let req = request(url).await;
// If the Reddit API returns an error, exit this function
if req.is_err() {
return Err(req.err().unwrap());
}
// Otherwise, grab the JSON output from the request
let res = req.unwrap();
// Fetch the list of posts from the JSON response
let post_list = res["data"]["children"].as_array().unwrap();
let mut posts: Vec<Post> = Vec::new();
for post in post_list.iter() {
let img = if val(post, "thumbnail").await.starts_with("https:/") {
format_url(val(post, "thumbnail").await.as_str()).await
} else {
String::new()
};
let unix_time: i64 = post["data"]["created_utc"].as_f64().unwrap().round() as i64;
let score = post["data"]["score"].as_i64().unwrap();
let title = val(post, "title").await;
posts.push(Post {
title: if title.is_empty() { fallback_title.to_owned() } else { title },
community: val(post, "subreddit").await,
body: val(post, "body").await,
author: val(post, "author").await,
score: format_num(score),
post_type: "link".to_string(),
media: img,
url: val(post, "permalink").await,
time: Utc.timestamp(unix_time, 0).format("%b %e '%y").to_string(),
flair: Flair(
val(post, "link_flair_text").await,
val(post, "link_flair_background_color").await,
if val(post, "link_flair_text_color").await == "dark" {
"black".to_string()
} else {
"white".to_string()
},
),
});
}
Ok((posts, res["data"]["after"].as_str().unwrap_or("").to_string()))
pub fn val(j: &Value, k: &str) -> String {
j["data"][k].as_str().unwrap_or_default().to_string()
}
//
// NETWORKING
//
// Make a request to a Reddit API and parse the JSON response
#[allow(dead_code)]
pub async fn request(url: String) -> Result<serde_json::Value, &'static str> {
// --- actix-web::client ---
// let client = actix_web::client::Client::default();
// let res = client
// .get(url)
// .send()
// .await?
// .body()
// .limit(1000000)
// .await?;
pub fn template(t: impl Template) -> tide::Result {
Ok(Response::builder(200).content_type("text/html").body(t.render().unwrap_or_default()).build())
}
// let body = std::str::from_utf8(res.as_ref())?; // .as_ref converts Bytes to [u8]
pub fn redirect(path: String) -> Response {
Response::builder(302)
.content_type("text/html")
.header("Location", &path)
.body(format!("Redirecting to <a href=\"{0}\">{0}</a>...", path))
.build()
}
// --- surf ---
// let req = get(&url).header("User-Agent", "libreddit");
// let client = client().with(Redirect::new(5));
// let mut res = client.send(req).await.unwrap();
// let success = res.status().is_success();
// let body = res.body_string().await.unwrap();
pub async fn error(req: Request<()>, msg: String) -> tide::Result {
let body = ErrorTemplate {
msg,
prefs: Preferences::new(req),
}
.render()
.unwrap_or_default();
// --- reqwest ---
let res = reqwest::get(&url).await.unwrap();
// Read the status from the response
let success = res.status().is_success();
// Read the body of the response
let body = res.text().await.unwrap();
Ok(Response::builder(404).content_type("text/html").body(body).build())
}
// Parse the response from Reddit as JSON
let json: Value = from_str(body.as_str()).unwrap_or(Value::Null);
#[async_recursion]
async fn connect(path: String) -> io::Result<String> {
// Build reddit-compliant user agent for Libreddit
let user_agent = format!("web:libreddit:{}", env!("CARGO_PKG_VERSION"));
if !success {
println!("! {} - {}", url, "Page not found");
Err("Page not found")
} else if json == Value::Null {
println!("! {} - {}", url, "Failed to parse page JSON data");
Err("Failed to parse page JSON data")
// Construct an HTTP request body
let req = format!(
"GET {} HTTP/1.1\r\nHost: www.reddit.com\r\nAccept: */*\r\nConnection: close\r\nUser-Agent: {}\r\n\r\n",
path, user_agent
);
// Open a TCP connection
let tcp_stream = TcpStream::connect("www.reddit.com:443").await?;
// Initialize TLS connector for requests
let connector = TlsConnector::default();
// Use the connector to start the handshake process
let mut tls_stream = connector.connect("www.reddit.com", tcp_stream).await?;
// Write the crafted HTTP request to the stream
tls_stream.write_all(req.as_bytes()).await?;
// And read the response
let mut writer = Vec::new();
io::copy(&mut tls_stream, &mut writer).await?;
let response = String::from_utf8_lossy(&writer).to_string();
let split = response.split("\r\n\r\n").collect::<Vec<&str>>();
let headers = split[0].split("\r\n").collect::<Vec<&str>>();
let status: i16 = headers[0].split(' ').collect::<Vec<&str>>()[1].parse().unwrap_or(200);
let body = split[1].to_string();
if (300..400).contains(&status) {
let location = headers
.iter()
.find(|header| header.starts_with("location:"))
.map(|f| f.to_owned())
.unwrap_or_default()
.split(": ")
.collect::<Vec<&str>>()[1];
connect(location.replace("https://www.reddit.com", "")).await
} else {
Ok(json)
Ok(body)
}
}
// Make a request to a Reddit API and parse the JSON response
#[cached(size = 100, time = 30, result = true)]
pub async fn request(path: String) -> Result<Value, String> {
let url = format!("https://www.reddit.com{}", path);
let err = |msg: &str, e: String| -> Result<Value, String> {
eprintln!("{} - {}: {}", url, msg, e);
Err(msg.to_string())
};
match connect(path).await {
Ok(body) => {
// Parse the response from Reddit as JSON
let parsed: Result<Value, Error> = from_str(&body);
match parsed {
Ok(json) => {
// If Reddit returned an error
if json["error"].is_i64() {
Err(
json["reason"]
.as_str()
.unwrap_or_else(|| {
json["message"].as_str().unwrap_or_else(|| {
eprintln!("{} - Error parsing reddit error", url);
"Error parsing reddit error"
})
})
.to_string(),
)
} else {
Ok(json)
}
}
Err(e) => err("Failed to parse page JSON data", e.to_string()),
}
}
Err(e) => err("Couldn't send request to Reddit", e.to_string()),
}
}

BIN
static/apple-touch-icon.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 5.5 KiB

BIN
static/favicon.ico Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 4.2 KiB

BIN
static/favicon.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 969 B

BIN
static/logo.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 7.9 KiB

23
static/manifest.json Normal file
View File

@ -0,0 +1,23 @@
{
"name": "Libreddit",
"short_name": "Libreddit",
"display": "standalone",
"background_color": "#1f1f1f",
"description": "An alternative private front-end to Reddit",
"theme_color": "#1f1f1f",
"icons": [
{
"src": "logo.png",
"sizes": "512x512",
"type": "image/png"
},
{
"src": "apple-touch-icon.png",
"sizes": "180x180"
},
{
"src": "favicon.ico",
"sizes": "32x32"
}
]
}

View File

@ -1,2 +0,0 @@
User-Agent: *
Disallow: /

File diff suppressed because it is too large Load Diff

View File

@ -2,29 +2,57 @@
<html lang="en">
<head>
{% block head %}
<title>{% block title %}{% endblock %}</title>
<title>{% block title %}Libreddit{% endblock %}</title>
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
<meta name="description" content="View on Libreddit, an alternative private front-end to Reddit.">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
<link rel="stylesheet" href="/style.css">
{% block sortstyle %}
<style>
#sort > #sort_{{ sort }} {
background: aqua;
color: black;
}
</style>
{% endblock %}
{% endblock %}
</head>
<body>
{% block header %}
<header>
<a href="/"><span id="lib">lib</span>reddit. <span id="version">v{{ env!("CARGO_PKG_VERSION") }}</span></a>
<a id="github" href="https://github.com/spikecodes/libreddit">GITHUB</a>
</header>
<!-- General PWA -->
<meta name="theme-color" content="#1F1F1F">
<!-- iOS Application -->
<meta name="apple-mobile-web-app-title" content="Libreddit">
<meta name="apple-mobile-web-app-capable" content="yes">
<meta name="apple-mobile-web-app-status-bar-style" content="default">
<!-- Android -->
<meta name="mobile-web-app-capable" content="yes">
<!-- iOS Logo -->
<link href="/touch-icon-iphone.png/" rel="apple-touch-icon">
<!-- PWA Manifest -->
<link rel="manifest" type="application/json" href="/manifest.json/">
<link rel="shortcut icon" type="image/x-icon" href="/favicon.ico/">
<link rel="stylesheet" type="text/css" href="/style.css/">
{% endblock %}
</head>
<body class="
{% if prefs.layout != "" %}{{ prefs.layout }}{% endif %}
{% if prefs.wide == "on" %} wide{% endif %}
{% if prefs.theme != "system" %} {{ prefs.theme }}{% endif %}">
<!-- NAVIGATION BAR -->
<nav>
<div id="logo">
<a id="libreddit" href="/"><span id="lib">lib</span><span id="reddit">reddit.</span></a>
<span id="version">v{{ env!("CARGO_PKG_VERSION") }}</span>
{% block subscriptions %}{% endblock %}
</div>
{% block search %}{% endblock %}
<div id="links">
<a id="settings_link" href="/settings/">
<span>settings</span>
<svg xmlns="http://www.w3.org/2000/svg" width="24" height="24" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round">
<title>settings</title>
<circle cx="12" cy="12" r="3"/><path d="M19.4 15a1.65 1.65 0 0 0 .33 1.82l.06.06a2 2 0 0 1 0 2.83 2 2 0 0 1-2.83 0l-.06-.06a1.65 1.65 0 0 0-1.82-.33 1.65 1.65 0 0 0-1 1.51V21a2 2 0 0 1-2 2 2 2 0 0 1-2-2v-.09A1.65 1.65 0 0 0 9 19.4a1.65 1.65 0 0 0-1.82.33l-.06.06a2 2 0 0 1-2.83 0 2 2 0 0 1 0-2.83l.06-.06a1.65 1.65 0 0 0 .33-1.82 1.65 1.65 0 0 0-1.51-1H3a2 2 0 0 1-2-2 2 2 0 0 1 2-2h.09A1.65 1.65 0 0 0 4.6 9a1.65 1.65 0 0 0-.33-1.82l-.06-.06a2 2 0 0 1 0-2.83 2 2 0 0 1 2.83 0l.06.06a1.65 1.65 0 0 0 1.82.33H9a1.65 1.65 0 0 0 1-1.51V3a2 2 0 0 1 2-2 2 2 0 0 1 2 2v.09a1.65 1.65 0 0 0 1 1.51 1.65 1.65 0 0 0 1.82-.33l.06-.06a2 2 0 0 1 2.83 0 2 2 0 0 1 0 2.83l-.06.06a1.65 1.65 0 0 0-.33 1.82V9a1.65 1.65 0 0 0 1.51 1H21a2 2 0 0 1 2 2 2 2 0 0 1-2 2h-.09a1.65 1.65 0 0 0-1.51 1z"/>
</svg>
</a>
<a id="code" href="https://github.com/spikecodes/libreddit">
<span>code</span>
<svg xmlns="http://www.w3.org/2000/svg" width="24" height="24" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round">
<title>code</title>
<polyline points="16 18 22 12 16 6"/><polyline points="8 6 2 12 8 18"/>
</svg>
</a>
</div>
</nav>
<!-- MAIN CONTENT -->
{% block body %}
<main>
{% block content %}
@ -32,4 +60,4 @@
</main>
{% endblock %}
</body>
</html>
</html>

25
templates/comment.html Normal file
View File

@ -0,0 +1,25 @@
{% import "utils.html" as utils %}
{% if kind == "more" && parent_kind == "t1" %}
<a class="deeper_replies" href="{{ post_link }}{{ parent_id }}">&rarr; More replies</a>
{% else if kind == "t1" %}
<div id="{{ id }}" class="comment">
<div class="comment_left">
<p class="comment_score">{{ score }}</p>
<div class="line"></div>
</div>
<details class="comment_right" open>
<summary class="comment_data">
<a class="comment_author {{ author.distinguished }} {% if author.name == post_author %}op{% endif %}" href="/u/{{ author.name }}">u/{{ author.name }}</a>
{% if author.flair.flair_parts.len() > 0 %}
<small class="author_flair">{% call utils::render_flair(author.flair.flair_parts) %}</small>
{% endif %}
<a href="{{ post_link }}{{ id }}/?context=3" class="created" title="{{ created }}">{{ rel_time }}</a>
{% if edited.0 != "".to_string() %}<span class="edited" title="{{ edited.1 }}">edited {{ edited.0 }}</span>{% endif %}
</summary>
<div class="comment_body {% if highlighted %}highlighted{% endif %}">{{ body }}</div>
<blockquote class="replies">{% for c in replies -%}{{ c.render().unwrap() }}{%- endfor %}
</blockquote>
</details>
</div>
{% endif %}

View File

@ -1,6 +1,9 @@
{% extends "base.html" %}
{% block title %}Error: {{ message }}{% endblock %}
{% block title %}Error: {{ msg }}{% endblock %}
{% block sortstyle %}{% endblock %}
{% block content %}
<h1 style="text-align: center; font-size: 50px;">{{ message }}</h1>
<div id="error">
<h1>{{ msg }}</h1>
<h3>Head back <a href="/">home</a>?</h3>
</div>
{% endblock %}

View File

@ -1,43 +0,0 @@
{% extends "base.html" %}
{% block title %}Libreddit{% endblock %}
{% block content %}
<div id="sort">
<div id="sort_hot"><a href="?sort=hot">Hot</a></div>
<div id="sort_top"><a href="?sort=top">Top</a></div>
<div id="sort_new"><a href="?sort=new">New</a></div>
<div id="sort_rising"><a href="?sort=rising">Rising</a></div>
</div>
{% for post in posts %}
<div class="post">
<div class="post_left">
<h3 class="post_score">{{ post.score }}</h3>
</div>
<div class="post_right">
<h4>
<b><a class="post_subreddit" href="/r/{{ post.community }}">r/{{ post.community }}</a></b>
&bull;
Posted by
<a class="post_author" href="/u/{{ post.author }}">u/{{ post.author }}</a>
<span class="datetime" style="float: right;">{{ post.time }}</span>
</h4>
<h3 class="post_title">
{% if post.flair.0 != "" %}
<small style="color:{{ post.flair.2 }}; background:{{ post.flair.1 }}">{{ post.flair.0 }}</small>
{% endif %}
<a href="{{ post.url }}">{{ post.title }}</a>
</h3>
</div>
<img class="post_thumbnail" src="{{ post.media }}">
</div><br>
{% endfor %}
<footer>
{% if ends.0 != "" %}
<a href="?before={{ ends.0 }}">PREV</a>
{% endif %}
{% if ends.1 != "" %}
<a href="?after={{ ends.1 }}">NEXT</a>
{% endif %}
</footer>
{% endblock %}

View File

@ -1,76 +1,131 @@
{% extends "base.html" %}
{% import "utils.html" as utils %}
{% block title %}{{ post.title }} - r/{{ post.community }}{% endblock %}
{% block head %}
{% call super() %}
<meta name="author" content="u/{{ post.author }}">
{% block search %}
{% call utils::search(["/r/", post.community.as_str()].concat(), "") %}
{% endblock %}
{% macro comment(item) -%}
{% block root %}/r/{{ post.community }}{% endblock %}{% block location %}r/{{ post.community }}{% endblock %}
{% block head %}
{% call super() %}
<!-- Meta Tags -->
<meta name="author" content="u/{{ post.author.name }}">
<meta name="title" content="{{ post.title }} - r/{{ post.community }}">
<meta name="description" content="View on Libreddit, an alternative private front-end to Reddit.">
<meta property="og:type" content="website">
<meta property="og:url" content="{{ post.permalink }}">
<meta property="og:title" content="{{ post.title }} - r/{{ post.community }}">
<meta property="og:description" content="View on Libreddit, an alternative private front-end to Reddit.">
<meta property="og:image" content="{{ post.thumbnail.url }}">
<meta property="twitter:card" content="summary_large_image">
<meta property="twitter:url" content="{{ post.permalink }}">
<meta property="twitter:title" content="{{ post.title }} - r/{{ post.community }}">
<meta property="twitter:description" content="View on Libreddit, an alternative private front-end to Reddit.">
<meta property="twitter:image" content="{{ post.thumbnail.url }}">
{% endblock %}
<div class="comment">
<div class="comment_left">
<h3 class="comment_score">{{ item.score }}</h3>
<div class="line"></div>
</div>
<details class="comment_right" open>
<summary class="comment_data">
<a class="comment_author" href="/u/{{ item.author }}">u/{{ item.author }}</a><span class="datetime">{{ item.time }}</span>
</summary>
<h4 class="comment_body">{{ item.body }}</h4>
{%- endmacro %}
{% block subscriptions %}
{% call utils::sub_list(post.community.as_str()) %}
{% endblock %}
{% block content %}
<div class="post highlighted">
<div class="post_left">
<h3 class="post_score">{{ post.score }}</h3>
</div>
<div class="post_right">
<h4>
<b><a class="post_subreddit" href="/r/{{ post.community }}">r/{{ post.community }}</a></b>
&bull;
Posted by
<a class="post_author" href="/u/{{ post.author }}">u/{{ post.author }}</a>
<span class="datetime">{{ post.time }}</span>
</h4>
<a href="{{ post.url }}" class="post_title">
{{ post.title }}
{% if post.flair.0 != "" %}
<small style="color:{{ post.flair.2 }}; background:{{ post.flair.1 }}">{{ post.flair.0 }}</small>
<div id="column_one">
<!-- POST CONTENT -->
<div class="post highlighted">
<p class="post_header">
<a class="post_subreddit" href="/r/{{ post.community }}">r/{{ post.community }}</a>
<span class="dot">&bull;</span>
<a class="post_author" href="/u/{{ post.author.name }}">u/{{ post.author.name }}</a>
{% if post.author.flair.flair_parts.len() > 0 %}
<small class="author_flair">{% call utils::render_flair(post.author.flair.flair_parts) %}</small>
{% endif %}
</a>
<span class="dot">&bull;</span>
<span class="created" title="{{ post.created }}">{{ post.rel_time }}</span>
</p>
<p class="post_title">
<a href="{{ post.permalink }}">{{ post.title }}</a>
{% if post.flair.flair_parts.len() > 0 %}
<a href="/r/{{ post.community }}/search/?q=flair_name%3A%22{{ post.flair.text }}%22&restrict_sr=on"
class="post_flair"
style="color:{{ post.flair.foreground_color }}; background:{{ post.flair.background_color }};">{% call utils::render_flair(post.flair.flair_parts) %}</a>
{% endif %}
{% if post.flags.nsfw %} <small class="nsfw">NSFW</small>{% endif %}
</p>
<!-- POST MEDIA -->
{% if post.post_type == "image" %}
<img class="post_media" src="{{ post.media }}"/>
{% else if post.post_type == "video" %}
<video class="post_media" src="{{ post.media }}" controls autoplay loop>
<a href="{{ post.media.url }}" class="post_media_image" >
<svg
width="{{ post.media.width }}px"
height="{{ post.media.height }}px"
xmlns="http://www.w3.org/2000/svg">
<image width="100%" height="100%" href="{{ post.media.url }}"/>
<desc>
<img alt="Post image" src="{{ post.media.url }}"/>
</desc>
</svg>
</a>
{% else if post.post_type == "video" || post.post_type == "gif" %}
<video class="post_media_video" src="{{ post.media.url }}" controls autoplay loop><a href={{ post.media.url }}>Video</a></video>
{% else if post.post_type == "gallery" %}
<div class="gallery">
{% for image in post.gallery -%}
<figure>
<a href="{{ image.url }}" ><img alt="Gallery image" src="{{ image.url }}"/></a>
<figcaption>
<p>{{ image.caption }}</p>
{% if image.outbound_url.len() > 0 %}
<p><a class="outbound_url" href="{{ image.outbound_url }}">{{ image.outbound_url }}</a>
{% endif %}
</figcaption>
</figure>
{%- endfor %}
</div>
{% else if post.post_type == "link" %}
<a id="post_url" href="{{ post.media }}">{{ post.media }}</a>
<a id="post_url" href="{{ post.media.url }}">{{ post.media.url }}</a>
{% endif %}
<h4 class="post_body">{{ post.body }}</h4>
</div>
</div>
<div id="sort">
<div id="sort_confidence"><a href="?sort=confidence">Best</a></div>
<div id="sort_top"><a href="?sort=top">Top</a></div>
<div id="sort_new"><a href="?sort=new">New</a></div>
<div id="sort_controversial"><a href="?sort=controversial">Controversial</a></div>
<div id="sort_old"><a href="?sort=old">Old</a></div>
</div>
{% for c in comments -%}
<!-- POST BODY -->
<div class="post_body">{{ post.body }}</div>
<div class="post_score">{{ post.score }}<span class="label"> Upvotes</span></div>
<div class="post_footer">
<ul id="post_links">
<li><a href="/{{ post.id }}">permalink</a></li>
<li><a href="https://reddit.com/{{ post.id }}">reddit</a></li>
</ul>
<p>{{ post.upvote_ratio }}% Upvoted</p>
</div>
</div>
<!-- SORT FORM -->
<form id="sort">
<select name="sort" title="Sort comments by">
{% call utils::options(sort, ["confidence", "top", "new", "controversial", "old"], "confidence") %}
</select><button id="sort_submit" class="submit">
<svg width="15" viewBox="0 0 110 100" fill="none" stroke-width="10" stroke-linecap="round">
<path d="M20 50 H100" />
<path d="M75 15 L100 50 L75 85" />
&rarr;
</svg>
</button>
</form>
<!-- COMMENTS -->
{% for c in comments -%}
<div class="thread">
{% call comment(c) %}
<div class="replies">
{% for reply in c.replies %}
{% call comment(reply) %}
<div class="replies">
{% for response in reply.replies %}
{% call comment(response) %}</details></div>
{% endfor %}
</div></details></div>
{% endfor %}
</div></details></div>
{% if single_thread %}
<p class="thread_nav"><a href="/{{ post.id }}">View all comments</a></p>
{% if c.parent_kind == "t1" %}
<p class="thread_nav"><a href="?context=9999">Show parent comments</a></p>
{% endif %}
{% endif %}
{{ c.render().unwrap() }}
</div>
{%- endfor %}
{%- endfor %}
{% endblock %}
</div>
{% endblock %}

83
templates/search.html Normal file
View File

@ -0,0 +1,83 @@
{% extends "base.html" %}
{% import "utils.html" as utils %}
{% block title %}Libreddit: search results - {{ params.q }}{% endblock %}
{% block subscriptions %}
{% call utils::sub_list("") %}
{% endblock %}
{% block content %}
<div id="column_one">
<form id="search_sort">
<input id="search" type="text" name="q" placeholder="Search" value="{{ params.q }}" title="Search libreddit">
{% if sub != "" %}
<div id="inside">
<input type="checkbox" name="restrict_sr" id="restrict_sr" {% if params.restrict_sr != "" %}checked{% endif %}>
<label for="restrict_sr" class="search_label">in r/{{ sub }}</label>
</div>
{% endif %}
<select id="sort_options" name="sort" title="Sort results by">
{% call utils::options(params.sort, ["relevance", "hot", "top", "new", "comments"], "") %}
</select>{% if params.sort != "new" %}<select id="timeframe" name="t" title="Timeframe">
{% call utils::options(params.t, ["hour", "day", "week", "month", "year", "all"], "all") %}
</select>{% endif %}<button id="sort_submit" class="submit">
<svg width="15" viewBox="0 0 110 100" fill="none" stroke-width="10" stroke-linecap="round">
<path d="M20 50 H100" />
<path d="M75 15 L100 50 L75 85" />
&rarr;
</svg>
</button>
</form>
{% if subreddits.len() > 0 %}
<div id="search_subreddits">
{% for subreddit in subreddits %}
<a href="{{ subreddit.url }}" class="search_subreddit">
<p class="search_subreddit_header">
<span class="search_subreddit_name">{{ subreddit.name }}</span>
<span class="dot">&bull;</span>
<span class="search_subreddit_members">{{ subreddit.subscribers }} Members</span>
</p>
<p class="search_subreddit_description">{{ subreddit.description }}</p>
</a>
{% endfor %}
</div>
{% endif %}
{% for post in posts %}
{% if post.flags.nsfw && prefs.show_nsfw != "on" %}
{% else if post.title != "Comment" %}
{% call utils::post_in_list(post) %}
{% else %}
<div class="comment">
<div class="comment_left">
<p class="comment_score">{{ post.score }}</p>
<div class="line"></div>
</div>
<details class="comment_right" open>
<summary class="comment_data">
<a class="comment_link" href="{{ post.permalink }}">COMMENT</a>
<span class="created" title="{{ post.created }}">{{ post.rel_time }}</span>
</summary>
<p class="comment_body">{{ post.body }}</p>
</details>
</div>
{% endif %}
{% endfor %}
<footer>
{% if params.before != "" %}
<a href="?q={{ params.q }}&restrict_sr={{ params.restrict_sr }}
&sort={{ params.sort }}&t={{ params.t }}
&before={{ params.before }}">PREV</a>
{% endif %}
{% if params.after != "" %}
<a href="?q={{ params.q }}&restrict_sr={{ params.restrict_sr }}
&sort={{ params.sort }}&t={{ params.t }}
&after={{ params.after }}">NEXT</a>
{% endif %}
</footer>
</div>
{% endblock %}

72
templates/settings.html Normal file
View File

@ -0,0 +1,72 @@
{% extends "base.html" %}
{% import "utils.html" as utils %}
{% block title %}Libreddit Settings{% endblock %}
{% block search %}
{% call utils::search("".to_owned(), "", "") %}
{% endblock %}
{% block content %}
<div id="settings">
<form action="/settings/" method="POST">
<div class="prefs">
<p>Appearance</p>
<div id="theme">
<label for="theme">Theme:</label>
<select name="theme">
{% call utils::options(prefs.theme, ["system", "light", "dark", "black"], "system") %}
</select>
</div>
<p>Interface</p>
<div id="front_page">
<label for="front_page">Front page:</label>
<select name="front_page">
{% call utils::options(prefs.front_page, ["default", "popular", "all"], "default") %}
</select>
</div>
<div id="layout">
<label for="layout">Layout:</label>
<select name="layout">
{% call utils::options(prefs.layout, ["card", "clean", "compact"], "card") %}
</select>
</div>
<div id="wide">
<label for="wide">Wide UI:</label>
<input type="checkbox" name="wide" {% if prefs.wide == "on" %}checked{% endif %}>
</div>
<p>Content</p>
<div id="comment_sort">
<label for="comment_sort">Default comment sort:</label>
<select name="comment_sort">
{% call utils::options(prefs.comment_sort, ["confidence", "top", "new", "controversial", "old"], "confidence") %}
</select>
</div>
<div id="show_nsfw">
<label for="show_nsfw">Show NSFW posts:</label>
<input type="checkbox" name="show_nsfw" {% if prefs.show_nsfw == "on" %}checked{% endif %}>
</div>
<input id="save" type="submit" value="Save">
</div>
</form>
{% if prefs.subscriptions.len() > 0 %}
<div class="prefs" id="settings_subs">
<p>Subscribed Subreddits</p>
{% for sub in prefs.subscriptions %}
<div>
<span>{{ sub }}</span>
<form action="/r/{{ sub }}/unsubscribe/?redirect=settings" method="POST">
<button class="unsubscribe">Unsubscribe</button>
</form>
</div>
{% endfor %}
</div>
{% endif %}
<div id="settings_note">
<p><b>Note:</b> settings and subscriptions are saved in browser cookies. Clearing your cookies will reset them.</p><br>
<p>You can restore your current settings and subscriptions after clearing your cookies using <a href="/settings/restore/?theme={{ prefs.theme }}&front_page={{ prefs.front_page }}&layout={{ prefs.layout }}&wide={{ prefs.wide }}&comment_sort={{ prefs.comment_sort }}&show_nsfw={{ prefs.show_nsfw }}&subscriptions={{ prefs.subscriptions.join("%2B") }}">this link</a>.</p>
</div>
</div>
{% endblock %}

View File

@ -1,56 +1,107 @@
{% extends "base.html" %}
{% block title %}r/{{ sub.name }}: {{ sub.description }}{% endblock %}
{% import "utils.html" as utils %}
{% block title %}
{% if sub.title != "" %}{{ sub.title }}
{% else if sub.name != "" %}{{ sub.name }}
{% else %}Libreddit{% endif %}
{% endblock %}
{% block search %}
{% call utils::search(["/r/", sub.name.as_str()].concat(), "") %}
{% endblock %}
{% block subscriptions %}
{% call utils::sub_list(sub.name.as_str(), "wide") %}
{% endblock %}
{% block body %}
<div id="about">
<div class="subreddit">
<div class="subreddit_left">
{{ sub.icon }}
</div>
<div class="subreddit_right">
<h2 class="subreddit_name">r/{{ sub.name }}</h2>
<p class="subreddit_description">{{ sub.description }}</p>
<div id="stats">👤 {{ sub.members }} 🟢 {{ sub.active }}</div>
</div>
</div>
</div>
<main>
<div id="sort">
<div id="sort_hot"><a href="?sort=hot">Hot</a></div>
<div id="sort_top"><a href="?sort=top">Top</a></div>
<div id="sort_new"><a href="?sort=new">New</a></div>
</div>
{% for post in posts %}
<div class="post">
<div class="post_left">
<h3 class="post_score">{{ post.score }}</h3>
</div>
<div class="post_right">
<h4>
<b><a class="post_subreddit" href="/r/{{ post.community }}">r/{{ sub.name }}</a></b>
&bull;
Posted by
<a class="post_author" href="/u/{{ post.author }}">u/{{ post.author }}</a>
<span class="datetime">{{ post.time }}</span>
</h4>
<h3 class="post_title">
{% if post.flair.0 != "" %}
<small style="color:{{ post.flair.2 }}; background:{{ post.flair.1 }}">{{ post.flair.0 }}</small>
<div id="column_one">
<form id="sort">
<div id="sort_options">
{% if sub.name.is_empty() %}
{% call utils::sort("", ["hot", "new", "top", "rising", "controversial"], sort.0) %}
{% else %}
{% call utils::sort(["/r/", sub.name.as_str()].concat(), ["hot", "new", "top", "rising", "controversial"], sort.0) %}
{% endif %}
<a href="{{ post.url }}">{{ post.title }}</a>
</h3>
</div>
{% if sort.0 == "top" || sort.0 == "controversial" %}<select id="timeframe" name="t" title="Timeframe">
{% call utils::options(sort.1, ["hour", "day", "week", "month", "year", "all"], "day") %}
</select>
<button id="sort_submit" class="submit">
<svg width="15" viewBox="0 0 110 100" fill="none" stroke-width="10" stroke-linecap="round">
<path d="M20 50 H100" />
<path d="M75 15 L100 50 L75 85" />
&rarr;
</svg>
</button>
{% endif %}
</form>
{% if sub.name.contains("+") %}
<form action="/r/{{ sub.name }}/subscribe/" method="POST">
<button id="multisub" class="subscribe" title="Subscribe to each sub in this multireddit">Subscribe to Multireddit</button>
</form>
{% endif %}
<div id="posts">
{% for post in posts %}
{% if !(post.flags.nsfw && prefs.show_nsfw != "on") %}
<hr class="sep" />
{% call utils::post_in_list(post) %}
{% endif %}
{% endfor %}
</div>
<img class="post_thumbnail" src="{{ post.media }}">
</div><br>
{% endfor %}
<footer>
{% if ends.0 != "" %}
<a href="?before={{ ends.0 }}">PREV</a>
{% endif %}
<footer>
{% if ends.0 != "" %}
<a href="?sort={{ sort.0 }}&t={{ sort.1 }}&before={{ ends.0 }}">PREV</a>
{% endif %}
{% if ends.1 != "" %}
<a href="?after={{ ends.1 }}">NEXT</a>
{% endif %}
</footer>
{% if ends.1 != "" %}
<a href="?sort={{ sort.0 }}&t={{ sort.1 }}&after={{ ends.1 }}">NEXT</a>
{% endif %}
</footer>
</div>
{% if sub.name != "" && !sub.name.contains("+") %}
<aside>
<div class="panel" id="subreddit">
{% if sub.wiki %}
<div id="top">
<div>Posts</div>
<a href="/r/{{ sub.name }}/wiki/index">Wiki</a>
</div>
{% endif %}
<div id="sub_meta">
<img id="sub_icon" src="{{ sub.icon }}" alt="Icon for r/{{ sub.name }}">
<p id="sub_title">{{ sub.title }}</p>
<p id="sub_name">r/{{ sub.name }}</p>
<p id="sub_description">{{ sub.description }}</p>
<div id="sub_details">
<label>Members</label>
<label>Active</label>
<div>{{ sub.members }}</div>
<div>{{ sub.active }}</div>
</div>
<div id="sub_subscription">
{% if prefs.subscriptions.contains(sub.name) %}
<form action="/r/{{ sub.name }}/unsubscribe/" method="POST">
<button class="unsubscribe">Unsubscribe</button>
</form>
{% else %}
<form action="/r/{{ sub.name }}/subscribe/" method="POST">
<button class="subscribe">Subscribe</button>
</form>
{% endif %}
</div>
</div>
</div>
<details class="panel" id="sidebar">
<summary id="sidebar_label">Sidebar</summary>
<div id="sidebar_contents">{{ sub.info }}</div>
</details>
</aside>
{% endif %}
</main>
{% endblock %}
{% endblock %}

View File

@ -1,62 +1,81 @@
{% extends "base.html" %}
{% block title %}Libreddit: u/{{ user.name }}{% endblock %}
{% import "utils.html" as utils %}
{% block search %}
{% call utils::search("".to_owned(), "", "") %}
{% endblock %}
{% block title %}{{ user.name.replace("u/", "") }} (u/{{ user.name }}) - Libreddit{% endblock %}
{% block subscriptions %}
{% call utils::sub_list("") %}
{% endblock %}
{% block body %}
<div id="about">
<div class="user">
<div class="user_left">
<img class="user_icon" src="{{ user.icon }}">
</div>
<div class="user_right">
<h2 class="user_name">u/{{ user.name }}</h2>
<p class="user_description"><span>Karma:</span> {{ user.karma }} | <span>Description:</span> "{{ user.description }}"</p>
</div>
</div>
</div>
<main>
<div id="sort">
<div id="sort_hot"><a href="?sort=hot">Hot</a></div>
<div id="sort_top"><a href="?sort=top">Top</a></div>
<div id="sort_new"><a href="?sort=new">New</a></div>
<div id="column_one">
<form id="sort">
<select name="sort">
{% call utils::options(sort.0, ["hot", "new", "top"], "") %}
</select>{% if sort.0 == "top" %}<select id="timeframe" name="t">
{% call utils::options(sort.1, ["hour", "day", "week", "month", "year", "all"], "all") %}
</select>{% endif %}<button id="sort_submit" class="submit">
<svg width="15" viewBox="0 0 110 100" fill="none" stroke-width="10" stroke-linecap="round">
<path d="M20 50 H100" />
<path d="M75 15 L100 50 L75 85" />
&rarr;
</svg>
</button>
</form>
<div id="posts">
{% for post in posts %}
{% if post.flags.nsfw && prefs.show_nsfw != "on" %}
{% else if post.title != "Comment" %}
{% call utils::post_in_list(post) %}
{% else %}
<div class="comment">
<div class="comment_left">
<p class="comment_score">{{ post.score }}</p>
<div class="line"></div>
</div>
<details class="comment_right" open>
<summary class="comment_data">
<a class="comment_link" href="{{ post.permalink }}">COMMENT</a>
<span class="created" title="{{ post.created }}">{{ post.rel_time }}</span>
</summary>
<p class="comment_body">{{ post.body }}</p>
</details>
</div>
{% endif %}
{% endfor %}
</div>
<footer>
{% if ends.0 != "" %}
<a href="?sort={{ sort.0 }}&t={{ sort.1 }}&before={{ ends.0 }}">PREV</a>
{% endif %}
{% if ends.1 != "" %}
<a href="?sort={{ sort.0 }}&t={{ sort.1 }}&after={{ ends.1 }}">NEXT</a>
{% endif %}
</footer>
</div>
{% for post in posts %}
{% if post.title != "Comment" %}
<div class='post'>
<div class="post_left">
<h3 class="post_score">{{ post.score }}</h3>
<aside>
<div class="panel" id="user">
<img id="user_icon" src="{{ user.icon }}" alt="User icon">
<p id="user_title">{{ user.title }}</p>
<p id="user_name">u/{{ user.name }}</p>
<div id="user_description">{{ user.description }}</div>
<div id="user_details">
<label>Karma</label>
<label>Created</label>
<div>{{ user.karma }}</div>
<div>{{ user.created }}</div>
</div>
</div>
<div class="post_right">
<h4>
<b><a class="post_subreddit" href="/r/{{ post.community }}">r/{{ post.community }}</a></b>
&bull;
Posted by
<a class="post_author" href="/u/{{ post.author }}">u/{{ post.author }}</a>
<span class="datetime" style="float: right;">{{ post.time }}</span>
</h4>
<h3 class="post_title">
{% if post.flair.0 == "Comment" %}
{% else if post.flair.0 == "" %}
{% else %}
<small style="color:{{ post.flair.2 }}; background:{{ post.flair.1 }}">{{ post.flair.0 }}</small>
{% endif %}
<a href="{{ post.url }}">{{ post.title }}</a>
</h3>
</div>
<img class="post_thumbnail" src="{{ post.media }}">
</div><br>
{% else %}
<div class="comment">
<div class="comment_left">
<h3 class="comment_score">{{ post.score }}</h3>
</div>
<div class="comment_right">
<h4>
COMMENT
<span class="datetime">{{ post.time }}</span>
</h4>
<h4 class="comment_body">{{ post.body }}</h4>
</div>
</div><br>
{% endif %}
{% endfor %}
</aside>
</main>
{% endblock %}
{% endblock %}

116
templates/utils.html Normal file
View File

@ -0,0 +1,116 @@
{% macro options(current, values, default) -%}
{% for value in values %}
<option value="{{ value }}" {% if current == value || (current == "" && value == default) %}selected{% endif %}>
{{ format!("{}{}", value.get(0..1).unwrap_or_default().to_uppercase(), value.get(1..).unwrap_or_default()) }}
</option>
{% endfor %}
{%- endmacro %}
{% macro sort(root, methods, selected) -%}
{% for method in methods %}
<a {% if method == selected %}class="selected"{% endif %} href="{{ root }}/{{ method }}">
{{ format!("{}{}", method.get(0..1).unwrap_or_default().to_uppercase(), method.get(1..).unwrap_or_default()) }}
</a>
{% endfor %}
{%- endmacro %}
{% macro search(root, search) -%}
<form action="{% if root != "/r/" && !root.is_empty() %}{{ root }}{% endif %}/search/" id="searchbox">
<input id="search" type="text" name="q" placeholder="Search" title="Search libreddit" value="{{ search }}">
{% if root != "/r/" && !root.is_empty() %}
<div id="inside">
<input type="checkbox" name="restrict_sr" id="restrict_sr">
<label for="restrict_sr" class="search_label" title="Restrict search to this subreddit">in {{ root }}</label>
</div>
{% endif %}
<button class="submit">
<svg width="15" viewBox="0 0 110 100" fill="none" stroke-width="10" stroke-linecap="round">
<path d="M20 50 H100" />
<path d="M75 15 L100 50 L75 85" />
&rarr;
</svg>
</button>
</form>
{%- endmacro %}
{% macro render_flair(flair_parts) -%}
{% for flair_part in flair_parts %}{% if flair_part.flair_part_type == "emoji" %}<span class="emoji" style="background-image:url('{{ flair_part.value }}');"></span>{% else if flair_part.flair_part_type == "text" && !flair_part.value.is_empty() %}<span>{{ flair_part.value }}</span>{% endif %}{% endfor %}
{%- endmacro %}
{% macro sub_list(current) -%}
{% if prefs.subscriptions.len() > 0 %}
<details id="feeds">
<summary>Feeds</summary>
<div id="feed_list">
<p>MAIN FEEDS</p>
<a href="/">Home</a>
<a href="/r/popular">Popular</a>
<a href="/r/all">All</a>
<p>REDDIT FEEDS</p>
{% for sub in prefs.subscriptions %}
<a href="/r/{{ sub }}" {% if sub == current %}class="selected"{% endif %}>{{ sub }}</a>
{% endfor %}
</div>
</details>
{% endif %}
{%- endmacro %}
{% macro post_in_list(post) -%}
<div class="post {% if post.flags.stickied %}stickied{% endif %}">
<p class="post_header">
<a class="post_subreddit" href="/r/{{ post.community }}">r/{{ post.community }}</a>
<span class="dot">&bull;</span>
<a class="post_author" href="/u/{{ post.author.name }}">u/{{ post.author.name }}</a>
<span class="dot">&bull;</span>
<span class="created" title="{{ post.created }}">{{ post.rel_time }}</span>
</p>
<p class="post_title">
{% if post.flair.flair_parts.len() > 0 %}
<a href="/r/{{ post.community }}/search/?q=flair_name%3A%22{{ post.flair.text }}%22&restrict_sr=on"
class="post_flair"
style="color:{{ post.flair.foreground_color }}; background:{{ post.flair.background_color }};">{% call render_flair(post.flair.flair_parts) %}</a>
{% endif %}
<a href="{{ post.permalink }}">{{ post.title }}</a>{% if post.flags.nsfw %} <small class="nsfw">NSFW</small>{% endif %}
</p>
<!-- POST MEDIA/THUMBNAIL -->
{% if (prefs.layout.is_empty() || prefs.layout == "card") && post.post_type == "image" %}
<a href="{{ post.media.url }}" class="post_media_image {% if post.media.height / post.media.width < 2 %}short{% endif %}" >
<svg
width="{{ post.media.width }}px"
height="{{ post.media.height }}px"
xmlns="http://www.w3.org/2000/svg">
<image width="100%" height="100%" href="{{ post.media.url }}"/>
<desc>
<img alt="Post image" src="{{ post.media.url }}"/>
</desc>
</svg>
</a>
{% else if (prefs.layout.is_empty() || prefs.layout == "card") && post.post_type == "gif" %}
<video class="post_media_video short" src="{{ post.media.url }}" width="{{ post.media.width }}" height="{{ post.media.height }}" controls loop autoplay><a href={{ post.media.url }}>Video</a></video>
{% else if (prefs.layout.is_empty() || prefs.layout == "card") && post.post_type == "video" %}
<video class="post_media_video short" src="{{ post.media.url }}" width="{{ post.media.width }}" height="{{ post.media.height }}" poster="{{ post.media.poster }}" preload="none" controls autoplay><a href={{ post.media.url }}>Video</a></video>
{% else if post.post_type != "self" %}
<a class="post_thumbnail {% if post.thumbnail.url.is_empty() %}no_thumbnail{% endif %}" href="{% if post.post_type == "link" %}{{ post.media.url }}{% else %}{{ post.permalink }}{% endif %}">
{% if post.thumbnail.url.is_empty() %}
<svg viewBox="0 0 100 106" width="140" height="53" xmlns="http://www.w3.org/2000/svg">
<title>Thumbnail</title>
<path d="M35,15h-15a10,10 0,0,0 0,20h25a10,10 0,0,0 10,-10m-12.5,0a10, 10 0,0,1 10, -10h25a10,10 0,0,1 0,20h-15" fill="none" stroke-width="5" stroke-linecap="round"/>
</svg>
{% else %}
<svg width="{{ post.thumbnail.width }}px" height="{{ post.thumbnail.height }}px" xmlns="http://www.w3.org/2000/svg">
<image width="100%" height="100%" href="{{ post.thumbnail.url }}"/>
<desc>
<img alt="Thumbnail" src="{{ post.thumbnail.url }}"/>
</desc>
</svg>
{% endif %}
<span>{% if post.post_type == "link" %}{{ post.domain }}{% else %}{{ post.post_type }}{% endif %}</span>
</a>
{% endif %}
<div class="post_score">{{ post.score }}<span class="label"> Upvotes</span></div>
<div class="post_footer">
<a href="{{ post.permalink }}" class="post_comments">{{ post.comments }} comments</a>
</div>
</div>
{%- endmacro %}

29
templates/wiki.html Normal file
View File

@ -0,0 +1,29 @@
{% extends "base.html" %}
{% import "utils.html" as utils %}
{% block title %}
{% if sub != "" %}{{ page }} - {{ sub }}
{% else %}Libreddit{% endif %}
{% endblock %}
{% block search %}
{% call utils::search(["/r/", sub.as_str()].concat(), "") %}
{% endblock %}
{% block subscriptions %}
{% call utils::sub_list(sub.as_str()) %}
{% endblock %}
{% block body %}
<main>
<div class="panel" id="column_one">
<div id="top">
<a href="/r/{{ sub }}">Posts</a>
<div>Wiki</div>
</div>
<div id="wiki">
{{ wiki }}
</div>
</div>
</main>
{% endblock %}