158 lines
5.2 KiB
Rust
158 lines
5.2 KiB
Rust
// CRATES
|
|
use crate::utils::{cookie, error, format_num, format_url, media, parse_rich_flair, param, prefs, request, rewrite_url, val, Comment, Flags, Flair, Post, Preferences};
|
|
use actix_web::{HttpRequest, HttpResponse};
|
|
|
|
use async_recursion::async_recursion;
|
|
|
|
use askama::Template;
|
|
use time::OffsetDateTime;
|
|
|
|
// STRUCTS
|
|
#[derive(Template)]
|
|
#[template(path = "post.html", escape = "none")]
|
|
struct PostTemplate {
|
|
comments: Vec<Comment>,
|
|
post: Post,
|
|
sort: String,
|
|
prefs: Preferences,
|
|
}
|
|
|
|
pub async fn item(req: HttpRequest) -> HttpResponse {
|
|
// Build Reddit API path
|
|
let mut path: String = format!("{}.json?{}&raw_json=1", req.path(), req.query_string());
|
|
|
|
// Set sort to sort query parameter
|
|
let mut sort: String = param(&path, "sort");
|
|
|
|
// Grab default comment sort method from Cookies
|
|
let default_sort = cookie(&req, "comment_sort");
|
|
|
|
// If there's no sort query but there's a default sort, set sort to default_sort
|
|
if sort.is_empty() && !default_sort.is_empty() {
|
|
sort = default_sort;
|
|
path = format!("{}.json?{}&sort={}&raw_json=1", req.path(), req.query_string(), sort);
|
|
}
|
|
|
|
// Log the post ID being fetched in debug mode
|
|
#[cfg(debug_assertions)]
|
|
dbg!(req.match_info().get("id").unwrap_or(""));
|
|
|
|
// Send a request to the url, receive JSON in response
|
|
match request(&path).await {
|
|
// Otherwise, grab the JSON output from the request
|
|
Ok(res) => {
|
|
// Parse the JSON into Post and Comment structs
|
|
let post = parse_post(&res[0]).await;
|
|
let comments = parse_comments(&res[1]).await;
|
|
|
|
// Use the Post and Comment structs to generate a website to show users
|
|
let s = PostTemplate {
|
|
comments,
|
|
post,
|
|
sort,
|
|
prefs: prefs(req),
|
|
}
|
|
.render()
|
|
.unwrap();
|
|
HttpResponse::Ok().content_type("text/html").body(s)
|
|
}
|
|
// If the Reddit API returns an error, exit and send error page to user
|
|
Err(msg) => error(msg.to_string()).await,
|
|
}
|
|
}
|
|
|
|
// POSTS
|
|
async fn parse_post(json: &serde_json::Value) -> Post {
|
|
// Retrieve post (as opposed to comments) from JSON
|
|
let post: &serde_json::Value = &json["data"]["children"][0];
|
|
|
|
// Grab UTC time as unix timestamp
|
|
let unix_time: i64 = post["data"]["created_utc"].as_f64().unwrap_or_default().round() as i64;
|
|
// Parse post score and upvote ratio
|
|
let score = post["data"]["score"].as_i64().unwrap_or_default();
|
|
let ratio: f64 = post["data"]["upvote_ratio"].as_f64().unwrap_or(1.0) * 100.0;
|
|
|
|
// Determine the type of media along with the media URL
|
|
let (post_type, media) = media(&post["data"]).await;
|
|
|
|
// Build a post using data parsed from Reddit post API
|
|
Post {
|
|
id: val(post, "id"),
|
|
title: val(post, "title"),
|
|
community: val(post, "subreddit"),
|
|
body: rewrite_url(&val(post, "selftext_html")),
|
|
author: val(post, "author"),
|
|
author_flair: Flair{
|
|
flair_parts: parse_rich_flair(val(post, "author_flair_type"), post["data"]["author_flair_richtext"].as_array(), post["data"]["author_flair_text"].as_str()),
|
|
background_color: val(post, "author_flair_background_color"),
|
|
foreground_color: val(post, "author_flair_text_color"),
|
|
},
|
|
permalink: val(post, "permalink"),
|
|
score: format_num(score),
|
|
upvote_ratio: ratio as i64,
|
|
post_type,
|
|
thumbnail: format_url(val(post, "thumbnail").as_str()),
|
|
flair: Flair{
|
|
flair_parts: parse_rich_flair(val(post, "link_flair_type"), post["data"]["link_flair_richtext"].as_array(), post["data"]["link_flair_text"].as_str()),
|
|
background_color: val(post, "link_flair_background_color"),
|
|
foreground_color: if val(post, "link_flair_text_color") == "dark" {
|
|
"black".to_string()
|
|
} else {
|
|
"white".to_string()
|
|
},
|
|
},
|
|
flags: Flags {
|
|
nsfw: post["data"]["over_18"].as_bool().unwrap_or(false),
|
|
stickied: post["data"]["stickied"].as_bool().unwrap_or(false),
|
|
},
|
|
media,
|
|
domain: val(post, "domain"),
|
|
time: OffsetDateTime::from_unix_timestamp(unix_time).format("%b %d %Y %H:%M UTC"),
|
|
}
|
|
}
|
|
|
|
// COMMENTS
|
|
#[async_recursion]
|
|
async fn parse_comments(json: &serde_json::Value) -> Vec<Comment> {
|
|
// Separate the comment JSON into a Vector of comments
|
|
let comment_data = match json["data"]["children"].as_array() {
|
|
Some(f) => f.to_owned(),
|
|
None => { let v = Vec::new(); v }
|
|
};
|
|
|
|
let mut comments: Vec<Comment> = Vec::new();
|
|
|
|
// For each comment, retrieve the values to build a Comment object
|
|
for comment in comment_data {
|
|
let unix_time: i64 = comment["data"]["created_utc"].as_f64().unwrap_or(0.0).round() as i64;
|
|
if unix_time == 0 {
|
|
continue;
|
|
}
|
|
|
|
let score = comment["data"]["score"].as_i64().unwrap_or(0);
|
|
let body = rewrite_url(&val(&comment, "body_html"));
|
|
|
|
let replies: Vec<Comment> = if comment["data"]["replies"].is_object() {
|
|
parse_comments(&comment["data"]["replies"]).await
|
|
} else {
|
|
Vec::new()
|
|
};
|
|
|
|
comments.push(Comment {
|
|
id: val(&comment, "id"),
|
|
body,
|
|
author: val(&comment, "author"),
|
|
score: format_num(score),
|
|
time: OffsetDateTime::from_unix_timestamp(unix_time).format("%b %d %Y %H:%M UTC"),
|
|
replies,
|
|
flair: Flair{
|
|
flair_parts: parse_rich_flair(val(&comment, "author_flair_type"), comment["data"]["author_flair_richtext"].as_array(), comment["data"]["author_flair_text"].as_str()),
|
|
background_color: val(&comment, "author_flair_background_color"),
|
|
foreground_color: val(&comment, "author_flair_text_color"),
|
|
},
|
|
});
|
|
}
|
|
|
|
comments
|
|
}
|