From 967d1e8143ab03c0d45c0575a065930768f78af3 Mon Sep 17 00:00:00 2001 From: Simon Date: Tue, 10 Mar 2026 17:48:45 +0000 Subject: [PATCH] removed spankbang from archive --- archive/spankbang.rs | 380 ------------------------------------------- 1 file changed, 380 deletions(-) delete mode 100644 archive/spankbang.rs diff --git a/archive/spankbang.rs b/archive/spankbang.rs deleted file mode 100644 index 393276c..0000000 --- a/archive/spankbang.rs +++ /dev/null @@ -1,380 +0,0 @@ -use std::vec; -use std::env; -use error_chain::error_chain; -use futures::future::join_all; -use htmlentity::entity::{decode, ICodedDataTrait}; -use crate::db; -use crate::providers::Provider; -use crate::util::cache::VideoCache; -use crate::util::flaresolverr::{FlareSolverrRequest, Flaresolverr}; -use crate::videos::ServerOptions; -use crate::videos::{VideoItem}; -use crate::DbPool; -use std::collections::HashMap; -use wreq::Client; -use wreq_util::Emulation; - -error_chain! { - foreign_links { - Io(std::io::Error); - HttpRequest(wreq::Error); - } -} - -#[derive(Debug, Clone)] -pub struct SpankbangProvider { - url: String, -} -impl SpankbangProvider { - pub fn new() -> Self { - SpankbangProvider { - url: "https://spankbang.com/".to_string() - } - } - async fn get(&self, cache:VideoCache, pool: DbPool, page: u8, sort: String) -> Result> { - - let url = format!("{}{}/{}/", self.url, sort, page); - - let old_items = match cache.get(&url) { - Some((time, items)) => { - if time.elapsed().unwrap_or_default().as_secs() < 60 * 60 { - // println!("Cache hit for URL: {}", url); - return Ok(items.clone()); - } - else{ - items.clone() - } - } - None => { - vec![] - } - }; - - - let client = Client::builder() - .emulation(Emulation::Firefox136) - .cert_verification(false) - .build()?; - - let response = client.get(url.clone()).send().await?; - let mut cookies_string = String::new(); - if let Some(_) = response.headers().get_all("set-cookie").iter().next() { - for _ in response.headers().get_all("set-cookie").iter() { - let mut cookies_map = HashMap::new(); - for value in response.headers().get_all("set-cookie").iter() { - if let Ok(cookie_str) = value.to_str() { - if let Some((k, v)) = cookie_str.split_once('=') { - let key = k.trim(); - let val = v.split(';').next().unwrap_or("").trim(); - cookies_map.insert(key.to_string(), val.to_string()); - } - } - } - cookies_string = cookies_map - .iter() - .map(|(k, v)| format!("{}={}", k, v)) - .collect::>() - .join("; "); - } - } - if response.status().is_success() { - let text = response.text().await?; - let video_items: Vec = self.get_video_items_from_html(text.clone(), &client, cookies_string, pool.clone()).await; - if !video_items.is_empty() { - cache.remove(&url); - cache.insert(url.clone(), video_items.clone()); - } else{ - return Ok(old_items); - } - Ok(video_items) - } else { - let flare_url = env::var("FLARE_URL").expect("FLARE_URL not set"); - let flare = Flaresolverr::new(flare_url); - let result = flare - .solve(FlareSolverrRequest { - cmd: "request.get".to_string(), - url: url.clone(), - maxTimeout: 60000, - }) - .await; - let video_items = match result { - Ok(res) => { - // println!("FlareSolverr response: {}", res); - self.get_video_items_from_html(res.solution.response, &client,String::new(), pool.clone()).await - } - Err(e) => { - println!("Error solving FlareSolverr: {}", e); - return Err("Failed to solve FlareSolverr".into()); - } - }; - if !video_items.is_empty() { - cache.remove(&url); - cache.insert(url.clone(), video_items.clone()); - } else { - return Ok(old_items); - } - Ok(video_items) - } - } - - async fn query(&self, cache: VideoCache, pool: DbPool, page: u8, query: &str) -> Result> { - let url = format!("{}s/{}/{}/", self.url, query.replace(" ", "+"), page); - - let old_items = match cache.get(&url) { - Some((time, items)) => { - if time.elapsed().unwrap_or_default().as_secs() < 60 * 60 { - // println!("Cache hit for URL: {}", url); - return Ok(items.clone()); - } - else{ - items.clone() - } - } - None => { - vec![] - } - }; - - - let client = Client::builder() - .emulation(Emulation::Firefox136) - .cert_verification(false) - .build()?; - - let response = client.get(url.clone()).send().await?; - let mut cookies_string = String::new(); - if let Some(_) = response.headers().get_all("set-cookie").iter().next() { - for _ in response.headers().get_all("set-cookie").iter() { - let mut cookies_map = HashMap::new(); - for value in response.headers().get_all("set-cookie").iter() { - if let Ok(cookie_str) = value.to_str() { - if let Some((k, v)) = cookie_str.split_once('=') { - let key = k.trim(); - let val = v.split(';').next().unwrap_or("").trim(); - cookies_map.insert(key.to_string(), val.to_string()); - } - } - } - cookies_string = cookies_map - .iter() - .map(|(k, v)| format!("{}={}", k, v)) - .collect::>() - .join("; "); - } - } - if response.status().is_success() { - let text = response.text().await?; - let video_items: Vec = self.get_video_items_from_html(text.clone(), &client, cookies_string, pool.clone()).await; - if !video_items.is_empty() { - cache.remove(&url); - cache.insert(url.clone(), video_items.clone()); - } else{ - return Ok(old_items); - } - Ok(video_items) - } else { - let flare_url = env::var("FLARE_URL").expect("FLARE_URL not set"); - let flare = Flaresolverr::new(flare_url); - let result = flare - .solve(FlareSolverrRequest { - cmd: "request.get".to_string(), - url: url.clone(), - maxTimeout: 60000, - }) - .await; - let video_items = match result { - Ok(res) => { - // println!("FlareSolverr response: {}", res); - self.get_video_items_from_html(res.solution.response, &client, String::new(), pool.clone()).await - } - Err(e) => { - println!("Error solving FlareSolverr: {}", e); - return Err("Failed to solve FlareSolverr".into()); - } - }; - if !video_items.is_empty() { - cache.remove(&url); - cache.insert(url.clone(), video_items.clone()); - } else { - return Ok(old_items); - } - Ok(video_items) - } - } - - async fn get_video_url(&self, url:String, client:&Client, cookies: String, pool: DbPool) -> Result { - - let mut conn = pool.get().expect("couldn't get db connection from pool"); - let db_result = db::get_video(&mut conn,url.clone()); - drop(conn); - match db_result { - Ok(Some(video_url)) => { - return Ok(video_url); - } - Ok(None) => (), - Err(e) => { - println!("Error fetching video from database: {}", e); - // return Err(format!("Error fetching video from database: {}", e).into()); - } - } - let response = client.get(url.clone()).header("Cookie", cookies.clone()).send().await?; - - let mut response = response; - while response.status().as_u16() == 429 { - // println!("Received 429 Too Many Requests. Waiting 10 seconds before retrying..."); - ntex::time::sleep(ntex::time::Seconds(60)).await; - response = client.get(url.clone()).header("Cookie", cookies.clone()).send().await?; - } - - if response.status().is_success() { - let text = response.text().await?; - let lines = text.split("\n").collect::>(); - let url_line = lines.iter() - .find(|s| s.trim_start().starts_with(">()[1].split("\"").collect::>()[0].to_string(); - let mut conn = pool.get().expect("couldn't get db connection from pool"); - let _ = db::insert_video(&mut conn, &url, &new_url); - drop(conn); - return Ok(new_url) - } - Err(Error::from("Failed to get video URL")) - } - - async fn parse_video_item( - &self, - mut html: String, - client: &Client, - cookies: String, - pool: DbPool - ) -> Result { - if html.contains("") { - html = html.split("").collect::>()[0].to_string(); - } - - let vid = html.split("\n").collect::>(); - if vid.len() > 200 { - return Err("Video item has too many lines".into()); - } - // for (index ,line) in vid.iter().enumerate() { - // println!("Line {}: {}", index, line); - // } - let title_line = vid.iter() - .find(|s| s.trim_start().starts_with(">()[1].split("\"").collect::>()[0].to_string(); - title = decode(title.as_bytes()).to_string().unwrap_or(title); - - let thumb_line = vid.iter() - .find(|s| s.trim_start().starts_with("data-src=") && s.contains(".jpg\"")) - .unwrap_or(&""); - let thumb = thumb_line.split("data-src=\"").collect::>()[1].split("\"").collect::>()[0].to_string(); - // let preview_line = vid.iter() - // .find(|s: &&&str| s.trim_start().starts_with(">()[1].split("\"").collect::>()[0].to_string(); - // } - // else{ - // preview = preview_line.split("data-src=\"").collect::>()[1].split("\"").collect::>()[0].to_string(); - // } - let duration_str = vid[64].split("m").collect::>()[0]; - let duration: u32 = duration_str.parse::().unwrap_or(0) * 60; - // let view_and_rating_str: Vec<&str> = vid.iter().copied().filter(|s| s.contains("")).collect(); - // let views_str = view_and_rating_str[0].split(">").collect::>()[1].split("K<").collect::>()[0]; - // let views = (views_str.parse::().unwrap_or(0.0) * 1000.0) as u32; - // let rate_str = view_and_rating_str[1].split(">").collect::>()[1].split("%<").collect::>()[0]; - // let rating = rate_str.parse::().unwrap_or(0.0); - let url_part = vid.iter().find(|s| s.contains(">()[1].split("\"").collect::>()[0]; - let url = match self.get_video_url(self.url.clone() + url_part, client, cookies, pool).await { - Ok(video_url) => video_url, - Err(e) => { - print!("Error fetching video URL: {}", e); - return Err("Failed to get video URL".into()); - } - }; - let id = url_part.split("/").collect::>()[0].to_string(); - - // let quality_str = match vid[25].contains("<"){ - // true => vid[25].split(">").collect::>()[1].split("<").collect::>()[0], - // false => "SD", - // }; - // let quality = match quality_str{ - // "HD" => "1080", - // "4k" => "2160", - // "SD" => "720", - // _ => "1080", - // }; - - let video_item = VideoItem::new(id, title, url.clone().to_string(), "spankbang".to_string(), thumb, duration) - // .views(views) - // .rating(rating) - // .formats(vec![format]) - // .preview(preview) - ; - Ok(video_item) - } - - async fn get_video_items_from_html(&self, html: String, client: &Client, cookies:String, pool: DbPool) -> Vec { - if html.is_empty() { - println!("HTML is empty"); - return vec![]; - } - let items: Vec = Vec::new(); - let split_html = html.split("\"video-list").collect::>(); - if split_html.len() < 2 { - println!("Could not find video-list in HTML"); - return items; - } - let video_listing_content = format!("{}{}", split_html[1], split_html.get(2).unwrap_or(&"")); - let raw_videos_vec = video_listing_content - .split("data-testid=\"video-item\"") - .collect::>(); - if raw_videos_vec.len() < 2 { - println!("Could not find video-item in HTML"); - return items; - } - let raw_videos = raw_videos_vec[1..].to_vec(); - let futures = raw_videos.into_iter().map(|el| self.parse_video_item(el.to_string(), client, cookies.clone(), pool.clone())); - let results: Vec> = join_all(futures).await; - let video_items: Vec = results - .into_iter() - .filter_map(Result::ok) - .collect(); - return video_items; - } - -} - -impl Provider for SpankbangProvider { - async fn get_videos( - &self, - cache: VideoCache, - pool: DbPool, - mut sort: String, - query: Option, - page: String, - per_page: String, - options: ServerOptions, - ) -> Vec { - let _ = options; - let _ = per_page; - let _ = pool; - - if sort == "date"{ - sort = "trending_videos".to_string(); - } - let videos: std::result::Result, Error> = match query { - Some(q) => self.query(cache, pool, page.parse::().unwrap_or(1), &q).await, - None => self.get(cache, pool, page.parse::().unwrap_or(1), sort).await, - }; - match videos { - Ok(v) => v, - Err(e) => { - println!("Error fetching videos: {}", e); - vec![] - } - } - } -}