use crate::DbPool; use crate::api::ClientVersion; use crate::providers::Provider; use crate::util::cache::VideoCache; use crate::util::parse_abbreviated_number; use crate::util::time::parse_time_to_seconds; use crate::videos::{ServerOptions, VideoItem}; use crate::{status::*, util}; use async_trait::async_trait; use error_chain::error_chain; use htmlentity::entity::{ICodedDataTrait, decode}; use std::sync::{Arc, RwLock}; use std::thread; use std::vec; error_chain! { foreign_links { Io(std::io::Error); HttpRequest(wreq::Error); } } #[derive(Debug, Clone)] pub struct OmgxxxProvider { url: String, sites: Arc>>, networks: Arc>>, } impl OmgxxxProvider { pub fn new() -> Self { println!("new"); let provider = OmgxxxProvider { url: "https://www.omg.xxx".to_string(), sites: Arc::new(RwLock::new(vec![])), networks: Arc::new(RwLock::new(vec![])), }; // Kick off the background load but return immediately provider.spawn_initial_load(); provider } fn spawn_initial_load(&self) { println!("spawn_initial_load"); let url = self.url.clone(); // let sites = Arc::clone(&self.sites); let networks = Arc::clone(&self.networks); thread::spawn(move || { // Create a tiny runtime just for these async tasks let rt = tokio::runtime::Builder::new_current_thread() .enable_all() .build() .expect("build tokio runtime"); rt.block_on(async move { // If you have a streaming sites loader, call it here too // if let Err(e) = Self::load_sites_into(&url, sites).await { // eprintln!("load_sites_into failed: {e}"); // } if let Err(e) = Self::load_networks(&url, networks).await { eprintln!("load_networks failed: {e}"); } }); }); } /// === separate, self-contained logic === /// Replace the bodies with your real fetching/parsing code. // async fn load_sites(base_url: &str) -> Result> { // // Example stub: // // let html = reqwest::blocking::get(format!("{}/sites/", base_url))?.text()?; // // Ok(parse_sites_from_html(&html)) // Ok(vec![ // FilterOption { // id: "site-a".into(), // title: "Site A".into(), // }, // FilterOption { // id: "site-b".into(), // title: "Site B".into(), // }, // ]) // } async fn load_networks(base_url: &str, networks: Arc>>) -> Result<()> { println!("load_networks"); let mut requester = util::requester::Requester::new(); let text = requester.get(&base_url).await.unwrap(); let networks_div = text.split("class=\"sites__list\"").collect::>()[1] .split("") .collect::>()[0]; let mut networks_vec: Vec = vec![]; for network_element in networks_div.split("sites__item").collect::>()[1..].to_vec() { if network_element.contains("sites__all"){continue;} let network_url = network_element.split("href=\"").collect::>()[1] .split("\"") .collect::>()[0]; let network_id = network_url.split("/").collect::>()[4].to_string(); let network_name = network_element.split(">").collect::>()[1] .split("<") .collect::>()[0] .to_string(); networks_vec.push(FilterOption { id: network_id.clone(), title: network_name.clone(), }); Self::push_unique( &networks, FilterOption { id: network_id, title: network_name, }, ); } return Ok(()); } // Push one item with minimal lock time and dedup by id fn push_unique(target: &Arc>>, item: FilterOption) { if let Ok(mut vec) = target.write() { if !vec.iter().any(|x| x.id == item.id) { vec.push(item); // Optional: keep it sorted for nicer UX // vec.sort_by(|a,b| a.title.cmp(&b.title)); } } } fn build_channel(&self, clientversion: ClientVersion) -> Channel { let _ = clientversion; let sites: Vec = self .sites .read() .map(|g| g.clone()) // or: .map(|g| g.to_vec()) .unwrap_or_default(); // or: .unwrap_or_else(|_| Vec::new()) let networks: Vec = self .networks .read() .map(|g| g.clone()) // or: .map(|g| g.to_vec()) .unwrap_or_default(); // or: .unwrap_or_else(|_| Vec::new()) Channel { id: "omgxxx".to_string(), name: "OMG XXX".to_string(), description: "Free Porn Site".to_string(), premium: false, favicon: "https://www.google.com/s2/favicons?sz=64&domain=www.omg.xxx".to_string(), status: "active".to_string(), categories: vec![], options: vec![ ChannelOption { id: "sort".to_string(), title: "Sort".to_string(), description: "Sort the Videos".to_string(), systemImage: "list.number".to_string(), colorName: "blue".to_string(), options: vec![ FilterOption { id: "latest-updates".into(), title: "Latest".into(), }, FilterOption { id: "most-popular".into(), title: "Most Viewed".into(), }, FilterOption { id: "top-rated".into(), title: "Top Rated".into(), }, ], multiSelect: false, }, ChannelOption { id: "sites".to_string(), title: "Sites".to_string(), description: "Sort the Videos".to_string(), systemImage: "list.bullet.indent".to_string(), colorName: "green".to_string(), options: sites, multiSelect: false, }, ChannelOption { id: "networks".to_string(), title: "Networks".to_string(), description: "Sort the Videos".to_string(), systemImage: "list.dash".to_string(), colorName: "purple".to_string(), options: networks, multiSelect: false, }, ], nsfw: true, cacheDuration: None, } } async fn get( &self, cache: VideoCache, page: u8, sort: &str, options: ServerOptions, ) -> Result> { let sort_string = match sort { "top-rated" => "top-rated", "most-popular" => "most-popular", _ => "latest-updates", }; let video_url = format!("{}/{}/{}/", self.url, sort_string, page); let old_items = match cache.get(&video_url) { Some((time, items)) => { if time.elapsed().unwrap_or_default().as_secs() < 60 * 5 { println!("Cache hit for URL: {}", video_url); return Ok(items.clone()); } else { items.clone() } } None => { vec![] } }; let mut requester = options.requester.clone().unwrap(); let text = requester.get(&video_url).await.unwrap(); let video_items: Vec = self.get_video_items_from_html(text.clone()); if !video_items.is_empty() { cache.remove(&video_url); cache.insert(video_url.clone(), video_items.clone()); } else { return Ok(old_items); } Ok(video_items) } async fn query( &self, cache: VideoCache, page: u8, query: &str, options: ServerOptions, ) -> Result> { let mut search_type = "search"; if query.starts_with("@models:") { search_type = "models"; } let video_url = format!( "{}/{}/{}/{}/", self.url, search_type, query .to_lowercase() .trim() .replace(" ", "-") .replace("@models:", ""), page ); // Check our Video Cache. If the result is younger than 1 hour, we return it. let old_items = match cache.get(&video_url) { Some((time, items)) => { if time.elapsed().unwrap_or_default().as_secs() < 60 * 5 { return Ok(items.clone()); } else { let _ = cache.check().await; return Ok(items.clone()); } } None => { vec![] } }; let mut requester = options.requester.clone().unwrap(); let text = requester.get(&video_url).await.unwrap(); let video_items: Vec = self.get_video_items_from_html(text.clone()); if !video_items.is_empty() { cache.remove(&video_url); cache.insert(video_url.clone(), video_items.clone()); } else { return Ok(old_items); } Ok(video_items) } fn get_video_items_from_html(&self, html: String) -> Vec { if html.is_empty() { println!("HTML is empty"); return vec![]; } let mut items: Vec = Vec::new(); let raw_videos = html.split("videos_list_pagination").collect::>()[0] .split(" class=\"pagination\" ") .collect::>()[0] .split("class=\"list-videos\"") .collect::>()[1] .split("class=\"item\"") .collect::>()[1..] .to_vec(); for video_segment in &raw_videos { // let vid = video_segment.split("\n").collect::>(); // for (index, line) in vid.iter().enumerate() { // println!("Line {}: {}", index, line); // } let video_url: String = video_segment.split(">()[1] .split("\"") .collect::>()[0] .to_string(); let mut title = video_segment.split(" title=\"").collect::>()[1] .split("\"") .collect::>()[0] .to_string(); // html decode title = decode(title.as_bytes()).to_string().unwrap_or(title); let id = video_url.split("/").collect::>()[4].to_string(); let thumb = match video_segment.split("img loading").collect::>()[1] .contains("data-src=\"") { true => video_segment.split("img loading").collect::>()[1] .split("data-src=\"") .collect::>()[1] .split("\"") .collect::>()[0] .to_string(), false => video_segment.split("img loading").collect::>()[1] .split("data-original=\"") .collect::>()[1] .split("\"") .collect::>()[0] .to_string(), }; let raw_duration = video_segment .split("") .collect::>()[1] .split("<") .collect::>()[0] .split(" ") .collect::>() .last() .unwrap_or(&"") .to_string(); let duration = parse_time_to_seconds(raw_duration.as_str()).unwrap_or(0) as u32; let views = parse_abbreviated_number( video_segment .split("
") .collect::>()[1] .split("<") .collect::>()[0] .to_string() .as_str(), ) .unwrap_or(0) as u32; let preview = video_segment .split("data-preview=\"") .collect::>()[1] .split("\"") .collect::>()[0] .to_string(); let tags = match video_segment.contains("class=\"models\">") { true => video_segment .split("class=\"models\">") .collect::>()[1] .split("
") .collect::>()[0] .split("href=\"") .collect::>()[1..] .into_iter() .map(|s| { format!( "@models:{}", s.split("/").collect::>()[4].to_string() ) }) .collect::>() .to_vec(), false => vec![], }; let video_item = VideoItem::new( id, title, video_url.to_string(), "omgxxx".to_string(), thumb, duration, ) .views(views) .preview(preview) .tags(tags); items.push(video_item); } return items; } } #[async_trait] impl Provider for OmgxxxProvider { async fn get_videos( &self, cache: VideoCache, pool: DbPool, sort: String, query: Option, page: String, per_page: String, options: ServerOptions, ) -> Vec { let _ = per_page; let _ = pool; let videos: std::result::Result, Error> = match query { Some(q) => { self.query(cache, page.parse::().unwrap_or(1), &q, options) .await } None => { self.get(cache, page.parse::().unwrap_or(1), &sort, options) .await } }; match videos { Ok(v) => v, Err(e) => { println!("Error fetching videos: {}", e); vec![] } } } fn get_channel(&self, clientversion: ClientVersion) -> crate::status::Channel { println!( "Getting channel for omgxxx with client version: {:?}", clientversion ); self.build_channel(clientversion) } }