//! Subscription and filter parsing use crate::error::Result; use crate::event::Event; use serde::de::Unexpected; use serde::ser::SerializeMap; use serde::{Deserialize, Deserializer, Serialize, Serializer}; use serde_json::Value; use std::collections::HashMap; use std::collections::HashSet; /// Subscription identifier and set of request filters #[derive(Serialize, PartialEq, Eq, Debug, Clone)] pub struct Subscription { pub id: String, pub filters: Vec, } /// Filter for requests /// /// Corresponds to client-provided subscription request elements. Any /// element can be present if it should be used in filtering, or /// absent ([`None`]) if it should be ignored. #[derive(PartialEq, Eq, Debug, Clone)] pub struct ReqFilter { /// Event hashes pub ids: Option>, /// Event kinds pub kinds: Option>, /// Events published after this time pub since: Option, /// Events published before this time pub until: Option, /// List of author public keys pub authors: Option>, /// Limit number of results pub limit: Option, /// Set of tags pub tags: Option>>, /// Force no matches due to malformed data // we can't represent it in the req filter, so we don't want to // erroneously match. This basically indicates the req tried to // do something invalid. pub force_no_match: bool, } impl Serialize for ReqFilter { fn serialize(&self, serializer: S) -> Result where S: Serializer, { let mut map = serializer.serialize_map(None)?; if let Some(ids) = &self.ids { map.serialize_entry("ids", &ids)?; } if let Some(kinds) = &self.kinds { map.serialize_entry("kinds", &kinds)?; } if let Some(until) = &self.until { map.serialize_entry("until", until)?; } if let Some(since) = &self.since { map.serialize_entry("since", since)?; } if let Some(limit) = &self.limit { map.serialize_entry("limit", limit)?; } if let Some(authors) = &self.authors { map.serialize_entry("authors", &authors)?; } // serialize tags if let Some(tags) = &self.tags { for (k, v) in tags { let vals: Vec<&String> = v.iter().collect(); map.serialize_entry(&format!("#{k}"), &vals)?; } } map.end() } } impl<'de> Deserialize<'de> for ReqFilter { fn deserialize(deserializer: D) -> Result where D: Deserializer<'de>, { let received: Value = Deserialize::deserialize(deserializer)?; let filter = received.as_object().ok_or_else(|| { serde::de::Error::invalid_type( Unexpected::Other("reqfilter is not an object"), &"a json object", ) })?; let mut rf = ReqFilter { ids: None, kinds: None, since: None, until: None, authors: None, limit: None, tags: None, force_no_match: false, }; let empty_string = "".into(); let mut ts = None; // iterate through each key, and assign values that exist for (key, val) in filter { // ids if key == "ids" { let raw_ids: Option> = Deserialize::deserialize(val).ok(); if let Some(a) = raw_ids.as_ref() { if a.contains(&empty_string) { return Err(serde::de::Error::invalid_type( Unexpected::Other("prefix matches must not be empty strings"), &"a json object", )); } } rf.ids = raw_ids; } else if key == "kinds" { rf.kinds = Deserialize::deserialize(val).ok(); } else if key == "since" { rf.since = Deserialize::deserialize(val).ok(); } else if key == "until" { rf.until = Deserialize::deserialize(val).ok(); } else if key == "limit" { rf.limit = Deserialize::deserialize(val).ok(); } else if key == "authors" { let raw_authors: Option> = Deserialize::deserialize(val).ok(); if let Some(a) = raw_authors.as_ref() { if a.contains(&empty_string) { return Err(serde::de::Error::invalid_type( Unexpected::Other("prefix matches must not be empty strings"), &"a json object", )); } } rf.authors = raw_authors; } else if key.starts_with('#') && key.len() > 1 && val.is_array() { if let Some(tag_search) = tag_search_char_from_filter(key) { if ts.is_none() { // Initialize the tag if necessary ts = Some(HashMap::new()); } if let Some(m) = ts.as_mut() { let tag_vals: Option> = Deserialize::deserialize(val).ok(); if let Some(v) = tag_vals { let hs = v.into_iter().collect::>(); m.insert(tag_search.to_owned(), hs); } }; } else { // tag search that is multi-character, don't add to subscription rf.force_no_match = true; continue; } } } rf.tags = ts; Ok(rf) } } /// Attempt to form a single-char identifier from a tag search filter fn tag_search_char_from_filter(tagname: &str) -> Option { let tagname_nohash = &tagname[1..]; // We return the tag character if and only if the tagname consists // of a single char. let mut tagnamechars = tagname_nohash.chars(); let firstchar = tagnamechars.next(); match firstchar { Some(_) => { // check second char if tagnamechars.next().is_none() { firstchar } else { None } } None => None, } } impl<'de> Deserialize<'de> for Subscription { /// Custom deserializer for subscriptions, which have a more /// complex structure than the other message types. fn deserialize(deserializer: D) -> Result where D: Deserializer<'de>, { let mut v: Value = Deserialize::deserialize(deserializer)?; // this should be a 3-or-more element array. // verify the first element is a String, REQ // get the subscription from the second element. // convert each of the remaining objects into filters // check for array let va = v .as_array_mut() .ok_or_else(|| serde::de::Error::custom("not array"))?; // check length if va.len() < 3 { return Err(serde::de::Error::custom("not enough fields")); } let mut i = va.iter_mut(); // get command ("REQ") and ensure it is a string let req_cmd_str: serde_json::Value = i.next().unwrap().take(); let req = req_cmd_str .as_str() .ok_or_else(|| serde::de::Error::custom("first element of request was not a string"))?; if req != "REQ" { return Err(serde::de::Error::custom("missing REQ command")); } // ensure sub id is a string let sub_id_str: serde_json::Value = i.next().unwrap().take(); let sub_id = sub_id_str .as_str() .ok_or_else(|| serde::de::Error::custom("missing subscription id"))?; let mut filters = vec![]; for fv in i { let f: ReqFilter = serde_json::from_value(fv.take()) .map_err(|_| serde::de::Error::custom("could not parse filter"))?; // create indexes filters.push(f); } filters.dedup(); Ok(Subscription { id: sub_id.to_owned(), filters, }) } } impl Subscription { /// Get a copy of the subscription identifier. #[must_use] pub fn get_id(&self) -> String { self.id.clone() } /// Determine if any filter is requesting historical (database) /// queries. If every filter has limit:0, we do not need to query the DB. #[must_use] pub fn needs_historical_events(&self) -> bool { self.filters.iter().any(|f| f.limit != Some(0)) } /// Determine if this subscription matches a given [`Event`]. Any /// individual filter match is sufficient. #[must_use] pub fn interested_in_event(&self, event: &Event) -> bool { for f in &self.filters { if f.interested_in_event(event) { return true; } } false } /// Is this subscription defined as a scraper query pub fn is_scraper(&self) -> bool { for f in &self.filters { let mut precision = 0; if f.ids.is_some() { precision += 2; } if f.authors.is_some() { precision += 1; } if f.kinds.is_some() { precision += 1; } if f.tags.is_some() { precision += 1; } if precision < 2 { return true; } } false } } fn prefix_match(prefixes: &[String], target: &str) -> bool { for prefix in prefixes { if target.starts_with(prefix) { return true; } } // none matched false } impl ReqFilter { fn ids_match(&self, event: &Event) -> bool { self.ids .as_ref() .map_or(true, |vs| prefix_match(vs, &event.id)) } fn authors_match(&self, event: &Event) -> bool { self.authors .as_ref() .map_or(true, |vs| prefix_match(vs, &event.pubkey)) } fn delegated_authors_match(&self, event: &Event) -> bool { if let Some(delegated_pubkey) = &event.delegated_by { self.authors .as_ref() .map_or(true, |vs| prefix_match(vs, delegated_pubkey)) } else { false } } fn tag_match(&self, event: &Event) -> bool { // get the hashset from the filter. if let Some(map) = &self.tags { for (key, val) in map.iter() { let tag_match = event.generic_tag_val_intersect(*key, val); // if there is no match for this tag, the match fails. if !tag_match { return false; } // if there was a match, we move on to the next one. } } // if the tag map is empty, the match succeeds (there was no filter) true } /// Check if this filter either matches, or does not care about the kind. fn kind_match(&self, kind: u64) -> bool { self.kinds.as_ref().map_or(true, |ks| ks.contains(&kind)) } /// Determine if all populated fields in this filter match the provided event. #[must_use] pub fn interested_in_event(&self, event: &Event) -> bool { // self.id.as_ref().map(|v| v == &event.id).unwrap_or(true) self.ids_match(event) && self.since.map_or(true, |t| event.created_at >= t) && self.until.map_or(true, |t| event.created_at <= t) && self.kind_match(event.kind) && (self.authors_match(event) || self.delegated_authors_match(event)) && self.tag_match(event) && !self.force_no_match } } #[cfg(test)] mod tests { use super::*; #[test] fn empty_request_parse() -> Result<()> { let raw_json = "[\"REQ\",\"some-id\",{}]"; let s: Subscription = serde_json::from_str(raw_json)?; assert_eq!(s.id, "some-id"); assert_eq!(s.filters.len(), 1); assert_eq!(s.filters.first().unwrap().authors, None); Ok(()) } #[test] fn incorrect_header() { let raw_json = "[\"REQUEST\",\"some-id\",\"{}\"]"; assert!(serde_json::from_str::(raw_json).is_err()); } #[test] fn req_missing_filters() { let raw_json = "[\"REQ\",\"some-id\"]"; assert!(serde_json::from_str::(raw_json).is_err()); } #[test] fn req_empty_authors_prefix() { let raw_json = "[\"REQ\",\"some-id\",{\"authors\": [\"\"]}]"; assert!(serde_json::from_str::(raw_json).is_err()); } #[test] fn req_empty_ids_prefix() { let raw_json = "[\"REQ\",\"some-id\",{\"ids\": [\"\"]}]"; assert!(serde_json::from_str::(raw_json).is_err()); } #[test] fn req_empty_ids_prefix_mixed() { let raw_json = "[\"REQ\",\"some-id\",{\"ids\": [\"\",\"aaa\"]}]"; assert!(serde_json::from_str::(raw_json).is_err()); } #[test] fn legacy_filter() { // legacy field in filter let raw_json = "[\"REQ\",\"some-id\",{\"kind\": 3}]"; assert!(serde_json::from_str::(raw_json).is_ok()); } #[test] fn dupe_filter() -> Result<()> { let raw_json = r#"["REQ","some-id",{"kinds": [1984]}, {"kinds": [1984]}]"#; let s: Subscription = serde_json::from_str(raw_json)?; assert_eq!(s.filters.len(), 1); Ok(()) } #[test] fn dupe_filter_many() -> Result<()> { // duplicate filters in different order let raw_json = r#"["REQ","some-id",{"kinds":[1984]},{"kinds":[1984]},{"kinds":[1984]},{"kinds":[1984]}]"#; let s: Subscription = serde_json::from_str(raw_json)?; assert_eq!(s.filters.len(), 1); Ok(()) } #[test] fn author_filter() -> Result<()> { let raw_json = r#"["REQ","some-id",{"authors": ["test-author-id"]}]"#; let s: Subscription = serde_json::from_str(raw_json)?; assert_eq!(s.id, "some-id"); assert_eq!(s.filters.len(), 1); let first_filter = s.filters.first().unwrap(); assert_eq!( first_filter.authors, Some(vec!("test-author-id".to_owned())) ); Ok(()) } #[test] fn interest_author_prefix_match() -> Result<()> { // subscription with a filter for ID let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"authors": ["abc"]}]"#)?; let e = Event { id: "foo".to_owned(), pubkey: "abcd".to_owned(), delegated_by: None, created_at: 0, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s.interested_in_event(&e)); Ok(()) } #[test] fn interest_id_prefix_match() -> Result<()> { // subscription with a filter for ID let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"ids": ["abc"]}]"#)?; let e = Event { id: "abcd".to_owned(), pubkey: "".to_owned(), delegated_by: None, created_at: 0, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s.interested_in_event(&e)); Ok(()) } #[test] fn interest_id_nomatch() -> Result<()> { // subscription with a filter for ID let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"ids": ["xyz"]}]"#)?; let e = Event { id: "abcde".to_owned(), pubkey: "".to_owned(), delegated_by: None, created_at: 0, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(!s.interested_in_event(&e)); Ok(()) } #[test] fn interest_until() -> Result<()> { // subscription with a filter for ID and time let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"ids": ["abc"], "until": 1000}]"#)?; let e = Event { id: "abc".to_owned(), pubkey: "".to_owned(), delegated_by: None, created_at: 50, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s.interested_in_event(&e)); Ok(()) } #[test] fn interest_range() -> Result<()> { // subscription with a filter for ID and time let s_in: Subscription = serde_json::from_str(r#"["REQ","xyz",{"ids": ["abc"], "since": 100, "until": 200}]"#)?; let s_before: Subscription = serde_json::from_str(r#"["REQ","xyz",{"ids": ["abc"], "since": 100, "until": 140}]"#)?; let s_after: Subscription = serde_json::from_str(r#"["REQ","xyz",{"ids": ["abc"], "since": 160, "until": 200}]"#)?; let e = Event { id: "abc".to_owned(), pubkey: "".to_owned(), delegated_by: None, created_at: 150, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s_in.interested_in_event(&e)); assert!(!s_before.interested_in_event(&e)); assert!(!s_after.interested_in_event(&e)); Ok(()) } #[test] fn interest_time_and_id() -> Result<()> { // subscription with a filter for ID and time let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"ids": ["abc"], "since": 1000}]"#)?; let e = Event { id: "abc".to_owned(), pubkey: "".to_owned(), delegated_by: None, created_at: 50, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(!s.interested_in_event(&e)); Ok(()) } #[test] fn interest_time_and_id2() -> Result<()> { // subscription with a filter for ID and time let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"id":"abc", "since": 1000}]"#)?; let e = Event { id: "abc".to_owned(), pubkey: "".to_owned(), delegated_by: None, created_at: 1001, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s.interested_in_event(&e)); Ok(()) } #[test] fn interest_id() -> Result<()> { // subscription with a filter for ID let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"id":"abc"}]"#)?; let e = Event { id: "abc".to_owned(), pubkey: "".to_owned(), delegated_by: None, created_at: 0, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s.interested_in_event(&e)); Ok(()) } #[test] fn authors_single() -> Result<()> { // subscription with a filter for ID let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"authors":["abc"]}]"#)?; let e = Event { id: "123".to_owned(), pubkey: "abc".to_owned(), delegated_by: None, created_at: 0, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s.interested_in_event(&e)); Ok(()) } #[test] fn authors_multi_pubkey() -> Result<()> { // check for any of a set of authors, against the pubkey let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"authors":["abc", "bcd"]}]"#)?; let e = Event { id: "123".to_owned(), pubkey: "bcd".to_owned(), delegated_by: None, created_at: 0, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(s.interested_in_event(&e)); Ok(()) } #[test] fn authors_multi_no_match() -> Result<()> { // check for any of a set of authors, against the pubkey let s: Subscription = serde_json::from_str(r#"["REQ","xyz",{"authors":["abc", "bcd"]}]"#)?; let e = Event { id: "123".to_owned(), pubkey: "xyz".to_owned(), delegated_by: None, created_at: 0, kind: 0, tags: Vec::new(), content: "".to_owned(), sig: "".to_owned(), tagidx: None, }; assert!(!s.interested_in_event(&e)); Ok(()) } #[test] fn serialize_filter() -> Result<()> { let s: Subscription = serde_json::from_str( r##"["REQ","xyz",{"authors":["abc", "bcd"], "since": 10, "until": 20, "limit":100, "#e": ["foo", "bar"], "#d": ["test"]}]"##, )?; let f = s.filters.first(); let serialized = serde_json::to_string(&f)?; let serialized_wrapped = format!(r##"["REQ", "xyz",{}]"##, serialized); let parsed: Subscription = serde_json::from_str(&serialized_wrapped)?; let parsed_filter = parsed.filters.first(); if let Some(pf) = parsed_filter { assert_eq!(pf.since, Some(10)); assert_eq!(pf.until, Some(20)); assert_eq!(pf.limit, Some(100)); } else { assert!(false, "filter could not be parsed"); } Ok(()) } #[test] fn is_scraper() -> Result<()> { assert!(serde_json::from_str::(r#"["REQ","some-id",{"kinds": [1984],"since": 123,"limit":1}]"#)?.is_scraper()); assert!(serde_json::from_str::(r#"["REQ","some-id",{"kinds": [1984]},{"kinds": [1984],"authors":["aaaa"]}]"#)?.is_scraper()); assert!(!serde_json::from_str::(r#"["REQ","some-id",{"kinds": [1984],"authors":["aaaa"]}]"#)?.is_scraper()); assert!(!serde_json::from_str::(r#"["REQ","some-id",{"ids": ["aaaa"]}]"#)?.is_scraper()); assert!(!serde_json::from_str::(r##"["REQ","some-id",{"#p": ["aaaa"],"kinds":[1,4]}]"##)?.is_scraper()); Ok(()) } }