mirror of
https://github.com/scsibug/nostr-rs-relay.git
synced 2024-11-22 00:59:07 -05:00
feat: nip-50 support (postgres)
This commit is contained in:
parent
c5fb16cd98
commit
e5da37b4f1
|
@ -68,6 +68,10 @@ impl From<Settings> for RelayInfo {
|
|||
fn from(c: Settings) -> Self {
|
||||
let mut supported_nips = vec![1, 2, 9, 11, 12, 15, 16, 20, 22, 33, 40];
|
||||
|
||||
if c.database.engine == "postgres" {
|
||||
supported_nips.push(50);
|
||||
}
|
||||
|
||||
if c.authorization.nip42_auth {
|
||||
supported_nips.push(42);
|
||||
supported_nips.sort();
|
||||
|
|
|
@ -59,7 +59,7 @@ async fn cleanup_expired(conn: PostgresPool, frequency: Duration) -> Result<()>
|
|||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
});
|
||||
Ok(())
|
||||
|
@ -157,7 +157,7 @@ ON CONFLICT (id) DO NOTHING"#,
|
|||
.and_then(|x| Utc.timestamp_opt(x as i64, 0).latest()),
|
||||
)
|
||||
.bind(e.kind as i64)
|
||||
.bind(event_str.into_bytes())
|
||||
.bind(event_str)
|
||||
.bind(delegator_blob)
|
||||
.execute(&mut tx)
|
||||
.await?
|
||||
|
@ -393,7 +393,7 @@ ON CONFLICT (id) DO NOTHING"#,
|
|||
}
|
||||
|
||||
row_count += 1;
|
||||
let event_json: Vec<u8> = row.unwrap().get(0);
|
||||
let event_json: String = row.unwrap().get(0);
|
||||
loop {
|
||||
if query_tx.capacity() != 0 {
|
||||
// we have capacity to add another item
|
||||
|
@ -421,7 +421,7 @@ ON CONFLICT (id) DO NOTHING"#,
|
|||
query_tx
|
||||
.send(QueryResult {
|
||||
sub_id: sub.get_id(),
|
||||
event: String::from_utf8(event_json).unwrap(),
|
||||
event: event_json,
|
||||
})
|
||||
.await
|
||||
.ok();
|
||||
|
@ -863,6 +863,16 @@ fn query_from_filter(f: &ReqFilter) -> Option<QueryBuilder<Postgres>> {
|
|||
.push_bind(Utc.timestamp_opt(f.until.unwrap() as i64, 0).unwrap());
|
||||
}
|
||||
|
||||
if let Some(search) = &f.search {
|
||||
if push_and {
|
||||
query.push(" AND ");
|
||||
}
|
||||
push_and = true;
|
||||
query.push("e.ts_content @@ websearch_to_tsquery(")
|
||||
.push_bind(search.clone())
|
||||
.push(")");
|
||||
}
|
||||
|
||||
// never display hidden events
|
||||
if push_and {
|
||||
query.push(" AND e.hidden != 1::bit(1)");
|
||||
|
|
|
@ -37,6 +37,7 @@ pub async fn run_migrations(db: &PostgresPool) -> crate::error::Result<usize> {
|
|||
run_migration(m003::migration(), db).await;
|
||||
run_migration(m004::migration(), db).await;
|
||||
run_migration(m005::migration(), db).await;
|
||||
run_migration(m006::migration(), db).await;
|
||||
Ok(current_version(db).await as usize)
|
||||
}
|
||||
|
||||
|
@ -286,7 +287,7 @@ mod m005 {
|
|||
|
||||
pub fn migration() -> impl Migration {
|
||||
SimpleSqlMigration {
|
||||
serial_number: VERSION,
|
||||
serial_number: crate::repo::postgres_migration::m005::VERSION,
|
||||
sql: vec![
|
||||
r#"
|
||||
-- Create account table
|
||||
|
@ -318,3 +319,28 @@ CREATE TABLE "invoice" (
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
mod m006 {
|
||||
use crate::repo::postgres_migration::{Migration, SimpleSqlMigration};
|
||||
|
||||
pub const VERSION: i64 = 6;
|
||||
|
||||
pub fn migration() -> impl Migration {
|
||||
SimpleSqlMigration {
|
||||
serial_number: VERSION,
|
||||
sql: vec![
|
||||
r#"
|
||||
--- Use text column for content
|
||||
ALTER TABLE "event" ALTER COLUMN "content" TYPE text USING convert_from("content", 'utf-8);
|
||||
|
||||
--- Create search col for event content
|
||||
ALTER TABLE event
|
||||
ADD COLUMN ts_content tsvector
|
||||
GENERATED ALWAYS AS (to_tsvector('english', "content"::json->>'content')) stored;
|
||||
|
||||
CREATE INDEX ts_content_idx ON event USING GIN (ts_content);
|
||||
"#,
|
||||
],
|
||||
}
|
||||
}
|
||||
}
|
|
@ -37,10 +37,12 @@ pub struct ReqFilter {
|
|||
/// Set of tags
|
||||
pub tags: Option<HashMap<char, HashSet<String>>>,
|
||||
/// Force no matches due to malformed data
|
||||
// we can't represent it in the req filter, so we don't want to
|
||||
// erroneously match. This basically indicates the req tried to
|
||||
// do something invalid.
|
||||
/// we can't represent it in the req filter, so we don't want to
|
||||
/// erroneously match. This basically indicates the req tried to
|
||||
/// do something invalid.
|
||||
pub force_no_match: bool,
|
||||
/// NIP-50 search query
|
||||
pub search: Option<String>,
|
||||
}
|
||||
|
||||
impl Serialize for ReqFilter {
|
||||
|
@ -67,6 +69,9 @@ impl Serialize for ReqFilter {
|
|||
if let Some(authors) = &self.authors {
|
||||
map.serialize_entry("authors", &authors)?;
|
||||
}
|
||||
if let Some(search) = &self.search {
|
||||
map.serialize_entry("search", &search)?;
|
||||
}
|
||||
// serialize tags
|
||||
if let Some(tags) = &self.tags {
|
||||
for (k, v) in tags {
|
||||
|
@ -98,6 +103,7 @@ impl<'de> Deserialize<'de> for ReqFilter {
|
|||
authors: None,
|
||||
limit: None,
|
||||
tags: None,
|
||||
search: None,
|
||||
force_no_match: false,
|
||||
};
|
||||
let empty_string = "".into();
|
||||
|
@ -124,6 +130,8 @@ impl<'de> Deserialize<'de> for ReqFilter {
|
|||
rf.until = Deserialize::deserialize(val).ok();
|
||||
} else if key == "limit" {
|
||||
rf.limit = Deserialize::deserialize(val).ok();
|
||||
} else if key == "search" {
|
||||
rf.search = Deserialize::deserialize(val).ok();
|
||||
} else if key == "authors" {
|
||||
let raw_authors: Option<Vec<String>> = Deserialize::deserialize(val).ok();
|
||||
if let Some(a) = raw_authors.as_ref() {
|
||||
|
|
Loading…
Reference in New Issue
Block a user