Lemmy’s PostgreSQL was developed with this philosophy, intentional or otherwise:

  1. Real-time client connection and notification via websocket that gets every single action that passes through PostgreSQL. One upvote, instantly sent to client. One new comment, instantly appeared on Lemmy-ui with version 0.17.4

  2. INSERT overhead for PostgreSQL. As soon as a Lemmy post or comment is done, a parallel post_aggregate and comment_aggregate row is created.

  3. INSERT counting overhead. Extreme effort is made by Lemmy to count things, all the time. Every new INSERT of a comment or post does a real-time update of the total server count. This is done via a SQL UPDATE and not by just issuing a COUNT(*) on the rows when the data is requested.

  4. No archiving or secondary storage concept. PostgreSQL has it in the main tables or nothing.

  5. Raw numbers, local and unique for each instance, for comment and post. But community name and username tend to be more known than raw numbers.

  6. Sorting choices presented on many things: communities, posts, comments. And new methods of sorting and slicing the data keep being added in 2023.

  7. No caching of data. The developers of lemmy have gone to extremes to avoid caching on either lemmy-ui or within the Rust code of lemmy_server. Lemmy philosophy favors putting constant connection to a single PostgreSQL.

  8. User preferences and customization are offloaded to PostgreSQL do do the heavy lifting. PostgreSQL has to look at the activity of each user to know if they have saved a post, previously read a post, upvoted that post, or even blocked the person who created the post.

  9. Language choice is built into the system early, but I see no evidence it has proven to be useful. But it carries a high overhead in how many PostgreSQL database rows each site carries - and is used in filtering More often than not, I’ve found end-users confused why they can’t find content when they accidentally turned off choices in lemmy-ui

  10. All fields on SELECT. Throughout the Rust Diesel ORM code, it’s every field in every table being touched.

  11. SELECT statements are almost always ORM machine generated. TRIGGER FUNCTION logic is hand-written.

  • RoundSparrowM
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    Counting is done on each new INSERT, but on the SELECT side of things there are some background batch jobs. Most notable:

    /// Update the hot_rank columns for the aggregates tables
    /// Runs in batches until all necessary rows are updated once
    fn update_hot_ranks(conn: &mut PgConnection) {
      info!("Updating hot ranks for all history...");
    
      process_hot_ranks_in_batches(
        conn,
        "post_aggregates",
        "a.hot_rank != 0 OR a.hot_rank_active != 0",
        "SET hot_rank = hot_rank(a.score, a.published),
             hot_rank_active = hot_rank(a.score, a.newest_comment_time_necro)",
      );
    
      process_hot_ranks_in_batches(
        conn,
        "comment_aggregates",
        "a.hot_rank != 0",
        "SET hot_rank = hot_rank(a.score, a.published)",
      );
    
      process_hot_ranks_in_batches(
        conn,
        "community_aggregates",
        "a.hot_rank != 0",
        "SET hot_rank = hot_rank(a.subscribers, a.published)",
      );
    
      info!("Finished hot ranks update!");
    }
    

    This is done every 15 minutes by the Rust code internally.

    • RoundSparrowM
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      function as of version 0.18.4

      CREATE FUNCTION public.hot_rank(score numeric, published timestamp without time zone) RETURNS integer
          LANGUAGE plpgsql IMMUTABLE PARALLEL SAFE
          AS $$
      DECLARE
          hours_diff numeric := EXTRACT(EPOCH FROM (timezone('utc', now()) - published)) / 3600;
      BEGIN
          IF (hours_diff > 0) THEN
              RETURN floor(10000 * log(greatest (1, score + 3)) / power((hours_diff + 2), 1.8))::integer;
          ELSE
              RETURN 0;
          END IF;
      END;
      $$;
      
    • RoundSparrowM
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Rust function:

      // Runs the hot rank update query in batches until all rows have been processed.
      /// In `where_clause` and `set_clause`, "a" will refer to the current aggregates table.
      /// Locked rows are skipped in order to prevent deadlocks (they will likely get updated on the next
      /// run)
      fn process_hot_ranks_in_batches(
        conn: &mut PgConnection,
        table_name: &str,
        where_clause: &str,
        set_clause: &str,
      ) {
        let process_start_time = NaiveDateTime::from_timestamp_opt(0, 0).expect("0 timestamp creation");
      
        let update_batch_size = 1000; // Bigger batches than this tend to cause seq scans
        let mut processed_rows_count = 0;
        let mut previous_batch_result = Some(process_start_time);
        while let Some(previous_batch_last_published) = previous_batch_result {
          // Raw `sql_query` is used as a performance optimization - Diesel does not support doing this
          // in a single query (neither as a CTE, nor using a subquery)
          let result = sql_query(format!(
            r#"WITH batch AS (SELECT a.id
                     FROM {aggregates_table} a
                     WHERE a.published > $1 AND ({where_clause})
                     ORDER BY a.published
                     LIMIT $2
                     FOR UPDATE SKIP LOCKED)
               UPDATE {aggregates_table} a {set_clause}
                   FROM batch WHERE a.id = batch.id RETURNING a.published;
          "#,
            aggregates_table = table_name,
            set_clause = set_clause,
            where_clause = where_clause
          ))
          .bind::(previous_batch_last_published)
          .bind::(update_batch_size)
          .get_results::(conn);
      
          match result {
            Ok(updated_rows) => {
              processed_rows_count += updated_rows.len();
              previous_batch_result = updated_rows.last().map(|row| row.published);
            }
            Err(e) => {
              error!("Failed to update {} hot_ranks: {}", table_name, e);
              break;
            }
          }
        }
        info!(
          "Finished process_hot_ranks_in_batches execution for {} (processed {} rows)",
          table_name, processed_rows_count
        );
      }