Click any tag below to further narrow down your results
Links
This article discusses how Recall.ai faced delays in PostgreSQL connections during high-load meeting spikes. The issue stemmed from the single-threaded nature of the postmaster process, which struggled to handle the surge in connection requests, leading to significant latency.
pg_tracing is a PostgreSQL extension that creates server-side spans for tracking query performance and execution. It supports various PostgreSQL events and allows trace context propagation through SQL comments or GUC parameters. The extension is currently in early development and works with PostgreSQL versions 14 to 16.
This article delves into Postgres' internal tables and catalog views, highlighting how they store critical information about database operations. It covers key psql commands and SQL queries to access valuable data for troubleshooting and performance monitoring. Understanding these internals can enhance your ability to manage and optimize your Postgres databases.
Allocating too much memory to Postgres can actually slow down performance, especially during index builds. The author explains how exceeding certain memory thresholds can lead to inefficient data processing and increased write operations, which negatively impact speed. It's better to use modest memory settings and adjust only based on proven benefits.
To efficiently insert large datasets into a Postgres database, combining Spark's parallel processing with Python's COPY command can significantly enhance performance. By repartitioning the data and utilizing multiple writers, the author was able to insert 22 million records in under 14 minutes, leveraging Postgres's bulk-loading capabilities over traditional JDBC methods.
The article discusses various methods to intentionally slow down PostgreSQL databases for testing purposes. It explores different configurations and practices to simulate performance degradation, aiding developers in understanding how their applications behave under stress. This approach helps in identifying potential bottlenecks and preparing for real-world scenarios.
The article discusses the advantages of using Redis for caching in applications, particularly in conjunction with Postgres for data storage. It highlights Redis's speed and efficiency in handling cache operations, which can significantly improve application performance. Additionally, it addresses potential pitfalls and best practices for integrating Redis with existing systems.
PostgreSQL 18 introduces significant enhancements, including a new asynchronous I/O subsystem for improved performance, native support for UUIDv7 for better indexing, and improved output for the EXPLAIN command. Additionally, it streamlines major version upgrades and offers new capabilities for handling NOT NULL constraints and RETURNING statements.
The podcast episode features Aaron Katz and Sai Krishna Srirampur discussing the transition from Postgres to ClickHouse, highlighting how this shift simplifies the modern data stack. They explore the benefits of ClickHouse's architecture for analytics and performance in data-driven environments.
The article discusses techniques for enhancing query performance in PostgreSQL by manipulating its statistics tables. It explains how to use these statistics effectively to optimize query planning and execution, ultimately leading to faster data retrieval. Practical examples and insights into the PostgreSQL system are provided to illustrate these methods.
The article discusses the advantages of indexing JSONB data types in PostgreSQL, emphasizing improved query performance and efficient data retrieval. It provides practical examples and techniques for creating indexes, as well as considerations for maintaining performance in applications that utilize JSONB fields.
The article discusses the exciting new features and improvements introduced in PostgreSQL 18, highlighting enhancements in performance, security, and usability. It emphasizes how these updates position PostgreSQL as a leading database solution for developers and businesses alike. Additionally, the blog encourages readers to explore the potential of PostgreSQL in their projects and applications.
Motion transitioned from CockroachDB to Postgres due to escalating costs and operational challenges, particularly with migrations and ETL processes. The migration revealed better performance with Postgres for many queries, despite some initial advantages of Cockroach’s query planner. The move ultimately streamlined operations and resolved numerous UI and support issues experienced with CockroachDB.
Recall.ai faced significant performance issues with their Postgres database due to the high concurrency of NOTIFY commands used during transactions, which caused global locks and serialized commits, leading to downtime. After investigating, they discovered that the LISTEN/NOTIFY feature did not scale well under their workload of tens of thousands of simultaneous writers. They advise against using LISTEN/NOTIFY in high-write scenarios to maintain database performance and scalability.
Unique indexes in PostgreSQL have a limitation where entries larger than 1/3 of a buffer page (~2.7KB) cannot be indexed, particularly affecting large text fields. To enforce uniqueness on large data, a common workaround is to create a hash of the data and index that instead, allowing for efficient comparisons without exceeding size constraints. The article explains the reasons behind these constraints and offers a practical solution using hash functions.
The article explores a mysterious issue related to PostgreSQL's handling of SIGTERM signals, which can lead to unexpected behavior during shutdown. It discusses the implications of this behavior on database performance and reliability, particularly in the context of modern cloud architectures. The author highlights the importance of understanding these nuances to avoid potential pitfalls in database management.