34 links
tagged with all of: performance + database
Click any tag below to further narrow down your results
Links
A slow database query caused significant downtime for the Placid app, highlighting the importance of monitoring and quickly addressing performance issues. The incident illustrates how rapid identification and resolution of such issues can minimize disruption and improve user experience. Implementing effective alerting systems and performance tracking can be crucial in preventing similar occurrences in the future.
The article compares the performance of ClickHouse and PostgreSQL, highlighting their strengths and weaknesses in handling analytical queries and data processing. It emphasizes ClickHouse's efficiency in large-scale data management and real-time analytics, making it a suitable choice for high-performance applications.
The article discusses the features and capabilities of DuckDB, a high-performance analytical database management system designed for data analytics. It highlights its integration with various data sources and its usability in data science workflows, emphasizing its efficiency and ease of use.
Wildcat is a high-performance embedded key-value database written in Go, featuring modern design principles such as LSM tree architecture and MVCC for efficient read and write operations. It supports ACID transactions, offers configurable durability levels, and provides comprehensive iteration capabilities, making it suitable for applications requiring immediate consistency and durability. Users can join a community for support and access resources for development and benchmarking.
DuckDB 0.14.0 has been released, featuring significant enhancements and new functionalities aimed at improving performance and usability. Key updates include support for new data types, optimizations for query execution, and better integration with various programming environments. This release continues DuckDB's commitment to providing a powerful analytical database for data science and analytics tasks.
The article discusses optimizing SQLite indexes to improve query performance, highlighting the importance of composite indexes over multiple single-column indexes and the significance of index column order. By understanding SQLite's query planner and utilizing techniques like partial indexes, the author achieved a 35% speedup in query execution for their application, Scour, which handles a rapidly increasing volume of content.
SQLite query optimization significantly improved the performance of the Matrix Rust SDK, boosting event processing from 19,000 to 4.2 million events per second. The article details the structure of data persistence using LinkedChunk and how identifying and addressing inefficiencies in SQL queries led to this enhancement. It emphasizes the importance of profiling tools and strategic indexing to optimize database interactions.
Efficient database connection management, particularly through connection pooling, is crucial for optimizing performance and scalability in applications. The article discusses the benefits of using a proxy-based connection pooler like AWS RDSProxy over application-based pooling methods, highlighting improved resource utilization, reduced overhead, and better management of concurrent connections. It also outlines the setup process for integrating RDSProxy with SQLAlchemy in a Flask application environment at Lyft.
The article discusses various methods to intentionally slow down PostgreSQL databases for testing purposes. It explores different configurations and practices to simulate performance degradation, aiding developers in understanding how their applications behave under stress. This approach helps in identifying potential bottlenecks and preparing for real-world scenarios.
Radar has developed HorizonDB, a high-performance geospatial database in Rust, to replace Elasticsearch and MongoDB for their geolocation services. This transition has significantly improved operational efficiency, reduced costs, and enhanced performance, allowing the platform to handle over 1 billion API calls daily with low latency and better scalability.
The article discusses performance improvements in pgstream, a tool used for taking snapshots of PostgreSQL databases. It highlights the underlying challenges and solutions implemented to enhance the speed and efficiency of database snapshots, ultimately benefiting users with faster data access and reduced operational overhead.
PostgreSQL 18 has been released, featuring significant performance improvements through a new asynchronous I/O subsystem, enhanced query execution capabilities, and easier major-version upgrades. The release also introduces new features such as virtual generated columns, OAuth 2.0 authentication support, and improved statistical handling during upgrades, solidifying PostgreSQL's position as a leading open source database solution.
The article discusses an innovative approach to database durability using async I/O on Linux with io_uring. By implementing a dual write-ahead log (WAL) system that separates intent and completion records, the author achieves significant improvements in transaction throughput while maintaining data consistency. This method allows for better utilization of modern storage hardware's parallelism, ultimately leading to a rethinking of traditional database architectures.
The article emphasizes the importance of database data fixtures in software development, arguing that they are both parallel-safe and efficient. It highlights how using these fixtures can improve testing speed and reliability, making them a valuable tool for developers.
The article discusses techniques for enhancing query performance in PostgreSQL by manipulating its statistics tables. It explains how to use these statistics effectively to optimize query planning and execution, ultimately leading to faster data retrieval. Practical examples and insights into the PostgreSQL system are provided to illustrate these methods.
The N+1 query problem arises when multiple database queries are triggered in a loop, leading to performance issues as data grows. By adopting efficient querying strategies, such as using JOINs or IN clauses, developers can significantly reduce unnecessary database traffic and improve application performance.
The article discusses the advantages of indexing JSONB data types in PostgreSQL, emphasizing improved query performance and efficient data retrieval. It provides practical examples and techniques for creating indexes, as well as considerations for maintaining performance in applications that utilize JSONB fields.
PostgreSQL 18 introduces significant enhancements, including a new asynchronous I/O subsystem for improved performance, native support for UUIDv7 for better indexing, and improved output for the EXPLAIN command. Additionally, it streamlines major version upgrades and offers new capabilities for handling NOT NULL constraints and RETURNING statements.
Pgline is a high-performance PostgreSQL driver for Node.js, developed in TypeScript, that implements Pipeline Mode, allowing for efficient concurrent queries with reduced CPU usage. Benchmark tests show Pgline outperforms competitors like Bun SQL, Postgresjs, and Node-postgres in terms of speed and resource efficiency. Installation and usage examples are provided to demonstrate its capabilities.
The article discusses the exciting new features and improvements introduced in PostgreSQL 18, highlighting enhancements in performance, security, and usability. It emphasizes how these updates position PostgreSQL as a leading database solution for developers and businesses alike. Additionally, the blog encourages readers to explore the potential of PostgreSQL in their projects and applications.
The article discusses the complexities and performance considerations of implementing a distributed database cache. It highlights the challenges of cache synchronization, data consistency, and the trade-offs between speed and accuracy in data retrieval. Additionally, it offers insights into strategies for optimizing caching methods to enhance overall system performance.
ClickHouse introduces its capabilities in full-text search, highlighting the efficiency and performance improvements it offers over traditional search solutions. The article discusses various features, including indexing and query optimization, that enhance the user experience for searching large datasets. Additionally, it covers practical use cases and implementation strategies for developers.
Pipelining in PostgreSQL allows clients to send multiple queries without waiting for the results of previous ones, significantly improving throughput. Introduced in PostgreSQL 18, this feature enhances the efficiency of query processing, especially when dealing with large batches of data across different network types. Performance tests indicate substantial speed gains, underscoring the benefits of utilizing pipelining in SQL operations.
PostgreSQL's Index Only Scan enhances query performance by allowing data retrieval without accessing the table heap, thus eliminating unnecessary delays. It requires specific index types and query conditions to function effectively, and the concept of a covering index, which includes fields in the index, further optimizes this process. Understanding these features is crucial for backend developers working with PostgreSQL databases.
uuidv47 enables the storage of sortable UUIDv7 in databases while presenting a UUIDv4-like facade at the API level. It employs a deterministic and invertible mapping through a keyed SipHash-2-4 stream, ensuring security and compatibility with RFC standards. The library includes a PostgreSQL extension and offers full testing and performance benchmarks.
Motion transitioned from CockroachDB to Postgres due to escalating costs and operational challenges, particularly with migrations and ETL processes. The migration revealed better performance with Postgres for many queries, despite some initial advantages of Cockroach’s query planner. The move ultimately streamlined operations and resolved numerous UI and support issues experienced with CockroachDB.
Arc is a high-performance time-series database capable of ingesting 2.4 million metrics per second, along with logs, traces, and events using a unified MessagePack columnar protocol. Currently in alpha release, it features a stable core with ongoing developments, including advanced SQL analytics via DuckDB, flexible storage options, and built-in token-based authentication, making it suitable for development and testing environments. The system is designed for high-throughput ingestion, low latency, and efficient data management, aiming to support observability across various telemetry types.
Recall.ai faced significant performance issues with their Postgres database due to the high concurrency of NOTIFY commands used during transactions, which caused global locks and serialized commits, leading to downtime. After investigating, they discovered that the LISTEN/NOTIFY feature did not scale well under their workload of tens of thousands of simultaneous writers. They advise against using LISTEN/NOTIFY in high-write scenarios to maintain database performance and scalability.
The article explores the differences in indexing between traditional relational databases and open table formats like Apache Iceberg and Delta Lake, emphasizing the challenges and limitations of adding secondary indexes to optimize query performance in analytical workloads. It highlights the importance of data organization and auxiliary structures in determining read efficiency, rather than relying solely on traditional indexing methods.
The article discusses common SQL anti-patterns that developers should avoid to improve database performance and maintainability. It highlights specific practices that can lead to inefficient queries and recommends better alternatives to enhance SQL code quality. Understanding and addressing these anti-patterns is crucial for effective database management.
The article discusses the temporal-spatial locality hypothesis in database design, highlighting its significance for optimizing performance in various database systems. It contrasts the behavior of streaming systems that benefit from this hypothesis with hash-based databases that do not, and explores the implications of different key assignment strategies on read and write performance. The author argues that while the hypothesis is often weakly true, its relevance varies across workloads and requires careful consideration in schema design.
The article discusses the advantages and practical applications of materialized views in database management, emphasizing their ability to enhance query performance and simplify complex data retrieval. It also addresses common misconceptions and highlights scenarios where their use is particularly beneficial for developers and data analysts.
The article explores a mysterious issue related to PostgreSQL's handling of SIGTERM signals, which can lead to unexpected behavior during shutdown. It discusses the implications of this behavior on database performance and reliability, particularly in the context of modern cloud architectures. The author highlights the importance of understanding these nuances to avoid potential pitfalls in database management.
The article discusses the implementation of a time series engine using Rust, focusing on optimizing performance under heavy load conditions. It highlights various techniques such as throttling to manage resource allocation effectively and ensure system stability during peak usage. Key challenges and solutions in developing a robust time series database are also addressed.