Crunchy Data is excited to release a new open source tool, Exporter2Perfdata. This connects Prometheus Exporters with Icinga2/Nagios. This also allows pgMonitor users to connect monitoring information from Prometheus to Icinga2 and Nagios.
David optimizes a slow query by splitting up a left join into two calls and using UNION ALL to combine them. The proof is in the pudding.
Which of my applications ran this SQL statement? Use SQL comments to track which application ran which queries.
For developers, databases are basically magic. Like Penn & Teller, this blog post is set to break the illusion. Databases are just like any other code, they have algorithms and processes. These algorithms and processes are meant to improve performance, but can cause limitations if they are not expected.
Craig's year end wrap-up of 5 things you can do for a healthier Postgres database next year.
A deep dive into Patroni, etcd, and the etcd consensus protocol to diagnose common communication and system issues.
Today I wanted to call some extra attention to those little things, the ones that don't get the spotlight, but simply make a developer's life better.
If you have insert-only tables in a version of PostgreSQL earlier that 13, you could benefit from running a regularly scheduled VACUUM.
By default Linux uses a controversial (for databases) memory extension feature called Overcommit. How that interacts with PostgreSQL is covered in the Managing Kernel Resources section of the PG manual.
Almost a decade after range types were introduced, Postgres 14 makes it easier to write "boring SQL" for range data. Meet the "multirange" data type.
A feature highlight of the new JSON subscript support within PostgreSQL 14.
Whether you are starting a new development project, launching an application modernization effort, or engaging in digital transformation, chances are you are evaluating Kubernetes. If you selected Kubernetes, chances are you will ultimately need a database.
An underappreciated element of PostgreSQL performance can be the data types chosen and their organization in tables. For sites that are always looking for that incremental performance improvement, managing the exact layout and utilization of every byte of a row (also known as a tuple) can be worthwhile.
The ultimate goal for any unplanned database interruption is to reduce data loss. So the advice here is aimed at someone in a production situation prioritizing minimizing data loss.
The pgBackRest team is pleased to announce the introduction of multiple repository support in v2.33. Backups already provide redundancy by creating an offline copy of your PostgreSQL cluster that can be used in disaster recovery.
Today we're announcing Crunchy Bridge support for Managed Postgres on Google Cloud. With Crunchy Bridge you can now have the same great PostgreSQL experience on any cloud and seamlessly migrate between cloud vendors as you see fit.
What follows is a summary of conversations I've had with customers on how to think about key tenants of database management, high availability and disaster recovery.
When Linux detects that the system is using too much memory, it will identify processes for termination and, well, assassinate them. The OOM killer has a noble role in ensuring a system does not run out of memory, but this can lead to unintended consequences.
Deciding to use Postgres is a great start, we like to think it’s the right start. As to how best to deploy it, you have a number of options.
A step-by-step guide for how to setup PostgreSQL change data capture with Debezium and Apache Kafka.