We just released a new version of Crunchy Postgres. Packed with security, access management, and monitoring updates. Giving you more flexibility and options for running production Postgres.
What happens when you query Postgres? Data can actually come from many different places like the application cache, buffer cache, and even down into the physical disk cache. This post surveys the data storage and flow of Postgres data.
Just kidding. This is not crypto mining tutorial! It is a very practical description of using pgcrypto for encrypting and decrypting data inside your Postgres database.
Checking on your collations is a must have stop on your migration path. You might just run a quick check and be on your way or you might need to add a few more steps to your cutover plans.
Seeing locks in your Postgres monitoring but don't know what it means? David takes a look at locks and what to take into consideration.
When it comes to database performance, many think they are solving technical problems, but are solving a money problem. There are 2 phases to database-cost management, and understanding which phase you are in will help you understand your options for growth.
Crunchy Data is excited to release a new open source tool, Exporter2Perfdata. This connects Prometheus Exporters with Icinga2/Nagios. This also allows pgMonitor users to connect monitoring information from Prometheus to Icinga2 and Nagios.
David optimizes a slow query by splitting up a left join into two calls and using UNION ALL to combine them. The proof is in the pudding.
Which of my applications ran this SQL statement? Use SQL comments to track which application ran which queries.
For developers, databases are basically magic. Like Penn & Teller, this blog post is set to break the illusion. Databases are just like any other code, they have algorithms and processes. These algorithms and processes are meant to improve performance, but can cause limitations if they are not expected.
Craig's year end wrap-up of 5 things you can do for a healthier Postgres database next year.
A deep dive into Patroni, etcd, and the etcd consensus protocol to diagnose common communication and system issues.
Today I wanted to call some extra attention to those little things, the ones that don't get the spotlight, but simply make a developer's life better.
If you have insert-only tables in a version of PostgreSQL earlier that 13, you could benefit from running a regularly scheduled VACUUM.
By default Linux uses a controversial (for databases) memory extension feature called Overcommit. How that interacts with PostgreSQL is covered in the Managing Kernel Resources section of the PG manual.
Almost a decade after range types were introduced, Postgres 14 makes it easier to write "boring SQL" for range data. Meet the "multirange" data type.
Whether you are starting a new development project, launching an application modernization effort, or engaging in digital transformation, chances are you are evaluating Kubernetes. If you selected Kubernetes, chances are you will ultimately need a database.
An underappreciated element of PostgreSQL performance can be the data types chosen and their organization in tables. For sites that are always looking for that incremental performance improvement, managing the exact layout and utilization of every byte of a row (also known as a tuple) can be worthwhile.
The ultimate goal for any unplanned database interruption is to reduce data loss. So the advice here is aimed at someone in a production situation prioritizing minimizing data loss.
The pgBackRest team is pleased to announce the introduction of multiple repository support in v2.33. Backups already provide redundancy by creating an offline copy of your PostgreSQL cluster that can be used in disaster recovery.