The simple story of spatial indexes is - if you are planning to do spatial queries (which, if you are storing spatial objects, you probably are) you should create a spatial index for your table.
Whether you are starting a new development project, launching an application modernization effort, or engaging in digital transformation, chances are you are evaluating Kubernetes. If you selected Kubernetes, chances are you will ultimately need a database.
An underappreciated element of PostgreSQL performance can be the data types chosen and their organization in tables. For sites that are always looking for that incremental performance improvement, managing the exact layout and utilization of every byte of a row (also known as a tuple) can be worthwhile.
Crunchy Data, the leading provider of trusted open source PostgreSQL, is proud to announce the addition of Google Cloud Platform (GCP) to its list of providers supported on Crunchy Bridge.
Connection pooling and management is one of those things most people ignore far too long when it comes to their database. As you grow into the hundreds, better connection management is a quick and easy win. Let's dig into the three variations of connection pooling and how to identify if you can benefit from a connection pooler and where.
The ultimate goal for any unplanned database interruption is to reduce data loss. So the advice here is aimed at someone in a production situation prioritizing minimizing data loss.
The pgBackRest team is pleased to announce the introduction of multiple repository support in v2.33. Backups already provide redundancy by creating an offline copy of your PostgreSQL cluster that can be used in disaster recovery.
Today we're announcing Crunchy Bridge support for Managed Postgres on Google Cloud. With Crunchy Bridge you can now have the same great PostgreSQL experience on any cloud and seamlessly migrate between cloud vendors as you see fit.
What follows is a summary of conversations I've had with customers on how to think about key tenants of database management, high availability and disaster recovery.
This post provides a framework for thinking through how you can confront "security questions" as you embark on your Postgres journey.
We at Crunchy Data put as much development effort into improving GEOS as we do improving PostGIS proper, because the GEOS library is so central to much geospatial processing.
The EXPLAIN command helps you look even closer into an individual query. If you're already proficient in EXPLAIN, great! Read on for an easy refresher. If you're less familiar with it, this will be a (hopefully) gentle introduction on what insights it might help provide.
Welcome to Episode 2 of the "Musings of a PostgreSQL Data Pontiff" series! In this installment I’m aiming to achieve three objectives.
This is the first in a series of blogs on the topic of using PostgreSQL for "data science". I put that in quotes because I would not consider myself to be a practicing "data scientist", per se. Of course I'm not sure there is a universally accepted definition of "data scientist". This article provides a nice illustration of my point.
With this release, we included features to streamline management of the Operator, added security features, and extra system metrics to enhance your high availability Kubernetes Postgres cluster. Let's take a look at what's new in the Postgres Operator 4.6!
Crunchy Data has recently announced an update to the CIS PostgreSQL Benchmark by the Center for Internet Security, a nonprofit organization that provides publications around standards and best practices for securing technologies systems.
How can you apply GitOps principles to running PostgreSQL on Kubernetes with Helm?
The page "Falsehoods Programmers Believe About Names" covers some of the ways names are hard to deal with in programming. This post will ignore most of those complexities, and deal with the problem of matching up loose user input to a database of names.
Today we are going to walk through some of the preliminary data shaping steps in data science using SQL in Postgres.
I want to work on optimizing all my queries all day long because it will definitely be worth the time and effort. That's a statement that has hopefully never been said. So when it comes to query optimizing, how should you pick your battles?