How to Reduce Your PostgreSQL Database Size
Your phone buzzes in the middle of the night. You pick it up. A monitor went off at work—your PostgreSQL database is slowly but steadily reaching its maximum storage space. You are the engineer in charge. What should you do?
Okay, if it comes down to that situation, you should probably remedy it ASAP by adding more storage. But you’re going to need a better strategy in the long run to optimize your PostgreSQL storage use, or you’ll keep paying more and more money.
Does your PostgreSQL database really need to be that large? Is there something you can do to optimize your storage use?
This article explores several strategies that will help you reduce your PostgreSQL database size considerably and sustainably.
Why Is PostgreSQL Storage Optimization Important?
Let’s spend a couple of minutes addressing this question first.
Perhaps you’re thinking:
“Storage is cheap these days, and optimizing a PostgreSQL database takes time and effort. I’ll just keep adding more storage.”
“My PostgreSQL provider is actually usage-based (like Timescale), and I don’t have the problem of being locked into a large disk.”
Indeed, resigning yourself to simply using more storage is the most straightforward way to tackle an increasingly growing PostgreSQL database. Are you running servers on-prem? Slap another hard drive on that bad boy. Are you running PostgreSQL in RDS? Raise the storage limits. But this comes with problems.
The first problem, and the most obvious, is the cost. If you’re running PostgreSQL in an EBS instance in AWS or in RDS, for example, you’ll be charged through an allocation-based model. This model assumes that you’ll need to predetermine how much disk space you’ll need in the future and then pay for it, regardless of whether you end up using it or not, and without the chance of downscaling.
In other PostgreSQL providers, when you run out of storage space, you must upgrade and pay for the next available plan or storage tier, meaning that you’ll see a considerably higher bill overnight.
In a way, these issues are mitigated by usage-based models. Timescale charges by the amount of storage you use: you don't need to worry about allocating storage or managing storage plans, which really simplifies things—and the less storage you use, the less it costs.
Usage-based models are a great incentive to actually optimize your PostgreSQL database size as much as possible since you’ll see immediate reductions in your bill. But yes, this also works the opposite way: if you don’t pay attention to managing your storage, your storage bill will increase.
The second problem that comes with not optimizing your PostgreSQL storage usage is that this situation can lead to bad performance, with queries running slower and your I/O operations increasing. This is something that often gets overlooked, but maintaining PostgreSQL storage usage is paramount to keeping large PostgreSQL tables fast.
This last point deserves a deeper dive into how data is actually stored in PostgreSQL and what is causing the problem, so let’s briefly cover some essential PostgreSQL storage concepts.
Essential PostgreSQL Storage Concepts
How does PostgreSQL store data?
At a high level, there are two terms you need to understand: tuples and pages.
- A tuple is the physical representation of an entry in a table. You'll generally see the terms tuple and row used interchangeably. Each element in a tuple corresponds to a specific column in that table, containing the actual data value for that column.
- A page is the unit of storage in PostgreSQL, typically 8 kB in size, that holds one or more tuples. PostgreSQL reads and writes data in page units.
Each page in PostgreSQL consists of a page header (which contains metadata about the page, such as page layout versions, page flags, and so on) and actual data (including tuples). There’s also a special area called the Line Pointer Array, which provides the offsets where each tuple begins.
What happens when querying data?
When querying data, PostgreSQL utilizes the metadata to quickly navigate to the relevant page and tuple. The PostgreSQL query planner examines the metadata to decide the optimal path for retrieving data, for example, estimating the cost of different query paths based on the metadata information about the tables, indexes, and data distribution.
What happens when we INSERT/ DELETE/ UPDATE a row in PostgreSQL?
When a new tuple is inserted into a PostgreSQL table, it gets added to a page with enough free space to accommodate the tuple. Each tuple within a page is identified and accessed using the offset provided in the Line Pointer Array.
If a tuple inserted is too big for the available space of a page, PostgreSQL doesn't split it between two 8kB pages. Instead, it employs TOAST to compress and/or break the large values into smaller pieces. These pieces are then stored in a separate TOAST table, while the original tuple retains a pointer to this external stored data.
What is a dead tuple?
A key aspect to understand (and this will influence our PostgreSQL database size, as we’ll see shortly) is that when you delete data in PostgreSQL via
DELETE FROM, you’re not actually deleting it but marking the rows as unavailable. These unavailable rows are usually referred to as “dead tuples.”
When you run
UPDATE, the row you’re updating will also be marked as a dead tuple. Then, PostgreSQL will insert a new tuple with the updated column.
You might be wondering why PostgreSQL does this. Dead tuples are actually a compromise to reduce excessive locks on tables during concurrent operations, multiple connections, and simplifying transactions. Imagine a transaction failing halfway through its execution; it is much easier to revert a change when the old data is still available than trying to rewind each action in an idempotent way. Furthermore, this mechanism supports the easy and efficient implementation of rollbacks, ensuring data consistency and integrity during transactions.
The trade-off, however, is the increased database size due to the accumulation of dead tuples, necessitating regular maintenance to reclaim space and maintain performance… What brings us to table bloat.
What is table bloat?
When a tuple is deleted or updated, its old instance is considered a dead tuple. The issue with dead tuples is that they’re effectively still a tuple on disk, taking up storage space—yes, that storage page that is costing you money every month.
Table bloat refers to this excess space that dead tuples occupy in your PostgreSQL database, which not only leads to an inflated table size but also to increased I/O and slower queries. Since PostgreSQL runs under the MVCC system, it doesn't immediately purge these dead tuples from the disk. Instead, they linger until a vacuum process reclaims their space.
Table bloat also occurs when a table contains unused pages, which can accumulate as a result of operations such as mass deletes.
Dead tuples get cleaned and deleted from storage when the
VACUUM command runs:
Vacuum has a lot of roles, but the relevant point for this article is that vacuum removes dead tuples once all connections using the dead tuples are closed.
VACUUM by itself will not delete pages, though. Any pages created by a table will stay allocated, although the memory in those pages is now usable space after running vacuum.
What is autovacuum?
Postgres conveniently includes a daemon to automatically run vacuum on tables that get heavy insert, update, and delete traffic. It operates in the background, monitoring the database to identify tables with accumulating dead tuples and then initiating the vacuum process autonomously.
Autovacuum comes enabled by default, although the threshold PostgreSQL uses to enable autovacuum is very conservative.
What is VACUUM FULL?
Autovacuum helps with dead tuples, but what about unused pages?
VACUUM FULL command is a more aggressive version of
VACUUM that locks the table, removes dead tuples and empty pages, and then returns the reclaimed space to the operating system.
VACUUM FULL can be resource-intensive and requires an exclusive lock on the table during the process. We’ll come back to this later.
Now that you have the necessary context, let’s jump into the advice.
How To Reduce Your PostgreSQL Database Size
Use Timescale compression
There are different ways we can compress our data to consistently save storage space. PostgreSQL has some compression mechanisms, but if you want to take data compression even further, especially for time-series data, you should use Timescale’s columnar compression.
It allows you to dramatically compress data through a provided
add_compression_policy() function. To achieve high compression rates, Timescale uses various compression techniques depending on data types to reduce your data footprint. Timescale also uses column stores to merge many rows into a single row, saving space.
Let's illustrate how this works with an example.
Let’s say we have a hypertable with a week's worth of data. Imagine that our application generally only needs data from the last day, but we must keep historical data around for reporting purposes. We could run
SELECT add_compression_policy('my_table', INTERVAL '24 hours'); which automatically compresses rows in the
my_table hypertable older than 24 hours.
Timescale’s compression would combine all the rows into a single row, where each column contains an array of all the row's data in segments of 1,000 rows. Visually, this would take a table that looks like this:
| time | location | temperature | |------------------------|----------|-------------| | 2023-09-20 00:16:00.00 | garage | 80 | | 2023-09-21 00:10:00.00 | attic | 92.3 | | 2023-09-22 00:5:00.00 | basement | 73.9 |
And compress it down to a table like this:
| time | location | temperature | |--------------------------------------------------------------------------|-----------------------------|---------------------------| | [2023-09-20 00:16:00.00, 2023-09-21 00:10:00.00, 2023-09-22 00:5:00.00] | [garage, attic, basement] | [80, 92.3, 73.9] |
To see exactly how much space we can save, let's run compression on a table with 400 rows, 50 rows per day for the last seven days, that looks like this:
CREATE TABLE conditions ( time TIMESTAMPTZ NOT NULL, location TEXT NOT NULL, temperature DOUBLE PRECISION NULL, ); SELECT create_hypertable('conditions', 'time');
Next, we'd add a compression policy to run compression on
conditions for rows older than one day:
SELECT add_compression_policy('conditions', INTERVAL '1 day')
In the Timescale platform, if we navigate to the Explorer tab under Services, we’d see our table shrink from 72 kB to 16 kB—78% savings!
This is a simple example, but it shows the potential that Timescale compression has to reduce storage space.
Monitor dead tuples
A great practice to ensure you’re using as little storage as possible is to consistently monitor the number of dead tuples in each table.This is the first step towards putting together an efficient PostgreSQL storage management strategy.
To see pages and tuples in action, you can use
pgstattuple(), an extension provided by the Postgres maintainers to gain insights into how our tables manage tuples:
CREATE EXTENSION IF NOT EXISTS pgstattuple;
If you run the following query,
Postgres would give you a table of helpful information in response:
table_len | tuple_count | tuple_len | tuple_percent | dead_tuple_count | dead_tuple_len | dead_tuple_percent | free_space | free_percent -----------+-------------+-----------+---------------+-----------------+----------------+--------------------+------------+-------------- 81920000 | 500000 | 40000000 | 48.8 | 10000 | 1000000 | 1.2 | 300000 | 0.4
table_lentells you how big your table is in bytes, including data, indexes, toast tables, and free space.
dead_tuple_lentells how much space is being occupied by dead tuples which can be reclaimed by vacuuming.
free_spaceindicates the unused space within the allocated pages of the table.. Take note that
free_spacewill reset for every new page created.
You can also perform calculations or transformations on the result to make the information more understandable. For example, this query calculates the ratios of dead tuples and free space to the total table length, giving you a clearer perspective on the storage efficiency of your table:
SELECT (dead_tuple_len * 100.0 / table_len) as dead_tuple_ratio, (free_space * 100.0 / table_len) as free_space_ratio FROM pgstattuple('my_table');
Run autovacuum more frequently
If your table is experiencing table bloat, having autovacuum run more frequently may help you free up wasted storage space.
The default thresholds and values for autovacuum are in
postgresql.conf will change the autovacuum behavior for the whole Postgres instance. However, this practice is generally not recommended, since some tables will have a higher affinity for dead tuples than others.
Instead, you should update autovacuum's settings per table. For example, consider the following query:
ALTER TABLE my_table SET (autovacuum_vacuum_scale_factor = 0, autovacuum_vacuum_threshold = 200)
This will update
my_table to have autovacuum run after 200 tuples have been updated or deleted.
More information about additional autovacuum settings are in the PostgreSQL documentation. Each database and table will require different settings for how often autovacuum should run, but running vacuum often is a great way to reduce storage space.
Also, keep an eye on long-running transactions that might block autovacuum, leading to issues. You can use PostgreSQL’s
pg_stat_activity view to identify such transactions, canceling them if necessary to allow autovacuum to complete its operations efficiently:
SELECT pid, NOW() - xact_start AS duration, query, state FROM pg_stat_activity WHERE (NOW() - xact_start) > INTERVAL '5 minutes'; #Cancelling SELECT pg_cancel_backend(pid);
You could also inspect long-running vacuum processes and adjust the
autovacuum_work_mem parameter to increase the memory allocation for each autovacuum invocation, as we discussed in our article about PostgreSQL fine tuning.
Reclaim unused pages
Autovacuum and vacuum will free up dead tuples, but you’ll need the big guns to clean up unused pages.
As we saw previously, running
VACUUM FULL my_table will reclaim pages, but it has a significant problem: it exclusively locks the entire table. A table running
VACUUM FULL cannot be read or written to while the vacuum has the lock, which can take a long time to finish. This is usually an instant no-go for any production database.
The PostgreSQL community has a solution, pg_repack.
pg_repack is an extension that will clean up unused pages and bloat from a table by cloning a given table, swapping the original table with the new table, and then deleting the old table. All these operations are done with minimal exclusive locks, leading to less downtime.
At the end of the
pg_repack process, the pages associated with the original table become deleted from storage, and the new table only has the absolute minimum number of pages to store its rows, thus freeing table bloat.
Find unused indexes
As we mention in this article on idexing design, over-indexing is a frequent issue in many large PostgreSQL databases. Indexes consume disk space, so removing unused or underutilized indexes will help you keep your PostgreSQL database lean.
You can use
pg_stat_user_indexes to spot opportunities:
SELECT relname AS table_name, indexrelname AS index_name, pg_size_pretty(pg_relation_size(indexrelid)) AS index_size, idx_scan AS index_scan_count FROM pg_stat_user_indexes WHERE idx_scan < 50 -- Choose a threshold that makes sense for your application. ORDER BY index_scan_count ASC, pg_relation_size(indexrelid) DESC;
(This query looks for indexes with fewer than 50 scans, but this is an arbitrary number. You should adjust it based on your own usage patterns.)
Arrange columns by data type (from largest to smallest)
In PostgreSQL, storage efficiency is significantly influenced by the ordering of columns, which is closely related to alignment padding determined by the size of the column types. Each data type is aligned at memory addresses that are multiples of their size.
This alignment is systematic, ensuring that data retrieval is efficient and that the architecture adheres to specific memory and storage management protocols. But this can also lead to unused spaces, as the alignment necessitates padding to meet the address multiple criteria.
The way to fix this is to strategically order you columns from the largest to the smallest data type in your table definitions. This practical tip will help you minimize wasted space. Check out this article for a more in-depth explanation.
Delete old data regularly
You should always ask yourself: how long should I keep data around? Setting up data retention policies is essential for managing storage appropriately. Your users may not need data older than a year ago. Deleting old, unused records and indexes regularly is an easy win to reduce your database size.
Timescale can automatically delete old data for us using retention policies. Timescale’s hypertables are automatically partitioned by time, which helps a lot with data retention. Retention policies automatically delete partitions (which are called chunks in Timescale) once the data contained in such partition is older than a given interval.
You can create a retention policy by running:
SELECT add_retention_policy('my_table', INTERVAL '24 hours');
In this snippet, Timescale would delete chunks older than 24 hours from
We examined how table bloat and dead tuples can contribute to wasted storage space, which not only affects your pocket but also the performance of your large PostgreSQL tables.
To make sure you’re reducing your PostgreSQL database size down to its minimum, make sure to enable Timescale compression, to use data retention policies, and to set up a maintenance routine to periodically and effectively delete your dead tuples and reclaim your unused pages.
All these techniques together provide a holistic approach to maintaining a healthy PostgreSQL database and keeping your PostgreSQL database costs low.