You know that a fast site == happier users, improved ranking from Google, and increased conversions. Maybe you even think your WordPress site is as fast as it can be – you’ve looked at site performance, from the best practices of setting up a server, to troubleshooting slow code, and offloading your images to a CDN, but is that everything?
With dynamic database-driven websites like WordPress, you might still have one problem on your hands: database queries slowing down your site.
In this post, I’ll take you through how to identify the queries causing bottlenecks, how to understand the problems with them, along with quick fixes and other approaches to speed sql query execution time. I’ll be using an actual query we recently tackled that was slowing things down on the customer portal of deliciousbrains.com.
Identification of Slow SQL Queries
The first step in fixing slow SQL queries is to find them. Ashley has sung the praises of the debugging plugin Query Monitor on the blog before, and it’s the databases queries feature of the plugin that really makes it an invaluable tool for identifying slow SQL queries and improving database performance. The plugin reports on all the database queries executed during the page request. It allows you to filter them by the code or component (the plugin, theme or WordPress core) calling them, and highlights duplicate and slow queries:
If you don’t want to install a debugging plugin on a production site (maybe you’re worried about adding some performance overhead) you can opt to turn on the MySQL Slow Query Log, which logs all queries that take a certain amount of time to execute. This is relatively simple to configure and set up where to log the queries to. As this is a server-level tweak, the performance hit will be less than a debugging plugin on the site, but should be turned off when not using it.
Understanding Slow SQL Queries
Once you have found an expensive query that you want to improve, the next step is to try to understand what is making the query slow. Recently during development to our site, we found a query that was taking around 8 seconds to execute!
SELECT
l.key_id,
l.order_id,
l.activation_email,
l.licence_key,
l.software_product_id,
l.software_version,
l.activations_limit,
l.created,
l.renewal_type,
l.renewal_id,
l.exempt_domain,
s.next_payment_date,
s.status,
pm2.post_id AS 'product_id',
pm.meta_value AS 'user_id'
FROM oiz6q8a_woocommerce_software_licences l
INNER JOIN oiz6q8a_woocommerce_software_subscriptions s
ON s.key_id = l.key_id
INNER JOIN oiz6q8a_posts p
ON p.ID = l.order_id
INNER JOIN oiz6q8a_postmeta pm
ON pm.post_id = p.ID
AND pm.meta_key = '_customer_user'
INNER JOIN oiz6q8a_postmeta pm2
ON pm2.meta_key = '_software_product_id'
AND pm2.meta_value = l.software_product_id
WHERE p.post_type = 'shop_order'
AND pm.meta_value = 279
ORDER BY s.next_payment_date
We use WooCommerce and a customized version of the WooCommerce Software Subscriptions plugin to run our plugins store. The purpose of this query is to get all subscriptions for a customer where we know their customer number. WooCommerce has a somewhat complex data model, in that even though an order is stored as a custom post type, the id of the customer (for stores where each customer gets a WordPress user created for them) is not stored as the post_author
, but instead as a piece of post meta data.
There are also a couple of inner joins to custom tables created by the software subscriptions plugin making this SQL statement more complex and harder to understand. Let’s dive in to understand the query more.
MySQL is your Friend for SQL Query Optimization
MySQL has a handy statement DESCRIBE
which can be used to output information about a table’s structure such as its columns, data types, defaults. So if you execute DESCRIBE wp_postmeta;
you will see the following results:
Field | Type | Null | Key | Default | Extra |
---|---|---|---|---|---|
meta_id | bigint(20) unsigned | NO | PRI | NULL | auto_increment |
post_id | bigint(20) unsigned | NO | MUL | 0 | |
meta_key | varchar(255) | YES | MUL | NULL | |
meta_value | longtext | YES | NULL |
That’s cool, but you may already know about it. But did you know that the DESCRIBE
statement prefix can actually be used on SELECT
, INSERT
, UPDATE
, REPLACE
and DELETE
statements? This is more commonly known by its synonym EXPLAIN
and will give us detailed information about how the statement will be executed.
Here are the results for our slow query:
id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra |
---|---|---|---|---|---|---|---|---|---|
1 | SIMPLE | pm2 | ref | meta_key | meta_key | 576 | const | 28 | Using where; Using temporary; Using filesort |
1 | SIMPLE | pm | ref | post_id,meta_key | meta_key | 576 | const | 37456 | Using where |
1 | SIMPLE | p | eq_ref | PRIMARY,type_status_date | PRIMARY | 8 | deliciousbrainsdev.pm.post_id | 1 | Using where |
1 | SIMPLE | l | ref | PRIMARY,order_id | order_id | 8 | deliciousbrainsdev.pm.post_id | 1 | Using index condition; Using where |
1 | SIMPLE | s | eq_ref | PRIMARY | PRIMARY | 8 | deliciousbrainsdev.l.key_id | 1 | NULL |
At first glance, this isn’t very easy to interpret. Luckily the folks over at SitePoint have put together a comprehensive guide to understanding the statement.
The most important column is type
, which describes how the tables are joined. If you see ALL
then that means MySQL is reading the whole table from disk, increasing I/O rates and putting load on the CPU. This is known as a “full table scan” – more on that later.
The rows
column is also a good indication of what MySQL is having to do, as this shows how many rows it has looked in to find a result.
Explain
also gives us more information we can use to optimize. For example, the pm2 table (wp_postmeta), it is telling us we are Using filesort
, because we are asking the results to be sorted using an ORDER BY
clause on the statement. If we were also grouping the query we would be adding overhead to the execution.
Visual Investigation with an Execution Plan
MySQL Workbench is another handy, free tool for this type of investigation. For databases running on MySQL 5.6 and above, the results of EXPLAIN
can be outputted as JSON, and MySQL Workbench turns that JSON into a visual execution plan of the statement:
It automatically draws your attention to issues by coloring parts of the query by cost. We can see straight away that the join to the wp_woocommerce_software_licences
(alias l) table has a serious issue. Let’s do some query tuning!
Solving Slow SQL Queries with Query Optimization
That part of the query is performing a full table scan, which you should try to avoid, as it uses a non-indexed column order_id
as the join between the wp_woocommerce_software_licences
table to the wp_posts
table. This is a common issue for slow queries and one that can be solved easily to improve query performance.
Indexes
order_id
is a pretty important piece of identifying data in the table, and if we are querying like this we really should have an index on the column, otherwise MySQL will literally scan each row of the table until it finds the rows needed. Let’s add an index and see what that does:
CREATE INDEX order_id ON wp_woocommerce_software_licences(order_id)
Wow, we’ve managed to shave over 5 seconds off the query by adding that index, good job! But why stop there, there are a few further optimization techniques we can use to reduce the execution time and reduce database server bottlenecks.
Know your Query
Examine the query – join by join, subquery by subquery. Does it do things it doesn’t need to? Can optimizations be made?
In this case we join the licenses table to the posts table using the order_id all the while restricting the statement to post types of shop_order
. This is to enforce data integrity to make sure we are only using the correct order records, however, it is actually a redundant part of the query. We know that it’s a safe bet that a software license row in the table has an order_id
relating to the WooCommerce order in the posts table, as this is enforced in PHP plugin code. Let’s remove the join and see if that improves things:
That’s not a huge saving but the query is now under 3 seconds.
Cache All The Things!
If your server doesn’t have MySQL query caching on by default then it is worth turning on. This means MySQL will keep a record of all statements executed with the result, and if an identical statement is subsequently executed the cached results are returned. The cache doesn’t get stale, as MySQL flushes the cache when tables are changed.
Query Monitor found our query to be running 4 times on a single page load, and although it’s good to have MySQL query caching on, duplicate reads to the database in one request should really be avoided full stop. Static caching in your PHP code is a simple and very effective way to solve this issue. Basically you are fetching the results of a query from the database the first time it is requested and storing them in static property of a class, and then subsequent calls will return the results from the static property:
class WC_Software_Subscription {
protected static $subscriptions = array();
public static function get_user_subscriptions( $user_id ) {
if ( isset( static::$subscriptions[ $user_id ] ) ) {
return static::$subscriptions[ $user_id ];
}
global $wpdb;
$sql = '...';
$results = $wpdb->get_results( $sql, ARRAY_A );
static::$subscriptions[ $user_id ] = $results;
return $results;
}
}
The cache has a lifespan of the request, more specifically that of the instantiated object. If you are looking at persisting query results across requests, then you would need to implement a persistent Object Cache. However, your code would need to be responsible for setting the cache, and invalidating the cache entry when the underlying data changes.
Thinking Outside the Box
There are other approaches we can take to try and speed up query execution that involve a bit more work than just tweaking the query or adding an index. One of the slowest parts of our query is the work done to join the tables to go from customer id to product id, and we have to do this for every customer. What if we did all that joining just once, so we could just grab the customer’s data when we need it?
You could denormalize the data by creating a table that stores the license data, along with the user id and product id for all licenses and just query against that for a specific customer. You would need to rebuild the table using MySQL triggers on INSERT/UPDATE/DELETE
to the licenses table (or others depending on how the data could change) but this would significantly improve the performance of querying that data.
Similarly, if a number of joins slow down your query in MySQL, it might be quicker to break the query into two or more statements and execute them separately in PHP and then collect and filter the results in code. Laravel does something similar by eager loading relationships in Eloquent.
WordPress can be prone to slower queries on the wp_posts
table, if you have a large amount of data, and many different custom post types. If you are finding querying for your post type slow, then consider moving away from the custom post type storage model and to a custom table.
We recently found the query getting slower and slower due to the ever-increasing size of our postmeta table (currently at 2.5 million rows). As the query joins twice to that table – once for the customer ID and once for the product ID, we decided to remove one of the joins.
We were already storing the ‘software_product_id’ in the licenses table, which is a string representation of the product the license is for, eg. WPMDB-DEV for the developer license of WP Migrate DB Pro. However, this abstraction from the actual WooCommerce product ID served no real purpose for our site, and so we replaced the ‘software_product_id’ column for a ‘product_id’ column, migrated the data, and updated all the code in our version of the WooCommerce software subscriptions addon to use the actual product ID.
This reduced some queries that were taking up to 4 seconds down to just 223 milliseconds! 💪
SQL Query Optimization Results
With these approaches to SQL query optimization we managed to take our query down from 8 seconds down to just over 2 seconds, and reduce the amount of times it was called from 4 to 1. As a note, those query times were recorded when running on our development environment and would be quicker on production. SQL query optimization is a part of a more complex process such as SQL Server performance tuning, learn more about it in this guide.
I hope this has been a helpful guide to tracking down slow queries and fixing them up. Query optimization might seem like a scary task, but as soon as you try it out and have some quick wins you’ll start to get the bug and want to improve things even further.
Do you have any tips for SQL query optimization or tools you like to use? Let us know in the comments.