In this article, I will establish how the SQL language and its implementations distinguish between
changed values and
modified values, where a
changed value is a value that has been “touched”, but not necessarily
modified, i.e. the value might be the same before and after the change.
Many ORMs, unfortunately, either update
all of a record’s values, or only the
modified ones. The first can be inefficient, and the latter can be wrong. Updating the
changed values would be correct.
Note that you may have a different definition of
changed and
modified. For this article, let’s just assume that the above definition is as valid as it is useful.
Introduction
A very interesting discussion was triggered recently by Vlad Mihalcea who was looking for an answer to this interesting question:
What’s the overhead of updating all columns, even the ones that haven’t changed?
Apart from the question being very interesting from a performance perspective,
the tweet also inspired functional aspects of a distinction between updating all columns vs. updating some columns, which I’ll summarise in this article.
What’s the Problem?
The problem is one that all ORM vendors need to solve: ORMs have a client side representation of the relational model, and that representation is cached (or “out of sync”) for a user to change and then persist again. The problem is now how to re-synchronise the client side representation with the server side representation in a consistent and correct way.
Sidenote: By ORM I understand any tool that maps from a client side representation of your database schema to the database schema itself, regardless if the product supports full-fledged JPA-style object graph persistence, or “merely” implements an “active record” pattern, such as jOOQ 3.x (I find that distinction a bit academic).
All such ORMs have a client side representation of a database record, for instance given the following table (I’m going to be using PostgreSQL syntax):
CREATE TABLE customer (
customer_id SERIAL8 NOT NULL PRIMARY KEY,
first_name VARCHAR(50) NOT NULL,
last_name VARCHAR(50) NOT NULL
)
You’re going to have a client side representation as the following (using Java, e.g. jOOQ or JPA):
// jOOQ generated UpdatableRecord
public class CustomerRecord
extends UpdatableRecordImpl<CustomerRecord> {
public CustomerRecord setCustomerId(Long customerId) { ... }
public Long getCustomerId() { ... }
public CustomerRecord setFirstName(String firstName) { ... }
public String getFirstName() { ... }
...
}
// JPA annotated entity
@Entity
public class Customer {
@Id
@GeneratedValue(strategy = IDENITITY)
public long customerId;
@Column
public String firstName;
...
}
In principle, these two approaches are the same thing with the distinction that jOOQ explicitly governs all
UpdatableRecord
interactions through type inheritance, whereas JPA makes this dependency more implicit through annotations:
- jOOQ – explicit behavioural dependency between entity and jOOQ logic
- JPA – implicit behavioural dependency between entity and JPA entity manager
In principle, the distinction is just a matter of taste, a programming style: Explicit vs. declarative.
But from a practical perspective, the JPA implementation lacks an important feature when it comes to synching the state back to the database. It cannot reflect
change, only
modification.
How to synch the state back to the database?
Let’s assume we have a customer called John Doe:
INSERT INTO customer (first_name, last_name)
VALUES ('John', 'Doe');
And that customer now changes their names to John Smith. We have several options of sending that update to the database, through “PATCH” or “PUT” semantics – terminology used by Morgan Tocker in another tweet in that discussion:
-- PATCH
UPDATE customer SET last_name = 'Smith' WHERE id = ?
-- PUT
UPDATE customer
SET first_name = 'John',
last_name = 'Smith'
WHERE customer_id = ?
A “PATCH” operation sends only the changed values back to the server, whereas a “PUT” operation sends the entire entity back to the server.
Discussion – Semantics.
In favour of PUT
The two operations are semantically very different. If another session attempts to rename this customer to Jane Doe concurrently (and without optimistic locking being in place), then the PATCH operation might result in an inconsistent outcome (Jane Smith), whereas the PUT operation would still produce one of the expected results, depending on what write is executed first:
-- PATCH result: Jane Smith
-- PATCH 1
UPDATE customer SET last_name = 'Smith' WHERE customer_id = ?
-- PATCH 2
UPDATE customer SET first_name = 'Jane' WHERE customer_id = ?
-- PUT result: Jane Doe
-- PUT 1
UPDATE customer
SET first_name = 'John',
last_name = 'Smith'
WHERE customer_id = ?
-- PUT 2
UPDATE customer
SET first_name = 'Jane',
last_name = 'Doe'
WHERE customer_id = ?
This is one of the reasons why Hibernate, as a JPA implementation, always implements PUT semantics by default, sending all the columns at once. You can opt out of this by using the
@DynamicUpdate
, which will only update
modified values (not “changed” values, I’ll explain this distinction later).
This makes perfect sense in such a trivial setup, but it is a short-sighted solution, when the table has many more columns. We’ll see right away why:
In favour of PATCH
One size doesn’t fit all. Sometimes, you do want concurrent updates to happen, and you do want to implement PATCH semantics, because sometimes, two concurrent updates do not work against each other. Take the following example using an enhancement of the customer table.
Business is asking us to collect some aggregate metrics for each customer. The number of clicks they made on our website, as well as the number of purchases they made:
CREATE TABLE customer (
customer_id SERIAL8 NOT NULL PRIMARY KEY,
first_name VARCHAR(50) NOT NULL,
last_name VARCHAR(50) NOT NULL,
clicks BIGINT NOT NULL DEFAULT 0,
purchases BIGINT NOT NULL DEFAULT 0
)
And, of course, once you agree that the above design is a suitable one, you’ll immediately agree that here, PATCH semantics is more desirable than PUT semantics:
-- Updating clicks
UPDATE customer SET clicks = clicks+1 WHERE customer_id = ?
-- Updating purchases
UPDATE customer SET purchases = purchases+1 WHERE customer_id = ?
Not only do we update only an individual column, we’re doing it entirely in SQL, including the calculation. With this approach, we do not even need optimistic locking to guarantee update correctness, as we’re not using any client side cached version of the customer record, which could be out of date and would need optimistic (or worse: pessimistic) locking.
If we implemented this differently, using client side calculation of the updated clicks / purchases counters…
-- Updating clicks
UPDATE customer
SET clicks = ?
WHERE customer_id = ?
-- Updating purchases
UPDATE customer
SET purchases = ?
WHERE customer_id = ?
… then we’d need one of these techniques:
- Pessimistic locking: Nope, won’t work. We could still get incorrect updates
- Optimistic locking: Indeed, any update would need to be done on a versioned customer record, so if there are two concurrent updates, one of them will fail and could try again. This guarantees data integrity, but will probably make this functionality very painful, because a lot of click updates are probably done in a short amount of time, and they would need to be repeated until they work!
- Client side synchronisation: Of course, we could prevent concurrency for these updates on the client side, making sure that only one concurrent process ever updates click counts (for a given customer). We could implement a click count update queue for this.
All of the above options have significant drawbacks, the easiest solution is really to just increment the counter directly in the database.
And don’t forget, if you choose a bind-variable based solution, and opt for updating
ALL the columns, rather than just the changed one, your first_name / last_name updates might conflict with these counter updates as well, making things even more complicated.
Partial PUT (or compound PATCH)
In fact, from a semantics perspective, if you do want to use an ORM to update an entity, you should think about a “partial PUT” semantics, which separates the different entity elements in “sub entities”. From a relational perspective, of course, no such thing as a subentity exists. The above example should be normalised into this, and we would have much less concurrency issues:
CREATE TABLE customer (
customer_id SERIAL8 NOT NULL PRIMARY KEY,
first_name VARCHAR(50) NOT NULL,
last_name VARCHAR(50) NOT NULL
);
CREATE TABLE customer_clicks
customer_id BIGINT NOT NULL PRIMARY KEY REFERENCES customer,
clicks BIGINT NOT NULL DEFAULT 0
);
CREATE TABLE customer_purchases
customer_id BIGINT NOT NULL PRIMARY KEY REFERENCES customer,
purchases BIGINT NOT NULL DEFAULT 0
);
This way, the previously mentioned PUT semantics would not create situations where individual, semantically unrelated updates (updates to names, updates to clicks) would interfere with each other. We would only need to make sure that e.g. two competing updates to clicks are correctly serialised.
Practically, we often don’t design our databases this way, either for convenience reasons, for optimised storage, for optimised querying (
see also our article when normalisation and surrogate keys hurt performance).
jOOQ’s “changed” value semantics
So that “sub entity” is really just a logical thing, which can be represented either as a logically separate entity in JPA, or we can use jOOQ, which works a bit differently here. In jOOQ, we can change an
UpdatableRecord
only partially, and that partial change is sent to the server:
CustomerRecord customer = ctx
.selectFrom(CUSTOMER)
.where(CUSTOMER.CUSTOMER_ID.eq(customerId))
.fetchOne();
customer.setFirstName("John");
customer.setLastName("Smith");
assertTrue(customer.changed(CUSTOMER.FIRST_NAME));
assertTrue(customer.changed(CUSTOMER.LAST_NAME));
assertFalse(customer.changed(CUSTOMER.CLICKS));
assertFalse(customer.changed(CUSTOMER.PURCHASES));
customer.store();
assertFalse(customer.changed(CUSTOMER.FIRST_NAME));
assertFalse(customer.changed(CUSTOMER.LAST_NAME));
assertFalse(customer.changed(CUSTOMER.CLICKS));
assertFalse(customer.changed(CUSTOMER.PURCHASES));
This will send the following statement to the server:
UPDATE customer
SET first_name = ?,
last_name = ?
WHERE customer_id = ?
Optionally, just as with JPA, you can turn on optimistic locking on this statement. The important thing here is that the
clicks
and
purchases
columns are left untouched, because they were not
changed by the client code. This is different from JPA, which either sends
all the values by default, or if you specify
@DynamicUpdate
in Hibernate, it would send only the
last_name
column, because while
first_name
was
changed it was not
modified.
My definition:
- changed: The value is “touched”, its state is “dirty” and the state needs to be synched to the database, regardless of modification.
- modified: The value is different from its previously known value. By necessity, a modified value is always changed.
As you can see, these are different things, and it is quite hard for a JPA-based API like Hibernate to implement
changed semantics because of the annotation-based declarative nature of how entities are defined. We’d need some sophisticated instrumentation to intercept all data
changes even when the values have not been
modified (I didn’t make those attributes public by accident).
Without this distinction, however, it is unreasonable to use
@DynamicUpdate
in Hibernate, as we might run into that situation we didn’t want to run into, where we get a customer called “Jane Smith” – or we use optimistic locking, in case of which there’s not much point in using
@DynamicUpdate
.
The database perspective
From a database perspective, it is also important to distinguish between
change and
modification semantics.
In the answer I gave on Stack Exchange, I’ve illustrated two situations:
INSERTs and DEFAULT values
Thus far, we’ve discussed only
UPDATE
statements, but similar reasoning may be made for
INSERT
as well. These two statements are the same:
INSERT INTO t (a, b) VALUES (?, ?);
INSERT INTO t (a, b, c) VALUES (?, ?, DEFAULT);
This one, however, is different:
INSERT INTO t (a, b, c) VALUES (?, ?, ?);
In the first case, a
DEFAULT
clause (e.g. timestamp generation, identity generation, trigger value generation, etc.) may apply to the column
c
. In the second case, the value
c
is provided explicitly by the client.
Languages like Java do not have any way to represent this distinction between
NULL
(which is usually, but not always, the DEFAULT
) in SQL
- an actual
DEFAULT
This can only be achieved when an ORM implements
changed semantics, like jOOQ does. When you create a customer with jOOQ, then
clicks
and
purchases
will have their
DEFAULT
applied:
CustomerRecord c1 = ctx.newRecord(CUSTOMER);
c1.setFirstName("John");
c1.setLastName("Doe");
c1.store();
CustomerRecord c2 = ctx.newRecord(CUSTOMER);
c2.setFirstName("Jane");
c2.setLastName("Smith");
c2.setClicks(1);
c2.setPurchases(1);
c2.store();
Resulting SQL:
-- c1.store();
INSERT INTO customer (first_name, last_name)
VALUES (?, ?);
-- c2.store();
INSERT INTO customer (first_name, last_name, clicks, purchases)
VALUES (?, ?, ?, ?);
In both cases, that’s what the user tells jOOQ to do, so jOOQ will generate a query accordingly.
Back to UPDATE statements
Consider the following example using Oracle triggers:
CREATE TABLE x (a INT PRIMARY KEY, b INT, c INT, d INT);
INSERT INTO x VALUES (1, 1, 1, 1);
CREATE OR REPLACE TRIGGER t
BEFORE UPDATE OF c, d -- Doesn't fire on UPDATE OF b!
ON x
BEGIN
IF updating('c') THEN
dbms_output.put_line('Updating c');
END IF;
IF updating('d') THEN
dbms_output.put_line('Updating d');
END IF;
END;
/
SET SERVEROUTPUT ON
UPDATE x SET b = 1 WHERE a = 1;
UPDATE x SET c = 1 WHERE a = 1;
UPDATE x SET d = 1 WHERE a = 1;
UPDATE x SET b = 1, c = 1, d = 1 WHERE a = 1;
It results in the following output:
table X created.
1 rows inserted.
TRIGGER T compiled
1 rows updated.
1 rows updated.
Updating c
1 rows updated.
Updating d
1 rows updated.
Updating c
Updating d
As you can see, the trigger doesn’t fire when we update only column
b
, which it is not interested in. Again, this goes in the direction of distinguishing between
changed and
modified values, where a trigger fires only when a value is
changed (but not necessarily
modified).
Now, if an ORM will always update
all the columns, this trigger will not work correctly. Sure, we can compare
:OLD.b
and
:NEW.b
, but that would check for
modification, not
change, and it might be costly to do so for large strings!
Speaking of costs…
Performance
Statement caching: Weakly in favour of PUT
While one of the reasons the Hibernate team mentioned in favour of updating
all the columns is improved cursor cache performance (fewer distinct SQL statements need to be parsed by the database as there are fewer distinct update configurations), I suggest that this “premature optimisation” is negligible. If a client application runs dynamic updates (in the jOOQ sense, where
changed values are updated, not just
modified values), then chances that the possible SQL statements that need to be parsed will explode are slim to non-existent.
I would definitely like to see real-world benchmarks on this topic!
Batching: Weakly in favour of PUT
When you want to batch tons of update statements from JDBC, then indeed, you will need to ensure that they all have the exact same SQL string. However, this is not a good argument in favour of using PUT semantics and updating all columns.
I’m saying “not good”, because such a batched update should still only consider a subset of the columns for update, not all the columns. And that subset should be determined on aggregated
changed flags, not data
modification.
Index updates: In favour of PATCH (depending on the database)
Most databases optimise index updates to ignore indexes whose columns have not been
changed. Oracle also doesn’t update indexes whose columns have not been
modified, in case of which PUT and PATCH semantics both work the same way from an indexing perspective. Other databases may not work this way, where PATCH semantics is favourable.
But even if the optimisation is in place, the old and the new values have to be compared for equality (i.e. to see if a
modification took place). You don’t want to compare millions of strings per second if there’s no need to do so! Check out
Morgan Tocker’s interesting answer on Stack Exchange, from a MySQL perspective
So, why not just prevent expensive
modification checks by telling the database what has
changed, instead?
UNDO overhead: In favour of PATCH
Every statement has a footprint on the UNDO / REDO logs. As I’ve shown above, the statements are semantically different in many ways, so if your statement is bigger (more columns are updated), then the impact on the UNDO / REDO log is bigger as well. This can have drastic effects depending on the size of your table / columns:
Don’t forget that this can also affect backup performance!
More performance related information in this blog post:
https://jonathanlewis.wordpress.com/2007/01/02/superfluous-updates
Note: While these bits of information were mostly Oracle-specific, common sense dictates that other RDBMS will behave in similar ways.
Conclusion
With all these negative aspects to including unnecessary columns for update through an ORM compared to the almost negligible benefits, I’d say that users should move forward and completely avoid this mess. Here’s how:
- jOOQ optimises this out of the box, if users set the changed values explicitly. Beware that when you “load” a POJO into a Record, it will set all the columns to changed, which may or may not be the desired effect!
- Hibernate allows for
@DynamicUpdate
, which may work incorrectly as we have minimal “PATCH” semantics based on modified values, not on changed values. However, JPA allows for declaring more than one entity per table, which might certainly be a valid option for this kind of problem
- Normalisation is always an option, with its own trade offs. The
clicks
and purchases
columns could be externalised in separate tables, if this benefits the overall design.
- More often than not, writing an UPDATE with SQL directly is the best choice. As we’ve seen in this article, the counters should be updated with expressions of the form
clicks = clicks + 1
, which circumvents most problems exposed in this article.
In short, as Michael Simons said:
And we all do feel very dirty when we write SELECT *
, right? So we should at least be wary of updating all the columns as well.
Like this:
Like Loading...
Very interesting! Some questions:
Couldn’t Hibernate generate proxies even for fully loaded entities and keep track of what has been changed instead of it determining what has been modified by comparing the values? The proxies do have some overhead, but so does the comparison.
It’d work using a write lock on the row for the whole transaction, wouldn’t it? This would seriously limit concurrency, but with some non-DB update aggregation, it should be no big deal (and losing a few seconds worth of statistics in case of a server crash or blackout is usually no big deal). Doing no aggregation sounds like a bad idea, anyway. For example, I’m storing counts of accessed web pages and doing one additional update per page sounds too crazy, doesn’t it?
Why can’t the redo logs be optimized as well, once the non-modification has been detected? Maybe because of the redo logs being written before the test can be done?
I’d also say, that the changed semantics isn’t always that nice. In introduces a difference between this simple statement
and the more complicated version
A more complicated case is a data import, I’m actually using. The server gets a CSV with maybe thousand rows and maybe ten columns describing the current catalog. It fetches the old version from the database, uses the itemNo for matching them, issues DELETEs for no more existing items, INSERTs for new items and updates the common ones which makes Hibernate issue the UPDATEs. If there was a changed semantics, it might lead to hundreds of different queries as oftentimes a few columns change. Luckily, the import is the only source of truth, so there are no concurrency issues.
No clue if this is possible with proxies. I mean, can you really intercept direct field access from client code (as in my example, where attributes are public and there aren’t any getters / setters)?
The lock wouldn’t prevent the update with “PATCH” semantics to be wrong. The only two solutions here are: “PUT” semantics, and/or optimistic locking. Think about it. An UPDATE already performs pessimistic locking, and blocks other writers. That doesn’t mean the other writers will run a correct write if they update only one of the first_name/last_name columns, but not the other.
Redo-logs: In principle, they could, but the assumption here (as assumed in https://jonathanlewis.wordpress.com/2007/01/02/superfluous-updates) is that this whole logging code is just way too messy to untangle for such an optimisation that can be implemented so easily with a smarter UPDATE statement. I wouldn’t be surprised…
I’m not sure why those two statements you’ve mentioned should be the same. Remember. We really WANT to be able to distinguish between changed and modified. If you want to avoid changed semantics, fine, you’ve shown how to do it. Again, the SQL
UPDATE
statement works exactly the same way, as I’ve shown with the trigger example. Changing a value (firing the trigger) and modifying the value aren’t the same thing in SQL.And again, the article does not imply at all that batch updates have to consider individual changed semantics on a row-by-row basis. The batch could easily aggregate all the changed flags into a global view. That’s what happens when you use SQL MERGE, btw. You’ll do “unify” the changing columns explicitly, regardless of individual value modifications.