Tips & tricks for optimising Spring Data & JPA queries
When you code enterprise applications on top of the Java Platform, most of the time, you use ORMs to interface them with relational databases. They bring a lot of simplicity which make you forget SQL queries syntax. Furthermore, most of the time, Java developers don’t really care/know what is under the hood of Spring Data and Java Persistence API (JPA) or such a facility. In my opinion, it’s mainly due to all the features provided by these specifications and frameworks.
Unfortunately, when your dataset is coming to grow, querying against your database could be difficult. Among other things, the different queries run by your Java application may potentially break your SLOs.
In this article, I have tried to write down a bunch of tips & tricks to tackle this issue. Even if some are related to Spring Data, I think you can use most of them if you use JPA in a standard way.
You will see that even if we can consider using JPA easy at first glance, it can bring a lot of complexity.
I would like to thank my colleagues Max Beckers, David Pequegnot & Peter Steiner for reviewing my article and giving their advices, useful links & tips.
Thanks also to Alexis Hassler for the tip about Hibernate 6 logging.
1 Observe your application
1.1 Observe your persistence layer
First and foremost, you MUST trace and monitor your persistence layer usage. If you use Hibernate (without Spring, Quarkus,…), you can get useful information configuring the logger:
<logger name="org.hibernate.SQL">
<level value="debug"/>
</logger>
If you use Spring (and JPA, Hibernate), you can also get them adding these configuration properties:
logging.level.org.hibernate.stat=TRACE
logging.level.org.hibernate.SQL=DEBUG
logging.level.org.hibernate.orm.jdbc.bind=TRACE
logging.level.org.hibernate.SQL_SLOW=TRACE
spring.jpa.properties.hibernate.generate_statistics=true
spring.jpa.properties.hibernate.format_sql=false
After getting all the queries and operations done by your persistence layer, you will be able to pinpoint which component is responsible for slowing down your queries. To cut long story short, which one is guilty? The database or the ORM.
In the case of huge SQL queries, I usually execute them directly in SQL using the database tools to check if I have the same behaviour.
Example of such an output:
2024-03-21T22:14:46.853+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select b1_0.id,b1_0.description,b1_0.isbn_10,b1_0.isbn_13,b1_0.medium_image_url,b1_0.nb_of_pages,b1_0.price,b1_0.rank,b1_0.small_image_url,b1_0.store_id,b1_0.title,b1_0.year_of_publication from book b1_0
2024-03-21T22:14:46.875+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select s1_0.id,s1_0.name,b1_0.store_id,b1_0.id,b1_0.description,b1_0.isbn_10,b1_0.isbn_13,b1_0.medium_image_url,b1_0.nb_of_pages,b1_0.price,b1_0.rank,b1_0.small_image_url,b1_0.title,b1_0.year_of_publication from store s1_0 left join book b1_0 on s1_0.id=b1_0.store_id where s1_0.id=?
2024-03-21T22:14:46.897+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.900+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.902+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.904+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.906+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.908+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.909+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.911+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.913+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.914+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.916+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.917+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.918+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-21T22:14:46.919+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] o.h.stat.internal.StatisticsImpl : HHH000117: HQL: [CRITERIA] select b1_0.id,b1_0.description,b1_0.isbn_10,b1_0.isbn_13,b1_0.medium_image_url,b1_0.nb_of_pages,b1_0.price,b1_0.rank,b1_0.small_image_url,b1_0.store_id,b1_0.title,b1_0.year_of_publication from book b1_0, time: 79ms, rows: 13
2024-03-21T22:14:47.007+01:00 DEBUG 39814 --- [optimization-jpa] [nio-8080-exec-1] org.hibernate.SQL : select b1_0.authors_id,b1_1.id,b1_1.description,b1_1.isbn_10,b1_1.isbn_13,b1_1.medium_image_url,b1_1.nb_of_pages,b1_1.price,b1_1.rank,b1_1.small_image_url,s1_0.id,s1_0.name,b1_1.title,b1_1.year_of_publication from book_authors b1_0 join book b1_1 on b1_1.id=b1_0.books_id left join store s1_0 on s1_0.id=b1_1.store_id where b1_0.authors_id=?
2024-03-21T22:14:47.038+01:00 WARN 39814 --- [optimization-jpa] [nio-8080-exec-1] .w.s.m.s.DefaultHandlerExceptionResolver : Ignoring exception, response committed already: org.springframework.http.converter.HttpMessageNotWritableException: Could not write JSON: Infinite recursion (StackOverflowError)
2024-03-21T22:14:47.039+01:00 WARN 39814 --- [optimization-jpa] [nio-8080-exec-1] .w.s.m.s.DefaultHandlerExceptionResolver : Resolved [org.springframework.http.converter.HttpMessageNotWritableException: Could not write JSON: Infinite recursion (StackOverflowError)]
2024-03-21T22:14:47.040+01:00 INFO 39814 --- [optimization-jpa] [nio-8080-exec-1] i.StatisticalLoggingSessionEventListener : Session Metrics {
869000 nanoseconds spent acquiring 1 JDBC connections;
0 nanoseconds spent releasing 0 JDBC connections;
2700503 nanoseconds spent preparing 16 JDBC statements;
16624802 nanoseconds spent executing 16 JDBC statements;
0 nanoseconds spent executing 0 JDBC batches;
0 nanoseconds spent performing 0 L2C puts;
0 nanoseconds spent performing 0 L2C hits;
0 nanoseconds spent performing 0 L2C misses;
0 nanoseconds spent executing 0 flushes (flushing a total of 0 entities and 0 collections);
190300 nanoseconds spent executing 1 partial-flushes (flushing a total of 0 entities and 0 collec
1.1.1 Dig into you datasource connection pool configuration
If you want to dive into your datasource and get clear insights of your database connection pool, you can also add Prometheus metrics to your application to observe it.
For a Spring Boot application, you can easily enable it adding two dependencies:
implementation 'org.springframework.boot:spring-boot-starter-actuator'
runtimeOnly 'io.micrometer:micrometer-registry-prometheus'
and these properties:
management.endpoint.prometheus.enabled=true
management.endpoints.web.exposure.include=health,info,prometheus
After rebooting your application, you will be able to get the status of your connection pool:
❯ http :8080/actuator/prometheus | grep hikari
# HELP hikaricp_connections Total connections
# TYPE hikaricp_connections gauge
hikaricp_connections{pool="HikariPool-1",} 10.0
# HELP hikaricp_connections_min Min connections
# TYPE hikaricp_connections_min gauge
hikaricp_connections_min{pool="HikariPool-1",} 10.0
# HELP hikaricp_connections_creation_seconds_max Connection creation time
# TYPE hikaricp_connections_creation_seconds_max gauge
hikaricp_connections_creation_seconds_max{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_creation_seconds Connection creation time
# TYPE hikaricp_connections_creation_seconds summary
hikaricp_connections_creation_seconds_count{pool="HikariPool-1",} 0.0
hikaricp_connections_creation_seconds_sum{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_timeout_total Connection timeout total count
# TYPE hikaricp_connections_timeout_total counter
hikaricp_connections_timeout_total{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_active Active connections
# TYPE hikaricp_connections_active gauge
hikaricp_connections_active{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_max Max connections
# TYPE hikaricp_connections_max gauge
hikaricp_connections_max{pool="HikariPool-1",} 10.0
# HELP hikaricp_connections_usage_seconds Connection usage time
# TYPE hikaricp_connections_usage_seconds summary
hikaricp_connections_usage_seconds_count{pool="HikariPool-1",} 0.0
hikaricp_connections_usage_seconds_sum{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_usage_seconds_max Connection usage time
# TYPE hikaricp_connections_usage_seconds_max gauge
hikaricp_connections_usage_seconds_max{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_pending Pending threads
# TYPE hikaricp_connections_pending gauge
hikaricp_connections_pending{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_idle Idle connections
# TYPE hikaricp_connections_idle gauge
hikaricp_connections_idle{pool="HikariPool-1",} 10.0
# HELP hikaricp_connections_acquire_seconds Connection acquire time
# TYPE hikaricp_connections_acquire_seconds summary
hikaricp_connections_acquire_seconds_count{pool="HikariPool-1",} 0.0
hikaricp_connections_acquire_seconds_sum{pool="HikariPool-1",} 0.0
# HELP hikaricp_connections_acquire_seconds_max Connection acquire time
# TYPE hikaricp_connections_acquire_seconds_max gauge
hikaricp_connections_acquire_seconds_max{pool="HikariPool-1",} 0.0
Obviously, it’s not recommended to use these metrics as is. Scrap them with Prometheus and Grafana to gather these metrics and create dashboards.
At this stage, you did the easiest part. You must now dig into the database documentation and measure, regarding your use case and the volumetry what are the good figures for every parameter.
If you use HikariCP, you can refer yourself to this guide delving into Pool sizing configuration.
1.2 Observe your database
As Java developers, we usually forget that database platforms provide valuable tools to analyse your queries. Once you have pointed out the time/resource consuming queries, you must check if your database query is time-consuming by, for instance, running a full scan of your table.
In this purpose, you can check the SQL queries execution plan.
If you use PostgreSQL (what else), you can get these insights using the EXPLAIN
command.
2 Check your entities associations
Let’s go back to our Java application. One of the main points of attention of any JPA (and SQL) queries is how your entity is joined with others. Every jointure brings costs and complexity.
For JPA queries, you must check first if your relationship between two objects should be either EAGER
or LAZY
.
You probably understood: there is no free lunch. You must measure first the JPA queries and mapping time-consumption and check which solution is the best.
By default, EAGER relationship are set up for @ManyToOne
and @OneToOne
. LAZY are for @OneToMany
.
Most of the time, I keep using the default configuration.
However, you must take care of the whole entity graph loaded by your query.
Does your entity loaded by a @OneToOne
relationship loads also a @OneToMany
relationship in a EAGER
way?
It’s the kind of question you will need to answer.
2.1 The famous N+1 issue
In this example, we will look into a 1-N
relationship:
@Entity
public class Store{
[...]
@OneToMany(fetch = FetchType.EAGER,mappedBy = "store")
private List<Book> books;
[...]
@Entity
public class Book {
[...]
@ManyToOne(targetEntity = Store.class)
private Store store;
[...]
If you remember well, this relationship is fetched in a EAGER way.
When I try to get all the stores using a findAll()
method:
For example:
public List<Store> findAllStores() {
return storeRepository.findStores().stream().toList();
}
Hibernate will query the database in this way:
- 1 query to select the main entity
- N queries for the entities linked by the jointure
In our case, we can see the following queries in the logs:
Hibernate: select s1_0.id,s1_0.name from store s1_0
Hibernate: select b1_0.store_id,b1_0.id,b1_0.description,b1_0.isbn_10,b1_0.isbn_13,b1_0.medium_image_url,b1_0.nb_of_pages,b1_0.price,b1_0.rank,b1_0.small_image_url,b1_0.title,b1_0.year_of_publication from book b1_0 where b1_0.store_id=?
Hibernate: select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
Hibernate: select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
Hibernate: select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
Hibernate: select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
[...]
It’s unfortunately not finished yet.
Imagine now, your book entity is related to another one in a EAGER way.
@ManyToMany(fetch = FetchType.EAGER)
private List<Author> authors;
You will execute then another SQL queries.
For instance, in this case:
Hibernate: select b1_0.authors_id,b1_1.id,b1_1.description,b1_1.isbn_10,b1_1.isbn_13,b1_1.medium_image_url,b1_1.nb_of_pages,b1_1.price,b1_1.rank,b1_1.small_image_url,s1_0.id,s1_0.name,b1_1.title,b1_1.year_of_publication from book_authors b1_0 join book b1_1 on b1_1.id=b1_0.books_id left join store s1_0 on s1_0.id=b1_1.store_id where b1_0.authors_id=?
2.2 Use a dedicated entity graph
If you are still struggling with the way Hibernate loads your Entity graph, you can also try to specify the graph of entities to load by yourself.
This feature introduced in JPA 2.1 can help you avoid retrieving specific useless attributes or optimise the loading of the linked entities.
Let’s go back to our application. Imagine that in one use case, when we fetch a list of books, we don’t need the list of authors. Using this API we can avoid fetching it in this way
@Entity
@NamedEntityGraph(name = "store[books]",
attributeNodes = @NamedAttributeNode("books")
)
public class Store implements Serializable {
[...]
@Repository
public interface StoreRepository extends JpaRepository<Store,Long> {
@EntityGraph(value = "store[books]")
Optional<Store> findByName(String name);
}
You will get the following output:
2024-03-22T14:35:17.515+01:00 DEBUG 74072 --- [optimization-jpa] [nio-8080-exec-3] org.hibernate.SQL : select s1_0.id,b1_0.store_id,b1_0.id,b1_0.description,b1_0.isbn_10,b1_0.isbn_13,b1_0.medium_image_url,b1_0.nb_of_pages,b1_0.price,b1_0.rank,b1_0.small_image_url,b1_0.title,b1_0.year_of_publication,s1_0.name from store s1_0 left join book b1_0 on s1_0.id=b1_0.store_id where s1_0.name=?
2024-03-22T14:35:17.537+01:00 DEBUG 74072 --- [optimization-jpa] [nio-8080-exec-3] o.h.stat.internal.StatisticsImpl : HHH000117: HQL: [CRITERIA] select s1_0.id,b1_0.store_id,b1_0.id,b1_0.description,b1_0.isbn_10,b1_0.isbn_13,b1_0.medium_image_url,b1_0.nb_of_pages,b1_0.price,b1_0.rank,b1_0.small_image_url,b1_0.title,b1_0.year_of_publication,s1_0.name from store s1_0 left join book b1_0 on s1_0.id=b1_0.store_id where s1_0.name=?, time: 21ms, rows: 1
2024-03-22T14:35:17.559+01:00 DEBUG 74072 --- [optimization-jpa] [nio-8080-exec-3] org.hibernate.SQL : select a1_0.books_id,a1_1.id,a1_1.firstname,a1_1.lastname,a1_1.public_id from book_authors a1_0 join author a1_1 on a1_1.id=a1_0.authors_id where a1_0.books_id=?
2024-03-22T14:35:17.565+01:00 DEBUG 74072 --- [optimization-jpa] [nio-8080-exec-3] org.hibernate.SQL : select b1_0.authors_id,b1_1.id,b1_1.description,b1_1.isbn_10,b1_1.isbn_13,b1_1.medium_image_url,b1_1.nb_of_pages,b1_1.price,b1_1.rank,b1_1.small_image_url,s1_0.id,s1_0.name,b1_1.title,b1_1.year_of_publication from book_authors b1_0 join book b1_1 on b1_1.id=b1_0.books_id left join store s1_0 on s1_0.id=b1_1.store_id where b1_0.authors_id=?
2024-03-22T14:35:17.582+01:00 WARN 74072 --- [optimization-jpa] [nio-8080-exec-3] .w.s.m.s.DefaultHandlerExceptionResolver : Ignoring exception, response committed already: org.springframework.http.converter.HttpMessageNotWritableException: Could not write JSON: Infinite recursion (StackOverflowError)
2024-03-22T14:35:17.583+01:00 WARN 74072 --- [optimization-jpa] [nio-8080-exec-3] .w.s.m.s.DefaultHandlerExceptionResolver : Resolved [org.springframework.http.converter.HttpMessageNotWritableException: Could not write JSON: Infinite recursion (StackOverflowError)]
2024-03-22T14:35:17.583+01:00 INFO 74072 --- [optimization-jpa] [nio-8080-exec-3] i.StatisticalLoggingSessionEventListener : Session Metrics {
571600 nanoseconds spent acquiring 1 JDBC connections;
0 nanoseconds spent releasing 0 JDBC connections;
954800 nanoseconds spent preparing 3 JDBC statements;
2433201 nanoseconds spent executing 3 JDBC statements;
0 nanoseconds spent executing 0 JDBC batches;
0 nanoseconds spent performing 0 L2C puts;
0 nanoseconds spent performing 0 L2C hits;
0 nanoseconds spent performing 0 L2C misses;
0 nanoseconds spent executing 0 flushes (flushing a total of 0 entities and 0 collections);
0 nanoseconds spent executing 0 partial-flushes (flushing a total of 0 entities and 0 collections)
}
2.3 Create a dedicated entity to reduce the number of attributes
One another misconception about JPA is to always fully map all the properties of table to its corresponding entity!
A gentle reminder: we don’t need to map all the columns in an entity! For instance, if your table has 30 columns, and you only need 10 in your use case, why querying, fetching and storing in memory all of these data?
That’s why I usually recommend to have, when it’s relevant, a dedicated entity for specific use cases. It could be lighter than the regular one and enhance the performances of your application.
For instance, if we have a regular Book
entity:
@Entity
public class Book implements Serializable {
@NotNull
private String title;
@Column(name = "isbn_13")
private String isbn13;
@Column(name = "isbn_10")
private String isbn10;
@ManyToMany(fetch = FetchType.EAGER, cascade = {CascadeType.PERSIST, CascadeType.MERGE, CascadeType.REFRESH, CascadeType.DETACH})
private List<Author> authors;
@Column(name = "year_of_publication")
private Integer yearOfPublication;
@Column(name = "nb_of_pages")
private Integer nbOfPages;
@Min(1)
@Max(10)
private Integer rank;
private BigDecimal price;
@Column(name = "small_image_url")
private URL smallImageUrl;
@Column(name = "medium_image_url")
private URL mediumImageUrl;
@Column(length = 10000)
private String description;
@Id
@GeneratedValue(strategy = GenerationType.IDENTITY)
private Long id;
@ManyToOne(targetEntity = Store.class)
private Store store;
@Transient
private transient String excerpt;
[...]
We can shrink it with only the required attributes:
@Entity
public class Book implements Serializable {
@NotNull
private String title;
@Column(name = "isbn_13")
private String isbn13;
@Column(name = "isbn_10")
private String isbn10;
@Column(name = "year_of_publication")
private Integer yearOfPublication;
@Column(name = "nb_of_pages")
private Integer nbOfPages;
@Min(1)
@Max(10)
private Integer rank;
private BigDecimal price;
private String description;
@Id
@GeneratedValue(strategy = GenerationType.IDENTITY)
private Long id;
@Transient
private transient String excerpt;
2.4 Use JOIN FETCH in your queries
Now, one another strategy is to manually control/declare the jointures and how different entities would be fetched by your queries.
To do that, you can use the JOIN FETCH
instruction:
For example:
@Query(value = "from Store store JOIN FETCH store.books books")
Set<Store> findStores();
In this manner, you can shrink the number of queries done from N+1 to only one. However, you MUST check and measure if it’s worth it. Sometimes, this kind of query can be more time-consuming in either database or in the JVM than several small ones.
2.5 Use a DTO or a tuple
Imagine we have a screen with of list of data coming from several entities. Instead of fetching all of these, and struggling with fetching strategies, we can also run DTO (or tuple) projections.
You can therefore select all (and only) the data you need with only ONE query. To get your code even clearer, you can also use records to make your data immutable.
public record BookDto (Long id, String description) {
}
You can get a set of this record writing the query:
@Query(value = "select new info.touret.query.optimizationjpa.BookDto(b.id, b.description) from Book b")
Set<BookDto> findAllDto();
3 Other optimisation tracks
3.1 Avoid transactions while reading our database with the annotation @Transactional(readonly=true)
One thing we often (again) remember: read-only database operations don’t need transactions!
In the good old days, it was already a good practice to set up two different datasources for the persistence context: one read-only avoiding database transactions and one which allowed it.
Anyway, you can now declare your service only reads data and doesn’t need to open a database transaction using the @Transactional(readonly=true)
annotation.
By the way, this feature goes well with using dedicated entities as mentioned above. For a specific search/query use case, you can use both of them to make your code even more straightforward.
3.2 Pagination w/ Spring Data
When you browse a large dataset, it’s usually a good practice to paginate results. The good news when you use Spring Data, is you have all the features included by default. The bad news is you may have time/cpu-consuming queries run for calculating the number of elements, pages and the position of the current result’s page.
If getting the number of pages is useless for you, you can switch to Slices instead of Pages.
When using this feature, you will only know if there is another slice available onwards or backwards through the methods hasNext()
and hasPrevious()
.
You will find below good links talking about it on StackOverflow:
- https://stackoverflow.com/questions/49918979/page-vs-slice-when-to-use-which
- https://stackoverflow.com/questions/12644749/way-to-disable-count-query-from-pagerequest-for-getting-total-pages
3.3 Caching specific data
You may use and query specific which is not daily (or monthly) updated. For instance, the department, country tables. In this case, you may want to cache them in memory (i.e., Second-Level cache.
With JPA you can quickly cache specific entities using the @Cacheable
annotation.
For instance, in a Spring Boot application, you must configure your cache first:
@SpringBootApplication
@EnableCaching
public class OptimizationJpaApplication {
public static void main(String[] args) {
SpringApplication.run(OptimizationJpaApplication.class, args);
}
@Bean
public Caffeine caffeineConfig() {
return Caffeine.newBuilder().expireAfterWrite(1, TimeUnit.DAYS);
}
@Bean
public CacheManager cacheManager(Caffeine caffeine) {
CaffeineCacheManager caffeineCacheManager = new CaffeineCacheManager();
caffeineCacheManager.setCaffeine(caffeine);
return caffeineCacheManager;
}
}
And declare your methods which use your cache:
@Query(value = "from Store store JOIN FETCH store.books books")
@Cacheable("stores")
Set<Store> findStores();
You will use your cache then and get the following logs after the second try:
2024-03-23T22:01:41.299+01:00 WARN 65315 --- [optimization-jpa] [nio-8080-exec-3] .w.s.m.s.DefaultHandlerExceptionResolver : Ignoring exception, response committed already: org.springframework.http.converter.HttpMessageNotWritableException: Could not write JSON: Infinite recursion (StackOverflowError)
2024-03-23T22:01:41.300+01:00 WARN 65315 --- [optimization-jpa] [nio-8080-exec-3] .w.s.m.s.DefaultHandlerExceptionResolver : Resolved [org.springframework.http.converter.HttpMessageNotWritableException: Could not write JSON: Infinite recursion (StackOverflowError)]
2024-03-23T22:01:41.300+01:00 INFO 65315 --- [optimization-jpa] [nio-8080-exec-3] i.StatisticalLoggingSessionEventListener : Session Metrics {
0 nanoseconds spent acquiring 0 JDBC connections;
0 nanoseconds spent releasing 0 JDBC connections;
0 nanoseconds spent preparing 0 JDBC statements;
0 nanoseconds spent executing 0 JDBC statements;
0 nanoseconds spent executing 0 JDBC batches;
0 nanoseconds spent performing 0 L2C puts;
0 nanoseconds spent performing 0 L2C hits;
0 nanoseconds spent performing 0 L2C misses;
0 nanoseconds spent executing 0 flushes (flushing a total of 0 entities and 0 collections);
0 nanoseconds spent executing 0 partial-flushes (flushing a total of 0 entities and 0 collections)
}
If you want to dig into the differences between Spring cache support & the JSR 107, you can check out this documentation.
3.4 In case of emergency: break the glass!
OK, none of all the tips exposed in this article has worked?
Now, remember that, at the end of the day, you use a database. It comes with many tools which may run your queries at lightning speed.
You can use SQL views or SQL materialized views to specify the data you want to fetch. In addition, feel free to use Native queries , Named Native Queries or Stored Procedure Queries (ONLY FOR) for the 10-20% of your most time-consuming queries. At the end of the day, you won’t be faster using an ORM!
For instance, when you use a SQL view, you can, with no effort, run either a native query or fetch a DTO or a tuple (see above):
Here is a trivial example to illustrate it:
@Query(value="select * from Store s where s.name= :name", nativeQuery=true)
Optional<Store> findByName(String name);
4 Conclusion
If you reached this last chapter, you would see there are plenty of solutions to fix ORM/JPA performance issues. I aimed at a summary of the most efficient solutions for type of problem. As a matter of fact, I’m pretty sure there are other ones.
Anyway, the first thing to put in place, is the whole observability stack: through logging, traces or prometheus metrics you will get deep insights of your application. Check also your database to see if you have a “full table scan” when you run specific SQL queries. It will help you find where is the bottleneck.
Last but not least, don’t rush into such optimisations (e.g., native queries)! Observe your application first to figure out if it’s worth it.
Don’t forget that any Premature optimisation is the root of all evil!