C++ Object Persistence with ODB

Copyright © 2009-2015 Code Synthesis

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, version 1.3; with no Invariant Sections, no Front-Cover Texts and no Back-Cover Texts.

Revision 2.4, February 2015

This revision of the manual describes ODB 2.4.0 and is available in the following formats: XHTML, PDF, and PostScript.


Table of Contents

Preface
About This Document
More Information
PART I OBJECT-RELATIONAL MAPPING
1Introduction
1.1Architecture and Workflow
1.2Benefits
1.3Supported C++ Standards
2Hello World Example
2.1Declaring Persistent Classes
2.2Generating Database Support Code
2.3Compiling and Running
2.4Making Objects Persistent
2.5Querying the Database for Objects
2.6Updating Persistent Objects
2.7Defining and Using Views
2.8Deleting Persistent Objects
2.9Changing Persistent Classes
2.10Accessing Multiple Databases
2.11Summary
3Working with Persistent Objects
3.1Concepts and Terminology
3.2Declaring Persistent Objects and Values
3.3Object and View Pointers
3.4Database
3.5Transactions
3.6Connections
3.7Error Handling and Recovery
3.8Making Objects Persistent
3.9Loading Persistent Objects
3.10Updating Persistent Objects
3.11Deleting Persistent Objects
3.12Executing Native SQL Statements
3.13Tracing SQL Statement Execution
3.14ODB Exceptions
4Querying the Database
4.1ODB Query Language
4.2Parameter Binding
4.3Executing a Query
4.4Query Result
4.5Prepared Queries
5Containers
5.1Ordered Containers
5.2Set and Multiset Containers
5.3Map and Multimap Containers
5.4Change-Tracking Containers
5.4.1Change-Tracking vector
5.5Using Custom Containers
6Relationships
6.1Unidirectional Relationships
6.1.1To-One Relationships
6.1.2To-Many Relationships
6.2Bidirectional Relationships
6.2.1One-to-One Relationships
6.2.2One-to-Many Relationships
6.2.3Many-to-Many Relationships
6.3Circular Relationships
6.4Lazy Pointers
6.5Using Custom Smart Pointers
7Value Types
7.1Simple Value Types
7.2Composite Value Types
7.2.1Composite Object Ids
7.2.2Composite Value Column and Table Names
7.3Pointers and NULL Value Semantics
8Inheritance
8.1Reuse Inheritance
8.2Polymorphism Inheritance
8.2.1Performance and Limitations
8.3Mixed Inheritance
9Sections
9.1Sections and Inheritance
9.2Sections and Optimistic Concurrency
9.3Sections and Lazy Pointers
9.4Sections and Change-Tracking Containers
10Views
10.1Object Views
10.2Object Loading Views
10.3Table Views
10.4Mixed Views
10.5View Query Conditions
10.6Native Views
10.7Other View Features and Limitations
11Session
11.1Object Cache
11.2Custom Sessions
12Optimistic Concurrency
13Database Schema Evolution
13.1Object Model Version and Changelog
13.2Schema Migration
13.3Data Migration
13.3.1Immediate Data Migration
13.3.2Gradual Data Migration
13.4Soft Object Model Changes
13.4.1Reuse Inheritance Changes
13.4.2Polymorphism Inheritance Changes
14ODB Pragma Language
14.1Object Type Pragmas
14.1.1table
14.1.2pointer
14.1.3abstract
14.1.4readonly
14.1.5optimistic
14.1.6no_id
14.1.7callback
14.1.8schema
14.1.9polymorphic
14.1.10session
14.1.11definition
14.1.12transient
14.1.13sectionable
14.1.14deleted
14.1.15bulk
14.2View Type Pragmas
14.2.1object
14.2.2table
14.2.3query
14.2.4pointer
14.2.5callback
14.2.6definition
14.2.7transient
14.3Value Type Pragmas
14.3.1type
14.3.2id_type
14.3.3null/not_null
14.3.4default
14.3.5options
14.3.6readonly
14.3.7definition
14.3.8transient
14.3.9unordered
14.3.10index_type
14.3.11key_type
14.3.12value_type
14.3.13value_null/value_not_null
14.3.14id_options
14.3.15index_options
14.3.16key_options
14.3.17value_options
14.3.18id_column
14.3.19index_column
14.3.20key_column
14.3.21value_column
14.4Data Member Pragmas
14.4.1id
14.4.2auto
14.4.3type
14.4.4id_type
14.4.5get/set/access
14.4.6null/not_null
14.4.7default
14.4.8options
14.4.9column (object, composite value)
14.4.10column (view)
14.4.11transient
14.4.12readonly
14.4.13virtual
14.4.14inverse
14.4.15on_delete
14.4.16version
14.4.17index
14.4.18unique
14.4.19unordered
14.4.20table
14.4.21load/update
14.4.22section
14.4.23added
14.4.24deleted
14.4.25index_type
14.4.26key_type
14.4.27value_type
14.4.28value_null/value_not_null
14.4.29id_options
14.4.30index_options
14.4.31key_options
14.4.32value_options
14.4.33id_column
14.4.34index_column
14.4.35key_column
14.4.36value_column
14.5Namespace Pragmas
14.5.1pointer
14.5.2table
14.5.3schema
14.5.4session
14.6Object Model Pragmas
14.6.1version
14.7Index Definition Pragmas
14.8Database Type Mapping Pragmas
14.9C++ Compiler Warnings
14.9.1GNU C++
14.9.2Visual C++
14.9.3Sun C++
14.9.4IBM XL C++
14.9.5HP aC++
14.9.6Clang
15Advanced Techniques and Mechanisms
15.1Transaction Callbacks
15.2Persistent Class Template Instantiations
15.3Bulk Database Operations
PART II DATABASE SYSTEMS
16Multi-Database Support
16.1Static Multi-Database Support
16.2Dynamic Multi-Database Support
16.2.216.2.2 Dynamic Loading of Database Support Code
17MySQL Database
17.1MySQL Type Mapping
17.1.1String Type Mapping
17.1.2Binary Type Mapping
17.2MySQL Database Class
17.3MySQL Connection and Connection Factory
17.4MySQL Exceptions
17.5MySQL Limitations
17.5.1Foreign Key Constraints
17.6MySQL Index Definition
17.7MySQL Stored Procedures
18SQLite Database
18.1SQLite Type Mapping
18.1.1String Type Mapping
18.1.2Binary Type Mapping
18.2SQLite Database Class
18.3SQLite Connection and Connection Factory
18.4SQLite Exceptions
18.5SQLite Limitations
18.5.1Query Result Caching
18.5.2Automatic Assignment of Object Ids
18.5.3Foreign Key Constraints
18.5.4Constraint Violations
18.5.5Sharing of Queries
18.5.6Forced Rollback
18.5.7Database Schema Evolution
18.6SQLite Index Definition
19PostgreSQL Database
19.1PostgreSQL Type Mapping
19.1.1String Type Mapping
19.1.2Binary Type and UUID Mapping
19.2PostgreSQL Database Class
19.3PostgreSQL Connection and Connection Factory
19.4PostgreSQL Exceptions
19.5PostgreSQL Limitations
19.5.1Query Result Caching
19.5.2Foreign Key Constraints
19.5.3Unique Constraint Violations
19.5.4Date-Time Format
19.5.5Timezones
19.5.6NUMERIC Type Support
19.6PostgreSQL Index Definition
20Oracle Database
20.1Oracle Type Mapping
20.1.1String Type Mapping
20.1.2Binary Type Mapping
20.2Oracle Database Class
20.3Oracle Connection and Connection Factory
20.4Oracle Exceptions
20.5Oracle Limitations
20.5.1Identifier Truncation
20.5.2Query Result Caching
20.5.3Foreign Key Constraints
20.5.4Unique Constraint Violations
20.5.5Large FLOAT and NUMBER Types
20.5.6Timezones
20.5.7LONG Types
20.5.8LOB Types and By-Value Accessors/Modifiers
20.5.9Database Schema Evolution
20.6Oracle Index Definition
21Microsoft SQL Server Database
21.1SQL Server Type Mapping
21.1.1String Type Mapping
21.1.2Binary Type and UNIQUEIDENTIFIER Mapping
21.1.3ROWVERSION Mapping
21.1.4Long String and Binary Types
21.2SQL Server Database Class
21.3SQL Server Connection and Connection Factory
21.4SQL Server Exceptions
21.5SQL Server Limitations
21.5.1Query Result Caching
21.5.2Foreign Key Constraints
21.5.3Unique Constraint Violations
21.5.4Multi-threaded Windows Applications
21.5.5Affected Row Count and DDL Statements
21.5.6Long Data and Auto Object Ids, ROWVERSION
21.5.7Long Data and By-Value Accessors/Modifiers
21.5.8Bulk Update and ROWVERSION
21.6SQL Server Index Definition
21.7SQL Server Stored Procedures
PART III PROFILES
22Profiles Introduction
23Boost Profile
23.1Smart Pointers Library
23.2Unordered Containers Library
23.3Multi-Index Container Library
23.4Optional Library
23.5Date Time Library
23.5.1MySQL Database Type Mapping
23.5.2SQLite Database Type Mapping
23.5.3PostgreSQL Database Type Mapping
23.5.4Oracle Database Type Mapping
23.5.5SQL Server Database Type Mapping
23.6Uuid Library
23.6.1MySQL Database Type Mapping
23.6.2SQLite Database Type Mapping
23.6.3PostgreSQL Database Type Mapping
23.6.4Oracle Database Type Mapping
23.6.5SQL Server Database Type Mapping
24Qt Profile
24.1Basic Types Library
24.1.1MySQL Database Type Mapping
24.1.2SQLite Database Type Mapping
24.1.3PostgreSQL Database Type Mapping
24.1.4Oracle Database Type Mapping
24.1.5SQL Server Database Type Mapping
24.2Smart Pointers Library
24.3Containers Library
24.3.1Change-Tracking QList
24.4Date Time Library
24.4.1MySQL Database Type Mapping
24.4.2SQLite Database Type Mapping
24.4.3PostgreSQL Database Type Mapping
24.4.4Oracle Database Type Mapping
24.4.5SQL Server Database Type Mapping

Preface

As more critical aspects of our lives become dependant on software systems, more and more applications are required to save the data they work on in persistent and reliable storage. Database management systems and, in particular, relational database management systems (RDBMS) are commonly used for such storage. However, while the application development techniques and programming languages have evolved significantly over the past decades, the relational database technology in this area stayed relatively unchanged. In particular, this led to the now infamous mismatch between the object-oriented model used by many modern applications and the relational model still used by RDBMS.

While relational databases may be inconvenient to use from modern programming languages, they are still the main choice for many applications due to their maturity, reliability, as well as the availability of tools and alternative implementations.

To allow application developers to utilize relational databases from their object-oriented applications, a technique called object-relational mapping (ORM) is often used. It involves a conversion layer that maps between objects in the application's memory and their relational representation in the database. While the object-relational mapping code can be written manually, automated ORM systems are available for most object-oriented programming languages in use today.

ODB is an ORM system for the C++ programming language. It was designed and implemented with the following main goals:

About This Document

The goal of this manual is to provide you with an understanding of the object persistence model and APIs which are implemented by ODB. As such, this document is intended for C++ application developers and software architects who are looking for a C++ object persistence solution. Prior experience with C++ is required to understand this document. A basic understanding of relational database systems is advantageous but not expected or required.

More Information

Beyond this manual, you may also find the following sources of information useful:


PART I   OBJECT-RELATIONAL MAPPING

Part I describes the essential database concepts, APIs, and tools that together comprise the object-relational mapping for C++ as implemented by ODB. It consists of the following chapters.

1Introduction
2Hello World Example
3Working with Persistent Objects
4Querying the Database
5Containers
6Relationships
7Value Types
8Inheritance
10Views
11Session
12Optimistic Concurrency
13Database Schema Evolution
14ODB Pragma Language

1 Introduction

ODB is an object-relational mapping (ORM) system for C++. It provides tools, APIs, and library support that allow you to persist C++ objects to a relational database (RDBMS) without having to deal with tables, columns, or SQL and without manually writing any of the mapping code.

ODB is highly flexible and customizable. It can either completely hide the relational nature of the underlying database or expose some of the details as required. For example, you can automatically map basic C++ types to suitable SQL types, generate the relational database schema for your persistent classes, and use simple, safe, and yet powerful object query language instead of SQL. Or you can assign SQL types to individual data members, use the existing database schema, run native SQL SELECT queries, and call stored procedures. In fact, at an extreme, ODB can be used as just a convenient way to handle results of native SQL queries.

ODB is not a framework. It does not dictate how you should write your application. Rather, it is designed to fit into your style and architecture by only handling object persistence and not interfering with any other functionality. There is no common base type that all persistent classes should derive from nor are there any restrictions on the data member types in persistent classes. Existing classes can be made persistent with a few or no modifications.

ODB has been designed for high performance and low memory overhead. Prepared statements are used to send and receive object state in binary format instead of text which reduces the load on the application and the database server. Extensive caching of connections, prepared statements, and buffers saves time and resources on connection establishment, statement parsing, and memory allocations. For each supported database system the native C API is used instead of ODBC or higher-level wrapper APIs to reduce overhead and provide the most efficient implementation for each database operation. Finally, persistent classes have zero memory overhead. There are no hidden "database" members that each class must have nor are there per-object data structures allocated by ODB.

In this chapter we present a high-level overview of ODB. We will start with the ODB architecture and then outline the workflow of building an application that uses ODB. We will then continue by contrasting the drawbacks of the traditional way of saving C++ objects to relational databases with the benefits of using ODB for object persistence. We conclude the chapter by discussing the C++ standards supported by ODB. The next chapter takes a more hands-on approach and shows the concrete steps necessary to implement object persistence in a simple "Hello World" application.

1.1 Architecture and Workflow

From the application developer's perspective, ODB consists of three main components: the ODB compiler, the common runtime library, called libodb, and the database-specific runtime libraries, called libodb-<database>, where <database> is the name of the database system this runtime is for, for example, libodb-mysql. For instance, if the application is going to use the MySQL database for object persistence, then the three ODB components that this application will use are the ODB compiler, libodb and libodb-mysql.

The ODB compiler generates the database support code for persistent classes in your application. The input to the ODB compiler is one or more C++ header files defining C++ classes that you want to make persistent. For each input header file the ODB compiler generates a set of C++ source files implementing conversion between persistent C++ classes defined in this header and their database representation. The ODB compiler can also generate a database schema file that creates tables necessary to store the persistent classes.

The ODB compiler is a real C++ compiler except that it produces C++ instead of assembly or machine code. In particular, it is not an ad-hoc header pre-processor that is only capable of recognizing a subset of C++. ODB is capable of parsing any standard C++ code.

The common runtime library defines database system-independent interfaces that your application can use to manipulate persistent objects. The database-specific runtime library provides implementations of these interfaces for a concrete database as well as other database-specific utilities that are used by the generated code. Normally, the application does not use the database-specific runtime library directly but rather works with it via the common interfaces from libodb. The following diagram shows the object persistence architecture of an application that uses MySQL as the underlying database system:

The ODB system also defines two special-purpose languages: the ODB Pragma Language and ODB Query Language. The ODB Pragma Language is used to communicate various properties of persistent classes to the ODB compiler by means of special #pragma directives embedded in the C++ header files. It controls aspects of the object-relational mapping such as names of tables and columns that are used for persistent classes and their members or mapping between C++ types and database types.

The ODB Query Language is an object-oriented database query language that can be used to search for objects matching certain criteria. It is modeled after and is integrated into C++ allowing you to write expressive and safe queries that look and feel like ordinary C++.

The use of the ODB compiler to generate database support code adds an additional step to your application build sequence. The following diagram outlines the typical build workflow of an application that uses ODB:

1.2 Benefits

The traditional way of saving C++ objects to relational databases requires that you manually write code which converts between the database and C++ representations of each persistent class. The actions that such code usually performs include conversion between C++ values and strings or database types, preparation and execution of SQL queries, as well as handling the result sets. Writing this code manually has the following drawbacks:

In contrast, using ODB for C++ object persistence has the following benefits:

Overall, ODB provides an easy to use yet flexible and powerful object-relational mapping (ORM) system for C++. Unlike other ORM implementations for C++ that still require you to write database conversion or member registration code for each persistent class, ODB keeps persistent classes purely declarative. The functional part, the database conversion code, is automatically generated by the ODB compiler from these declarations.

1.3 Supported C++ Standards

ODB provides support for ISO/IEC C++ 1998/2003 (C++98/03), ISO/IEC TR 19768 C++ Library Extensions (C++ TR1), and ISO/IEC C++ 2011 (C++11). While the majority of the examples in this manual use C++98/03, support for the new functionality and library components introduced in TR1 and C++11 are discussed throughout the document. The c++11 example in the odb-examples package also shows ODB support for various C++11 features.


2 Hello World Example

In this chapter we will show how to create a simple C++ application that relies on ODB for object persistence using the traditional "Hello World" example. In particular, we will discuss how to declare persistent classes, generate database support code, as well as compile and run our application. We will also learn how to make objects persistent, load, update and delete persistent objects, as well as query the database for persistent objects that match certain criteria. The example also shows how to define and use views, a mechanism that allows us to create projections of persistent objects, database tables, or to handle results of native SQL queries or stored procedure calls.

The code presented in this chapter is based on the hello example which can be found in the odb-examples package of the ODB distribution.

2.1 Declaring Persistent Classes

In our "Hello World" example we will depart slightly from the norm and say hello to people instead of the world. People in our application will be represented as objects of C++ class person which is saved in person.hxx:

// person.hxx
//

#include <string>

class person
{
public:
  person (const std::string& first,
          const std::string& last,
          unsigned short age);

  const std::string& first () const;
  const std::string& last () const;

  unsigned short age () const;
  void age (unsigned short);

private:
  std::string first_;
  std::string last_;
  unsigned short age_;
};
  

In order not to miss anyone whom we need to greet, we would like to save the person objects in a database. To achieve this we declare the person class as persistent:

// person.hxx
//

#include <string>

#include <odb/core.hxx>     // (1)

#pragma db object           // (2)
class person
{
  ...

private:
  person () {}              // (3)

  friend class odb::access; // (4)

  #pragma db id auto        // (5)
  unsigned long id_;        // (5)

  std::string first_;
  std::string last_;
  unsigned short age_;
};
  

To be able to save the person objects in the database we had to make five changes, marked with (1) to (5), to the original class definition. The first change is the inclusion of the ODB header <odb/core.hxx>. This header provides a number of core ODB declarations, such as odb::access, that are used to define persistent classes.

The second change is the addition of db object pragma just before the class definition. This pragma tells the ODB compiler that the class that follows is persistent. Note that making a class persistent does not mean that all objects of this class will automatically be stored in the database. You would still create ordinary or transient instances of this class just as you would before. The difference is that now you can make such transient instances persistent, as we will see shortly.

The third change is the addition of the default constructor. The ODB-generated database support code will use this constructor when instantiating an object from the persistent state. Just as we have done for the person class, you can make the default constructor private or protected if you don't want to make it available to the users of your class. Note also that with some limitations it is possible to have a persistent class without the default constructor.

With the fourth change we make the odb::access class a friend of our person class. This is necessary to make the default constructor and the data members accessible to the database support code. If your class has a public default constructor and either public data members or public accessors and modifiers for the data members, then the friend declaration is unnecessary.

The final change adds a data member called id_ which is preceded by another pragma. In ODB every persistent object normally has a unique, within its class, identifier. Or, in other words, no two persistent instances of the same type have equal identifiers. While it is possible to define a persistent class without an object id, the number of database operations that can be performed on such a class is limited. For our class we use an integer id. The db id auto pragma that precedes the id_ member tells the ODB compiler that the following member is the object's identifier. The auto specifier indicates that it is a database-assigned id. A unique id will be automatically generated by the database and assigned to the object when it is made persistent.

In this example we chose to add an identifier because none of the existing members could serve the same purpose. However, if a class already has a member with suitable properties, then it is natural to use that member as an identifier. For example, if our person class contained some form of personal identification (SSN in the United States or ID/passport number in other countries), then we could use that as an id. Or, if we stored an email associated with each person, then we could have used that if each person is presumed to have a unique email address.

As another example, consider the following alternative version of the person class. Here we use one of the existing data members as id. Also the data members are kept private and are instead accessed via public accessor and modifier functions. Finally, the ODB pragmas are grouped together and are placed after the class definition. They could have also been moved into a separate header leaving the original class completely unchanged (for more information on such a non-intrusive conversion refer to Chapter 14, "ODB Pragma Language").

class person
{
public:
  person ();

  const std::string& email () const;
  void email (const std::string&);

  const std::string& get_name () const;
  std::string& set_name ();

  unsigned short getAge () const;
  void setAge (unsigned short);

private:
  std::string email_;
  std::string name_;
  unsigned short age_;
};

#pragma db object(person)
#pragma db member(person::email_) id
  

Now that we have the header file with the persistent class, let's see how we can generate that database support code.

2.2 Generating Database Support Code

The persistent class definition that we created in the previous section was particularly light on any code that could actually do the job and store the person's data to a database. There was no serialization or deserialization code, not even data member registration, that you would normally have to write by hand in other ORM libraries for C++. This is because in ODB code that translates between the database and C++ representations of an object is automatically generated by the ODB compiler.

To compile the person.hxx header we created in the previous section and generate the support code for the MySQL database, we invoke the ODB compiler from a terminal (UNIX) or a command prompt (Windows):

odb -d mysql --generate-query person.hxx
  

We will use MySQL as the database of choice in the remainder of this chapter, though other supported database systems can be used instead.

If you haven't installed the common ODB runtime library (libodb) or installed it into a directory where C++ compilers don't search for headers by default, then you may get the following error:

person.hxx:10:24: fatal error: odb/core.hxx: No such file or directory
  

To resolve this you will need to specify the libodb headers location with the -I preprocessor option, for example:

odb -I.../libodb -d mysql --generate-query person.hxx
  

Here .../libodb represents the path to the libodb directory.

The above invocation of the ODB compiler produces three C++ files: person-odb.hxx, person-odb.ixx, person-odb.cxx. You normally don't use types or functions contained in these files directly. Rather, all you have to do is include person-odb.hxx in C++ files where you are performing database operations with classes from person.hxx as well as compile person-odb.cxx and link the resulting object file to your application.

You may be wondering what the --generate-query option is for. It instructs the ODB compiler to generate optional query support code that we will use later in our "Hello World" example. Another option that we will find useful is --generate-schema. This option makes the ODB compiler generate a fourth file, person.sql, which is the database schema for the persistent classes defined in person.hxx:

odb -d mysql --generate-query --generate-schema person.hxx
  

The database schema file contains SQL statements that creates tables necessary to store the persistent classes. We will learn how to use it in the next section.

If you would like to see a list of all the available ODB compiler options, refer to the ODB Compiler Command Line Manual.

Now that we have the persistent class and the database support code, the only part that is left is the application code that does something useful with all of this. But before we move on to the fun part, let's first learn how to build and run an application that uses ODB. This way when we have some application code to try, there are no more delays before we can run it.

2.3 Compiling and Running

Assuming that the main() function with the application code is saved in driver.cxx and the database support code and schema are generated as described in the previous section, to build our application we will first need to compile all the C++ source files and then link them with two ODB runtime libraries.

On UNIX, the compilation part can be done with the following commands (substitute c++ with your C++ compiler name; for Microsoft Visual Studio setup, see the odb-examples package):

c++ -c driver.cxx
c++ -c person-odb.cxx
  

Similar to the ODB compilation, if you get an error stating that a header in odb/ or odb/mysql directory is not found, you will need to use the -I preprocessor option to specify the location of the common ODB runtime library (libodb) and MySQL ODB runtime library (libodb-mysql).

Once the compilation is done, we can link the application with the following command:

c++ -o driver driver.o person-odb.o -lodb-mysql -lodb
  

Notice that we link our application with two ODB libraries: libodb which is a common runtime library and libodb-mysql which is a MySQL runtime library (if you use another database, then the name of this library will change accordingly). If you get an error saying that one of these libraries could not be found, then you will need to use the -L linker option to specify their locations.

Before we can run our application we need to create a database schema using the generated person.sql file. For MySQL we can use the mysql client program, for example:

mysql --user=odb_test --database=odb_test < person.sql
  

The above command will log in to a local MySQL server as user odb_test without a password and use the database named odb_test. Beware that after executing this command, all the data stored in the odb_test database will be deleted.

Note also that using a standalone generated SQL file is not the only way to create a database schema in ODB. We can also embed the schema directly into our application or use custom schemas that were not generated by the ODB compiler. Refer to Section 3.4, "Database" for details.

Once the database schema is ready, we run our application using the same login and database name:

./driver --user odb_test --database odb_test
  

2.4 Making Objects Persistent

Now that we have the infrastructure work out of the way, it is time to see our first code fragment that interacts with the database. In this section we will learn how to make person objects persistent:

// driver.cxx
//

#include <memory>   // std::auto_ptr
#include <iostream>

#include <odb/database.hxx>
#include <odb/transaction.hxx>

#include <odb/mysql/database.hxx>

#include "person.hxx"
#include "person-odb.hxx"

using namespace std;
using namespace odb::core;

int
main (int argc, char* argv[])
{
  try
  {
    auto_ptr<database> db (new odb::mysql::database (argc, argv));

    unsigned long john_id, jane_id, joe_id;

    // Create a few persistent person objects.
    //
    {
      person john ("John", "Doe", 33);
      person jane ("Jane", "Doe", 32);
      person joe ("Joe", "Dirt", 30);

      transaction t (db->begin ());

      // Make objects persistent and save their ids for later use.
      //
      john_id = db->persist (john);
      jane_id = db->persist (jane);
      joe_id = db->persist (joe);

      t.commit ();
    }
  }
  catch (const odb::exception& e)
  {
    cerr << e.what () << endl;
    return 1;
  }
}
  

Let's examine this code piece by piece. At the beginning we include a bunch of headers. After the standard C++ headers we include <odb/database.hxx> and <odb/transaction.hxx> which define database system-independent odb::database and odb::transaction interfaces. Then we include <odb/mysql/database.hxx> which defines the MySQL implementation of the database interface. Finally, we include person.hxx and person-odb.hxx which define our persistent person class.

Then we have two using namespace directives. The first one brings in the names from the standard namespace and the second brings in the ODB declarations which we will use later in the file. Notice that in the second directive we use the odb::core namespace instead of just odb. The former only brings into the current namespace the essential ODB names, such as the database and transaction classes, without any of the auxiliary objects. This minimizes the likelihood of name conflicts with other libraries. Note also that you should continue using the odb namespace when qualifying individual names. For example, you should write odb::database, not odb::core::database.

Once we are in main(), the first thing we do is create the MySQL database object. Notice that this is the last line in driver.cxx that mentions MySQL explicitly; the rest of the code works through the common interfaces and is database system-independent. We use the argc/argv mysql::database constructor which automatically extract the database parameters, such as login name, password, database name, etc., from the command line. In your own applications you may prefer to use other mysql::database constructors which allow you to pass this information directly (Section 17.2, "MySQL Database Class").

Next, we create three person objects. Right now they are transient objects, which means that if we terminate the application at this point, they will be gone without any evidence of them ever existing. The next line starts a database transaction. We discuss transactions in detail later in this manual. For now, all we need to know is that all ODB database operations must be performed within a transaction and that a transaction is an atomic unit of work; all database operations performed within a transaction either succeed (committed) together or are automatically undone (rolled back).

Once we are in a transaction, we call the persist() database function on each of our person objects. At this point the state of each object is saved in the database. However, note that this state is not permanent until and unless the transaction is committed. If, for example, our application crashes at this point, there will still be no evidence of our objects ever existing.

In our case, one more thing happens when we call persist(). Remember that we decided to use database-assigned identifiers for our person objects. The call to persist() is where this assignment happens. Once this function returns, the id_ member contains this object's unique identifier. As a convenience, the persist() function also returns a copy of the object's identifier that it made persistent. We save the returned identifier for each object in a local variable. We will use these identifiers later in the chapter to perform other database operations on our persistent objects.

After we have persisted our objects, it is time to commit the transaction and make the changes permanent. Only after the commit() function returns successfully, are we guaranteed that the objects are made persistent. Continuing with the crash example, if our application terminates after the commit for whatever reason, the objects' state in the database will remain intact. In fact, as we will discover shortly, our application can be restarted and load the original objects from the database. Note also that a transaction must be committed explicitly with the commit() call. If the transaction object leaves scope without the transaction being explicitly committed or rolled back, it will automatically be rolled back. This behavior allows you not to worry about exceptions being thrown within a transaction; if they cross the transaction boundary, the transaction will automatically be rolled back and all the changes made to the database undone.

The final bit of code in our example is the catch block that handles the database exceptions. We do this by catching the base ODB exception (Section 3.14, "ODB Exceptions") and printing the diagnostics.

Let's now compile (Section 2.3, "Compiling and Running") and then run our first ODB application:

mysql --user=odb_test --database=odb_test < person.sql
./driver --user odb_test --database odb_test
  

Our first application doesn't print anything except for error messages so we can't really tell whether it actually stored the objects' state in the database. While we will make our application more entertaining shortly, for now we can use the mysql client to examine the database content. It will also give us a feel for how the objects are stored:

mysql --user=odb_test --database=odb_test

Welcome to the MySQL monitor.

mysql> select * from person;

+----+-------+------+-----+
| id | first | last | age |
+----+-------+------+-----+
|  1 | John  | Doe  |  33 |
|  2 | Jane  | Doe  |  32 |
|  3 | Joe   | Dirt |  30 |
+----+-------+------+-----+
3 rows in set (0.00 sec)

mysql> quit
  

Another way to get more insight into what's going on under the hood, is to trace the SQL statements executed by ODB as a result of each database operation. Here is how we can enable tracing just for the duration of our transaction:

    // Create a few persistent person objects.
    //
    {
      ...

      transaction t (db->begin ());

      t.tracer (stderr_tracer);

      // Make objects persistent and save their ids for later use.
      //
      john_id = db->persist (john);
      jane_id = db->persist (jane);
      joe_id = db->persist (joe);

      t.commit ();
    }
  

With this modification our application now produces the following output:

INSERT INTO `person` (`id`,`first`,`last`,`age`) VALUES (?,?,?,?)
INSERT INTO `person` (`id`,`first`,`last`,`age`) VALUES (?,?,?,?)
INSERT INTO `person` (`id`,`first`,`last`,`age`) VALUES (?,?,?,?)
  

Note that we see question marks instead of the actual values because ODB uses prepared statements and sends the data to the database in binary form. For more information on tracing, refer to Section 3.13, "Tracing SQL Statement Execution". In the next section we will see how to access persistent objects from our application.

2.5 Querying the Database for Objects

So far our application doesn't resemble a typical "Hello World" example. It doesn't print anything except for error messages. Let's change that and teach our application to say hello to people from our database. To make it a bit more interesting, let's say hello only to people over 30:

// driver.cxx
//

...

int
main (int argc, char* argv[])
{
  try
  {
    ...

    // Create a few persistent person objects.
    //
    {
      ...
    }

    typedef odb::query<person> query;
    typedef odb::result<person> result;

    // Say hello to those over 30.
    //
    {
      transaction t (db->begin ());

      result r (db->query<person> (query::age > 30));

      for (result::iterator i (r.begin ()); i != r.end (); ++i)
      {
        cout << "Hello, " << i->first () << "!" << endl;
      }

      t.commit ();
    }
  }
  catch (const odb::exception& e)
  {
    cerr << e.what () << endl;
    return 1;
  }
}
  

The first half of our application is the same as before and is replaced with "..." in the above listing for brevity. Again, let's examine the rest of it piece by piece.

The two typedefs create convenient aliases for two template instantiations that will be used a lot in our application. The first is the query type for the person objects and the second is the result type for that query.

Then we begin a new transaction and call the query() database function. We pass a query expression (query::age > 30) which limits the returned objects only to those with the age greater than 30. We also save the result of the query in a local variable.

The next few lines perform a standard for-loop iteration over the result sequence printing hello for every returned person. Then we commit the transaction and that's it. Let's see what this application will print:

mysql --user=odb_test --database=odb_test < person.sql
./driver --user odb_test --database odb_test

Hello, John!
Hello, Jane!
  

That looks about right, but how do we know that the query actually used the database instead of just using some in-memory artifacts of the earlier persist() calls? One way to test this would be to comment out the first transaction in our application and re-run it without re-creating the database schema. This way the objects that were persisted during the previous run will be returned. Alternatively, we can just re-run the same application without re-creating the schema and notice that we now show duplicate objects:

./driver --user odb_test --database odb_test

Hello, John!
Hello, Jane!
Hello, John!
Hello, Jane!
  

What happens here is that the previous run of our application persisted a set of person objects and when we re-run the application, we persist another set with the same names but with different ids. When we later run the query, matches from both sets are returned. We can change the line where we print the "Hello" string as follows to illustrate this point:

cout << "Hello, " << i->first () << " (" << i->id () << ")!" << endl;
  

If we now re-run this modified program, again without re-creating the database schema, we will get the following output:

./driver --user odb_test --database odb_test

Hello, John (1)!
Hello, Jane (2)!
Hello, John (4)!
Hello, Jane (5)!
Hello, John (7)!
Hello, Jane (8)!
  

The identifiers 3, 6, and 9 that are missing from the above list belong to the "Joe Dirt" objects which are not selected by this query.

2.6 Updating Persistent Objects

While making objects persistent and then selecting some of them using queries are two useful operations, most applications will also need to change the object's state and then make these changes persistent. Let's illustrate this by updating Joe's age who just had a birthday:

// driver.cxx
//

...

int
main (int argc, char* argv[])
{
  try
  {
    ...

    unsigned long john_id, jane_id, joe_id;

    // Create a few persistent person objects.
    //
    {
      ...

      // Save object ids for later use.
      //
      john_id = john.id ();
      jane_id = jane.id ();
      joe_id = joe.id ();
    }

    // Joe Dirt just had a birthday, so update his age.
    //
    {
      transaction t (db->begin ());

      auto_ptr<person> joe (db->load<person> (joe_id));
      joe->age (joe->age () + 1);
      db->update (*joe);

      t.commit ();
    }

    // Say hello to those over 30.
    //
    {
      ...
    }
  }
  catch (const odb::exception& e)
  {
    cerr << e.what () << endl;
    return 1;
  }
}
  

The beginning and the end of the new transaction are the same as the previous two. Once within a transaction, we call the load() database function to instantiate a person object with Joe's persistent state. We pass Joe's object identifier that we stored earlier when we made this object persistent. While here we use std::auto_ptr to manage the returned object, we could have also used another smart pointer, for example std::unique_ptr from C++11 or shared_ptr from TR1, C++11, or Boost. For more information on the object lifetime management and the smart pointers that we can use for that, see Section 3.3, "Object and View Pointers".

With the instantiated object in hand we increment the age and call the update() function to update the object's state in the database. Once the transaction is committed, the changes are made permanent.

If we now run this application, we will see Joe in the output since he is now over 30:

mysql --user=odb_test --database=odb_test < person.sql
./driver --user odb_test --database odb_test

Hello, John!
Hello, Jane!
Hello, Joe!
  

What if we didn't have an identifier for Joe? Maybe this object was made persistent in another run of our application or by another application altogether. Provided that we only have one Joe Dirt in the database, we can use the query facility to come up with an alternative implementation of the above transaction:

    // Joe Dirt just had a birthday, so update his age. An
    // alternative implementation without using the object id.
    //
    {
      transaction t (db->begin ());

      // Here we know that there can be only one Joe Dirt in our
      // database so we use the query_one() shortcut instead of
      // manually iterating over the result returned by query().
      //
      auto_ptr<person> joe (
        db->query_one<person> (query::first == "Joe" &&
                               query::last == "Dirt"));

      if (joe.get () != 0)
      {
        joe->age (joe->age () + 1);
        db->update (*joe);
      }

      t.commit ();
    }
  

2.7 Defining and Using Views

Suppose that we need to gather some basic statistics about the people stored in our database. Things like the total head count, as well as the minimum and maximum ages. One way to do it would be to query the database for all the person objects and then calculate this information as we iterate over the query result. While this approach may work fine for our database with just three people in it, it would be very inefficient if we had a large number of objects.

While it may not be conceptually pure from the object-oriented programming point of view, a relational database can perform some computations much faster and much more economically than if we performed the same operations ourselves in the application's process.

To support such cases ODB provides the notion of views. An ODB view is a C++ class that embodies a light-weight, read-only projection of one or more persistent objects or database tables or the result of a native SQL query execution or stored procedure call.

Some of the common applications of views include loading a subset of data members from objects or columns database tables, executing and handling results of arbitrary SQL queries, including aggregate queries, as well as joining multiple objects and/or database tables using object relationships or custom join conditions.

While you can find a much more detailed description of views in Chapter 10, "Views", here is how we can define the person_stat view that returns the basic statistics about the person objects:

#pragma db view object(person)
struct person_stat
{
  #pragma db column("count(" + person::id_ + ")")
  std::size_t count;

  #pragma db column("min(" + person::age_ + ")")
  unsigned short min_age;

  #pragma db column("max(" + person::age_ + ")")
  unsigned short max_age;
};
  

Normally, to get the result of a view we use the same query() function as when querying the database for an object. Here, however, we are executing an aggregate query which always returns exactly one element. Therefore, instead of getting the result instance and then iterating over it, we can use the shortcut query_value() function. Here is how we can load and print our statistics using the view we have just created:

    // Print some statistics about all the people in our database.
    //
    {
      transaction t (db->begin ());

      // The result of this query always has exactly one element.
      //
      person_stat ps (db->query_value<person_stat> ());

      cout << "count  : " << ps.count << endl
           << "min age: " << ps.min_age << endl
           << "max age: " << ps.max_age << endl;

      t.commit ();
    }
  

If we now add the person_stat view to the person.hxx header, the above transaction to driver.cxx, as well as re-compile and re-run our example, then we will see the following additional lines in the output:

count  : 3
min age: 31
max age: 33
  

2.8 Deleting Persistent Objects

The last operation that we will discuss in this chapter is deleting the persistent object from the database. The following code fragment shows how we can delete an object given its identifier:

    // John Doe is no longer in our database.
    //
    {
      transaction t (db->begin ());
      db->erase<person> (john_id);
      t.commit ();
    }
  

To delete John from the database we start a transaction, call the erase() database function with John's object id, and commit the transaction. After the transaction is committed, the erased object is no longer persistent.

If we don't have an object id handy, we can use queries to find and delete the object:

    // John Doe is no longer in our database. An alternative
    // implementation without using the object id.
    //
    {
      transaction t (db->begin ());

      // Here we know that there can be only one John Doe in our
      // database so we use the query_one() shortcut again.
      //
      auto_ptr<person> john (
        db->query_one<person> (query::first == "John" &&
                               query::last == "Doe"));

      if (john.get () != 0)
        db->erase (*john);

      t.commit ();
    }
  

2.9 Changing Persistent Classes

When the definition of a transient C++ class is changed, for example by adding or deleting a data member, we don't have to worry about any existing instances of this class not matching the new definition. After all, to make the class changes effective we have to restart the application and none of the transient instances will survive this.

Things are not as simple for persistent classes. Because they are stored in the database and therefore survive application restarts, we have a new problem: what happens to the state of existing objects (which correspond to the old definition) once we change our persistent class?

The problem of working with old objects, called database schema evolution, is a complex issue and ODB provides comprehensive support for handling it. While this support is covered in detail in Chapter 13, "Database Schema Evolution", let us consider a simple example that should give us a sense of the functionality provided by ODB in this area.

Suppose that after using our person persistent class for some time and creating a number of databases containing its instances, we realized that for some people we also need to store their middle name. If we go ahead and just add the new data member, everything will work fine with new databases. Existing databases, however, have a table that does not correspond to the new class definition. Specifically, the generated database support code now expects there to be a column to store the middle name. But such a column was never created in the old databases.

ODB can automatically generate SQL statements that will migrate old databases to match the new class definitions. But first, we need to enable schema evolution support by defining a version for our object model:

// person.hxx
//

#pragma db model version(1, 1)

class person
{
  ...

  std::string first_;
  std::string last_;
  unsigned short age_;
};
  

The first number in the version pragma is the base model version. This is the lowest version we will be able to migrate from. The second number is the current model version. Since we haven't made any changes yet to our persistent class, both of these values are 1.

Next we need to re-compile our person.hxx header file with the ODB compiler, just as we did before:

odb -d mysql --generate-query --generate-schema person.hxx
  

If we now look at the list of files produced by the ODB compiler, we will notice a new file: person.xml. This is a changelog file where the ODB compiler keeps track of the database changes corresponding to our class changes. Note that this file is automatically maintained by the ODB compiler and all we have to do is keep it around between re-compilations.

Now we are ready to add the middle name to our person class. We also give it a default value (empty string) which is what will be assigned to existing objects in old databases. Notice that we have also incremented the current version:

// person.hxx
//

#pragma db model version(1, 2)

class person
{
  ...

  std::string first_;

  #pragma db default("")
  std::string middle_;

  std::string last_;
  unsigned short age_;
};
  

If we now recompile the person.hxx header again, we will see two extra generated files: person-002-pre.sql and person-002-post.sql. These two files contain schema migration statements from version 1 to version 2. Similar to schema creation, schema migration statements can also be embedded into the generated C++ code.

person-002-pre.sql and person-002-post.sql are the pre and post schema migration files. To migrate one of our old databases, we first execute the pre migration file:

mysql --user=odb_test --database=odb_test < person-002-pre.sql
  

Between the pre and post schema migrations we can run data migration code, if required. At this stage, we can both access the old and store the new data. In our case we don't need any data migration code since we assigned the default value to the middle name for all the existing objects.

To finish the migration process we execute the post migration statements:

mysql --user=odb_test --database=odb_test < person-002-post.sql
  

2.10 Working with Multiple Databases

Accessing multiple databases (that is, data stores) is simply a matter of creating multiple odb::<db>::database instances representing each database. For example:

odb::mysql::database db1 ("john", "secret", "test_db1");
odb::mysql::database db2 ("john", "secret", "test_db2");
  

Some database systems also allow attaching multiple databases to the same instance. A more interesting question is how we access multiple database systems (that is, database implementations) from the same application. For example, our application may need to store some objects in a remote MySQL database and others in a local SQLite file. Or, our application may need to be able to store its objects in a database system that is selected by the user at runtime.

ODB provides comprehensive multi-database support that ranges from tight integration with specific database systems to being able to write database-agnostic code and loading individual database systems support dynamically. While all these aspects are covered in detail in Chapter 16, "Multi-Database Support", in this section we will get a taste of this functionality by extending our "Hello World" example to be able to store its data either in MySQL or PostgreSQL (other database systems supported by ODB can be added in a similar manner).

The first step in adding multi-database support is to re-compile our person.hxx header to generate database support code for additional database systems:

odb --multi-database dynamic -d common -d mysql -d pgsql \
--generate-query --generate-schema person.hxx
  

The --multi-database ODB compiler option turns on multi-database support. For now it is not important what the dynamic value that we passed to this option means, but if you are curious, see Chapter 16. The result of this command are three sets of generated files: person-odb.?xx (common interface; corresponds to the common database), person-odb-mysql.?xx (MySQL support code), and person-odb-pgsql.?xx (PostgreSQL support code). There are also two schema files: person-mysql.sql and person-pgsql.sql.

The only part that we need to change in driver.cxx is how we create the database instance. Specifically, this line:

auto_ptr<database> db (new odb::mysql::database (argc, argv));
  

Now our example is capable of storing its data either in MySQL or PostgreSQL so we need to somehow allow the caller to specify which database we must use. To keep things simple, we will make the first command line argument specify the database system we must use while the rest will contain the database-specific options which we will pass to the odb::<db>::database constructor as before. Let's put all this logic into a separate function which we will call create_database(). Here is what the beginning of our modified driver.cxx will look like (the remainder is unchanged):

// driver.cxx
//

#include <string>
#include <memory>   // std::auto_ptr
#include <iostream>

#include <odb/database.hxx>
#include <odb/transaction.hxx>

#include <odb/mysql/database.hxx>
#include <odb/pgsql/database.hxx>

#include "person.hxx"
#include "person-odb.hxx"

using namespace std;
using namespace odb::core;

auto_ptr<database>
create_database (int argc, char* argv[])
{
  auto_ptr<database> r;

  if (argc < 2)
  {
    cerr << "error: database system name expected" << endl;
    return r;
  }

  string db (argv[1]);

  if (db == "mysql")
    r.reset (new odb::mysql::database (argc, argv));
  else if (db == "pgsql")
    r.reset (new odb::pgsql::database (argc, argv));
  else
    cerr << "error: unknown database system " << db << endl;

  return r;
}

int
main (int argc, char* argv[])
{
  try
  {
    auto_ptr<database> db (create_database (argc, argv));

    if (db.get () == 0)
      return 1; // Diagnostics has already been issued.

    ...
  

And that's it. The only thing left is to build and run our example:

c++ -c driver.cxx
c++ -c person-odb.cxx
c++ -c person-odb-mysql.cxx
c++ -c person-odb-pgsql.cxx
c++ -o driver driver.o person-odb.o person-odb-mysql.o \
person-odb-pgsql.o -lodb-mysql -lodb-pgsql -lodb
  

Here is how we can access a MySQL database:

mysql --user=odb_test --database=odb_test < person-mysql.sql
./driver mysql --user odb_test --database odb_test
  

Or a PostgreSQL database:

psql --user=odb_test --dbname=odb_test -f person-pgsql.sql
./driver pgsql --user odb_test --database odb_test
  

2.11 Summary

This chapter presented a very simple application which, nevertheless, exercised all of the core database functions: persist(), query(), load(), update(), and erase(). We also saw that writing an application that uses ODB involves the following steps:

  1. Declare persistent classes in header files.
  2. Compile these headers to generate database support code.
  3. Link the application with the generated code and two ODB runtime libraries.

Do not be concerned if, at this point, much appears unclear. The intent of this chapter is to give you only a general idea of how to persist C++ objects with ODB. We will cover all the details throughout the remainder of this manual.


3 Working with Persistent Objects

The previous chapters gave us a high-level overview of ODB and showed how to use it to store C++ objects in a database. In this chapter we will examine the ODB object persistence model as well as the core database APIs in greater detail. We will start with basic concepts and terminology in Section 3.1 and Section 3.3 and continue with the discussion of the odb::database class in Section 3.4, transactions in Section 3.5, and connections in Section 3.6. The remainder of this chapter deals with the core database operations and concludes with the discussion of ODB exceptions.

In this chapter we will continue to use and expand the person persistent class that we have developed in the previous chapter.

3.1 Concepts and Terminology

The term database can refer to three distinct things: a general notion of a place where an application stores its data, a software implementation for managing this data (for example MySQL), and, finally, some database software implementations may manage several data stores which are usually distinguished by name. This name is also commonly referred to as a database.

In this manual, when we use the word database, we refer to the first meaning above, for example, "The update() function saves the object's state to the database." The term Database Management System (DBMS) is often used to refer to the second meaning of the word database. In this manual we will use the term database system for short, for example, "Database system-independent application code." Finally, to distinguish the third meaning from the other two, we will use the term database name, for example, "The second option specifies the database name that the application should use to store its data."

In C++ there is only one notion of a type and an instance of a type. For example, a fundamental type, such as int, is, for the most part, treated the same as a user defined class type. However, when it comes to persistence, we have to place certain restrictions and requirements on certain C++ types that can be stored in the database. As a result, we divide persistent C++ types into two groups: object types and value types. An instance of an object type is called an object and an instance of a value type — a value.

An object is an independent entity. It can be stored, updated, and deleted in the database independent of other objects. Normally, an object has an identifier, called object id, that is unique among all instances of an object type within a database. In contrast, a value can only be stored in the database as part of an object and doesn't have its own unique identifier.

An object consists of data members which are either values (Chapter 7, "Value Types"), pointers to other objects (Chapter 6, "Relationships"), or containers of values or pointers to other objects (Chapter 5, "Containers"). Pointers to other objects and containers can be viewed as special kinds of values since they also can only be stored in the database as part of an object.

An object type is a C++ class. Because of this one-to-one relationship, we will use terms object type and object class interchangeably. In contrast, a value type can be a fundamental C++ type, such as int or a class type, such as std::string. If a value consists of other values, then it is called a composite value and its type — a composite value type (Section 7.2, "Composite Value Types"). Otherwise, the value is called simple value and its type — a simple value type (Section 7.1, "Simple Value Types"). Note that the distinction between simple and composite values is conceptual rather than representational. For example, std::string is a simple value type because conceptually string is a single value even though the representation of the string class may contain several data members each of which could be considered a value. In fact, the same value type can be viewed (and mapped) as both simple and composite by different applications.

While not strictly necessary in a purely object-oriented application, practical considerations often require us to only load a subset of an object's data members or a combination of members from several objects. We may also need to factor out some computations to the relational database instead of performing them in the application's process. To support such requirements ODB distinguishes a third kind of C++ types, called views (Chapter 10, "Views"). An ODB view is a C++ class that embodies a light-weight, read-only projection of one or more persistent objects or database tables or the result of a native SQL query execution.

Understanding how all these concepts map to the relational model will hopefully make these distinctions clearer. In a relational database an object type is mapped to a table and a value type is mapped to one or more columns. A simple value type is mapped to a single column while a composite value type is mapped to several columns. An object is stored as a row in this table and a value is stored as one or more cells in this row. A simple value is stored in a single cell while a composite value occupies several cells. A view is not a persistent entity and it is not stored in the database. Rather, it is a data structure that is used to capture a single row of an SQL query result.

Going back to the distinction between simple and composite values, consider a date type which has three integer members: year, month, and day. In one application it can be considered a composite value and each member will get its own column in a relational database. In another application it can be considered a simple value and stored in a single column as a number of days from some predefined date.

Until now, we have been using the term persistent class to refer to object classes. We will continue to do so even though a value type can also be a class. The reason for this asymmetry is the subordinate nature of value types when it comes to database operations. Remember that values are never stored directly but rather as part of an object that contains them. As a result, when we say that we want to make a C++ class persistent or persist an instance of a class in the database, we invariably refer to an object class rather than a value class.

Normally, you would use object types to model real-world entities, things that have their own identity. For example, in the previous chapter we created a person class to model a person, which is a real-world entity. Name and age, which we used as data members in our person class are clearly values. It is hard to think of age 31 or name "Joe" as having their own identities.

A good test to determine whether something is an object or a value, is to consider if other objects might reference it. A person is clearly an object because it can be referred to by other objects such as a spouse, an employer, or a bank. On the other hand, a person's age or name is not something that other objects would normally refer to.

Also, when an object represents a real entity, it is easy to choose a suitable object id. For example, for a person there is an established notion of an identifier (SSN, student id, passport number, etc). Another alternative is to use a person's email address as an identifier.

Note, however, that these are only guidelines. There could be good reasons to make something that would normally be a value an object. Consider, for example, a database that stores a vast number of people. Many of the person objects in this database have the same names and surnames and the overhead of storing them in every object may negatively affect the performance. In this case, we could make the first name and last name each an object and only store pointers to these objects in the person class.

An instance of a persistent class can be in one of two states: transient and persistent. A transient instance only has a representation in the application's memory and will cease to exist when the application terminates, unless it is explicitly made persistent. In other words, a transient instance of a persistent class behaves just like an instance of any ordinary C++ class. A persistent instance has a representation in both the application's memory and the database. A persistent instance will remain even after the application terminates unless and until it is explicitly deleted from the database.

3.2 Declaring Persistent Objects and Values

To make a C++ class a persistent object class we declare it as such using the db object pragma, for example:

#pragma db object
class person
{
  ...
};
  

The other pragma that we often use is db id which designates one of the data members as an object id, for example:

#pragma db object
class person
{
  ...

  #pragma db id
  unsigned long id_;
};
  

The object id can be of a simple or composite (Section 7.2.1, "Composite Object Ids") value type. This type should be default-constructible, copy-constructible, and copy-assignable. It is also possible to declare a persistent class without an object id, however, such a class will have limited functionality (Section 14.1.6, "no_id").

The above two pragmas are the minimum required to declare a persistent class with an object id. Other pragmas can be used to fine-tune the database-related properties of a class and its members (Chapter 14, "ODB Pragma Language").

Normally, a persistent class should define the default constructor. The generated database support code uses this constructor when instantiating an object from the persistent state. If we add the default constructor only for the database support code, then we can make it private provided we also make the odb::access class, defined in the <odb/core.hxx> header, a friend of this object class. For example:

#include <odb/core.hxx>

#pragma db object
class person
{
  ...

private:
  friend class odb::access;
  person () {}
};
  

It is also possible to have an object class without the default constructor. However, in this case, the database operations will only be able to load the persistent state into an existing instance (Section 3.9, "Loading Persistent Objects", Section 4.4, "Query Result").

The ODB compiler also needs access to the non-transient (Section 14.4.11, "transient") data members of a persistent class. The ODB compiler can access such data members directly if they are public. It can also do so if they are private or protected and the odb::access class is declared a friend of the object type. For example:

#include <odb/core.hxx>

#pragma db object
class person
{
  ...

private:
  friend class odb::access;
  person () {}

  #pragma db id
  unsigned long id_;

  std::string name_;
};
  

If data members are not accessible directly, then the ODB compiler will try to automatically find suitable accessor and modifier functions. To accomplish this, the ODB compiler will try to lookup common accessor and modifier names derived from the data member name. Specifically, for the name_ data member in the above example, the ODB compiler will look for accessor functions with names: get_name(), getName(), getname(), and just name() as well as for modifier functions with names: set_name(), setName(), setname(), and just name(). You can also add support for custom name derivations with the --accessor-regex and --modifier-regex ODB compiler options. Refer to the ODB Compiler Command Line Manual for details on these options. The following example illustrates automatic accessor and modifier discovery:

#pragma db object
class person
{
public:
  person () {}

  ...

  unsigned long id () const;
  void id (unsigned long);

  const std::string& get_name () const;
  std::string& set_name ();

private:
  #pragma db id
  unsigned long id_; // Uses id() for access.

  std::string name_; // Uses get_name()/set_name() for access.
};
  

Finally, if a data member is not directly accessible and the ODB compiler was unable to discover suitable accessor and modifier functions, then we can provide custom accessor and modifier expressions using the db get and db set pragmas. For more information on custom accessor and modifier expressions refer to Section 14.4.5, "get/set/access".

Data members of a persistent class can also be split into separately-loaded and/or separately-updated sections. For more information on this functionality, refer to Chapter 9, "Sections".

You may be wondering whether we also have to declare value types as persistent. We don't need to do anything special for simple value types such as int or std::string since the ODB compiler knows how to map them to suitable database types and how to convert between the two. On the other hand, if a simple value is unknown to the ODB compiler then we will need to provide the mapping to the database type and, possibly, the code to convert between the two. For more information on how to achieve this refer to the db type pragma description in Section 14.3.1, "type".

Similar to object classes, composite value types have to be explicitly declared as persistent using the db value pragma, for example:

#pragma db value
class name
{
  ...

  std::string first_;
  std::string last_;
};
  

Note that a composite value cannot have a data member designated as an object id since, as we have discussed earlier, values do not have a notion of identity. A composite value type also doesn't have to define the default constructor, unless it is used as an element of a container. The ODB compiler uses the same mechanisms to access data members in composite value types as in object types. Composite value types are discussed in more detail in Section 7.2, "Composite Value Types".

3.3 Object and View Pointers

As we have seen in the previous chapter, some database operations create dynamically allocated instances of persistent classes and return pointers to these instances. As we will see in later chapters, pointers are also used to establish relationships between objects (Chapter 6, "Relationships") as well as to cache persistent objects in a session (Chapter 11, "Session"). While in most cases you won't need to deal with pointers to views, it is possible to a obtain a dynamically allocated instance of a view using the result_iterator::load() function (Section 4.4, "Query Results").

By default, all these mechanisms use raw pointers to return objects and views as well as to pass and cache objects. This is normally sufficient for applications that have simple object lifetime requirements and do not use sessions or object relationships. In particular, a dynamically allocated object or view that is returned as a raw pointer from a database operation can be assigned to a smart pointer of our choice, for example std::auto_ptr, std::unique_ptr from C++11, or shared_ptr from TR1, C++11, or Boost.

However, to avoid any possibility of a mistake, such as forgetting to use a smart pointer for a returned object or view, as well as to simplify the use of more advanced ODB functionality, such as sessions and bidirectional object relationships, it is recommended that you use smart pointers with the sharing semantics as object pointers. The shared_ptr smart pointer from TR1, C++11, or Boost is a good default choice. However, if sharing is not required and sessions are not used, then std::unique_ptr or std::auto_ptr can be used just as well.

ODB provides several mechanisms for changing the object or view pointer type. To specify the pointer type on the per object or per view basis we can use the db pointer pragma, for example:

#pragma db object pointer(std::tr1::shared_ptr)
class person
{
  ...
};
  

We can also specify the default pointer for a group of objects or views at the namespace level:

#pragma db namespace pointer(std::tr1::shared_ptr)
namespace accounting
{
  #pragma db object
  class employee
  {
    ...
  };

  #pragma db object
  class employer
  {
    ...
  };
}
  

Finally, we can use the --default-pointer option to specify the default pointer for the whole file. Refer to the ODB Compiler Command Line Manual for details on this option's argument. The typical usage is shown below:

--default-pointer std::tr1::shared_ptr
  

An alternative to this method with the same effect is to specify the default pointer for the global namespace:

#pragma db namespace() pointer(std::tr1::shared_ptr)
  

Note that we can always override the default pointer specified at the namespace level or with the command line option using the db pointer object or view pragma. For example:

#pragma db object pointer(std::shared_ptr)
namespace accounting
{
  #pragma db object
  class employee
  {
    ...
  };

  #pragma db object pointer(std::unique_ptr)
  class employer
  {
    ...
  };
}
  

Refer to Section 14.1.2, "pointer (object)", Section 14.2.4, "pointer (view)", and Section 14.5.1, "pointer (namespace)" for more information on these mechanisms.

Built-in support that is provided by the ODB runtime library allows us to use shared_ptr (TR1 or C++11), std::unique_ptr (C++11), or std::auto_ptr as pointer types. Plus, ODB profile libraries, that are available for commonly used frameworks and libraries (such as Boost and Qt), provide support for smart pointers found in these frameworks and libraries (Part III, "Profiles"). It is also easy to add support for our own smart pointers, as described in Section 6.5, "Using Custom Smart Pointers".

3.4 Database

Before an application can make use of persistence services offered by ODB, it has to create a database class instance. A database instance is the representation of the place where the application stores its persistent objects. We create a database instance by instantiating one of the database system-specific classes. For example, odb::mysql::database would be such a class for the MySQL database system. We will also normally pass a database name as an argument to the class' constructor. The following code fragment shows how we can create a database instance for the MySQL database system:

#include <odb/database.hxx>
#include <odb/mysql/database.hxx>

auto_ptr<odb::database> db (
  new odb::mysql::database (
    "test_user"     // database login name
    "test_password" // database password
    "test_database" // database name
    ));
  

The odb::database class is a common interface for all the database system-specific classes provided by ODB. You would normally work with the database instance via this interface unless there is a specific functionality that your application depends on and which is only exposed by a particular system's database class. You will need to include the <odb/database.hxx> header file to make this class available in your application.

The odb::database interface defines functions for starting transactions and manipulating persistent objects. These are discussed in detail in the remainder of this chapter as well as the next chapter which is dedicated to the topic of querying the database for persistent objects. For details on the system-specific database classes, refer to Part II, "Database Systems".

Before we can persist our objects, the corresponding database schema has to be created in the database. The schema contains table definitions and other relational database artifacts that are used to store the state of persistent objects in the database.

There are several ways to create the database schema. The easiest is to instruct the ODB compiler to generate the corresponding schema from the persistent classes (--generate-schema option). The ODB compiler can generate the schema as a standalone SQL file, embedded into the generated C++ code, or as a separate C++ source file (--schema-format option). If we are using the SQL file to create the database schema, then this file should be executed, normally only once, before the application is started.

Alternatively, if the schema is embedded directly into the generated code or produced as a separate C++ source file, then we can use the odb::schema_catalog class to create it in the database from within our application, for example:

#include <odb/schema-catalog.hxx>

odb::transaction t (db->begin ());
odb::schema_catalog::create_schema (*db);
t.commit ();
  

Refer to the next section for information on the odb::transaction class. The complete version of the above code fragment is available in the schema/embedded example in the odb-examples package.

The odb::schema_catalog class has the following interface. You will need to include the <odb/schema-catalog.hxx> header file to make this class available in your application.

namespace odb
{
  class schema_catalog
  {
  public:
    static void
    create_schema (database&,
                   const std::string& name = "",
                   bool drop = true);

    static void
    drop_schema (database&, const std::string& name = "");

    static bool
    exists (database_id, const std::string& name = "");

    static bool
    exists (const database&, const std::string& name = "")
  };
}
  

The first argument to the create_schema() function is the database instance that we would like to create the schema in. The second argument is the schema name. By default, the ODB compiler generates all embedded schemas with the default schema name (empty string). However, if your application needs to have several separate schemas, you can use the --schema-name ODB compiler option to assign custom schema names and then use these names as a second argument to create_schema(). By default, create_schema() will also delete all the database objects (tables, indexes, etc.) if they exist prior to creating the new ones. You can change this behavior by passing false as the third argument. The drop_schema() function allows you to delete all the database objects without creating the new ones.

If the schema is not found, the create_schema() and drop_schema() functions throw the odb::unknown_schema exception. You can use the exists() function to check whether a schema for the specified database and with the specified name exists in the catalog. Note also that the create_schema() and drop_schema() functions should be called within a transaction.

ODB also provides support for database schema evolution. Similar to schema creation, schema migration statements can be generated either as standalone SQL files or embedded into the generated C++ code. For more information on schema evolution support, refer to Chapter 13, "Database Schema Evolution".

Finally, we can also use a custom database schema with ODB. This approach can work similarly to the standalone SQL file described above except that the database schema is hand-written or produced by another program. Or we could execute custom SQL statements that create the schema directly from our application. To map persistent classes to custom database schemas, ODB provides a wide range of mapping customization pragmas, such as db table, db column, and db type (Chapter 14, "ODB Pragma Language"). For sample code that shows how to perform such mapping for various C++ constructs, refer to the schema/custom example in the odb-examples package.

3.5 Transactions

A transaction is an atomic, consistent, isolated and durable (ACID) unit of work. Database operations can only be performed within a transaction and each thread of execution in an application can have only one active transaction at a time.

By atomicity we mean that when it comes to making changes to the database state within a transaction, either all the changes are applied or none at all. Consider, for example, a transaction that transfers funds between two objects representing bank accounts. If the debit function on the first object succeeds but the credit function on the second fails, the transaction is rolled back and the database state of the first object remains unchanged.

By consistency we mean that a transaction must take all the objects stored in the database from one consistent state to another. For example, if a bank account object must reference a person object as its owner and we forget to set this reference before making the object persistent, the transaction will be rolled back and the database will remain unchanged.

By isolation we mean that the changes made to the database state during a transaction are only visible inside this transaction until and unless it is committed. Using the above example with the bank transfer, the results of the debit operation performed on the first object is not visible to other transactions until the credit operation is successfully completed and the transaction is committed.

By durability we mean that once the transaction is committed, the changes that it made to the database state are permanent and will survive failures such as an application crash. From now on the only way to alter this state is to execute and commit another transaction.

A transaction is started by calling either the database::begin() or connection::begin() function. The returned transaction handle is stored in an instance of the odb::transaction class. You will need to include the <odb/transaction.hxx> header file to make this class available in your application. For example:

#include <odb/transaction.hxx>

transaction t (db.begin ())

// Perform database operations.

t.commit ();
  

The odb::transaction class has the following interface:

namespace odb
{
  class transaction
  {
  public:
    typedef odb::database database_type;
    typedef odb::connection connection_type;

    explicit
    transaction (transaction_impl*, bool make_current = true);

    transaction ();

    void
    reset (transaction_impl*, bool make_current = true);

    void
    commit ();

    void
    rollback ();

    database_type&
    database ();

    connection_type&
    connection ();

    bool
    finilized () const;

  public:
    static bool
    has_current ();

    static transaction&
    current ();

    static void
    current (transaction&);

    static bool
    reset_current ();

    // Callback API.
    //
  public:
    ...
  };
}
  

The commit() function commits a transaction and rollback() rolls it back. Unless the transaction has been finalized, that is, explicitly committed or rolled back, the destructor of the transaction class will automatically roll it back when the transaction instance goes out of scope. If we try to commit or roll back a finalized transaction, the odb::transaction_already_finalized exception is thrown.

The database() accessor returns the database this transaction is working on. Similarly, the connection() accessor returns the database connection this transaction is on (Section 3.6, "Connections").

The static current() accessor returns the currently active transaction for this thread. If there is no active transaction, this function throws the odb::not_in_transaction exception. We can check whether there is a transaction in effect in this thread using the has_current() static function.

The make_current argument in the transaction constructor as well as the static current() modifier and reset_current() function give us additional control over the nomination of the currently active transaction. If we pass false as the make_current argument, then the newly created transaction will not automatically be made the active transaction for this thread. Later, we can use the static current() modifier to set this transaction as the active transaction. The reset_current() static function clears the currently active transaction. Together, these mechanisms allow for more advanced use cases, such as multiplexing two or more transactions on the same thread. For example:

transaction t1 (db1.begin ());        // Active transaction.
transaction t2 (db2.begin (), false); // Not active.

// Perform database operations on db1.

transaction::current (t2);            // Deactivate t1, activate t2.

// Perform database operations on db2.

transaction::current (t1);            // Switch back to t1.

// Perform some more database operations on db1.

t1.commit ();

transaction::current (t2);            // Switch to t2.

// Perform some more database operations on db2.

t2.commit ();
  

The reset() modifier allows us to reuse the same transaction instance to complete several database transactions. Similar to the destructor, reset() will roll the current transaction back if it hasn't been finalized. The default transaction constructor creates a finalized transaction which can later be initialized using reset(). The finilized() accessor can be used to check whether the transaction has been finalized. Here is how we can use this functionality to commit the current transaction and start a new one every time a certain number of database operations has been performed:

transaction t (db.begin ());

for (size_t i (0); i < n; ++i)
{
  // Perform a database operation, such as persist an object.

  // Commit the current transaction and start a new one after
  // every 100 operations.
  //
  if (i % 100 == 0)
  {
    t.commit ();
    t.reset (db.begin ());
  }
}

t.commit ();
  

For more information on the transaction callback support, refer to Section 15.1, "Transaction Callbacks".

Note that in the above discussion of atomicity, consistency, isolation, and durability, all of those guarantees only apply to the object's state in the database as opposed to the object's state in the application's memory. It is possible to roll a transaction back but still have changes from this transaction in the application's memory. An easy way to avoid this potential inconsistency is to instantiate persistent objects only within the transaction scope. Consider, for example, these two implementations of the same transaction:

void
update_age (database& db, person& p)
{
  transaction t (db.begin ());

  p.age (p.age () + 1);
  db.update (p);

  t.commit ();
}
  

In the above implementation, if the update() call fails and the transaction is rolled back, the state of the person object in the database and the state of the same object in the application's memory will differ. Now consider an alternative implementation which only instantiates the person object for the duration of the transaction:

void
update_age (database& db, unsigned long id)
{
  transaction t (db.begin ());

  auto_ptr<person> p (db.load<person> (id));
  p.age (p.age () + 1);
  db.update (p);

  t.commit ();
}
  

Of course, it may not always be possible to write the application in this style. Oftentimes we need to access and modify the application's state of persistent objects out of transactions. In this case it may make sense to try to roll back the changes made to the application state if the transaction was rolled back and the database state remains unchanged. One way to do this is to re-load the object's state from the database, for example:

void
update_age (database& db, person& p)
{
  try
  {
    transaction t (db.begin ());

    p.age (p.age () + 1);
    db.update (p);

    t.commit ();
  }
  catch (...)
  {
    transaction t (db.begin ());
    db.load (p.id (), p);
    t.commit ();

    throw;
  }
}
  

See also Section 15.1, "Transaction Callbacks" for an alternative approach.

3.6 Connections

The odb::connection class represents a connection to the database. Normally, you wouldn't work with connections directly but rather let the ODB runtime obtain and release connections as needed. However, certain use cases may require obtaining a connection manually. For completeness, this section describes the connection class and discusses some of its use cases. You may want to skip this section if you are reading through the manual for the first time.

Similar to odb::database, the odb::connection class is a common interface for all the database system-specific classes provided by ODB. For details on the system-specific connection classes, refer to Part II, "Database Systems".

To make the odb::connection class available in your application you will need to include the <odb/connection.hxx> header file. The odb::connection class has the following interface:

namespace odb
{
  class connection
  {
  public:
    typedef odb::database database_type;

    transaction
    begin () = 0;

    unsigned long long
    execute (const char* statement);

    unsigned long long
    execute (const std::string& statement);

    unsigned long long
    execute (const char* statement, std::size_t length);

    database_type&
    database ();
  };

  typedef details::shared_ptr<connection> connection_ptr;
}
  

The begin() function is used to start a transaction on the connection. The execute() functions allow us to execute native database statements on the connection. Their semantics are equivalent to the database::execute() functions (Section 3.12, "Executing Native SQL Statements") except that they can be legally called outside a transaction. Finally, the database() accessor returns a reference to the odb::database instance to which this connection corresponds.

To obtain a connection we call the database::connection() function. The connection is returned as odb::connection_ptr, which is an implementation-specific smart pointer with the shared pointer semantics. This, in particular, means that the connection pointer can be copied and returned from functions. Once the last instance of connection_ptr pointing to the same connection is destroyed, the connection is returned to the database instance. The following code fragment shows how we can obtain, use, and release a connection:

using namespace odb::core;

database& db = ...
connection_ptr c (db.connection ());

// Temporarily disable foreign key constraints.
//
c->execute ("SET FOREIGN_KEY_CHECKS = 0");

// Start a transaction on this connection.
//
transaction t (c->begin ());
...
t.commit ();

// Restore foreign key constraints.
//
c->execute ("SET FOREIGN_KEY_CHECKS = 1");

// When 'c' goes out of scope, the connection is returned to 'db'.
  

Some of the use cases which may require direct manipulation of connections include out-of-transaction statement execution, such as the execution of connection configuration statements, the implementation of a connection-per-thread policy, and making sure that a set of transactions is executed on the same connection.

3.7 Error Handling and Recovery

ODB uses C++ exceptions to report database operation errors. Most ODB exceptions signify hard errors or errors that cannot be corrected without some intervention from the application. For example, if we try to load an object with an unknown object id, the odb::object_not_persistent exception is thrown. Our application may be able to correct this error, for instance, by obtaining a valid object id and trying again. The hard errors and corresponding ODB exceptions that can be thrown by each database function are described in the remainder of this chapter with Section 3.14, "ODB Exceptions" providing a quick reference for all the ODB exceptions.

The second group of ODB exceptions signify soft or recoverable errors. Such errors are temporary failures which normally can be corrected by simply re-executing the transaction. ODB defines three such exceptions: odb::connection_lost, odb::timeout, and odb::deadlock. All recoverable ODB exceptions are derived from the common odb::recoverable base exception which can be used to handle all the recoverable conditions with a single catch block.

The odb::connection_lost exception is thrown if a connection to the database is lost in the middle of a transaction. In this situation the transaction is aborted but it can be re-tried without any changes. Similarly, the odb::timeout exception is thrown if one of the database operations or the whole transaction has timed out. Again, in this case the transaction is aborted but can be re-tried as is.

If two or more transactions access or modify more than one object and are executed concurrently by different applications or by different threads within the same application, then it is possible that these transactions will try to access objects in an incompatible order and deadlock. The canonical example of a deadlock are two transactions in which the first has modified object1 and is waiting for the second transaction to commit its changes to object2 so that it can also update object2. At the same time the second transaction has modified object2 and is waiting for the first transaction to commit its changes to object1 because it also needs to modify object1. As a result, none of the two transactions can be completed.

The database system detects such situations and automatically aborts the waiting operation in one of the deadlocked transactions. In ODB this translates to the odb::deadlock recoverable exception being thrown from one of the database functions.

The following code fragment shows how to handle the recoverable exceptions by restarting the affected transaction:

const unsigned short max_retries = 5;

for (unsigned short retry_count (0); ; retry_count++)
{
  try
  {
    transaction t (db.begin ());

    ...

    t.commit ();
    break;
  }
  catch (const odb::recoverable& e)
  {
    if (retry_count > max_retries)
      throw retry_limit_exceeded (e.what ());
    else
      continue;
  }
}
  

3.8 Making Objects Persistent

A newly created instance of a persistent class is transient. We use the database::persist() function template to make a transient instance persistent. This function has four overloaded versions with the following signatures:

  template <typename T>
  typename object_traits<T>::id_type
  persist (const T& object);

  template <typename T>
  typename object_traits<T>::id_type
  persist (const object_traits<T>::const_pointer_type& object);

  template <typename T>
  typename object_traits<T>::id_type
  persist (T& object);

  template <typename T>
  typename object_traits<T>::id_type
  persist (const object_traits<T>::pointer_type& object);
  

Here and in the rest of the manual, object_traits<T>::pointer_type and object_traits<T>::const_pointer_type denote the unrestricted and constant object pointer types (Section 3.3, "Object and View Pointers"), respectively. Similarly, object_traits<T>::id_type denotes the object id type. The odb::object_traits template is part of the database support code generated by the ODB compiler.

The first persist() function expects a constant reference to an instance being persisted. The second function expects a constant object pointer. Both of these functions can only be used on objects with application-assigned object ids (Section 14.4.2, "auto").

The second and third persist() functions are similar to the first two except that they operate on unrestricted references and object pointers. If the identifier of the object being persisted is assigned by the database, these functions update the id member of the passed instance with the assigned value. All four functions return the object id of the newly persisted object.

If the database already contains an object of this type with this identifier, the persist() functions throw the odb::object_already_persistent exception. This should never happen for database-assigned object ids as long as the number of objects persisted does not exceed the value space of the id type.

When calling the persist() functions, we don't need to explicitly specify the template type since it will be automatically deduced from the argument being passed. The following example shows how we can call these functions:

person john ("John", "Doe", 33);
shared_ptr<person> jane (new person ("Jane", "Doe", 32));

transaction t (db.begin ());

db.persist (john);
unsigned long jane_id (db.persist (jane));

t.commit ();

cerr << "Jane's id: " << jane_id << endl;
  

Notice that in the above code fragment we have created instances that we were planning to make persistent before starting the transaction. Likewise, we printed Jane's id after we have committed the transaction. As a general rule, you should avoid performing operations within the transaction scope that can be performed before the transaction starts or after it terminates. An active transaction consumes both your application's resources, such as a database connection, as well as the database server's resources, such as object locks. By following the above rule you make sure these resources are released and made available to other threads in your application and to other applications as soon as possible.

Some database systems support persisting multiple objects with a single underlying statement execution which can result in significantly improved performance. For such database systems ODB provides bulk persist() functions. For details, refer to Section 15.3, "Bulk Database Operations".

3.9 Loading Persistent Objects

Once an object is made persistent, and you know its object id, it can be loaded by the application using the database::load() function template. This function has two overloaded versions with the following signatures:

  template <typename T>
  typename object_traits<T>::pointer_type
  load (const typename object_traits<T>::id_type& id);

  template <typename T>
  void
  load (const typename object_traits<T>::id_type& id, T& object);
  

Given an object id, the first function allocates a new instance of the object class in the dynamic memory, loads its state from the database, and returns the pointer to the new instance. The second function loads the object's state into an existing instance. Both functions throw odb::object_not_persistent if there is no object of this type with this id in the database.

When we call the first load() function, we need to explicitly specify the object type. We don't need to do this for the second function because the object type will be automatically deduced from the second argument, for example:

transaction t (db.begin ());

auto_ptr<person> jane (db.load<person> (jane_id));

db.load (jane_id, *jane);

t.commit ();
  

In certain situations it may be necessary to reload the state of an object from the database. While this is easy to achieve using the second load() function, ODB provides the database::reload() function template that has a number of special properties. This function has two overloaded versions with the following signatures:

  template <typename T>
  void
  reload (T& object);

  template <typename T>
  void
  reload (const object_traits<T>::pointer_type& object);
  

The first reload() function expects an object reference, while the second expects an object pointer. Both functions expect the id member in the passed object to contain a valid object identifier and, similar to load(), both will throw odb::object_not_persistent if there is no object of this type with this id in the database.

The first special property of reload() compared to the load() function is that it does not interact with the session's object cache (Section 11.1, "Object Cache"). That is, if the object being reloaded is already in the cache, then it will remain there after reload() returns. Similarly, if the object is not in the cache, then reload() won't put it there either.

The second special property of the reload() function only manifests itself when operating on an object with the optimistic concurrency model. In this case, if the states of the object in the application memory and in the database are the same, then no reloading will occur. For more information on optimistic concurrency, refer to Chapter 12, "Optimistic Concurrency".

If we don't know for sure whether an object with a given id is persistent, we can use the find() function instead of load(), for example:

  template <typename T>
  typename object_traits<T>::pointer_type
  find (const typename object_traits<T>::id_type& id);

  template <typename T>
  bool
  find (const typename object_traits<T>::id_type& id, T& object);
  

If an object with this id is not found in the database, the first find() function returns a NULL pointer while the second function leaves the passed instance unmodified and returns false.

If we don't know the object id, then we can use queries to find the object (or objects) matching some criteria (Chapter 4, "Querying the Database"). Note, however, that loading an object's state using its identifier can be significantly faster than executing a query.

3.10 Updating Persistent Objects

If a persistent object has been modified, we can store the updated state in the database using the database::update() function template. This function has three overloaded versions with the following signatures:

  template <typename T>
  void
  update (const T& object);

  template <typename T>
  void
  update (const object_traits<T>::const_pointer_type& object);

  template <typename T>
  void
  update (const object_traits<T>::pointer_type& object);
  

The first update() function expects an object reference, while the other two expect object pointers. If the object passed to one of these functions does not exist in the database, update() throws the odb::object_not_persistent exception (but see a note on optimistic concurrency below).

Below is an example of the funds transfer that we talked about in the earlier section on transactions. It uses the hypothetical bank_account persistent class:

void
transfer (database& db,
          unsigned long from_acc,
          unsigned long to_acc,
          unsigned int amount)
{
  bank_account from, to;

  transaction t (db.begin ());

  db.load (from_acc, from);

  if (from.balance () < amount)
    throw insufficient_funds ();

  db.load (to_acc, to);

  to.balance (to.balance () + amount);
  from.balance (from.balance () - amount);

  db.update (to);
  db.update (from);

  t.commit ();
}
  

The same can be accomplished using dynamically allocated objects and the update() function with object pointer argument, for example:

transaction t (db.begin ());

shared_ptr<bank_account> from (db.load<bank_account> (from_acc));

if (from->balance () < amount)
  throw insufficient_funds ();

shared_ptr<bank_account> to (db.load<bank_account> (to_acc));

to->balance (to->balance () + amount);
from->balance (from->balance () - amount);

db.update (to);
db.update (from);

t.commit ();
  

If any of the update() functions are operating on a persistent class with the optimistic concurrency model, then they will throw the odb::object_changed exception if the state of the object in the database has changed since it was last loaded into the application memory. Furthermore, for such classes, update() no longer throws the object_not_persistent exception if there is no such object in the database. Instead, this condition is treated as a change of object state and object_changed is thrown instead. For a more detailed discussion of optimistic concurrency, refer to Chapter 12, "Optimistic Concurrency".

In ODB, persistent classes, composite value types, as well as individual data members can be declared read-only (see Section 14.1.4, "readonly (object)", Section 14.3.6, "readonly (composite value)", and Section 14.4.12, "readonly (data member)").

If an individual data member is declared read-only, then any changes to this member will be ignored when updating the database state of an object using any of the above update() functions. A const data member is automatically treated as read-only. If a composite value is declared read-only then all its data members are treated as read-only.

If the whole object is declared read-only then the database state of this object cannot be changed. Calling any of the above update() functions for such an object will result in a compile-time error.

Similar to persist(), for database systems that support this functionality, ODB provides bulk update() functions. For details, refer to Section 15.3, "Bulk Database Operations".

3.11 Deleting Persistent Objects

To delete a persistent object's state from the database we use the database::erase() or database::erase_query() function templates. If the application still has an instance of the erased object, this instance becomes transient. The erase() function has the following overloaded versions:

  template <typename T>
  void
  erase (const T& object);

  template <typename T>
  void
  erase (const object_traits<T>::const_pointer_type& object);

  template <typename T>
  void
  erase (const object_traits<T>::pointer_type& object);

  template <typename T>
  void
  erase (const typename object_traits<T>::id_type& id);
  

The first erase() function uses an object itself, in the form of an object reference, to delete its state from the database. The next two functions accomplish the same result but using object pointers. Note that all three functions leave the passed object unchanged. It simply becomes transient. The last function uses the object id to identify the object to be deleted. If the object does not exist in the database, then all four functions throw the odb::object_not_persistent exception (but see a note on optimistic concurrency below).

We have to specify the object type when calling the last erase() function. The same is unnecessary for the first three functions because the object type will be automatically deduced from their arguments. The following example shows how we can call these functions:

person& john = ...
shared_ptr<jane> jane = ...
unsigned long joe_id = ...

transaction t (db.begin ());

db.erase (john);
db.erase (jane);
db.erase<person> (joe_id);

t.commit ();
  

If any of the erase() functions except the last one are operating on a persistent class with the optimistic concurrency model, then they will throw the odb::object_changed exception if the state of the object in the database has changed since it was last loaded into the application memory. Furthermore, for such classes, erase() no longer throws the object_not_persistent exception if there is no such object in the database. Instead, this condition is treated as a change of object state and object_changed is thrown instead. For a more detailed discussion of optimistic concurrency, refer to Chapter 12, "Optimistic Concurrency".

Similar to persist() and update(), for database systems that support this functionality, ODB provides bulk erase() functions. For details, refer to Section 15.3, "Bulk Database Operations".

The erase_query() function allows us to delete the state of multiple objects matching certain criteria. It uses the query expression of the database::query() function (Chapter 4, "Querying the Database") and, because the ODB query facility is optional, it is only available if the --generate-query ODB compiler option was specified. The erase_query() function has the following overloaded versions:

  template <typename T>
  unsigned long long
  erase_query ();

  template <typename T>
  unsigned long long
  erase_query (const odb::query<T>&);
  

The first erase_query() function is used to delete the state of all the persistent objects of a given type stored in the database. The second function uses the passed query instance to only delete the state of objects matching the query criteria. Both functions return the number of objects erased. When calling the erase_query() function, we have to explicitly specify the object type we are erasing. For example:

typedef odb::query<person> query;

transaction t (db.begin ());

db.erase_query<person> (query::last == "Doe" && query::age < 30);

t.commit ();
  

Unlike the query() function, when calling erase_query() we cannot use members from pointed-to objects in the query expression. However, we can still use a member corresponding to a pointer as an ordinary object member that has the id type of the pointed-to object (Chapter 6, "Relationships"). This allows us to compare object ids as well as test the pointer for NULL. As an example, the following transaction makes sure that all the employee objects that reference an employer object that is about to be deleted are deleted as well. Here we assume that the employee class contains a pointer to the employer class. Refer to Chapter 6, "Relationships" for complete definitions of these classes.

typedef odb::query<employee> query;

transaction t (db.begin ());

employer& e = ... // Employer object to be deleted.

db.erase_query<employee> (query::employer == e.id ());
db.erase (e);

t.commit ();
  

3.12 Executing Native SQL Statements

In some situations we may need to execute native SQL statements instead of using the object-oriented database API described above. For example, we may want to tune the database schema generated by the ODB compiler or take advantage of a feature that is specific to the database system we are using. The database::execute() function, which has three overloaded versions, provides this functionality:

  unsigned long long
  execute (const char* statement);

  unsigned long long
  execute (const std::string& statement);

  unsigned long long
  execute (const char* statement, std::size_t length)
  

The first execute() function expects the SQL statement as a zero-terminated C-string. The last version expects the explicit statement length as the second argument and the statement itself may contain '\0' characters, for example, to represent binary data, if the database system supports it. All three functions return the number of rows that were affected by the statement. For example:

transaction t (db.begin ());

db.execute ("DROP TABLE test");
db.execute ("CREATE TABLE test (n INT PRIMARY KEY)");

t.commit ();
  

While these functions must always be called within a transaction, it may be necessary to execute a native statement outside a transaction. This can be done using the connection::execute() functions as described in Section 3.6, "Connections".

3.13 Tracing SQL Statement Execution

Oftentimes it is useful to understand what SQL statements are executed as a result of high-level database operations. For example, we can use this information to figure out why certain transactions don't produce desired results or why they take longer than expected.

While this information can usually be obtained from the database logs, ODB provides an application-side SQL statement tracing support that is both more convenient and finer-grained. For example, in a typical situation that calls for tracing we would like to see the SQL statements executed as a result of a specific transaction. While it may be difficult to extract such a subset of statements from the database logs, it is easy to achieve with ODB tracing support:

transaction t (db.begin ());
t.tracer (stderr_tracer);

...

t.commit ();
  

ODB allows us to specify a tracer on the database, connection, and transaction levels. If specified for the database, then all the statements executed on this database will be traced. On the other hand, if a tracer is specified for the connection, then only the SQL statements executed on this connection will be traced. Similarly, a tracer specified for a transaction will only show statements that are executed as part of this transaction. All three classes (odb::database, odb::connection, and odb::transaction) provide the identical tracing API:

  void
  tracer (odb::tracer&);

  void
  tracer (odb::tracer*);

  odb::tracer*
  tracer () const;
  

The first two tracer() functions allow us to set the tracer object with the second one allowing us to clear the current tracer by passing a NULL pointer. The last tracer() function allows us to get the current tracer object. It returns a NULL pointer if there is no tracer in effect. Note that the tracing API does not manage the lifetime of the tracer object. The tracer should be valid for as long as it is being used. Furthermore, the tracing API is not thread-safe. Trying to set a tracer from multiple threads simultaneously will result in undefined behavior.

The odb::tracer class defines a callback interface that can be used to create custom tracer implementations. The odb::stderr_tracer and odb::stderr_full_tracer are built-in tracer implementations provided by the ODB runtime. They both print SQL statements being executed to the standard error stream. The full tracer, in addition to tracing statement executions, also traces their preparations and deallocations. One situation where the full tracer can be particularly useful is if a statement (for example a custom query) contains a syntax error. In this case the error will be detected during preparation and, as a result, the statement will never be executed. The only way to see such a statement is by using the full tracing.

The odb::tracer class is defined in the <odb/tracer.hxx> header file which you will need to include in order to make this class available in your application. The odb::tracer interface provided the following callback functions:

namespace odb
{
  class tracer
  {
  public:
    virtual void
    prepare (connection&, const statement&);

    virtual void
    execute (connection&, const statement&);

    virtual void
    execute (connection&, const char* statement) = 0;

    virtual void
    deallocate (connection&, const statement&);
  };
}
  

The prepare() and deallocate() functions are called when a prepared statement is created and destroyed, respectively. The first execute() function is called when a prepared statement is executed while the second one is called when a normal statement is executed. The default implementations for the prepare() and deallocate() functions do nothing while the first execute() function calls the second one passing the statement text as the second argument. As a result, if all you are interested in are the SQL statements being executed, then you only need to override the second execute() function.

In addition to the common odb::tracer interface, each database runtime provides a database-specific version as odb::<database>::tracer. It has exactly the same interface as the common version except that the connection and statement types are database-specific, which gives us access to additional, database-specific information.

As an example, consider a more elaborate, PostgreSQL-specific tracer implementation. Here we rely on the fact that the PostgreSQL ODB runtime uses names to identify prepared statements and this information can be obtained from the odb::pgsql::statement object:

#include <odb/pgsql/tracer.hxx>
#include <odb/pgsql/database.hxx>
#include <odb/pgsql/connection.hxx>
#include <odb/pgsql/statement.hxx>

class pgsql_tracer: public odb::pgsql::tracer
{
  virtual void
  prepare (odb::pgsql::connection& c, const odb::pgsql::statement& s)
  {
    cerr << c.database ().db () << ": PREPARE " << s.name ()
         << " AS " << s.text () << endl;
  }

  virtual void
  execute (odb::pgsql::connection& c, const odb::pgsql::statement& s)
  {
    cerr << c.database ().db () << ": EXECUTE " << s.name () << endl;
  }

  virtual void
  execute (odb::pgsql::connection& c, const char* statement)
  {
    cerr << c.database ().db () << ": " << statement << endl;
  }

  virtual void
  deallocate (odb::pgsql::connection& c, const odb::pgsql::statement& s)
  {
    cerr << c.database ().db () << ": DEALLOCATE " << s.name () << endl;
  }
};
  

Note also that you can only set a database-specific tracer object using a database-specific database instance, for example:

pgsql_tracer tracer;

odb::database& db = ...;
db.tracer (tracer); // Compile error.

odb::pgsql::database& db = ...;
db.tracer (tracer); // Ok.
  

3.14 ODB Exceptions

In the previous sections we have already mentioned some of the exceptions that can be thrown by the database functions. In this section we will discuss the ODB exception hierarchy and document all the exceptions that can be thrown by the common ODB runtime.

The root of the ODB exception hierarchy is the abstract odb::exception class. This class derives from std::exception and has the following interface:

namespace odb
{
  struct exception: std::exception
  {
    virtual const char*
    what () const throw () = 0;
  };
}
  

Catching this exception guarantees that we will catch all the exceptions thrown by ODB. The what() function returns a human-readable description of the condition that triggered the exception.

The concrete exceptions that can be thrown by ODB are presented in the following listing:

namespace odb
{
  struct null_pointer: exception
  {
    virtual const char*
    what () const throw ();
  };

  // Transaction exceptions.
  //
  struct already_in_transaction: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct not_in_transaction: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct transaction_already_finalized: exception
  {
    virtual const char*
    what () const throw ();
  };

  // Session exceptions.
  //
  struct already_in_session: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct not_in_session: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct session_required: exception
  {
    virtual const char*
    what () const throw ();
  };

  // Database operations exceptions.
  //
  struct recoverable: exception
  {
  };

  struct connection_lost: recoverable
  {
    virtual const char*
    what () const throw ();
  };

  struct timeout: recoverable
  {
    virtual const char*
    what () const throw ();
  };

  struct deadlock: recoverable
  {
    virtual const char*
    what () const throw ();
  };

  struct object_not_persistent: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct object_already_persistent: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct object_changed: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct result_not_cached: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct database_exception: exception
  {
  };

  // Polymorphism support exceptions.
  //
  struct abstract_class: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct no_type_info: exception
  {
    virtual const char*
    what () const throw ();
  };

  // Prepared query support exceptions.
  //
  struct prepared_already_cached: exception
  {
    const char*
    name () const;

    virtual const char*
    what () const throw ();
  };

  struct prepared_type_mismatch: exception
  {
    const char*
    name () const;

    virtual const char*
    what () const throw ();
  };

  // Schema catalog exceptions.
  //
  struct unknown_schema: exception
  {
    const std::string&
    name () const;

    virtual const char*
    what () const throw ();
  };

  struct unknown_schema_version: exception
  {
    schema_version
    version () const;

    virtual const char*
    what () const throw ();
  };

  // Section exceptions.
  //
  struct section_not_loaded: exception
  {
    virtual const char*
    what () const throw ();
  };

  struct section_not_in_object: exception
  {
    virtual const char*
    what () const throw ();
  };

  // Bulk operation exceptions.
  //
  struct multiple_exceptions: exception
  {
    ...

    virtual const char*
    what () const throw ();
  };
}
  

The null_pointer exception is thrown when a pointer to a persistent object declared non-NULL with the db not_null or db value_not_null pragma has the NULL value. See Chapter 6, "Relationships" for details.

The next three exceptions (already_in_transaction, not_in_transaction, transaction_already_finalized) are thrown by the odb::transaction class and are discussed in Section 3.5, "Transactions".

The next two exceptions (already_in_session, and not_in_session) are thrown by the odb::session class and are discussed in Chapter 11, "Session".

The session_required exception is thrown when ODB detects that correctly loading a bidirectional object relationship requires a session but one is not used. See Section 6.2, "Bidirectional Relationships" for more information on this exception.

The recoverable exception serves as a common base for all the recoverable exceptions, which are: connection_lost, timeout, and deadlock. The connection_lost exception is thrown when a connection to the database is lost. Similarly, the timeout exception is thrown if one of the database operations or the whole transaction has timed out. The deadlock exception is thrown when a transaction deadlock is detected by the database system. These exceptions can be thrown by any database function. See Section 3.7, "Error Handling and Recovery" for details.

The object_already_persistent exception is thrown by the persist() database function. See Section 3.8, "Making Objects Persistent" for details.

The object_not_persistent exception is thrown by the load(), update(), and erase() database functions. Refer to Section 3.9, "Loading Persistent Objects", Section 3.10, "Updating Persistent Objects", and Section 3.11, "Deleting Persistent Objects" for more information.

The object_changed exception is thrown by the update() database function and certain erase() database functions when operating on objects with the optimistic concurrency model. See Chapter 12, "Optimistic Concurrency" for details.

The result_not_cached exception is thrown by the query result class. Refer to Section 4.4, "Query Result" for details.

The database_exception exception is a base class for all database system-specific exceptions that are thrown by the database system-specific runtime library. Refer to Part II, "Database Systems" for more information.

The abstract_class exception is thrown by the database functions when we attempt to persist, update, load, or erase an instance of a polymorphic abstract class. For more information on abstract classes, refer to Section 14.1.3, "abstract".

The no_type_info exception is thrown by the database functions when we attempt to persist, update, load, or erase an instance of a polymorphic class for which no type information is present in the application. This normally means that the generated database support code for this class has not been linked (or dynamically loaded) into the application or the discriminator value has not been mapped to a persistent class. For more information on polymorphism support, refer to Section 8.2, "Polymorphism Inheritance".

The prepared_already_cached exception is thrown by the cache_query() function if a prepared query with the specified name is already cached. The prepared_type_mismatch exception is thrown by the lookup_query() function if the specified prepared query object type or parameters type does not match the one in the cache. Refer to Section 4.5, "Prepared Queries" for details.

The unknown_schema exception is thrown by the odb::schema_catalog class if a schema with the specified name is not found. Refer to Section 3.4, "Database" for details. The unknown_schema_version exception is thrown by the schema_catalog functions that deal with database schema evolution if the passed version is unknow. Refer to Chapter 13, "Database Schema Evolution" for details.

The section_not_loaded exception is thrown if we attempt to update an object section that hasn't been loaded. The section_not_in_object exception is thrown if the section instance being loaded or updated does not belong to the corresponding object. See Chapter 9, "Sections" for more information on these exceptions.

The multiple_exceptions exception is thrown by the bulk API functions. Refer to Section 15.3, "Bulk Database Operations" for details.

The odb::exception class is defined in the <odb/exception.hxx> header file. All the concrete ODB exceptions are defined in <odb/exceptions.hxx> which also includes <odb/exception.hxx>. Normally you don't need to include either of these two headers because they are automatically included by <odb/database.hxx>. However, if the source file that handles ODB exceptions does not include <odb/database.hxx>, then you will need to explicitly include one of these headers.


4 Querying the Database

If we don't know the identifiers of the objects that we are looking for, we can use queries to search the database for objects matching certain criteria. The ODB query facility is optional and we need to explicitly request the generation of the necessary database support code with the --generate-query ODB compiler option.

ODB provides a flexible query API that offers two distinct levels of abstraction from the database system query language such as SQL. At the high level we are presented with an easy to use yet powerful object-oriented query language, called ODB Query Language. This query language is modeled after and is integrated into C++ allowing us to write expressive and safe queries that look and feel like ordinary C++. We have already seen examples of these queries in the introductory chapters. Below is another, more interesting, example:

  typedef odb::query<person> query;
  typedef odb::result<person> result;

  unsigned short age;
  query q (query::first == "John" && query::age < query::_ref (age));

  for (age = 10; age < 100; age += 10)
  {
    result r (db.query<person> (q));
    ...
  }
  

At the low level, queries can be written as predicates using the database system-native query language such as the WHERE predicate from the SQL SELECT statement. This language will be referred to as native query language. At this level ODB still takes care of converting query parameters from C++ to the database system format. Below is the re-implementation of the above example using SQL as the native query language:

  query q ("first = 'John' AND age = " + query::_ref (age));
  

Note that at this level we lose the static typing of query expressions. For example, if we wrote something like this:

  query q (query::first == 123 && query::agee < query::_ref (age));
  

We would get two errors during the C++ compilation. The first would indicate that we cannot compare query::first to an integer and the second would pick the misspelling in query::agee. On the other hand, if we wrote something like this:

  query q ("first = 123 AND agee = " + query::_ref (age));
  

It would compile fine and would trigger an error only when executed by the database system.

We can also combine the two query languages in a single query, for example:

  query q ("first = 'John' AND" + (query::age < query::_ref (age)));
  

4.1 ODB Query Language

An ODB query is an expression that tells the database system whether any given object matches the desired criteria. As such, a query expression always evaluates as true or false. At the higher level, an expression consists of other expressions combined with logical operators such as && (AND), || (OR), and ! (NOT). For example:

  typedef odb::query<person> query;

  query q (query::first == "John" || query::age == 31);
  

At the core of every query expression lie simple expressions which involve one or more object members, values, or parameters. To refer to an object member we use an expression such as query::first above. The names of members in the query class are derived from the names of data members in the object class by removing the common member name decorations, such as leading and trailing underscores, the m_ prefix, etc.

In a simple expression an object member can be compared to a value, parameter, or another member using a number of predefined operators and functions. The following table gives an overview of the available expressions:

Operator Description Example
== equal query::age == 31
!= unequal query::age != 31
< less than query::age < 31
> greater than query::age > 31
<= less than or equal query::age <= 31
>= greater than or equal query::age >= 31
in() one of the values query::age.in (30, 32, 34)
in_range() one of the values in range query::age.in_range (begin, end)
like() matches a pattern query::first.like ("J%")
is_null() value is NULL query::age.is_null ()
is_not_null() value is NOT NULL query::age.is_not_null ()

The in() function accepts a maximum of five arguments. Use the in_range() function if you need to compare to more than five values. This function accepts a pair of standard C++ iterators and compares to all the values from the begin position inclusive and until and excluding the end position. The following code fragment shows how we can use these functions:

  std::vector<string> names;

  names.push_back ("John");
  names.push_back ("Jack");
  names.push_back ("Jane");

  query q1 (query::first.in ("John", "Jack", "Jane"));
  query q2 (query::first.in_range (names.begin (), names.end ()));
  

Note that the like() function does not perform any translation of the database system-specific extensions of the SQL LIKE operator. As a result, if you would like your application to be portable among various database systems, then limit the special characters used in the pattern to % (matches zero or more characters) and _ (matches exactly one character). It is also possible to specify the escape character as a second argument to the like() function. This character can then be used to escape the special characters (% and _) in the pattern. For example, the following query will match any two characters separated by an underscore:

  query q (query::name.like ("_!__", "!"));
  

The operator precedence in the query expressions are the same as for equivalent C++ operators. We can use parentheses to make sure the expression is evaluated in the desired order. For example:

  query q ((query::first == "John" || query::first == "Jane") &&
           query::age < 31);
  

4.2 Parameter Binding

An instance of the odb::query class encapsulates two parts of information about the query: the query expression and the query parameters. Parameters can be bound to C++ variables either by value or by reference.

If a parameter is bound by value, then the value for this parameter is copied from the C++ variable to the query instance at the query construction time. On the other hand, if a parameter is bound by reference, then the query instance stores a reference to the bound variable. The actual value of the parameter is only extracted at the query execution time. Consider, for example, the following two queries:

  string name ("John");

  query q1 (query::first == query::_val (name));
  query q2 (query::first == query::_ref (name));

  name = "Jane";

  db.query<person> (q1); // Find John.
  db.query<person> (q2); // Find Jane.
  

The odb::query class provides two special functions, _val() and _ref(), that allow us to bind the parameter either by value or by reference, respectively. In the ODB query language, if the binding is not specified explicitly, the value semantic is used by default. In the native query language, binding must always be specified explicitly. For example:

  query q1 (query::age < age);                // By value.
  query q2 (query::age < query::_val (age));  // By value.
  query q3 (query::age < query::_ref (age));  // By reference.

  query q4 ("age < " + age);                  // Error.
  query q5 ("age < " + query::_val (age));    // By value.
  query q6 ("age < " + query::_ref (age));    // By reference.
  

A query that only has by-value parameters does not depend on any other variables and is self-sufficient once constructed. A query that has one or more by-reference parameters depends on the bound variables until the query is executed. If one such variable goes out of scope and we execute the query, the behavior is undefined.

4.3 Executing a Query

Once we have the query instance ready and by-reference parameters initialized, we can execute the query using the database::query() function template. It has two overloaded versions:

  template <typename T>
  result<T>
  query (bool cache = true);

  template <typename T>
  result<T>
  query (const odb::query<T>&, bool cache = true);
  

The first query() function is used to return all the persistent objects of a given type stored in the database. The second function uses the passed query instance to only return objects matching the query criteria. The cache argument determines whether the objects' states should be cached in the application's memory or if they should be returned by the database system one by one as the iteration over the result progresses. The result caching is discussed in detail in the next section.

When calling the query() function, we have to explicitly specify the object type we are querying. For example:

  typedef odb::query<person> query;
  typedef odb::result<person> result;

  result all (db.query<person> ());
  result johns (db.query<person> (query::first == "John"));
  

Note that it is not required to explicitly create a named query variable before executing it. For example, the following two queries are equivalent:

  query q (query::first == "John");

  result r1 (db.query<person> (q));
  result r1 (db.query<person> (query::first == "John"));
  

Normally, we would create a named query instance if we are planning to run the same query multiple times and would use the in-line version for those that are executed only once (see also Section 4.5, "Prepared Queries" for a more optimal way to re-execute the same query multiple times). A named query instance that does not have any by-reference parameters is immutable and can be shared between multiple threads without synchronization. On the other hand, a query instance with by-reference parameters is modified every time it is executed. If such a query is shared among multiple threads, then access to this query instance must be synchronized from the execution point and until the completion of the iteration over the result.

It is also possible to create queries from other queries by combining them using logical operators. For example:

result
find_minors (database& db, const query& name_query)
{
  return db.query<person> (name_query && query::age < 18);
}

result r (find_minors (db, query::first == "John"));
  

The result of executing a query is zero, one, or more objects matching the query criteria. The query() function returns this result as an instance of the odb::result class template, which provides a stream-like interface and is discussed in detail in the next section.

In situations where we know that a query produces at most one element, we can instead use the database::query_one() and database::query_value() shortcut functions, for example:

  typedef odb::query<person> query;

  auto_ptr<person> p (
    db.query_one<person> (
      query::email == "jon@example.com"));
  

The shortcut query functions have the following signatures:

  template <typename T>
  typename object_traits<T>::pointer_type
  query_one ();

  template <typename T>
  bool
  query_one (T&);

  template <typename T>
  T
  query_value ();

  template <typename T>
  typename object_traits<T>::pointer_type
  query_one (const odb::query<T>&);

  template <typename T>
  bool
  query_one (const odb::query<T>&, T&);

  template <typename T>
  T
  query_value (const odb::query<T>&);
  

Similar to query(), the first three functions are used to return the only persistent object of a given type stored in the database. The second three versions use the passed query instance to only return the object matching the query criteria.

Similar to the database::find() functions (Section 3.9, "Loading Persistent Objects"), query_one() can either allocate a new instance of the object class in the dynamic memory or it can load the object's state into an existing instance. The query_value() function allocates and returns the object by value.

The query_one() function allows us to determine if the query result contains zero or one element. If no objects matching the query criteria were found in the database, the first version of query_one() returns the NULL pointer while the second — false. If the second version returns false, then the passed object remains unchanged. For example:

  if (unique_ptr<person> p = db.query_one<person> (
        query::email == "jon@example.com"))
  {
    ...
  }

  person p;
  if (db.query_one<person> (query::email == "jon@example.com", p))
  {
    ...
  }
  

If the query executed using query_one() or query_value() returns more than one element, then these functions fail with an assertion. Additionally, query_value() also fails with an assertion if the query returned no elements.

Common situations where we can use the shortcut functions are a query condition that uses a data member with the unique constraint (at most one element returned; see Section 14.7, "Index Definition Pragmas") as well as aggregate queries (exactly one element returned; see Chapter 10, "Views").

4.4 Query Result

The database::query() function returns the result of executing a query as an instance of the odb::result class template, for example:

  typedef odb::query<person> query;
  typedef odb::result<person> result;

  result johns (db.query<person> (query::first == "John"));
  

It is best to view an instance of odb::result as a handle to a stream, such as a file stream. While we can make a copy of a result or assign one result to another, the two instances will refer to the same result stream. Advancing the current position in one instance will also advance it in another. The result instance is only usable within the transaction it was created in. Trying to manipulate the result after the transaction has terminated leads to undefined behavior.

The odb::result class template conforms to the standard C++ sequence requirements and has the following interface:

namespace odb
{
  template <typename T>
  class result
  {
  public:
    typedef odb::result_iterator<T> iterator;

  public:
    result ();

    result (const result&);

    result&
    operator= (const result&);

    void
    swap (result&)

  public:
    iterator
    begin ();

    iterator
    end ();

  public:
    void
    cache ();

    bool
    empty () const;

    std::size_t
    size () const;
  };
}
  

The default constructor creates an empty result set. The cache() function caches the returned objects' state in the application's memory. We have already mentioned result caching when we talked about query execution. As you may remember the database::query() function caches the result unless instructed not to by the caller. The cache() function allows us to cache the result at a later stage if it wasn't already cached during query execution.

If the result is cached, the database state of all the returned objects is stored in the application's memory. Note that the actual objects are still only instantiated on demand during result iteration. It is the raw database state that is cached in memory. In contrast, for uncached results the object's state is sent by the database system one object at a time as the iteration progresses.

Uncached results can improve the performance of both the application and the database system in situations where we have a large number of objects in the result or if we will only examine a small portion of the returned objects. However, uncached results have a number of limitations. There can only be one uncached result in a transaction. Creating another result (cached or uncached) by calling database::query() will invalidate the existing uncached result. Furthermore, calling any other database functions, such as update() or erase() will also invalidate the uncached result. It also follows that uncached results cannot be used on objects with containers (Chapter 5, "Containers") since loading a container would invalidate the uncached result.

The empty() function returns true if there are no objects in the result and false otherwise. The size() function can only be called for cached results. It returns the number of objects in the result. If we call this function on an uncached result, the odb::result_not_cached exception is thrown.

To iterate over the objects in a result we use the begin() and end() functions together with the odb::result<T>::iterator type, for example:

  result r (db.query<person> (query::first == "John"));

  for (result::iterator i (r.begin ()); i != r.end (); ++i)
  {
    ...
  }
  

In C++11 we can use the auto-typed variabe instead of spelling the iterator type explicitly, for example:

  for (auto i (r.begin ()); i != r.end (); ++i)
  {
    ...
  }
  

The C++11 range-based for-loop can be used to further simplify the iteration:

  for (person& p: r)
  {
    ...
  }
  

The result iterator is an input iterator which means that the only two position operations that it supports are to move to the next object and to determine whether the end of the result stream has been reached. In fact, the result iterator can only be in two states: the current position and the end position. If we have two iterators pointing to the current position and then we advance one of them, the other will advance as well. This, for example, means that it doesn't make sense to store an iterator that points to some object of interest in the result stream with the intent of dereferencing it after the iteration is over. Instead, we would need to store the object itself.

The result iterator has the following dereference functions that can be used to access the pointed-to object:

namespace odb
{
  template <typename T>
  class result_iterator
  {
  public:
    T*
    operator-> () const;

    T&
    operator* () const;

    typename object_traits<T>::pointer_type
    load ();

    void
    load (T& x);

    typename object_traits<T>::id_type
    id ();
  };
}
  

When we call the * or -> operator, the iterator will allocate a new instance of the object class in the dynamic memory, load its state from the database state, and return a reference or pointer to the new instance. The iterator maintains the ownership of the returned object and will return the same pointer for subsequent calls to either of these operators until it is advanced to the next object or we call the first load() function (see below). For example:

  result r (db.query<person> (query::first == "John"));

  for (result::iterator i (r.begin ()); i != r.end ();)
  {
    cout << i->last () << endl; // Create an object.
    person& p (*i);             // Reference to the same object.
    cout << p.age () << endl;
    ++i;                        // Free the object.
  }
  

The overloaded result_iterator::load() functions are similar to database::load(). The first function returns a dynamically allocated instance of the current object. As an optimization, if the iterator already owns an object as a result of an earlier call to the * or -> operator, then it relinquishes the ownership of this object and returns it instead. This allows us to write code like this without worrying about a double allocation:

  result r (db.query<person> (query::first == "John"));

  for (result::iterator i (r.begin ()); i != r.end (); ++i)
  {
    if (i->last == "Doe")
    {
      auto_ptr p (i.load ());
      ...
    }
  }
  

Note, however, that because of this optimization, a subsequent to load() call to the * or -> operator results in the allocation of a new object.

The second load() function allows us to load the current object's state into an existing instance. For example:

  result r (db.query<person> (query::first == "John"));

  person p;
  for (result::iterator i (r.begin ()); i != r.end (); ++i)
  {
    i.load (p);
    cout << p.last () << endl;
    cout << i.age () << endl;
  }
  

The id() function return the object id of the current object. While we can achieve the same by loading the object and getting its id, this function is more efficient since it doesn't actually create the object. This can be useful when all we need is the object's identifier. For example:

  std::set<unsigned long> set = ...; // Persons of interest.

  result r (db.query<person> (query::first == "John"));

  for (result::iterator i (r.begin ()); i != r.end (); ++i)
  {
    if (set.find (i.id ()) != set.end ()) // No object loaded.
    {
      cout << i->first () << endl; // Object loaded.
    }
  }
  

4.5 Prepared Queries

Most modern relational database systems have the notion of a prepared statement. Prepared statements allow us to perform the potentially expensive tasks of parsing SQL, preparing the query execution plan, etc., once and then executing the same query multiple times, potentially using different values for parameters in each execution.

In ODB all the non-query database operations such as persist(), load(), update(), etc., are implemented in terms of prepared statements that are cached and reused. While the query(), query_one(), and query_one() database operations also use prepared statements, these statements are not cached or reused by default since ODB has no knowledge of whether a query will be executed multiple times or only once. Instead, ODB provides a mechanism, called prepared queries, that allows us to prepare a query once and execute it multiple times. In other words, ODB prepared queries are a thin wrapper around the underlying database's prepared statement functionality.

In most cases ODB shields the application developer from database connection management and multi-threading issues. However, when it comes to prepared queries, a basic understanding of how ODB manages these aspects is required. Conceptually, the odb::database class represents a specific database, that is, a data store. However, underneath, it maintains one or more connections to this database. A connection can be used only by a single thread at a time. When we start a transaction (by calling database::begin()), the transaction instance obtains a connection and holds on to it until the transaction is committed or rolled back. During this time no other thread can use this connection. When the transaction releases the connection, it may be closed or reused by another transaction in this or another thread. What exactly happens to a connection after it has been released depends on the connection factory that is used by the odb::database instance. For more information on connection factories, refer to Part II, "Database Systems".

A query prepared on one connection cannot be executed on another. In other words, a prepared query is associated with the connection. One important implication of this restriction is that we cannot prepare a query in one transaction and then try to execute it in another without making sure that both transactions use the same connection.

To enable the prepared query functionality we need to specify the --generate-prepared ODB compiler option. If we are planning to always prepare our queries, then we can disable the once-off query execution support by also specifying the --omit-unprepared option.

To prepare a query we use the prepare_query() function template. This function can be called on both the odb::database and odb::connection instances. The odb::database version simply obtains the connection used by the currently active transaction and calls the corresponding odb::connection version. If no transaction is currently active, then this function throws the odb::not_in_transaction exception (Section 3.5, "Transactions"). The prepare_query() function has the following signature:

  template <typename T>
  prepared_query<T>
  prepare_query (const char* name, const odb::query<T>&);
  

The first argument to the prepare_query() function is the prepared query name. This name is used as a key for prepared query caching (discussed later) and must be unique. For some databases, notably PostgreSQL, it is also used as a name of the underlying prepared statement. The name "object_query" (for example, "person_query") is reserved for the once-off queries executed by the database::query() function. Note that the prepare_query() function makes only a shallow copy of this argument, which means that the name must be valid for the lifetime of the returned prepared_query instance.

The second argument to the prepare_query() function is the query criteria. It has the same semantics as in the query() function discussed in Section 4.3, "Executing a Query". Similar to query(), we also have to explicitly specify the object type that we will be querying. For example:

typedef odb::query<person> query;
typedef odb::prepared_query<person> prep_query;

prep_query pq (
  db.prepare_query<person> ("person-age-query", query::age > 50));
  

The result of executing the prepare_query() function is the prepared_query instance that represent the prepared query. It is best to view prepared_query as a handle to the underlying prepared statement. While we can make a copy of it or assign one prepared_query to another, the two instances will refer to the same prepared statement. Once the last instance of prepared_query referencing a specific prepared statement is destroyed, this statement is released. The prepared_query class template has the following interface:

namespace odb
{
  template <typename T>
  struct prepared_query
  {
    prepared_query ();

    prepared_query (const prepared_query&)
    prepared_query& operator= (const prepared_query&)

    result<T>
    execute (bool cache = true);

    typename object_traits<T>::pointer_type
    execute_one ();

    bool
    execute_one (T& object);

    T
    execute_value ();

    const char*
    name () const;

    statement&
    statement () const;

    operator unspecified_bool_type () const;
  };
}
  

The default constructor creates an empty prepared_query instance, that is, an instance that does not reference a prepared statement and therefore cannot be executed. The only way to create a non-empty prepared query is by calling the prepare_query() function discussed above. To test whether the prepared query is empty, we can use the implicit conversion operator to a boolean type. For example:

  prepared_query<person> pq;

  if (pq)
  {
    // Not empty.
    ...
  }
  

The execute() function executes the query and returns the result instance. The cache argument indicates whether the result should be cached and has the same semantics as in the query() function. In fact, conceptually, prepare_query() and execute() are just the query() function split into two: prepare_query() takes the first query() argument (the query condition) while execute() takes the second (the cache flag). Note also that re-executing a prepared query invalidates the previous execution result, whether cached or uncached.

The execute_one() and execute_value() functions can be used as shortcuts to execute a query that is known to return at most one or exactly one object, respectively. The arguments and return values in these functions have the same semantics as in query_one() and query_value(). And similar to execute() above, prepare_query() and execute_one/value() can be seen as the query_one/value() function split into two: prepare_query() takes the first query_one/value() argument (the query condition) while execute_one/value() takes the second argument (if any) and returns the result. Note also that execute_one/value() never caches its result but invalidates the result of any previous execute() call on the same prepared query.

The name() function returns the prepared query name. This is the same name as was passed as the first argument in the prepare_query() call. The statement() function returns a reference to the underlying prepared statement. Note also that calling any of these functions on an empty prepared_query instance results in undefined behavior.

The simplest use-case for a prepared query is the need to execute the same query multiple times within a single transaction. Consider the following example that queries for people that are older than a number of different ages. This and subsequent code fragments are taken from the prepared example in the odb-examples package.

typedef odb::query<person> query;
typedef odb::prepared_query<person> prep_query;
typedef odb::result<person> result;

transaction t (db.begin ());

unsigned short age;
query q (query::age > query::_ref (age));
prep_query pq (db.prepare_query<person> ("person-age-query", q));

for (age = 90; age > 40; age -= 10)
{
  result r (pq.execute ());
  ...
}

t.commit ();
  

Another scenario is the need to reuse the same query in multiple transactions that are executed at once. As was mentioned above, in this case we need to make sure that the prepared query and all the transactions use the same connection. Consider an alternative version of the above example that executes each query in a separate transaction:

connection_ptr conn (db.connection ());

unsigned short age;
query q (query::age > query::_ref (age));
prep_query pq (conn->prepare_query<person> ("person-age-query", q));

for (age = 90; age > 40; age -= 10)
{
  transaction t (conn->begin ());

  result r (pq.execute ());
  ...

  t.commit ();
}
  

Note that with this approach we hold on to the database connection until all the transactions involving the prepared query are executed. In particular, this means that while we are busy, the connection cannot be reused by another thread. Therefore, this approach is only recommended if all the transactions are executed close to each other. Also note that an uncached (see below) prepared query is invalidated once we release the connection on which it was prepared.

If we need to reuse a prepared query in transactions that are executed at various times, potentially in different threads, then the recommended approach is to cache the prepared query on the connection. To support this functionality the odb::database and odb::connection classes provide the following function templates. Similar to prepare_query(), the odb::database versions of the below functions call the corresponding odb::connection versions using the currently active transaction to resolve the connection.

  template <typename T>
  void
  cache_query (const prepared_query<T>&);

  template <typename T, typename P>
  void
  cache_query (const prepared_query<T>&,
               std::[auto|unique]_ptr<P> params);

  template <typename T>
  prepared_query<T>
  lookup_query (const char* name) const;

  template <typename T, typename P>
  prepared_query<T>
  lookup_query (const char* name, P*& params) const;
  

The cache_query() function caches the passed prepared query on the connection. The second overloaded version of cache_query() also takes a pointer to the by-reference query parameters. In C++98/03 it should be std::auto_ptr while in C++11 std::auto_ptr or std::unique_ptr can be used. The cache_query() function assumes ownership of the passed params argument. If a prepared query with the same name is already cached on this connection, then the odb::prepared_already_cached exception is thrown.

The lookup_query() function looks up a previously cached prepared query given its name. The second overloaded version of lookup_query() also returns a pointer to the by-reference query parameters. If a prepared query with this name has not been cached, then an empty prepared_query instance is returned. If a prepared query with this name has been cached but either the object type or the parameters type does not match that which was cached, then the odb::prepared_type_mismatch exception is thrown.

As a first example of the prepared query cache functionality, consider the case that does not use any by-reference parameters:

for (unsigned short i (0); i < 5; ++i)
{
  transaction t (db.begin ());

  prep_query pq (db.lookup_query<person> ("person-age-query"));

  if (!pq)
  {
    pq = db.prepare_query<person> (
      "person-val-age-query", query::age > 50);
    db.cache_query (pq);
  }

  result r (pq.execute ());
  ...

  t.commit ();

  // Do some other work.
  //
  ...
}
  

The following example shows how to do the same but for a query that includes by-reference parameters. In this case the parameters are cached together with the prepared query.

for (unsigned short age (90); age > 40; age -= 10)
{
  transaction t (db.begin ());

  unsigned short* age_param;
  prep_query pq (
    db.lookup_query<person> ("person-age-query", age_param));

  if (!pq)
  {
    auto_ptr<unsigned short> p (new unsigned short);
    age_param = p.get ();
    query q (query::age > query::_ref (*age_param));
    pq = db.prepare_query<person> ("person-age-query", q);
    db.cache_query (pq, p); // Assumes ownership of p.
  }

  *age_param = age; // Initialize the parameter.
  result r (pq.execute ());
  ...

  t.commit ();

  // Do some other work.
  //
  ...
}
  

As is evident from the above examples, when we use a prepared query cache, each transaction that executes a query must also include code that prepares and caches this query if it hasn't already been done. If a prepared query is used in a single place in the application, then this is normally not an issue since all the relevant code is kept in one place. However, if the same query is used in several different places in the application, then we may end up duplicating the same preparation and caching code, which makes it hard to maintain.

To resolve this issue ODB allows us to register a prepared query factory that will be called to prepare and cache a query during the call to lookup_query(). To register a factory we use the database::query_factory() function. In C++98/03 it has the following signature:

  void
  query_factory (const char* name,
                 void (*factory) (const char* name, connection&));
  

While in C++11 it uses the std::function class template:

  void
  query_factory (const char* name,
                 std::function<void (const char* name, connection&)>);
  

The first argument to the query_factory() function is the prepared query name that this factory will be called to prepare and cache. An empty name is treated as a fallback wildcard factory that is capable of preparing any query. The second argument is the factory function or, in C++11, function object or lambda.

The example fragment shows how we can use the prepared query factory:

struct params
{
  unsigned short age;
  string first;
};

static void
query_factory (const char* name, connection& c)
{
  auto_ptr<params> p (new params);
  query q (query::age > query::_ref (p->age) &&
           query::first == query::_ref (p->first));
  prep_query pq (c.prepare_query<person> (name, q));
  c.cache_query (pq, p);
}

db.query_factory ("person-age-name-query", &query_factory);

for (unsigned short age (90); age > 40; age -= 10)
{
  transaction t (db.begin ());

  params* p;
  prep_query pq (db.lookup_query<person> ("person-age-name-query", p));
  assert (pq);

  p->age = age;
  p->first = "John";
  result r (pq.execute ());
  ...

  t.commit ();
}
  

In C++11 we could have instead used a lambda function as well as unique_ptr rather than auto_ptr:

db.query_factory (
  "person-age-name-query",
  [] (const char* name, connection& c)
  {
    unique_ptr<params> p (new params);
    query q (query::age > query::_ref (p->age) &&
             query::first == query::_ref (p->first));
    prep_query pq (c.prepare_query<person> (name, q));
    c.cache_query (pq, std::move (p));
  });
  

5 Containers

The ODB runtime library provides built-in persistence support for all the commonly used standard C++98/03 containers, namely, std::vector, std::list, std::deque, std::set, std::multiset, std::map, and std::multimap as well as C++11 std::array, std::forward_list, std::unordered_set, std::unordered_multiset, std::unordered_map, and std::unordered_multimap. Plus, ODB profile libraries, that are available for commonly used frameworks and libraries (such as Boost and Qt), provide persistence support for containers found in these frameworks and libraries (Part III, "Profiles"). Both the ODB runtime library and profile libraries also provide a number of change-tracking container equivalents which can be used to minimize the number of database operations necessary to synchronize the container state with the database (Section 5.4, "Change-Tracking Containers"). It is also easy to persist custom container types as discussed later in Section 5.5, "Using Custom Containers".

We don't need to do anything special to declare a member of a container type in a persistent class. For example:

#pragma db object
class person
{
  ...
private:
  std::vector<std::string> nicknames_;
  ...
};
  

The complete version of the above code fragment and the other code samples presented in this chapter can be found in the container example in the odb-examples package.

A data member in a persistent class that is of a container type behaves like a value type. That is, when an object is made persistent, the elements of the container are stored in the database. Similarly, when a persistent object is loaded from the database, the contents of the container are automatically loaded as well. A data member of a container type can also use a smart pointer, as discussed in Section 7.3, "Pointers and NULL Value Semantics".

While an ordinary member is mapped to one or more columns in the object's table, a member of a container type is mapped to a separate table. The exact schema of such a table depends on the kind of container. ODB defines the following container kinds: ordered, set, multiset, map, and multimap. The container kinds and the contents of the tables to which they are mapped are discussed in detail in the following sections.

Containers in ODB can contain simple value types (Section 7.1, "Simple Value Types"), composite value types (Section 7.2, "Composite Value Types"), and pointers to objects (Chapter 6, "Relationships"). Containers of containers, either directly or indirectly via a composite value type, are not allowed. A key in a map or multimap container can be a simple or composite value type but not a pointer to an object. An index in the ordered container should be a simple integer value type.

The value type in the ordered, set, and map containers as well as the key type in the map containers should be default-constructible. The default constructor in these types can be made private in which case the odb::access class should be made a friend of the value or key type. For example:

#pragma db value
class name
{
public:
  name (const std::string&, const std::string&);
  ...
private:
  friend class odb::access;
  name ();
  ...
};

#pragma db object
class person
{
  ...
private:
  std::vector<name> aliases_;
  ...
};
  

5.1 Ordered Containers

In ODB an ordered container is any container that maintains (explicitly or implicitly) an order of its elements in the form of an integer index. Standard C++ containers that are ordered include std::vector std::list, and std::deque as well as C++11 std::array and std::forward_list. While elements in std::set are also kept in a specific order, this order is not based on an integer index but rather on the relationship between elements. As a result, std::set is not considered an ordered container for the purpose of persistence.

The database table for an ordered container consists of at least three columns. The first column contains the object id of a persistent class instance of which the container is a member. The second column contains the element index within a container. And the last column contains the element value. If the object id or element value are composite, then, instead of a single column, they can occupy multiple columns. For an ordered container table the ODB compiler also defines two indexes: one for the object id column(s) and the other for the index column. Refer to Section 14.7, "Index Definition Pragmas" for more information on how to customize these indexes.

Consider the following persistent object as an example:

#pragma db object
class person
{
  ...
private:
  #pragma db id auto
  unsigned long id_;

  std::vector<std::string> nicknames_;
  ...
};
  

The resulting database table (called person_nicknames) will contain the object id column of type unsigned long (called object_id), the index column of an integer type (called index), and the value column of type std::string (called value).

A number of ODB pragmas allow us to customize the table name, column names, and native database types of an ordered container both, on the per-container and per-member basis. For more information on these pragmas, refer to Chapter 14, "ODB Pragma Language". The following example shows some of the possible customizations:

#pragma db object
class person
{
  ...
private:
  #pragma db table("nicknames")              \
             id_column("person_id")          \
             index_type("SMALLINT UNSIGNED") \
             index_column("nickname_number") \
             value_type("VARCHAR(255)")      \
             value_column("nickname")
  std::vector<std::string> nicknames_;
  ...
};
  

While the C++ container used in a persistent class may be ordered, sometimes we may wish to store such a container in the database without the order information. In the example above, for instance, the order of person's nicknames is probably not important. To instruct the ODB compiler to ignore the order in ordered containers we can use the db unordered pragma (Section 14.3.9, "unordered", Section 14.4.19, "unordered"). For example:

#pragma db object
class person
{
  ...
private:
  #pragma db unordered
  std::vector<std::string> nicknames_;
  ...
};
  

The table for an ordered container that is marked unordered won't have the index column and the order in which elements are retrieved from the database may not be the same as the order in which they were stored.

5.2 Set and Multiset Containers

In ODB set and multiset containers (referred to as just set containers) are associative containers that contain elements based on some relationship between them. A set container may or may not guarantee a particular order of the elements that it stores. Standard C++ containers that are considered set containers for the purpose of persistence include std::set and std::multiset as well as C++11 std::unordered_set and std::unordered_multiset.

The database table for a set container consists of at least two columns. The first column contains the object id of a persistent class instance of which the container is a member. And the second column contains the element value. If the object id or element value are composite, then, instead of a single column, they can occupy multiple columns. ODB compiler also defines an index on a set container table for the object id column(s). Refer to Section 14.7, "Index Definition Pragmas" for more information on how to customize this index.

Consider the following persistent object as an example:

#pragma db object
class person
{
  ...
private:
  #pragma db id auto
  unsigned long id_;

  std::set<std::string> emails_;
  ...
};
  

The resulting database table (called person_emails) will contain the object id column of type unsigned long (called object_id) and the value column of type std::string (called value).

A number of ODB pragmas allow us to customize the table name, column names, and native database types of a set container, both on the per-container and per-member basis. For more information on these pragmas, refer to Chapter 14, "ODB Pragma Language". The following example shows some of the possible customizations:

#pragma db object
class person
{
  ...
private:
  #pragma db table("emails")            \
             id_column("person_id")     \
             value_type("VARCHAR(255)") \
             value_column("email")
  std::set<std::string> emails_;
  ...
};
  

5.3 Map and Multimap Containers

In ODB map and multimap containers (referred to as just map containers) are associative containers that contain key-value elements based on some relationship between keys. A map container may or may not guarantee a particular order of the elements that it stores. Standard C++ containers that are considered map containers for the purpose of persistence include std::map and std::multimap as well as C++11 std::unordered_map and std::unordered_multimap.

The database table for a map container consists of at least three columns. The first column contains the object id of a persistent class instance of which the container is a member. The second column contains the element key. And the last column contains the element value. If the object id, element key, or element value are composite, then instead of a single column they can occupy multiple columns. ODB compiler also defines an index on a map container table for the object id column(s). Refer to Section 14.7, "Index Definition Pragmas" for more information on how to customize this index.

Consider the following persistent object as an example:

#pragma db object
class person
{
  ...
private:
  #pragma db id auto
  unsigned long id_;

  std::map<unsigned short, float> age_weight_map_;
  ...
};
  

The resulting database table (called person_age_weight_map) will contain the object id column of type unsigned long (called object_id), the key column of type unsigned short (called key), and the value column of type float (called value).

A number of ODB pragmas allow us to customize the table name, column names, and native database types of a map container, both on the per-container and per-member basis. For more information on these pragmas, refer to Chapter 14, "ODB Pragma Language". The following example shows some of the possible customizations:

#pragma db object
class person
{
  ...
private:
  #pragma db table("weight_map")      \
             id_column("person_id")   \
             key_type("INT UNSIGNED") \
             key_column("age")        \
             value_type("DOUBLE")     \
             value_column("weight")
  std::map<unsigned short, float> age_weight_map_;
  ...
};
  

5.4 Change-Tracking Containers

When a persistent object containing one of the standard containers is updated in the database, ODB has no knowledge of which elements were inserted, erased, or modified. As a result, ODB has no choice but to assume the whole container has changed and update the state of every single element. This can result in a significant overhead if a container contains a large number of elements and we only changed a small subset of them.

To eliminate this overhead, ODB provides a notion of change-tracking containers. A change-tracking container, besides containing its elements, just like an ordinary container, also includes the change state for each element. When it is time to update such a container in the database, ODB can use this change information to perform a minimum number of database operations necessary to synchronize the container state with the database.

The current version of the ODB runtime library provides a change-tracking equivalent of std::vector (Section 5.4.1, "Change-Tracking vector") with support for other standard container equivalents planned for future releases. ODB profile libraries also provide change-tracking equivalents for some containers found in the corresponding frameworks and libraries (Part III, "Profiles").

A change-tracking container equivalent can normally be used as a drop-in replacement for an ordinary container except for a few minor interface differences (discussed in the corresponding sub-sections). In particular, we don't need to do anything extra to effect change tracking. ODB will automatically start, stop, and reset change tracking when necessary. The following example illustrates this point using odb::vector as a replacement for std::vector.

#pragma db object
class person
{
  ...

  odb::vector<std::string> names;
};

person p; // No change tracking (not persistent).
p.names.push_back ("John Doe");

{
  transaction t (db.begin ());
  db.persist (p); // Start change tracking (persistent).
  t.commit ();
}

p.names.push_back ("Johnny Doo");

{
  transaction t (db.begin ());
  db.update (p); // One INSERT; reset change state.
  t.commit ();
}

p.names.modify (0) = "Doe, John"; // Instead of operator[].
p.names.pop_back ();

{
  transaction t (db.begin ());
  db.update (p); // One UPDATE, one DELETE; reset change state.
  t.commit ();
}

{
  transaction t (db.begin ());
  auto_ptr<person> p1 (db.load<person> (...)); // Start change tracking.
  p1->names.insert (p1->names.begin (), "Joe Do");
  db.update (*p1); // One UPDATE, one INSERT; reset change state.
  t.commit ();
}

{
  transaction t (db.begin ());
  db.erase (p); // One DELETE; stop change tracking (not persistent).
  t.commit ();
}
  

One interesting aspect of change tracking is what happens when a transaction that contains an update is later rolled back. In this case, while the change-tracking container has reset the change state (after update), actual changes were not committed to the database. Change-tracking containers handle this case by automatically registering a rollback callback and then, if it is called, marking the container as "completely changed". In this state, the container no longer tracks individual element changes and, when updated, falls back to the complete state update, just like an ordinary container. The following example illustrates this point:

person p;
p.names.push_back ("John Doe");

{
  transaction t (db.begin ());
  db.persist (p); // Start change tracking (persistent).
  t.commit ();
}

p.names.push_back ("Johnny Doo");

for (;;)
{
  try
  {
    transaction t (db.begin ());

    // First try: one INSERT.
    // Next try: one DELETE, two INSERTs.
    //
    db.update (p); // Reset change state.

    t.commit (); // If throws (rollback), mark as completely changed.
    break;
  }
  catch (const odb::recoverable&)
  {
    continue;
  }
}
  

For the interaction of change-tracking containers with change-updated object sections, refer to Section 9.4, "Sections and Change-Tracking Containers".

5.4.1 Change-Tracking vector

Class template odb::vector, defined in <odb/vector.hxx>, is a change-tracking equivalent for std::vector. It is implemented in terms of std::vector and is implicit-convertible to and implicit-constructible from const std::vector&. In particular, this means that we can use odb::vector instance anywhere const std::vector& is expected. In addition, odb::vector constant iterator (const_iterator) is the same type as that of std::vector.

odb::vector incurs 2-bit per element overhead in order to store the change state. It cannot be stored unordered in the database (Section 14.4.19 "unordered") but can be used as an inverse side of a relationship (6.2 "Bidirectional Relationships"). In this case, no change tracking is performed since no state for such a container is stored in the database.

The number of database operations required to update the state of odb::vector corresponds well to the complexity of std::vector functions. In particular, adding or removing an element from the back of the vector (for example, with push_back() and pop_back()), requires only a single database statement execution. In contrast, inserting or erasing an element somewhere in the middle of the vector will require a database statement for every element that follows it.

odb::vector replicates most of the std::vector interface as defined in both C++98/03 and C++11 standards. However, functions and operators that provide direct write access to the elements had to be altered or disabled in order to support change tracking. Additional functions used to interface with std::vector and to control the change tracking state were also added. The following listing summarizes the differences between the odb::vector and std::vector interfaces. Any std::vector function or operator not mentioned in this listing has exactly the same signature and semantics in odb::vector. Functions and operators that were disabled are shown as commented out and are followed by functions/operators that replace them.

namespace odb
{
  template <class T, class A = std::allocator<T> >
  class vector
  {
    ...

    // Element access.
    //

    //reference operator[] (size_type);
      reference modify (size_type);

    //reference at (size_type);
      reference modify_at (size_type);

    //reference front ();
      reference modify_front ();

    //reference back ();
      reference modify_back ();

    //T*        data () noexcept;
      T*        modify_data () noexcept; // C++11 only.

    // Iterators.
    //
    typedef typename std::vector<T, A>::const_iterator const_iterator;

    class iterator
    {
      ...

      // Element Access.
      //

      //reference       operator* () const;
        const_reference operator* () const;
        reference       modify () const;

      //pointer       operator-> () const;
        const_pointer operator-> () const;

      //reference       operator[] (difference_type);
        const_reference operator[] (difference_type);
        reference       modify (difference_type) const;

      // Interfacing with std::vector::iterator.
      //
      typename std::vector<T, A>::iterator base () const;
    };

    // Return std::vector iterators. The begin() functions mark
    // all the elements as modified.
    //
    typename std::vector<T, A>::iterator         mbegin ();
    typename std::vector<T, A>::iterator         mend ();
    typename std::vector<T, A>::reverse_iterator mrbegin ();
    typename std::vector<T, A>::reverse_iterator mrend ();

    // Interfacing with std::vector.
    //
    vector (const std::vector<T, A>&);
    vector (std::vector<T, A>&&); // C++11 only.

    vector& operator= (const std::vector<T, A>&);
    vector& operator= (std::vector<T, A>&&); // C++11 only.

    operator const std::vector<T, A>& () const;
    std::vector<T, A>& base ();
    const std::vector<T, A>& base ();

    // Change tracking.
    //
    bool _tracking () const;
    void _start () const;
    void _stop () const;
    void _arm (transaction&) const;
  };
}
  

The following example highlights some of the differences between the two interfaces. std::vector versions are commented out.

#include <vector>
#include <odb/vector.hxx>

void f (const std::vector<int>&);

odb::vector<int> v ({1, 2, 3});

f (v); // Ok, implicit conversion.

if (v[1] == 2) // Ok, const access.
  //v[1]++;
  v.modify (1)++;

//v.back () = 4;
v.modify_back () = 4;

for (auto i (v.begin ()); i != v.end (); ++i)
{
  if (*i != 0) // Ok, const access.
    //*i += 10;
    i.modify () += 10;
}

std::sort (v.mbegin (), v.mend ());
  

Note also the subtle difference between copy/move construction and copy/move assignment of odb::vector instances. While copy/move constructor will copy/move both the elements as well as their change state, in contrast, assignment is tracked as any other change to the vector content.

5.5 Using Custom Containers

While the ODB runtime and profile libraries provide support for a wide range of containers, it is also easy to persist custom container types or make a change-tracking version out of one.

To achieve this you will need to implement the container_traits class template specialization for your container. First, determine the container kind (ordered, set, multiset, map, or multimap) for your container type. Then use a specialization for one of the standard C++ containers found in the common ODB runtime library (libodb) as a base for your own implementation.

Once the container traits specialization is ready for your container, you will need to include it into the ODB compilation process using the --odb-epilogue option and into the generated header files with the --hxx-prologue option. As an example, suppose we have a hash table container for which we have the traits specialization implemented in the hashtable-traits.hxx file. Then, we can create an ODB compiler options file for this container and save it to hashtable.options:

# Options file for the hash table container.
#
--odb-epilogue '#include "hashtable-traits.hxx"'
--hxx-prologue '#include "hashtable-traits.hxx"'
  

Now, whenever we compile a header file that uses the hashtable container, we can specify the following command line option to make sure it is recognized by the ODB compiler as a container and the traits file is included in the generated code:

--options-file hashtable.options
  

6 Relationships

Relationships between persistent objects are expressed with pointers or containers of pointers. The ODB runtime library provides built-in support for shared_ptr/weak_ptr (TR1 or C++11), std::unique_ptr (C++11), std::auto_ptr, and raw pointers. Plus, ODB profile libraries, that are available for commonly used frameworks and libraries (such as Boost and Qt), provide support for smart pointers found in these frameworks and libraries (Part III, "Profiles"). It is also easy to add support for a custom smart pointer as discussed later in Section 6.5, "Using Custom Smart Pointers". Any supported smart pointer can be used in a data member as long as it can be explicitly constructed from the canonical object pointer (Section 3.3, "Object and View Pointers"). For example, we can use weak_ptr if the object pointer is shared_ptr.

When an object containing a pointer to another object is loaded, the pointed-to object is loaded as well. In some situations this eager loading of the relationships is undesirable since it can lead to a large number of otherwise unused objects being instantiated from the database. To support finer control over relationships loading, the ODB runtime and profile libraries provide the so-called lazy versions of the supported pointers. An object pointed-to by a lazy pointer is not loaded automatically when the containing object is loaded. Instead, we have to explicitly request the instantiation of the pointed-to object. Lazy pointers are discussed in detail in Section 6.4, "Lazy Pointers".

As a simple example, consider the following employee-employer relationship. Code examples presented in this chapter will use the shared_ptr and weak_ptr smart pointers from the TR1 (std::tr1) namespace.

#pragma db object
class employer
{
  ...

  #pragma db id
  std::string name_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  std::string first_name_;
  std::string last_name_;

  shared_ptr<employer> employer_;
};
  

By default, an object pointer can be NULL. To specify that a pointer always points to a valid object we can use the not_null pragma (Section 14.4.6, "null/not_null") for single object pointers and the value_not_null pragma (Section 14.4.28, "value_null/value_not_null") for containers of object pointers. For example:

#pragma db object
class employee
{
  ...

  #pragma db not_null
  shared_ptr<employer> current_employer_;

  #pragma db value_not_null
  std::vector<shared_ptr<employer> > previous_employers_;
};
  

In this case, if we call either persist() or update() database function on the employee object and the current_employer_ pointer or one of the pointers stored in the previous_employers_ container is NULL, then the odb::null_pointer exception will be thrown.

We don't need to do anything special to establish or navigate a relationship between two persistent objects, as shown in the following code fragment:

// Create an employer and a few employees.
//
unsigned long john_id, jane_id;
{
  shared_ptr<employer> er (new employer ("Example Inc"));
  shared_ptr<employee> john (new employee ("John", "Doe"));
  shared_ptr<employee> jane (new employee ("Jane", "Doe"));

  john->employer_ = er;
  jane->employer_ = er;

  transaction t (db.begin ());

  db.persist (er);
  john_id = db.persist (john);
  jane_id = db.persist (jane);

  t.commit ();
}

// Load a few employee objects and print their employer.
//
{
  session s;
  transaction t (db.begin ());

  shared_ptr<employee> john (db.load<employee> (john_id));
  shared_ptr<employee> jane (db.load<employee> (jane_id));

  cout << john->employer_->name_ << endl;
  cout << jane->employer_->name_ << endl;

  t.commit ();
}
  

The only notable line in the above code is the creation of a session before the second transaction starts. As discussed in Chapter 11, "Session", a session acts as a cache of persistent objects. By creating a session before loading the employee objects we make sure that their employer_ pointers point to the same employer object. Without a session, each employee would have ended up pointing to its own, private instance of the Example Inc employer.

As a general guideline, you should use a session when loading objects that have pointers to other persistent objects. A session makes sure that for a given object id, a single instance is shared among all other objects that relate to it.

We can also use data members from pointed-to objects in database queries (Chapter 4, "Querying the Database"). For each pointer in a persistent class, the query class defines a smart pointer-like member that contains members corresponding to the data members in the pointed-to object. We can then use the access via a pointer syntax (->) to refer to data members in pointed-to objects. For example, the query class for the employee object contains the employer member (its name is derived from the employer_ pointer) which in turn contains the name member (its name is derived from the employer::name_ data member of the pointed-to object). As a result, we can use the query::employer->name expression while querying the database for the employee objects. For example, the following transaction finds all the employees of Example Inc that have the Doe last name:

typedef odb::query<employee> query;
typedef odb::result<employee> result;

session s;
transaction t (db.begin ());

result r (db.query<employee> (
  query::employer->name == "Example Inc" && query::last == "Doe"));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
  cout << i->first_ << " " << i->last_ << endl;

t.commit ();
  

A query class member corresponding to a non-inverse (Section 6.2, "Bidirectional Relationships") object pointer can also be used as a normal member that has the id type of the pointed-to object. For example, the following query locates all the employee objects that don't have an associated employer object:

result r (db.query<employee> (query::employer.is_null ()));
  

An important concept to keep in mind when working with object relationships is the independence of persistent objects. In particular, when an object containing a pointer to another object is made persistent or is updated, the pointed-to object is not automatically persisted or updated. Rather, only a reference to the object (in the form of the object id) is stored for the pointed-to object in the database. The pointed-to object itself is a separate entity and should be made persistent or updated independently. By default, the same principle also applies to erasing pointed-to objects. That is, we have to make sure all the pointing objects are updated accordingly. However, in the case of erase, we can specify an alternative on-delete semantic as discussed in Section 14.4.15, "on_delete".

When persisting or updating an object containing a pointer to another object, the pointed-to object must have a valid object id. This, however, may not always be easy to achieve in complex relationships that involve objects with automatically assigned identifiers. In such cases it may be necessary to first persist an object with a pointer set to NULL and then, once the pointed-to object is made persistent and its identifier assigned, set the pointer to the correct value and update the object in the database.

Persistent object relationships can be divided into two groups: unidirectional and bidirectional. Each group in turn contains several configurations that vary depending on the cardinality of the sides of the relationship. All possible unidirectional and bidirectional configurations are discussed in the following sections.

6.1 Unidirectional Relationships

In unidirectional relationships we are only interested in navigating from object to object in one direction. Because there is no interest in navigating in the opposite direction, the cardinality of the other end of the relationship is unimportant. As a result, there are only two possible unidirectional relationships: to-one and to-many. Each of these relationships is described in the following sections. For sample code that shows how to work with these relationships, refer to the relationship example in the odb-examples package.

6.1.1 To-One Relationships

An example of a unidirectional to-one relationship is the employee-employer relationship (an employee has one employer). The following persistent C++ classes model this relationship:

#pragma db object
class employer
{
  ...

  #pragma db id
  std::string name_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db not_null
  shared_ptr<employer> employer_;
};
  

The corresponding database tables look like this:

CREATE TABLE employer (
  name VARCHAR (255) NOT NULL PRIMARY KEY);

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  employer VARCHAR (255) NOT NULL REFERENCES employer (name));
  

6.1.2 To-Many Relationships

An example of a unidirectional to-many relationship is the employee-project relationship (an employee can be involved in multiple projects). The following persistent C++ classes model this relationship:

#pragma db object
class project
{
  ...

  #pragma db id
  std::string name_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db value_not_null unordered
  std::vector<shared_ptr<project> > projects_;
};
  

The corresponding database tables look like this:

CREATE TABLE project (
  name VARCHAR (255) NOT NULL PRIMARY KEY);

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY);

CREATE TABLE employee_projects (
  object_id BIGINT UNSIGNED NOT NULL,
  value VARCHAR (255) NOT NULL REFERENCES project (name));
  

To obtain a more canonical database schema, the names of tables and columns above can be customized using ODB pragmas (Chapter 14, "ODB Pragma Language"). For example:

#pragma db object
class employee
{
  ...

  #pragma db value_not_null unordered \
             id_column("employee_id") value_column("project_name")
  std::vector<shared_ptr<project> > projects_;
};
  

The resulting employee_projects table would then look like this:

CREATE TABLE employee_projects (
  employee_id BIGINT UNSIGNED NOT NULL,
  project_name VARCHAR (255) NOT NULL REFERENCES project (name));
  

6.2 Bidirectional Relationships

In bidirectional relationships we are interested in navigating from object to object in both directions. As a result, each object class in a relationship contains a pointer to the other object. If smart pointers are used, then a weak pointer should be used as one of the pointers to avoid ownership cycles. For example:

class employee;

#pragma db object
class position
{
  ...

  #pragma db id
  unsigned long id_;

  weak_ptr<employee> employee_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db not_null
  shared_ptr<position> position_;
};
  

Note that when we establish a bidirectional relationship, we have to set both pointers consistently. One way to make sure that a relationship is always in a consistent state is to provide a single function that updates both pointers at the same time. For example:

#pragma db object
class position: public enable_shared_from_this<position>
{
  ...

  void
  fill (shared_ptr<employee> e)
  {
    employee_ = e;
    e->positions_ = shared_from_this ();
  }

private:
  weak_ptr<employee> employee_;
};

#pragma db object
class employee
{
  ...

private:
  friend class position;

  #pragma db not_null
  shared_ptr<position> position_;
};
  

At the beginning of this chapter we examined how to use a session to make sure a single object is shared among all other objects pointing to it. With bidirectional relationships involving weak pointers the use of a session becomes even more crucial. Consider the following transaction that tries to load the position object from the above example without using a session:

transaction t (db.begin ())
shared_ptr<position> p (db.load<position> (1));
...
t.commit ();
  

When we load the position object, the employee object, which it points to, is also loaded. While employee is initially stored as shared_ptr, it is then assigned to the employee_ member which is weak_ptr. Once the assignment is complete, the shared pointer goes out of scope and the only pointer that points to the newly loaded employee object is the employee_ weak pointer. And that means the employee object is deleted immediately after being loaded. To help avoid such pathological situations ODB detects cases where a newly loaded object will immediately be deleted and throws the odb::session_required exception.

As the exception name suggests, the easiest way to resolve this problem is to use a session:

session s;
transaction t (db.begin ())
shared_ptr<position> p (db.load<position> (1));
...
t.commit ();
  

In our example, the session will maintain a shared pointer to the loaded employee object preventing its immediate deletion. Another way to resolve this problem is to avoid immediate loading of the pointed-to objects using lazy weak pointers. Lazy pointers are discussed in Section 6.4, "Lazy Pointers" later in this chapter.

Above, to model a bidirectional relationship in persistent classes, we used two pointers, one in each object. While this is a natural representation in C++, it does not translate to a canonical relational model. Consider the database schema generated for the above two classes:

CREATE TABLE position (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  employee BIGINT UNSIGNED REFERENCES employee (id));

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  position BIGINT UNSIGNED NOT NULL REFERENCES position (id));
  

While this database schema is valid, it is unconventional. We have a reference from a row in the position table to a row in the employee table. We also have a reference from this same row in the employee table back to the row in the position table. From the relational point of view, one of these references is redundant since in SQL we can easily navigate in both directions using just one of these references.

To eliminate redundant database schema references we can use the inverse pragma (Section 14.4.14, "inverse") which tells the ODB compiler that a pointer is the inverse side of a bidirectional relationship. Either side of a relationship can be made inverse. For example:

#pragma db object
class position
{
  ...

  #pragma db inverse(position_)
  weak_ptr<employee> employee_;
};

#pragma db object
class employee
{
  ...

  #pragma db not_null
  shared_ptr<position> position_;
};
  

The resulting database schema looks like this:

CREATE TABLE position (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY);

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  position BIGINT UNSIGNED NOT NULL REFERENCES position (id));
  

As you can see, an inverse member does not have a corresponding column (or table, in case of an inverse container of pointers) and, from the point of view of database operations, is effectively read-only. The only way to change a bidirectional relationship with an inverse side is to set its direct (non-inverse) pointer. Also note that an ordered container (Section 5.1, "Ordered Containers") of pointers that is an inverse side of a bidirectional relationship is always treated as unordered (Section 14.4.19, "unordered") because the contents of such a container are implicitly built from the direct side of the relationship which does not contain the element order (index).

There are three distinct bidirectional relationships that we will cover in the following sections: one-to-one, one-to-many, and many-to-many. We will only talk about bidirectional relationships with inverse sides since they result in canonical database schemas. For sample code that shows how to work with these relationships, refer to the inverse example in the odb-examples package.

6.2.1 One-to-One Relationships

An example of a bidirectional one-to-one relationship is the presented above employee-position relationship (an employee fills one position and a position is filled by one employee). The following persistent C++ classes model this relationship:

class employee;

#pragma db object
class position
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db inverse(position_)
  weak_ptr<employee> employee_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db not_null
  shared_ptr<position> position_;
};
  

The corresponding database tables look like this:

CREATE TABLE position (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY);

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  position BIGINT UNSIGNED NOT NULL REFERENCES position (id));
  

If instead the other side of this relationship is made inverse, then the database tables will change as follows:

CREATE TABLE position (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  employee BIGINT UNSIGNED REFERENCES employee (id));

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY);
  

6.2.2 One-to-Many Relationships

An example of a bidirectional one-to-many relationship is the employer-employee relationship (an employer has multiple employees and an employee is employed by one employer). The following persistent C++ classes model this relationship:

class employee;

#pragma db object
class employer
{
  ...

  #pragma db id
  std::string name_;

  #pragma db value_not_null inverse(employer_)
  std::vector<weak_ptr<employee> > employees_
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db not_null
  shared_ptr<employer> employer_;
};
  

The corresponding database tables differ significantly depending on which side of the relationship is made inverse. If the one side (employer) is inverse as in the code above, then the resulting database schema looks like this:

CREATE TABLE employer (
  name VARCHAR (255) NOT NULL PRIMARY KEY);

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  employer VARCHAR (255) NOT NULL REFERENCES employer (name));
  

If instead the many side (employee) of this relationship is made inverse, then the database tables will change as follows:

CREATE TABLE employer (
  name VARCHAR (255) NOT NULL PRIMARY KEY);

CREATE TABLE employer_employees (
  object_id VARCHAR (255) NOT NULL,
  value BIGINT UNSIGNED NOT NULL REFERENCES employee (id));

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY);
  

6.2.3 Many-to-Many Relationships

An example of a bidirectional many-to-many relationship is the employee-project relationship (an employee can work on multiple projects and a project can have multiple participating employees). The following persistent C++ classes model this relationship:

class employee;

#pragma db object
class project
{
  ...

  #pragma db id
  std::string name_;

  #pragma db value_not_null inverse(projects_)
  std::vector<weak_ptr<employee> > employees_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db value_not_null unordered
  std::vector<shared_ptr<project> > projects_;
};
  

The corresponding database tables look like this:

CREATE TABLE project (
  name VARCHAR (255) NOT NULL PRIMARY KEY);

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY);

CREATE TABLE employee_projects (
  object_id BIGINT UNSIGNED NOT NULL,
  value VARCHAR (255) NOT NULL REFERENCES project (name));
  

If instead the other side of this relationship is made inverse, then the database tables will change as follows:

CREATE TABLE project (
  name VARCHAR (255) NOT NULL PRIMARY KEY);

CREATE TABLE project_employees (
  object_id VARCHAR (255) NOT NULL,
  value BIGINT UNSIGNED NOT NULL REFERENCES employee (id));

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY);
  

6.3 Circular Relationships

A relationship between two persistent classes is circular if each of them references the other. Bidirectional relationships are always circular. A unidirectional relationship combined with inheritance (Chapter 8, "Inheritance") can also be circular. For example, the employee class could derive from person which, in turn, could contain a pointer to employee.

We don't need to do anything extra if persistent classes with circular dependencies are defined in the same header file. Specifically, ODB will make sure that the database tables and foreign key constraints are created in the correct order. As a result, unless you have good reasons not to, it is recommended that you keep persistent classes with circular dependencies in the same header file.

If you have to keep such classes in separate header files, then there are two extra steps that you may need to take in order to use these classes with ODB. Consider again the example from Section 6.2.1, "One-to-One Relationships" but this time with the classes defined in separate headers:

// position.hxx
//
class employee;

#pragma db object
class position
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db inverse(position_)
  weak_ptr<employee> employee_;
};
  
// employee.hxx
//
#include "position.hxx"

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db not_null
  shared_ptr<position> position_;
};
  

Note that the position.hxx header contains only the forward declaration for employee. While this is sufficient to define a valid, from the C++ point of view, position class, the ODB compiler needs to "see" the definitions of the pointed-to persistent classes. There are several ways we can fulfil this requirement. The easiest is to simply include employee.hxx at the end of position.hxx:

// position.hxx
//
class employee;

#pragma db object
class position
{
  ...
};

#include "employee.hxx"
  

We can also limit this inclusion only to the time when position.hxx is compiled with the ODB compiler:

// position.hxx
//

...

#ifdef ODB_COMPILER
#  include "employee.hxx"
#endif
  

Finally, if we don't want to modify position.hxx, then we can add employee.hxx to the ODB compilation process with the --odb-epilogue option. For example:

odb ... --odb-epilogue "#include \"employee.hxx\"" position.hxx
  

Note also that in this example we didn't have to do anything extra for employee.hxx because it already includes position.hxx. However, if instead it relied only on the forward declaration of the position class, then we would have to handle it in the same way as position.hxx.

The other difficulty with separately defined classes involving circular relationships has to do with the correct order of foreign key constraint creation in the generated database schema. In the above example, if we generate the database schema as standalone SQL files, then we will end up with two such files: position.sql and employee.sql. If we try to execute employee.sql first, then we will get an error indicating that the table corresponding to the position class and referenced by the foreign key constraint corresponding to the position_ pointer does not yet exist.

Note that there is no such problem if the database schema is embedded in the generated C++ code instead of being produced as standalone SQL files. In this case, the ODB compiler is able to ensure the correct creation order even if the classes are defined in separate header files.

In certain cases, for example, a bidirectional relationship with an inverse side, this problem can be resolved by executing the database schema creation files in the correct order. In our example, this would be position.sql first and employee.sql second. However, this approach doesn't scale beyond simple object models.

A more robust solution to this problem is to generate the database schema for all the persistent classes into a single SQL file. This way, the ODB compiler can again ensure the correct creation order of tables and foreign keys. To instruct the ODB compiler to produce a combined schema file for several headers we can use the --generate-schema-only and --at-once options. For example:

odb ... --generate-schema-only --at-once --input-name company \
position.hxx employee.hxx
  

The result of the above command is a single company.sql file (the name is derived from the --input-name value) that contains the database creation code for both position and employee classes.

6.4 Lazy Pointers

Consider again the bidirectional, one-to-many employer-employee relationship that was presented earlier in this chapter:

class employee;

#pragma db object
class employer
{
  ...

  #pragma db id
  std::string name_;

  #pragma db value_not_null inverse(employer_)
  std::vector<weak_ptr<employee> > employees_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  #pragma db not_null
  shared_ptr<employer> employer_;
};
  

Consider also the following transaction which obtains the employer name given the employee id:

unsigned long id = ...
string name;

session s;
transaction t (db.begin ());

shared_ptr<employee> e (db.load<employee> (id));
name = e->employer_->name_;

t.commit ();
  

While this transaction looks very simple, it actually does a lot more than what meets the eye and is necessary. Consider what happens when we load the employee object: the employer_ pointer is also automatically loaded which means the employer object corresponding to this employee is also loaded. But the employer object in turn contains the list of pointers to all the employees, which are also loaded. A a result, when object relationships are involved, a simple transaction like the above can load many more objects than is necessary.

To overcome this problem ODB offers finer grained control over the relationship loading in the form of lazy pointers. A lazy pointer does not automatically load the pointed-to object when the containing object is loaded. Instead, we have to explicitly load the pointed-to object if and when we need to access it.

The ODB runtime library provides lazy counterparts for all the supported pointers, namely: odb::lazy_shared_ptr/lazy_weak_ptr for C++11 std::shared_ptr/weak_ptr, odb::tr1::lazy_shared_ptr/lazy_weak_ptr for TR1 std::tr1::shared_ptr/weak_ptr, odb::lazy_unique_ptr for C++11 std::unique_ptr, odb::lazy_auto_ptr for std::auto_ptr, and odb::lazy_ptr for raw pointers. The TR1 lazy pointers are defined in the <odb/tr1/lazy-ptr.hxx> header while all the others — in <odb/lazy-ptr.hxx>. The ODB profile libraries also provide lazy pointer implementations for smart pointers from popular frameworks and libraries (Part III, "Profiles").

While we will discuss the interface of lazy pointers in more detail shortly, the most commonly used extra function provided by these pointers is load(). This function loads the pointed-to object if it hasn't already been loaded. After the call to this function, the lazy pointer can be used in the the same way as its eager counterpart. The load() function also returns the eager pointer, in case you need to pass it around. For a lazy weak pointer, the load() function also locks the pointer.

The following example shows how we can change our employer-employee relationship to use lazy pointers. Here we choose to use lazy pointers for both sides of the relationship.

class employee;

#pragma db object
class employer
{
  ...

  #pragma db value_not_null inverse(employer_)
  std::vector<lazy_weak_ptr<employee> > employees_;
};

#pragma db object
class employee
{
  ...

  #pragma db not_null
  lazy_shared_ptr<employer> employer_;
};
  

And the transaction is changed like this:

unsigned long id = ...
string name;

session s;
transaction t (db.begin ());

shared_ptr<employee> e (db.load<employee> (id));
e->employer_.load ();
name = e->employer_->name_;

t.commit ();
  

As a general guideline we recommend that you make at least one side of a bidirectional relationship lazy, especially for relationships with a many side.

A lazy pointer implementation mimics the interface of its eager counterpart which can be used once the pointer is loaded. It also adds a number of additional functions that are specific to the lazy loading functionality. Overall, the interface of a lazy pointer follows this general outline:

template <class T>
class lazy_ptr
{
public:
  //
  // The eager pointer interface.
  //

  // Initialization/assignment from an eager pointer.
  //
public:
  template <class Y> lazy_ptr (const eager_ptr<Y>&);
  template <class Y> lazy_ptr& operator= (const eager_ptr<Y>&);

  // Lazy loading interface.
  //
public:
  //  NULL      loaded()
  //
  //  true       true      NULL pointer to transient object
  //  false      true      valid pointer to persistent object
  //  true       false     unloaded pointer to persistent object
  //  false      false     valid pointer to transient object
  //
  bool loaded () const;

  eager_ptr<T> load () const;

  // Unload the pointer. For transient objects this function is
  // equivalent to reset().
  //
  void unload () const;

  // Get the underlying eager pointer. If this is an unloaded pointer
  // to a persistent object, then the returned pointer will be NULL.
  //
  eager_ptr<T> get_eager () const;

  // Initialization with a persistent loaded object.
  //
  template <class Y> lazy_ptr (database&, Y*);
  template <class Y> lazy_ptr (database&, const eager_ptr<Y>&);

  template <class Y> void reset (database&, Y*);
  template <class Y> void reset (database&, const eager_ptr<Y>&);

  // Initialization with a persistent unloaded object.
  //
  template <class ID> lazy_ptr (database&, const ID&);

  template <class ID> void reset (database&, const ID&);

  // Query object id and database of a persistent object.
  //
  template <class O /* = T */>
  // C++11: template <class O = T>
  object_traits<O>::id_type object_id () const;

  odb::database& database () const;
};
  

In a lazy weak pointer interface, the load() function returns the strong (shared) eager pointer. The following transaction demonstrates the use of a lazy weak pointer based on the employer and employee classes presented earlier.

typedef std::vector<lazy_weak_ptr<employee> > employees;

session s;
transaction t (db.begin ());

shared_ptr<employer> er (db.load<employer> ("Example Inc"));
employees& es (er->employees ());

for (employees::iterator i (es.begin ()); i != es.end (); ++i)
{
  // We are only interested in employees with object id less than
  // 100.
  //
  lazy_weak_ptr<employee>& lwp (*i);

  if (lwp.object_id<employee> () < 100)
  // C++11: if (lwp.object_id () < 100)
  {
    shared_ptr<employee> e (lwp.load ()); // Load and lock.
    cout << e->first_ << " " << e->last_ << endl;
  }
}

t.commit ();
  

Notice that inside the for-loop we use a reference to the lazy weak pointer instead of making a copy. This is not merely to avoid a copy. When a lazy pointer is loaded, all other lazy pointers that point to the same object do not automatically become loaded (though an attempt to load such copies will result in them pointing to the same object, provided the same session is still in effect). By using a reference in the above transaction we make sure that we load the pointer that is contained in the employer object. This way, if we later need to re-examine this employee object, the pointer will already be loaded.

As another example, suppose we want to add an employee to Example Inc. The straightforward implementation of this transaction is presented below:

session s;
transaction t (db.begin ());

shared_ptr<employer> er (db.load<employer> ("Example Inc"));
shared_ptr<employee> e (new employee ("John", "Doe"));

e->employer_ = er;
er->employees ().push_back (e);

db.persist (e);
t.commit ();
  

Notice here that we didn't have to update the employer object in the database since the employees_ list of pointers is an inverse side of a bidirectional relationship and is effectively read-only, from the persistence point of view.

A faster implementation of this transaction, that avoids loading the employer object, relies on the ability to initialize an unloaded lazy pointer with the database where the object is stored as well as its identifier:

lazy_shared_ptr<employer> er (db, std::string ("Example Inc"));
shared_ptr<employee> e (new employee ("John", "Doe"));

e->employer_ = er;

session s;
transaction t (db.begin ());

db.persist (e);

t.commit ();
  

For the interaction of lazy pointers with lazy-loaded object sections, refer to Section 9.3, "Sections and Lazy Pointers".

6.5 Using Custom Smart Pointers

While the ODB runtime and profile libraries provide support for the majority of widely-used pointers, it is also easy to add support for a custom smart pointer.

To achieve this you will need to implement the pointer_traits class template specialization for your pointer. The first step is to determine the pointer kind since the interface of the pointer_traits specialization varies depending on the pointer kind. The supported pointer kinds are: raw (raw pointer or equivalent, that is, unmanaged), unique (smart pointer that doesn't support sharing), shared (smart pointer that supports sharing), and weak (weak counterpart of the shared pointer). Any of these pointers can be lazy, which also affects the interface of the pointer_traits specialization.

Once you have determined the pointer kind for your smart pointer, use a specialization for one of the standard pointers found in the common ODB runtime library (libodb) as a base for your own implementation.

Once the pointer traits specialization is ready, you will need to include it into the ODB compilation process using the --odb-epilogue option and into the generated header files with the --hxx-prologue option. As an example, suppose we have the smart_ptr smart pointer for which we have the traits specialization implemented in the smart-ptr-traits.hxx file. Then, we can create an ODB compiler options file for this pointer and save it to smart-ptr.options:

# Options file for smart_ptr.
#
--odb-epilogue '#include "smart-ptr-traits.hxx"'
--hxx-prologue '#include "smart-ptr-traits.hxx"'
  

Now, whenever we compile a header file that uses smart_ptr, we can specify the following command line option to make sure it is recognized by the ODB compiler as a smart pointer and the traits file is included in the generated code:

--options-file smart-ptr.options
  

It is also possible to implement a lazy counterpart for your smart pointer. The ODB runtime library provides a class template that encapsulates the object id management and loading functionality that is needed to implement a lazy pointer. All you need to do is wrap it with an interface that mimics your smart pointer. Using one of the existing lazy pointer implementations (either from the ODB runtime library or one of the profile libraries) as a base for your implementation is the easiest way to get started.


7 Value Types

In Section 3.1, "Concepts and Terminology" we have already discussed the notion of values and value types as well as the distinction between simple and composite values. This chapter covers simple and composite value types in more detail.

7.1 Simple Value Types

A simple value type is a fundamental C++ type or a class type that is mapped to a single database column. For each supported database system the ODB compiler provides a default mapping to suitable database types for most fundamental C++ types, such as int or float as well as some class types, such as std::string. For more information about the default mapping for each database system refer to Part II, Database Systems. We can also provide a custom mapping for these or our own value types using the db type pragma (Section 14.3.1, "type").

7.2 Composite Value Types

A composite value type is a class or struct type that is mapped to more than one database column. To declare a composite value type we use the db value pragma, for example:

#pragma db value
class basic_name
{
  ...

  std::string first_;
  std::string last_;
};
  

The complete version of the above code fragment and the other code samples presented in this section can be found in the composite example in the odb-examples package.

A composite value type does not have to define a default constructor, unless it is used as an element of a container. In this case the default constructor can be made private provided we also make the odb::access class, defined in the <odb/core.hxx> header, a friend of this value type. For example:

#include <odb/core.hxx>

#pragma db value
class basic_name
{
public:
  basic_name (const std::string& first, const std::string& last);

  ...

private:
  friend class odb::access;

  basic_name () {} // Needed for storing basic_name in containers.

  ...
};
  

The ODB compiler also needs access to the non-transient (Section 14.4.11, "transient") data members of a composite value type. It uses the same mechanisms as for persistent classes which are discussed in Section 3.2, "Declaring Persistent Objects and Values".

The members of a composite value can be other value types (either simple or composite), containers (Chapter 5, "Containers"), and pointers to objects (Chapter 6, "Relationships"). Similarly, a composite value type can be used in object members, as an element of a container, and as a base for another composite value type. In particular, composite value types can be used as element types in set containers (Section 5.2, "Set and Multiset Containers") and as key types in map containers (Section 5.3, "Map and Multimap Containers"). A composite value type that is used as an element of a container cannot contain other containers since containers of containers are not allowed. The following example illustrates some of the possible use cases:

#pragma db value
class basic_name
{
  ...

  std::string first_;
  std::string last_;
};

typedef std::vector<basic_name> basic_names;

#pragma db value
class name_extras
{
  ...

  std::string nickname_;
  basic_names aliases_;
};

#pragma db value
class name: public basic_name
{
  ...

  std::string title_;
  name_extras extras_;
};

#pragma db object
class person
{
  ...

  name name_;
};
  

A composite value type can be defined inside a persistent class, view, or another composite value and even made private, provided we make odb::access a friend of the containing class, for example:

#pragma db object
class person
{
  ...

  #pragma db value
  class name
  {
    ...

    std::string first_;
    std::string last_;
  };

  name name_;
};
  

A composite value type can also be defined as an instantiation of a C++ class template, for example:

template <typename T>
struct point
{
  T x;
  T y;
  T z;
};

typedef point<int> int_point;
#pragma db value(int_point)

#pragma db object
class object
{
  ...

  int_point center_;
};
  

Note that the database support code for such a composite value type is generated when compiling the header containing the db value pragma and not the header containing the template definition or the typedef name. This allows us to use templates defined in other files, such as std::pair defined in the utility standard header file:

#include <utility> // std::pair

typedef std::pair<std::string, std::string> phone_numbers;
#pragma db value(phone_numbers)

#pragma db object
class person
{
  ...

  phone_numbers phone_;
};
  

We can also use data members from composite value types in database queries (Chapter 4, "Querying the Database"). For each composite value in a persistent class, the query class defines a nested member that contains members corresponding to the data members in the value type. We can then use the member access syntax (.) to refer to data members in value types. For example, the query class for the person object presented above contains the name member (its name is derived from the name_ data member) which in turn contains the extras member (its name is derived from the name::extras_ data member of the composite value type). This process continues recursively for nested composite value types and, as a result, we can use the query::name.extras.nickname expression while querying the database for the person objects. For example:

typedef odb::query<person> query;
typedef odb::result<person> result;

transaction t (db.begin ());

result r (db.query<person> (
  query::name.extras.nickname == "Squeaky"));

...

t.commit ();
  

7.2.1 Composite Object Ids

An object id can be of a composite value type, for example:

#pragma db value
class name
{
  ...

  std::string first_;
  std::string last_;
};

#pragma db object
class person
{
  ...

  #pragma db id
  name name_;
};
  

However, a value type that can be used as an object id has a number of restrictions. Such a value type cannot have container, object pointer, or read-only data members. It also must be default-constructible, copy-constructible, and copy-assignable. Furthermore, if the persistent class in which this composite value type is used as object id has session support enabled (Chapter 11, "Session"), then it must also implement the less-than comparison operator (operator<).

7.2.2 Composite Value Column and Table Names

Customizing a column name for a data member of a simple value type is straightforward: we simply specify the desired name with the db column pragma (Section 14.4.9, "column"). For composite value types things are slightly more complex since they are mapped to multiple columns. Consider the following example:

#pragma db value
class name
{
  ...

  std::string first_;
  std::string last_;
};

#pragma db object
class person
{
  ...

  #pragma db id auto
  unsigned long id_;

  name name_;
};
  

The column names for the first_ and last_ members are constructed by using the sanitized name of the person::name_ member as a prefix and the names of the members in the value type (first_ and last_) as suffixes. As a result, the database schema for the above classes will look like this:

CREATE TABLE person (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  name_first TEXT NOT NULL,
  name_last TEXT NOT NULL);
  

We can customize both the prefix and the suffix using the db column pragma as shown in the following example:

#pragma db value
class name
{
  ...

  #pragma db column("first_name")
  std::string first_;

  #pragma db column("last_name")
  std::string last_;
};

#pragma db object
class person
{
  ...

  #pragma db column("person_")
  name name_;
};
  

The database schema changes as follows:

CREATE TABLE person (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  person_first_name TEXT NOT NULL,
  person_last_name TEXT NOT NULL);
  

We can also make the column prefix empty, for example:

#pragma db object
class person
{
  ...

  #pragma db column("")
  name name_;
};
  

This will result in the following schema:

CREATE TABLE person (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  first_name TEXT NOT NULL,
  last_name TEXT NOT NULL);
  

The same principle applies when a composite value type is used as an element of a container, except that instead of db column, either the db value_column (Section 14.4.36, "value_column") or db key_column (Section 14.4.35, "key_column") pragmas are used to specify the column prefix.

When a composite value type contains a container, an extra table is used to store its elements (Chapter 5, "Containers"). The names of such tables are constructed in a way similar to the column names, except that by default both the object name and the member name are used as a prefix. For example:

#pragma db value
class name
{
  ...

  std::string first_;
  std::string last_;
  std::vector<std::string> nicknames_;
};

#pragma db object
class person
{
  ...

  name name_;
};
  

The corresponding database schema will look like this:

CREATE TABLE person_name_nicknames (
  object_id BIGINT UNSIGNED NOT NULL,
  index BIGINT UNSIGNED NOT NULL,
  value TEXT NOT NULL)

CREATE TABLE person (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  name_first TEXT NOT NULL,
  name_last TEXT NOT NULL);
  

To customize the container table name we can use the db table pragma (Section 14.4.20, "table"), for example:

#pragma db value
class name
{
  ...

  #pragma db table("nickname")
  std::vector<std::string> nicknames_;
};

#pragma db object
class person
{
  ...

  #pragma db table("person_")
  name name_;
};
  

This will result in the following schema changes:

CREATE TABLE person_nickname (
  object_id BIGINT UNSIGNED NOT NULL,
  index BIGINT UNSIGNED NOT NULL,
  value TEXT NOT NULL)
  

Similar to columns, we can make the table prefix empty.

7.3 Pointers and NULL Value Semantics

Relational database systems have a notion of the special NULL value that is used to indicate the absence of a valid value in a column. While by default ODB maps values to columns that do not allow NULL values, it is possible to change that with the db null pragma (Section 14.4.6, "null/not_null").

To properly support the NULL semantics, the C++ value type must have a notion of a NULL value or a similar special state concept. Most basic C++ types, such as int or std::string, do not have this notion and therefore cannot be used directly for NULL-enabled data members (in the case of a NULL value being loaded from the database, such data members will be default-initialized).

To allow the easy conversion of value types that do not support the NULL semantics into the ones that do, ODB provides the odb::nullable class template. It allows us to wrap an existing C++ type into a container-like class that can either be NULL or contain a value of the wrapped type. ODB also automatically enables the NULL values for data members of the odb::nullable type. For example:

#include <odb/nullable.hxx>

#pragma db object
class person
{
  ...

  std::string first_;                    // TEXT NOT NULL
  odb::nullable<std::string> middle_;    // TEXT NULL
  std::string last_;                     // TEXT NOT NULL
};
  

The odb::nullable class template is defined in the <odb/nullable.hxx> header file and has the following interface:

namespace odb
{
  template <typename T>
  class nullable
  {
  public:
    typedef T value_type;

    nullable ();
    nullable (const T&);
    nullable (const nullable&);
    template <typename Y> explicit nullable (const nullable<Y>&);

    nullable& operator= (const T&);
    nullable& operator= (const nullable&);
    template <typename Y> nullable& operator= (const nullable<Y>&);

    void swap (nullable&);

    // Accessor interface.
    //
    bool null () const;

    T&       get ();
    const T& get () const;

    // Pointer interface.
    //
    operator bool_convertible () const;

    T*       operator-> ();
    const T* operator-> () const;

    T&       operator* ();
    const T& operator* () const;

    // Reset to the NULL state.
    //
    void reset ();
  };
}
  

The following example shows how we can use this interface:

  nullable<string> ns;

  // Using the accessor interface.
  //
  if (ns.null ())
  {
    s = "abc";
  }
  else
  {
    string s (ns.get ());
    ns.reset ();
  }

  // The same using the pointer interface.
  //
  if (ns)
  {
    s = "abc";
  }
  else
  {
    string s (*ns);
    ns.reset ();
  }
  

The odb::nullable class template requires the wrapped type to have public default and copy constructors as well as the copy assignment operator. Note also that the odb::nullable implementation is not the most efficient in that it always contains a fully constructed value of the wrapped type. This is normally not a concern for simple types such as the C++ fundamental types or std::string. However, it may become an issue for more complex types. In such cases you may want to consider using a more efficient implementation of the optional value concept such as the optional class template from Boost (Section 23.4, "Optional Library").

Another common C++ representation of a value that can be NULL is a pointer. ODB will automatically handle data members that are pointers to values, however, it will not automatically enable NULL values for such data members, as is the case for odb::nullable. Instead, if the NULL value is desired, we will need to enable it explicitly using the db null pragma. For example:

#pragma db object
class person
{
  ...

  std::string first_;

  #pragma db null
  std::auto_ptr<std::string> middle_;

  std::string last_;
};
  

The ODB compiler includes built-in support for using std::auto_ptr, std::unique_ptr (C++11), and shared_ptr (TR1 or C++11) as pointers to values. Plus, ODB profile libraries, that are available for commonly used frameworks and libraries (such as Boost and Qt), provide support for smart pointers found in these frameworks and libraries (Part III, "Profiles").

ODB also supports the NULL semantics for composite values. In the relational database the NULL composite value is translated to NULL values for all the simple data members of this composite value. For example:

#pragma db value
struct name
{
  std::string first_;
  odb::nullable<std::string> middle_;
  std::string last_;
};

#pragma db object
class person
{
  ...
  odb::nullable<name> name_;
};
  

ODB does not support the NULL semantics for containers. This also means that a composite value that contains a container cannot be NULL. With this limitation in mind, we can still use smart pointers in data members of container types. The only restriction is that these pointers must not be NULL. For example:

#pragma db object
class person
{
  ...

  std::auto_ptr<std::vector<std::string> > aliases_;
};
  

8 Inheritance

In C++ inheritance can be used to achieve two different goals. We can employ inheritance to reuse common data and functionality in multiple classes. For example:

class person
{
public:
  const std::string& first () const;
  const std::string& last () const;

private:
  std::string first_;
  std::string last_;
};

class employee: public person
{
  ...
};

class contractor: public person
{
  ...
};
  

In the above example both the employee and contractor classes inherit the first_ and last_ data members as well as the first() and last() accessors from the person base class.

A common trait of this inheritance style, referred to as reuse inheritance from now on, is the lack of virtual functions and a virtual destructor in the base class. Also with this style the application code is normally written in terms of the derived classes instead of the base.

The second way to utilize inheritance in C++ is to provide polymorphic behavior through a common interface. In this case the base class defines a number of virtual functions and, normally, a virtual destructor while the derived classes provide specific implementations of these virtual functions. For example:

class person
{
public:
  enum employment_status
  {
    unemployed,
    temporary,
    permanent,
    self_employed
  };

  virtual employment_status
  employment () const = 0;

  virtual
  ~person ();
};

class employee: public person
{
public:
  virtual employment_status
  employment () const
  {
    return temporary_ ? temporary : permanent;
  }

private:
  bool temporary_;
};

class contractor: public person
{
public:
  virtual employment_status
  employment () const
  {
    return self_employed;
  }
};
  

With this inheritance style, which we will call polymorphism inheritance, the application code normally works with derived classes via the base class interface. Note also that it is very common to mix both styles in the same hierarchy. For example, the above two code fragments can be combined so that the person base class provides the common data members and functions as well as defines the polymorphic interface.

The following sections describe the available strategies for mapping reuse and polymorphism inheritance styles to a relational data model. Note also that the distinction between the two styles is conceptual rather than formal. For example, it is possible to treat a class hierarchy that defines virtual functions as a case of reuse inheritance if this results in the desired database mapping and semantics.

Generally, classes that employ reuse inheritance are mapped to completely independent entities in the database. They use different object id spaces and should always be passed to and returned from the database operations as pointers or references to derived types. In other words, from the persistence point of view, such classes behave as if the data members from the base classes were copied verbatim into the derived ones.

In contrast, classes that employ polymorphism inheritance share the object id space and can be passed to and returned from the database operations polymorphically as pointers or references to the base class.

For both inheritance styles it is sometimes desirable to prevent instances of a base class from being stored in the database. To achieve this a persistent class can be declared abstract using the db abstract pragma (Section 14.1.3, "abstract"). Note that a C++-abstract class, or a class that has one or more pure virtual functions and therefore cannot be instantiated, is also database-abstract. However, a database-abstract class is not necessarily C++-abstract. The ODB compiler automatically treats C++-abstract classes as database-abstract.

8.1 Reuse Inheritance

Each non-abstract class from the reuse inheritance hierarchy is mapped to a separate database table that contains all its data members, including those inherited from base classes. An abstract persistent class does not have to define an object id, nor a default constructor, and it does not have a corresponding database table. An abstract class cannot be a pointed-to object in a relationship. Multiple inheritance is supported as long as each base class is only inherited once. The following example shows a persistent class hierarchy employing reuse inheritance:

// Abstract person class. Note that it does not declare the
// object id.
//
#pragma db object abstract
class person
{
  ...

  std::string first_;
  std::string last_;
};

// Abstract employee class. It derives from the person class and
// declares the object id for all the concrete employee types.
//
#pragma db object abstract
class employee: public person
{
  ...

  #pragma db id auto
  unsigned long id_;
};

// Concrete permanent_employee class. Note that it doesn't define
// any data members of its own.
//
#pragma db object
class permanent_employee: public employee
{
  ...
};

// Concrete temporary_employee class. It adds the employment
// duration in months.
//
#pragma db object
class temporary_employee: public employee
{
  ...

  unsigned long duration_;
};

// Concrete contractor class. It derives from the person class
// (and not employee; an independent contractor is not considered
// an employee). We use the contractor's external email address
// as the object id.
//
#pragma db object
class contractor: public person
{
  ...

  #pragma db id
  std::string email_;
};
  

The sample database schema for this hierarchy is shown below.

CREATE TABLE permanent_employee (
  first TEXT NOT NULL,
  last TEXT NOT NULL,
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY AUTO_INCREMENT);

CREATE TABLE temporary_employee (
  first TEXT NOT NULL,
  last TEXT NOT NULL,
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY AUTO_INCREMENT,
  duration BIGINT UNSIGNED NOT NULL);

CREATE TABLE contractor (
  first TEXT NOT NULL,
  last TEXT NOT NULL,
  email VARCHAR (255) NOT NULL PRIMARY KEY);
  

The complete version of the code presented in this section is available in the inheritance/reuse example in the odb-examples package.

8.2 Polymorphism Inheritance

There are three general approaches to mapping a polymorphic class hierarchy to a relational database. These are table-per-hierarchy, table-per-difference, and table-per-class. With the table-per-hierarchy mapping, all the classes in a hierarchy are stored in a single, "wide" table. NULL values are stored in columns corresponding to data members of derived classes that are not present in any particular instance.

In the table-per-difference mapping, each class is mapped to a separate table. For a derived class, this table contains only columns corresponding to the data members added by this derived class.

Finally, in the table-per-class mapping, each class is mapped to a separate table. For a derived class, this table contains columns corresponding to all the data members, from this derived class all the way down to the root of the hierarchy.

The table-per-difference mapping is generally considered as having the best balance of flexibility, performance, and space efficiency. It also results in a more canonical relational database model compared to the other two approaches. As a result, this is the mapping currently implemented in ODB. Other mappings may be supported in the future.

A pointer or reference to an ordinary, non-polymorphic object has just one type — the class type of that object. When we start working with polymorphic objects, there are two types to consider: the static type, or the declaration type of a reference or pointer, and the object's actual or dynamic type. An example will help illustrate the difference:

class person {...};
class employee: public person {...};

person p;
employee e;

person& r1 (p);
person& r2 (e);

auto_ptr<person> p1 (new employee);
  

In the above example, the r1 reference's both static and dynamic types are person. In contrast, the r2 reference's static type is person while its dynamic type (the actual object that it refers to) is employee. Similarly, p1 points to the object of the person static type but employee dynamic type.

In C++, the primary mechanisms for working with polymorphic objects are virtual functions. We call a virtual function only knowing the object's static type, but the version corresponding to the object's dynamic type is automatically executed. This is the essence of runtime polymorphism support in C++: we can operate in terms of a base class interface but get the derived class' behavior. Similarly, the essence of the runtime polymorphism support in ODB is to allow us to persist, load, update, and query in terms of the base class interface but have the derived class actually stored in the database.

To declare a persistent class as polymorphic we use the db polymorphic pragma. We only need to declare the root class of a hierarchy as polymorphic; ODB will treat all the derived classes as polymorphic automatically. For example:

#pragma db object polymorphic
class person
{
  ...

  virtual
  ~person () = 0; // Automatically abstract.

  #pragma db id auto
  unsigned long id_;

  std::string first_;
  std::string last_;
};

#pragma db object
class employee: public person
{
  ...

  bool temporary_;
};

#pragma db object
class contractor: public person
{

  std::string email_;
};
  

A persistent class hierarchy declared polymorphic must also be polymorphic in the C++ sense, that is, the root class must declare or inherit at least one virtual function. It is recommended that the root class also declares a virtual destructor. The root class of the polymorphic hierarchy must contain the data member designated as object id (a persistent class without an object id cannot be polymorphic). Note also that, unlike reuse inheritance, abstract polymorphic classes have a table in the database, just like non-abstract classes.

Persistent classes in the same polymorphic hierarchy must use the same kind of object pointer (Section 3.3, "Object and View Pointers"). If the object pointer for the root class is specified as a template or using the special raw pointer syntax (*), then the ODB compiler will automatically use the same object pointer for all the derived classes. For example:

#pragma db object polymorphic pointer(std::shared_ptr)
class person
{
  ...
};

#pragma db object // Object pointer is std::shared_ptr<employee>.
class employee: public person
{
  ...
};

#pragma db object // Object pointer is std::shared_ptr<contractor>.
class contractor: public person
{
  ...
};
  

Similarly, if we enable or disable session support (Chapter 11, "Session") for the root class, then the ODB compiler will automatically enable or disable it for all the derived classes.

For polymorphic persistent classes, all the database operations can be performed on objects with different static and dynamic types. Similarly, operations that load persistent objects from the database (load(), query(), etc.), can return objects with different static and dynamic types. For example:

unsigned long id1, id2;

// Persist.
//
{
  shared_ptr<person> p1 (new employee (...));
  shared_ptr<person> p2 (new contractor (...));

  transaction t (db.begin ());
  id1 = db.persist (p1); // Stores employee.
  id2 = db.persist (p2); // Stores contractor.
  t.commit ();
}

// Load.
//
{
  shared_ptr<person> p;

  transaction t (db.begin ());
  p = db.load<person> (id1); // Loads employee.
  p = db.load<person> (id2); // Loads contractor.
  t.commit ();
}

// Query.
//
{
  typedef odb::query<person> query;
  typedef odb::result<person> result;

  transaction t (db.begin ());

  result r (db.query<person> (query::last == "Doe"));

  for (result::iterator i (r.begin ()); i != r.end (); ++i)
  {
    person& p (*i); // Can be employee or contractor.
  }

  t.commit ();
}

// Update.
//
{
  shared_ptr<person> p;
  shared_ptr<employee> e;

  transaction t (db.begin ());

  e = db.load<employee> (id1);
  e->temporary (false);
  p = e;
  db.update (p); // Updates employee.

  t.commit ();
}

// Erase.
//
{
  shared_ptr<person> p;

  transaction t (db.begin ());
  p = db.load<person> (id1); // Loads employee.
  db.erase (p);              // Erases employee.
  db.erase<person> (id2);    // Erases contractor.
  t.commit ();
}
  

The table-per-difference mapping, as supported by ODB, requires two extra columns, in addition to those corresponding to the data members. The first, called discriminator, is added to the table corresponding to the root class of the hierarchy. This column is used to determine the dynamic type of each object. The second column is added to tables corresponding to the derived classes and contains the object id. This column is used to form a foreign key constraint referencing the root class table.

When querying the database for polymorphic objects, it is possible to obtain the discriminator value without instantiating the object. For example:

typedef odb::query<person> query;
typedef odb::result<person> result;

transaction t (db.begin ());

result r (db.query<person> (query::last == "Doe"));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
{
  std::string d (i.discriminator ());
  ...
}

t.commit ();
  

In the current implementation, ODB has limited support for customizing names, types, and values of the extra columns. Currently, the discriminator column is always called typeid and contains a namespace-qualified class name (for example, "employee" or "hr::employee"). The id column in the derived class table has the same name as the object id column in the root class table. Future versions of ODB will add support for customizing these extra columns.

The sample database schema for the above polymorphic hierarchy is shown below.

CREATE TABLE person (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY AUTO_INCREMENT,
  typeid VARCHAR(255) NOT NULL,
  first TEXT NOT NULL,
  last TEXT NOT NULL);

CREATE TABLE employee (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  temporary TINYINT(1) NOT NULL,

  CONSTRAINT employee_id_fk
    FOREIGN KEY (id)
    REFERENCES person (id)
    ON DELETE CASCADE);

CREATE TABLE contractor (
  id BIGINT UNSIGNED NOT NULL PRIMARY KEY,
  email TEXT NOT NULL,

  CONSTRAINT contractor_id_fk
    FOREIGN KEY (id)
    REFERENCES person (id)
    ON DELETE CASCADE);
  

The complete version of the code presented in this section is available in the inheritance/polymorphism example in the odb-examples package.

8.2.1 Performance and Limitations

A database operation on a non-polymorphic object normally translates to a single database statement execution (objects with containers and eager object pointers can be the exception). Because polymorphic objects have their data members stored in multiple tables, some database operations on such objects may result in multiple database statements being executed while others may require more complex statements. There is also some functionality that is not available to polymorphic objects.

The first part of this section discusses the performance implications to keep in mind when designing and working with polymorphic hierarchies. The second part talks about limitations of polymorphic objects.

The most important aspect of a polymorphic hierarchy that affects database performance is its depth. The distance between the root of the hierarchy and the derived class translates directly to the number of database statements that will have to be executed in order to persist, update, or erase this derived class. It also translates directly to the number of SQL JOIN clauses that will be needed to load or query the database for this derived class. As a result, to achieve best performance, we should try to keep our polymorphic hierarchies as flat as possible.

When loading an object or querying the database for objects, ODB will need to execute two statements if this object's static and dynamic types are different but only one statement if they are the same. This example will help illustrate the difference:

unsigned long id;

{
  employee e (...);

  transaction t (db.begin ());
  id = db.persist (e);
  t.commit ();
}

{
  shared_ptr<person> p;

  transaction t (db.begin ());
  p = db.load<person> (id);   // Requires two statement.
  p = db.load<employee> (id); // Requires only one statement.
  t.commit ();
}
  

As a result, we should try to load and query using the most derived class possible.

Finally, for polymorphic objects, erasing via the object instance is faster than erasing via its object id. In the former case the object's dynamic type can be determined locally in the application while in the latter case an extra statement has to be executed to achieve the same result. For example:

shared_ptr<person> p = ...;

transaction t (db.begin ());
db.erase<person> (p.id ()); // Slower (executes extra statement).
db.erase (p);               // Faster.
t.commit ();
  

Polymorphic objects can use all the mechanisms that are available to ordinary objects. These include containers (Chapter 5, "Containers"), object relationships, including to polymorphic objects (Chapter 6, "Relationships"), views (Chapter 10, "Views"), session (Chapter 11, "Session"), and optimistic concurrency (Chapter 12, "Optimistic Concurrency"). There are, however, a few limitations, mainly due to the underlying use of SQL to access the data.

When a polymorphic object is "joined" in a view, and the join condition (either in the form of an object pointer or a custom condition) comes from the object itself (as opposed to one of the objects joined previously), then this condition must only use data members from the derived class. For example, consider the following polymorphic object hierarchy and a view:

#pragma db object polymorphic
class employee
{
  ...
};

#pragma db object
class permanent_employee: public employee
{
  ...
};

#pragma db object
class temporary_employee: public employee
{
  ...

  shared_ptr<permanent_employee> manager_;
};

#pragma db object
class contractor: public temporary_employee
{
  shared_ptr<permanent_employee> manager_;
};

#pragma db view object(permanent_employee) \
                object(contractor: contractor::manager_)
struct contractor_manager
{
  ...
};
  

This view will not function correctly because the join condition (manager_) comes from the base class (temporary_employee) instead of the derived (contractor). The reason for this limitation is the JOIN clause order in the underlying SQL SELECT statement. In the view presented above, the table corresponding to the base class (temporary_employee) will have to be joined first which will result in this view matching both the temporary_employee and contractor objects instead of just contractor. It is usually possible to resolve this issue by reordering the objects in the view. Our example, for instance, can be fixed by swapping the two objects:

#pragma db view object(contractor) \
                object(permanent_employee: contractor::manager_)
struct contractor_manager
{
  ...
};
  

The erase_query() database function (Section 3.11, "Deleting Persistent Objects") also has limited functionality when used on polymorphic objects. Because many database implementations do not support JOIN clauses in the SQL DELETE statement, only data members from the derived class being erased can be used in the query condition. For example:

typedef odb::query<employee> query;

transaction t (db.begin ());
db.erase_query<employee> (query::permanent);     // Ok.
db.erase_query<employee> (query::last == "Doe"); // Error.
t.commit ();
  

8.3 Mixed Inheritance

It is possible to mix the reuse and polymorphism inheritance styles in the same hierarchy. In this case, the reuse inheritance must be used for the "bottom" (base) part of the hierarchy while the polymorphism inheritance — for the "top" (derived) part. For example:

#pragma db object
class person
{
  ...
};

#pragma db object polymorphic
class employee: public person // Reuse inheritance.
{
  ...
};

#pragma db object
class temporary_employee: public employee // Polymorphism inheritance.
{
  ...
};

#pragma db object
class permanent_employee: public employee // Polymorphism inheritance.
{
  ...
};
  

9 Sections

ODB sections are an optimization mechanism that allows us to partition data members of a persistent class into groups that can be separately loaded and/or updated. This can be useful, for example, if an object contains expensive to load or update data members (such as BLOBs or containers) and that are accessed or modified infrequently. For example:

#include <odb/section.hxx>

#pragma db object
class person
{
  ...

  #pragma db load(lazy) update(manual)
  odb::section keys_;

  #pragma db section(keys_) type("BLOB")
  char public_key_[1024];

  #pragma db section(keys_) type("BLOB")
  char private_key_[1024];
};

transaction t (db.begin ());

auto_ptr<person> p (db.load<person> (...)); // Keys are not loaded.

if (need_keys)
{
  db.load (*p, p->keys_); // Load keys.
  ...
}

db.update (*p); // Keys are not updated.

if (update_keys)
{
  ...
  db.update (*p, p->keys_); // Update keys.
}

t.commit ();
  

A complete example that shows how to use sections is available in the section directory in the odb-examples package.

Why do we need to group data members into sections? Why can't each data member be loaded and updated independently if and when necessary? The reason for this requirement is that loading or updating a group of data members with a single database statement is significantly more efficient than loading or updating each data member with a separate statement. Because ODB prepares and caches statements used to load and update persistent objects, generating a custom statement for a specific set of data members that need to be loaded or updated together is not a viable approach either. To resolve this, ODB allows us to group data members that are often updated and/or loaded together into sections. To achieve the best performance, we should aim to find a balance between having too many sections with too few data members and too few sections with too many data members. We can use the access and modification patterns of our application as a base for this decision.

To add a new section to a persistent class we declare a new data member of the odb::section type. At this point we also need to specify the loading and updating behavior of this section with the db load and db update pragmas, respectively.

The loading behavior of a section can be either eager or lazy. An eager-loaded section is always loaded as part of the object load. A lazy-loaded section is not loaded as part of the object load and has to be explicitly loaded with the database::load() function (discussed below) if and when necessary.

The updating behavior of a section can be always, change, or manual. An always-updated section is always updated as part of the object update, provided it has been loaded. A change-updated section is only updated as part of the object update if it has been loaded and marked as changed. A manually-updated section is never updated as part of the object update and has to be explicitly updated with the database::update() function (discussed below) if and when necessary.

If no loading behavior is specified explicitly, then an eager-loaded section is assumed. Similarly, if no updating behavior is specified, then an always-updated section is assumed. An eager-loaded, always-updated section is pointless and therefore illegal. Only persistent classes with an object id can have sections.

To specify that a data member belongs to a section we use the db section pragma with the section's member name as its single argument. Except for special data members such as the object id and optimistic concurrency version, any direct, non-transient member of a persistent class can belong to a section, including composite values, containers, and pointers to objects. For example:

#pragma db value
class text
{
  std::string data;
  std::string lang;
};

#pragma db object
class person
{
  ...

  #pragma db load(lazy)
  odb::section extras_;

  #pragma db section(extras_)
  text bio_;

  #pragma db section(extras_)
  std::vector<std::string> nicknames_;

  #pragma db section(extras_)
  std::shared_ptr<person> emergency_contact_;
};
  

An empty section is pointless and therefore illegal, except in abstract or polymorphic classes where data members can be added to a section by derived classes (see Section 9.1, "Sections and Inheritance").

The odb::section class is defined in the <odb/section.hxx> header file and has the following interface:

namespace odb
{
  class section
  {
  public:
    // Load state.
    //
    bool
    loaded () const;

    void
    unload ();

    // Change state.
    //
    bool
    changed () const;

    void
    change ();

    // User data.
    //
    unsigned char
    user_data () const;

    void
    user_data (unsigned char);
  };
}
  

The loaded() accessor can be used to determine whether a section is already loaded. The unload() modifier marks a loaded section as not loaded. This, for example, can be useful if you don't want the section to be reloaded during the object reload.

The changed() accessor can be used to query the section's change state. The change() modifier marks the section as changed. It is valid to call this modifier for an unloaded (or transient) section, however, the state will be reset back to unchanged once the section (or object) is loaded. The change state is only relevant to sections with change-updated behavior and is ignored for all other sections.

The size of the section class is one byte with four bits available to store a custom state via the user_data() accessor and modifier.

The odb::database class provides special versions of the load() and update() functions that allow us to load and update sections of a persistent class. Their signatures are as follows:

  template <typename T>
  void
  load (T& object, section&);

  template <typename T>
  void
  update (const T& object, const section&);
  

Before calling the section load() function, the object itself must already be loaded. If the section is already loaded, then the call to load() will reload its data members. It is illegal to explicitly load an eager-loaded section.

Before calling the section update() function, the section (and therefore the object) must be in the loaded state. If the section is not loaded, the odb::section_not_loaded exception is thrown. The section update() function does not check but does clear the section's change state. In other words, section update() will always update section data members in the database and clear the change flag. Note also that any section, that is, always-, change-, or manually-updated, can be explicitly updated with this function.

Both section load() and update(), just like the rest of the database operations, must be performed within a transaction. Notice also that both load() and update() expect a reference to the section as their second argument. This reference must refer to the data member in the object passed as the first argument. If instead it refers to some other instance of the section class, for example, a local copy or a temporary, then the odb::section_not_in_object exception is thrown. For example:

#pragma db object
class person
{
public:
  ...

  odb::section
  keys () const {return keys_;}

private:
  odb::section keys_;

  ...
};

auto_ptr<person> p (db.load<person> (...));

section s (p->keys ());
db.load (*p, s);            // Throw section_not_in_object, copy.

db.update (*p, p->keys ()); // Throw section_not_in_object, copy.
  

At first glance it may seem more appropriate to make the section class non-copyable in order to prevent such errors from happening. However, it is perfectly reasonable to expect to be able to copy (or assign) sections as part of the object copying (or assignment). As a result, sections are left copyable and copy-assignable, however, this functionality should not be used in accessors or modifiers. Instead, section accessors and modifiers should always be by-reference. Here is how we can fix our previous example:

#pragma db object
class person
{
public:
  ...

  const odb::section&
  keys () const {return keys_;}

  odb::section&
  keys () {return keys_;}

private:
  odb::section keys_;

  ...
};

auto_ptr<person> p (db.load<person> (...));

section& s (p->keys ());
db.load (*p, s);            // Ok, reference.

db.update (*p, p->keys ()); // Ok, reference.
  

Several other database operations affect sections. The state of a section in a transient object is undefined. That is, before the call to object persist() or load() functions, or after the call to object erase() function, the values returned by the section::loaded() and section::changed() accessors are undefined.

After the call to persist(), all sections, including eager-loaded ones, are marked as loaded and unchanged. If instead we are loading an object with the load() call or as a result of a query, then eager-loaded sections are loaded and marked as loaded and unchanged while lazy-loaded ones are marked as unloaded. If a lazy-loaded section is later loaded with the section load() call, then it is marked as loaded and unchanged.

When we update an object with the update() call, manually-updated sections are ignored while always-updated sections are updated if they are loaded. Change-updated sections are only updated if they are both loaded and marked as changed. After the update, such sections are reset to the unchanged state. When we reload an object with the reload() call, sections that were loaded are automatically reloaded and reset to the unchanged state.

To further illustrate the state transitions of a section, consider this example:

#pragma db object
class person
{
  ...

  #pragma db load(lazy) update(change)
  odb::section keys_;

  ...
};

transaction t (db.begin ());

person p ("John", "Doe"); // Section state is undefined (transient).

db.persist (p);           // Section state: loaded, unchanged.

auto_ptr<person> l (
  db.load<person> (...)); // Section state: unloaded, unchanged.

db.update (*l);           // Section not updated since not loaded.
db.update (p);            // Section not updated since not changed.

p.keys_.change ();        // Section state: loaded, changed.
db.update (p);            // Section updated, state: loaded, unchanged.

db.update (*l, l->keys_); // Throw section_not_loaded.
db.update (p, p.keys_);   // Section updated even though not changed.

db.reload (*l);           // Section not reloaded since not loaded.
db.reload (p);            // Section reloaded, state: loaded, unchanged.

db.load (*l, l->keys_);   // Section loaded, state: loaded, unchanged.
db.load (p, p.keys_);     // Section reloaded, state: loaded, unchanged.

db.erase (p);             // Section state is undefined (transient).

t.commit ();
  

When using change-updated behavior, it is our responsibility to mark the section as changed when any of the data members belonging to this section is modified. A natural place to mark the section as changed is the modifiers for section data members, for example:

#pragma db object
class person
{
  ...

  typedef std::array<char, 1024> key_type;

  const key_type&
  public_key () const {return public_key_;}

  void
  public_key (const key_type& k)
  {
    public_key_ = k;
    keys_.change ();
  }

  const key_type&
  private_key () const {return private_key_;}

  void
  private_key (const key_type& k)
  {
    private_key_ = k;
    keys_.change ();
  }

private:
  #pragma db load(lazy) update(change)
  odb::section keys_;

  #pragma db section(keys_) type("BLOB")
  key_type public_key_;

  #pragma db section(keys_) type("BLOB")
  key_type private_key_;

  ...
};
  

One interesting aspect of change-updated sections is what happens when a transaction that performed an object or section update is later rolled back. In this case, while the change state of a section has been reset (after update), actual changes were not committed to the database. Change-updated sections handle this case by automatically registering a rollback callback and then, if it is called, restoring the original change state. The following code illustrates this semantics (continuing with the previous example):

auto_ptr<person> p;

try
{
  transaction t (db.begin ());
  p = db.load<person> (...);
  db.load (*p, p->keys_);

  p->private_key (new_key); // The section is marked changed.
  db.update (*p);           // The section is reset to unchanged.

  throw failed ();          // Triggers rollback.
  t.commit ();
}
catch (const failed&)
{
  // The section is restored back to changed.
}
  

9.1 Sections and Inheritance

With both reuse and polymorphism inheritance (Chapter 8, "Inheritance") it is possible to add new sections to derived classes. It is also possible to add data members from derived classes to sections declared in the base. For example:

#pragma db object polymorphic
class person
{
  ...

  virtual void
  print ();

  #pragma db load(lazy)
  odb::section print_;

  #pragma db section(print_)
  std::string bio_;
};

#pragma db object
class employee: public person
{
  ...

  virtual void
  print ();

  #pragma db section(print_)
  std::vector<std::string> employment_history_;
};

transaction t (db.begin ());

auto_ptr<person> p (db.load<person> (...)); // Person or employee.
db.load (*p, p->print_); // Load data members needed for print.
p->print ();

t.commit ();
  

When data members of a section are spread over several classes in a reuse inheritance hierarchy, both section load and update are performed with a single database statement. In contrast, with polymorphism inheritance, section load is performed with a single statement while update requires a separate statement for each class that adds to the section.

Note also that in polymorphism inheritance the section-to-object association is static. Or, in other words, you can load a section via an object only if its static type actually contains this section. The following example will help illustrate this point further:

#pragma db object polymorphic
class person
{
  ...
};

#pragma db object
class employee: public person
{
  ...

  #pragma db load(lazy)
  odb::section extras_;

  ...
};

#pragma db object
class manager: public employee
{
  ...
};

auto_ptr<manager> m (db.load<manager> (...));

person& p (*m);
employee& e (*m);
section& s (m->extras_);

db.load (p, s); // Error: extras_ is not in person.
db.load (e, s); // Ok: extras_ is in employee.
  

9.2 Sections and Optimistic Concurrency

When sections are used in a class with the optimistic concurrency model (Chapter 12, "Optimistic Concurrency"), both section update and load operations compare the object version to that in the database and throw the odb::object_changed exception if they do not match. In addition, the section update operation increments the version to indicate that the object state has changed. For example:

#pragma db object optimistic
class person
{
  ...

  #pragma db version
  unsigned long long version_;

  #pragma db load(lazy)
  odb::section extras_;

  #pragma db section(extras_)
  std::string bio_;
};

auto_ptr<person> p;

{
  transaction t (db.begin ());
  p = db.load<person> (...);
  t.commit ();
}

{
  transaction t (db.begin ());

  try
  {
    db.load (*p, p->extras_); // Throws if object state has changed.
  }
  catch (const object_changed&)
  {
    db.reload (*p);
    db.load (*p, p->extras_); // Cannot fail.
  }

  t.commit ();
}
  

Note also that if an object update triggers one or more section updates, then each such update will increment the object version. As a result, an update of an object that contains sections may result in a version increment by more than one.

When sections are used together with optimistic concurrency and inheritance, an extra step may be required to enable this functionality. If you plan to add new sections to derived classes, then the root class of the hierarchy (the one that declares the version data member) must be declared as sectionable with the db sectionable pragma. For example:

#pragma db object polymorphic sectionable
class person
{
  ...

  #pragma db version
  unsigned long long version_;
};

#pragma db object
class employee: public person
{
  ...

  #pragma db load(lazy)
  odb::section extras_;

  #pragma db section(extras_)
  std::vector<std::string> employment_history_;
};
  

This requirement has to do with the need to generate extra version increment code in the root class that will be used by sections added in the derived classes. If you forget to declare the root class as sectionable and later add a section to one of the derived classes, the ODB compiler will issue diagnostics.

9.3 Sections and Lazy Pointers

If a lazy pointer (Section 6.4, "Lazy Pointers") belongs to a lazy-loaded section, then we end up with two levels of lazy loading. Specifically, when the section is loaded, the lazy pointer is initialized with the object id but the object itself is not loaded. For example:

#pragma db object
class employee
{
  ...

  #pragma db load(lazy)
  odb::section extras_;

  #pragma db section(extras_)
  odb::lazy_shared_ptr<employer> employer_;
};

transaction t (db.begin ());

auto_ptr<employee> e (db.load<employee> (...)); // employer_ is NULL.

db.load (*e, e->extras_); // employer_ contains valid employer id.

e->employer_.load (); // employer_ points to employer object.

t.commit ();
  

9.4 Sections and Change-Tracking Containers

If a change-tracking container (Section 5.4, "Change-Tracking Containers") belongs to a change-updated section, then prior to an object update ODB will check if the container has been changed and if so, automatically mark the section as changed. For example:

#pragma db object
class person
{
  ...

  #pragma db load(lazy) update(change)
  odb::section extras_;

  #pragma db section(extras_)
  odb::vector<std::string> nicknames_;
};

transaction t (db.begin ());

auto_ptr<person> p (db.load<person> (...));
db.load (*p, p->extras_);

p->nicknames_.push_back ("JD");

db.update (*p); // Section is automatically updated even
                // though it was not marked as changed.
t.commit ();
  

10 Views

An ODB view is a C++ class or struct type that embodies a light-weight, read-only projection of one or more persistent objects or database tables or the result of a native SQL query execution.

Some of the common applications of views include loading a subset of data members from objects or columns from database tables, executing and handling results of arbitrary SQL queries, including aggregate queries and stored procedure calls, as well as joining multiple objects and/or database tables using object relationships or custom join conditions.

Many relational databases also define the concept of views. Note, however, that ODB views are not mapped to database views. Rather, by default, an ODB view is mapped to an SQL SELECT query. However, if desired, it is easy to create an ODB view that is based on a database view.

Usually, views are defined in terms of other persistent entities, such as persistent objects, database tables, sequences, etc. Therefore, before we can examine our first view, we need to define a few persistent objects and a database table. We will use this model in examples throughout this chapter. Here we assume that you are familiar with ODB object relationship support (Chapter 6, "Relationships").

#pragma db object
class country
{
  ...

  #pragma db id
  std::string code_; // ISO 2-letter country code.

  std::string name_;
};

#pragma db object
class employer
{
  ...

  #pragma db id
  unsigned long id_;

  std::string name_;
};

#pragma db object
class employee
{
  ...

  #pragma db id
  unsigned long id_;

  std::string first_;
  std::string last_;

  unsigned short age_;

  shared_ptr<country> residence_;
  shared_ptr<country> nationality_;

  shared_ptr<employer> employed_by_;
};
  

Besides these objects, we also have the legacy employee_extra table that is not mapped to any persistent class. It has the following definition:

CREATE TABLE employee_extra(
  employee_id INTEGER NOT NULL,
  vacation_days INTEGER NOT NULL,
  previous_employer_id INTEGER)
  

The above persistent objects and database table as well as many of the views shown in this chapter are based on the view example which can be found in the odb-examples package of the ODB distribution.

To declare a view we use the db view pragma, for example:

#pragma db view object(employee)
struct employee_name
{
  std::string first;
  std::string last;
};
  

The above example shows one of the simplest views that we can create. It has a single associated object (employee) and its purpose is to extract the employee's first and last names without loading any other data, such as the referenced country and employer objects.

Views use the same query facility (Chapter 4, "Querying the Database") as persistent objects. Because support for queries is optional and views cannot be used without this support, you need to compile any header that defines a view with the --generate-query ODB compiler option.

To query the database for a view we use the database::query(), database::query_one(), or database::query_value() functions in exactly the same way as we would use them to query the database for an object. For example, the following code fragment shows how we can find the names of all the employees that are younger than 31:

typedef odb::query<employee_name> query;
typedef odb::result<employee_name> result;

transaction t (db.begin ());

result r (db.query<employee_name> (query::age < 31));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
{
  const employee_name& en (*i);
  cout << en.first << " " << en.last << endl;
}

t.commit ();
  

A view can be defined as a projection of one or more objects, one or more tables, a combination of objects and tables, or it can be the result of a custom SQL query. The following sections discuss each of these kinds of view in more detail.

10.1 Object Views

To associate one or more objects with a view we use the db object pragma (Section 14.2.1, "object"). We have already seen a simple, single-object view in the introduction to this chapter. To associate the second and subsequent objects we repeat the db object pragma for each additional object, for example:

#pragma db view object(employee) object(employer)
struct employee_employer
{
  std::string first;
  std::string last;
  std::string name;
};
  

The complete syntax of the db object pragma is shown below:

object(name [= alias] [join-type] [: join-condition])

The name part is a potentially qualified persistent class name that has been defined previously. The optional alias part gives this object an alias. If provided, the alias is used in several contexts instead of the object's unqualified name. We will discuss aliases further as we cover each of these contexts below. The optional join-type part specifies the way this object is associated. It can be left, right, full, inner, and cross with left being the default. Finally, the optional join-condition part provides the criteria which should be used to associate this object with any of the previously associated objects or, as we will see in Section 10.4, "Mixed Views", tables. Note that while the first associated object can have an alias, it cannot have a join type or condition.

For each subsequent associated object the ODB compiler needs a join condition and there are several ways to specify it. The easiest way is to omit it altogether and let the ODB compiler try to come up with a join condition automatically. To do this the ODB compiler will examine each previously associated object for object relationships (Chapter 6, "Relationships") that may exist between these objects and the object being associated. If such a relationship exists and is unambiguous, that is there is only one such relationship, then the ODB compiler will automatically use it to come up with the join condition for this object. This is exactly what happens in the previous example: there is a single relationship (employee::employed_by) between the employee and employer objects.

On the other hand, consider this view:

#pragma db view object(employee) object(country)
struct employee_residence
{
  std::string first;
  std::string last;
  std::string name;
};
  

While there is a relationship between country and employee, it is ambiguous. It can be employee::residence_ (which is what we want) or it can be employee::nationality_ (which we don't want). As result, when compiling the above view, the ODB compiler will issue an error indicating an ambiguous object relationship. To resolve this ambiguity, we can explicitly specify the object relationship that should be used to create the join condition as the name of the corresponding data member. Here is how we can fix the employee_residence view:

#pragma db view object(employee) object(country: employee::residence_)
struct employee_residence
{
  std::string first;
  std::string last;
  std::string name;
};
  

It is possible to associate the same object with a single view more than once using different join conditions. However, in this case, we have to use aliases to assign different names for each association. For example:

#pragma db view object(employee) \
  object(country = res_country: employee::residence_) \
  object(country = nat_country: employee::nationality_)
struct employee_country
{
  ...
};
  

Note that correctly defining data members in this view requires the use of a mechanism that we haven't yet covered. We will see how to do this shortly.

If we assign an alias to an object and refer to a data member of this object in one of the join conditions, we have to use the unqualified alias name instead of the potentially qualified object name. For example:

#pragma db view object(employee = ee) object(country: ee::residence_)
struct employee_residence
{
  ...
};
  

The last way to specify a join condition is to provide a custom query expression. This method is primarily useful if you would like to associate an object using a condition that does not involve an object relationship. Consider, for example, a modified employee object from the beginning of the chapter with an added country of birth member. For one reason or another we have decided not to use a relationship to the country object, as we have done with residence and nationality.

#pragma db object
class employee
{
  ...

  std::string birth_place_; // Country name.
};
  

If we now want to create a view that returns the birth country code for an employee, then we have to use a custom join condition when associating the country object. For example:

#pragma db view object(employee) \
  object(country: employee::birth_place_ == country::name_)
struct employee_birth_code
{
  std::string first;
  std::string last;
  std::string code;
};
  

The syntax of the query expression in custom join conditions is the same as in the query facility used to query the database for objects (Chapter 4, "Querying the Database") except that for query members, instead of using odb::query<object>::member names, we refer directly to object members.

Looking at the views we have defined so far, you may be wondering how the ODB compiler knows which view data members correspond to which object data members. While the names are similar, they are not exactly the same, for example employee_name::first and employee::first_.

As with join conditions, when it comes to associating data members, the ODB compiler tries to do this automatically. It first searches all the associated objects for an exact name match. If no match is found, then the ODB compiler compares the so-called public names. A public name of a member is obtained by removing the common member name decorations, such as leading and trailing underscores, the m_ prefix, etc. In both of these searches the ODB compiler also makes sure that the types of the two members are the same or compatible.

If one of the above searches returned a match and it is unambiguous, that is there is only one match, then the ODB compiler will automatically associate the two members. On the other hand, if no match is found or the match is ambiguous, the ODB compiler will issue an error. To associate two differently-named members or to resolve an ambiguity, we can explicitly specify the member association using the db column pragma (Section 14.4.9, "column"). For example:

#pragma db view object(employee) object(employer)
struct employee_employer
{
  std::string first;
  std::string last;

  #pragma db column(employer::name_)
  std::string employer_name;
};
  

If an object data member specifies the SQL type with the db type pragma (Section 14.4.3, "type"), then this type is also used for the associated view data members.

Note also that similar to join conditions, if we assign an alias to an object and refer to a data member of this object in one of the db column pragmas, then we have to use the unqualified alias name instead of the potentially qualified object name. For example:

#pragma db view object(employee) \
  object(country = res_country: employee::residence_) \
  object(country = nat_country: employee::nationality_)
struct employee_country
{
  std::string first;
  std::string last;

  #pragma db column(res_country::name_)
  std::string res_country_name;

  #pragma db column(nat_country::name_)
  std::string nat_country_name;
};
  

Besides specifying just the object member, we can also specify a +-expression in the db column pragma. A +-expression consists of string literals and object member references connected using the + operator. It is primarily useful for defining aggregate views based on SQL aggregate functions, for example:

#pragma db view object(employee)
struct employee_count
{
  #pragma db column("count(" + employee::id_ + ")")
  std::size_t count;
};
  

When querying the database for a view, we may want to provide additional query criteria based on the objects associated with this view. To support this a view defines query members for all the associated objects which allows us to refer to such objects' members using the odb::query<view>::member expressions. This is similar to how we can refer to object members using the odb::query<object>::member expressions when querying the database for an object. For example:

typedef odb::query<employee_count> query;

transaction t (db.begin ());

// Find the number of employees with the Doe last name. Result of this
// aggregate query contains only one element so use the query_value()
// shortcut function.
//
employee_count ec (
  db.query_value<employee_count> (query::last == "Doe"));

cout << ec.count << endl;

t.commit ();
  

In the above query we used the last name data member from the associated employee object to only count employees with the specific name.

When a view has only one associated object, the query members corresponding to this object are defined directly in the odb::query<view> scope. For instance, in the above example, we referred to the last name member as odb::query<employee_count>::last. However, if a view has multiple associated objects, then query members corresponding to each such object are defined in a nested scope named after the object. As an example, consider the employee_employer view again:

#pragma db view object(employee) object(employer)
struct employee_employer
{
  std::string first;
  std::string last;

  #pragma db column(employer::name_)
  std::string employer_name;
};
  

Now, to refer to the last name data member from the employee object we use the odb::query<...>::employee::last expression. Similarly, to refer to the employer name, we use the odb::query<...>::employer::name expression. For example:

typedef odb::result<employee_employer> result;
typedef odb::query<employee_employer> query;

transaction t (db.begin ());

result r (db.query<employee_employer> (
  query::employee::last == "Doe" &&
  query::employer::name == "Simple Tech Ltd"));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
  cout << i->first << " " << i->last <<  " " << i->employer_name << endl;

t.commit ();
  

If we assign an alias to an object, then this alias is used to name the query members scope instead of the object name. As an example, consider the employee_country view again:

#pragma db view object(employee) \
  object(country = res_country: employee::residence_) \
  object(country = nat_country: employee::nationality_)
struct employee_country
{
  ...
};
  

And a query which returns all the employees that have the same country of residence and nationality:

typedef odb::query<employee_country> query;
typedef odb::result<employee_country> result;

transaction t (db.begin ());

result r (db.query<employee_country> (
  query::res_country::name == query::nat_country::name));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
  cout << i->first << " " << i->last << " " << i->res_country_name << endl;

t.commit ();
  

Note also that unlike object query members, view query members do no support referencing members in related objects. For example, the following query is invalid:

typedef odb::query<employee_name> query;
typedef odb::result<employee_name> result;

transaction t (db.begin ());

result r (db.query<employee_name> (
  query::employed_by->name == "Simple Tech Ltd"));

t.commit ();
  

To get this behavior, we would instead need to associate the employer object with this view and then use the query::employer::name expression instead of query::employed_by->name.

As we have discussed above, if specified, an object alias is used instead of the object name in the join condition, data member references in the db column pragma, as well as to name the query members scope. The object alias is also used as a table name alias in the underlying SELECT statement generated by the ODB compiler. Normally, you would not use the table alias directly with object views. However, if for some reason you need to refer to a table column directly, for example, as part of a native query expression, and you need to qualify the column with the table, then you will need to use the table alias instead.

10.2 Object Loading Views

A special variant of object views is object loading views. Object loading views allow us to load one or more complete objects instead of, or in addition to, a subset of data member. While we can often achieve the same end result by calling database::load(), using a view has several advantages.

If we need to load multiple objects, then using a view allows us to do this with a single SELECT statement execution instead of one for each object that would be necessary in case of load(). A view can also be useful for loading only a single object if the query criterion that we would like to use involves other, potentially unrelated, objects. We will examine concrete examples of these and other scenarios in the rest of this section.

To load a complete object as part of a view we use a data member of the pointer to object type, just like for object relationships (Chapter 6, "Relationships"). As an example, here is how we can load both the employee and employer objects from the previous section with a single statement:

#pragma db view object(employee) object(employer)
struct employee_employer
{
  shared_ptr<employee> ee;
  shared_ptr<employer> er;
};
  

We use an object loading view just like any other view. In the result of a query, as we would expect, the pointer data members point to the loaded objects. For example:

typedef odb::query<employee_employer> query;

transaction t (db.begin ());

for (const employee_employer& r:
       db.query<employee_employer> (query::employee::age < 31))
{
  cout << r.ee->age () << " " << r.er->name () << endl;
}

t.commit ();
  

As another example, consider a query that loads the employer objects using some condition based on its employees. For instance, we want to find all the employers that employ people over 65 years old. We can use this object loading view to implement such a query (notice the distinct result modifier discussed later in Section 10.5, "View Query Conditions"):

#pragma db view object(employer) object(employee) query(distinct)
struct employer_view
{
  shared_ptr<employer> er;
};
  

And this is how we can use this view to find all the employers that employ seniors:

typedef odb::query<employer_view> query;

db.query<employer_view> (query::employee::age > 65)
  

We can even use object loading views to load completely unrelated (from the ODB object relationships point of view) objects. For example, the following view will load all the employers that are named the same as a country (notice the inner join type):

#pragma db view object(employer) \
  object(country inner: employer::name == country::name)
struct employer_named_country
{
  shared_ptr<employer> e;
  shared_ptr<country> c;
};
  

An object loading view can contain ordinary data members in addition to object pointers. For example, if we are only interested in the country code in the above view, then we can reimplement it like this:

#pragma db view object(employer) \
  object(country inner: employer::name == country::name)
struct employer_named_country
{
  shared_ptr<employer> e;
  std::string code;
};
  

Object loading views also have a few rules and restrictions. Firstly, the pointed-to object in the data member must be associated with the view. Furthermore, if the associated object has an alias, then the data member name must be the same as the alias (more precisely, the public name derived from the data member must match the alias; which means we can use normal data member decorations such as trailing underscores, etc., see the previous section for more information on public names). The following view illustrates the use of aliases as data member names:

#pragma db view object(employee)               \
  object(country = res: employee::residence_)  \
  object(country = nat: employee::nationality_)
struct employee_country
{
  shared_ptr<country> res;
  shared_ptr<country> nat_;
};
  

Finally, the object pointers must be direct data members of the view. Using, for example, a composite value that contains pointers as a view data member is not supported. Note also that depending on the join type you are using, some of the resulting pointers might be NULL.

Up until now we have consistently used shared_ptr as an object pointer in our views. Can we use other pointers, such as unique_ptr or raw pointers? To answer this question we first need to discuss what happens with object pointers that may be inside objects that a view loads. As a concrete example, let us revisit the employee_employer view from the beginning of this section:

#pragma db view object(employee) object(employer)
struct employee_employer
{
  shared_ptr<employee> ee;
  shared_ptr<employer> er;
};
  

This view loads two objects: employee and employer. The employee object, however, also contains a pointer to employer (see the employed_by_ data member). In fact, this is the same object that the view loads since employer is associated with the view using this same relationship (ODB automatically uses it since it is the only one). The correct result of loading such a view is then clear: both er and er->employed_by_ must point to (or share) the same instance.

Just like object loading via the database class functions, views achieve this correct behavior of only loading a single instance of the same object with the help of session's object cache (Chapter 11, "Session"). In fact, object loading views enforce this by throwing the session_required exception if there is no current session and the view loads an object that is also indirectly loaded by one of the other objects. The ODB compiler will also issue diagnostics if such an object has session support disabled (Section 14.1.10, "session").

With this understanding we can now provide the correct implementation of our transaction that uses the employee_employer view:

typedef odb::query<employee_employer> query;

transaction t (db.begin ());
odb::session s;

for (const employee_employer& r:
       db.query<employee_employer> (query::employee::age < 31))
{
  assert (r.ee->employed_by_ == r.er);
  cout << r.ee->age () << " " << r.er->name () << endl;
}

t.commit ();
  

It might seem logical, then, to always load all the objects from all the eager relationships with the view. After all, this will lead to them all being loaded with a single statement. While this is theoretically true, the reality is slightly more nuanced. If there is a high probability of the object already have been loaded and sitting in the cache, then not loading the object as part of the view (and therefore not fetching all its data from the database) might result in better performance.

Now we can also answer the question about which pointers we can use in object loading views. From the above discussion it should be clear that if an object that we are loading is also part of a relationship inside another object that we are loading, then we should use some form of a shared ownership pointer. If, however, there are no relationships involved, as is the case, for example, in our employer_named_country and employee_country views above, then we can use a unique ownership pointer such as unique_ptr.

Note also that your choice of a pointer type can be limited by the "official" object pointer type assigned to the object (Section 3.3, "Object and View Pointers"). For example, if the object pointer type is shared_ptr, you will not be able to use unique_ptr to load such an object into a view since initializing unique_ptr from shared_ptr would be a mistake.

Unless you want to perform your own object cleanup, raw object pointers in views are not particularly useful. They do have one special semantics, however: If a raw pointer is used as a view member, then, before creating a new instance, the implementation will check if the member is NULL. If it is not, then it is assumed to point to an existing instance and the implementation will load the data into it instead of creating a new one. The primary use of this special functionality is to implement by-value loading with the ability to detect NULL values.

To illustrate this functionality, consider the following view that load the employee's residence country by value:

#pragma db view object(employee) \
  object(country = res: employee::residence_) transient
struct employee_res_country
{
  typedef country* country_ptr;

  #pragma db member(res_) virtual(country_ptr) get(&this.res) \
    set(this.res_null = ((?) == nullptr))

  country res;
  bool res_null;
};
  

Here we are using a virtual data member (Section 14.4.13, "virtual") to add an object pointer member to the view. Its accessor expression returns the pointer to the res member so that the implementation can load the data into it. The modifier expression checks the passed pointer to initialize the NULL value indicator. Here, the two possible values that can be passed to the modifier expression are the address of the res member that we returned earlier from the accessor and NULL (strictly speaking, there is a third possibility: the address of an object that was found in the session cache).

If we are not interested in the NULL indicator, then the above view can simplified to this:

#pragma db view object(employee) \
  object(country = res: employee::residence_) transient
struct employee_res_country
{
  typedef country* country_ptr;

  #pragma db member(res_) virtual(country_ptr) get(&this.res) set()

  country res;
};
  

That is, we specify an empty modifier expression which leads to the value being ignored.

As another example of by-value loading, consider a view that allows us to load objects into existing instances that have been allocated outside the view:

#pragma db view object(employee)               \
  object(country = res: employee::residence_)  \
  object(country = nat: employee::nationality_)
struct employee_country
{
  employee_country (country& r, country& n): res (&r), nat (&n) {}

  country* res;
  country* nat;
};
  

And here is how we can use this view:

typedef odb::result<employee_country> result;

transaction t (db.begin ());

result r (db.query<employee_country> (...);

for (result::iterator i (r.begin ()); i != r.end (); ++i)
{
  country res, nat;
  employee_country v (res, nat);
  i.load (v);

  if (v.res != nullptr)
    ... // Result is in res.

  if (v.nat != nullptr)
    ... // Result is in nat.
}

t.commit ();
  

As a final example of the by-value loading, consider the following view which implements a slightly more advanced logic: if the object is already in the session cache, then it sets the pointer data member in the view (er_p) to that. Otherwise, it loads the data into the by-value instance (er). We can also check whether the pointer data member points to the instance to distinguish between the two outcomes. And we can check it for nullptr to detect NULL values.

#pragma db view object(employer)
struct employer_view
{
  // Since we may be getting the pointer as both smart and raw, we
  // need to create a bit of support code to use in the modifier
  // expression.
  //
  void set_er (employer* p) {er_p = p;}                   // &er or NULL.
  void set_er (shared_ptr<employer> p) {er_p = p.get ();} // From cache.

  #pragma db get(&this.er) set(set_er(?))
  employer* er_p;

  #pragma db transient
  employer er;

  // Return-by-value support (e.g., query_value()).
  //
  employer_view (): er_p (0) {}
  employer_view (const employer_view& x)
    : er_p (x.er_p == &x.er ? &er : x.er_p), er (x.er) {}
};
  

We can use object loading views with polymorphic objects (Section 8.2, "Polymorphism Inheritance"). Note, however, that when loading a derived object via the base pointer in a view, a separate statement will be executed to load the dynamic part of the object. There is no support for by-value loading for polymorphic objects.

We can also use object loading views with objects without id (Section 14.1.6, "no_id"). Note, however, that for such objects, NULL values are not automatically detected (since there is no primary key, which is otherwise guaranteed to be not NULL, there might not be a column on which to base this detection). The workaround for this limitation is to load an otherwise not NULL column next to the object which will serve as an indicator. For example:

#pragma db object no_id
class object
{
  ...

  int n; // NOT NULL
  std::string s;
};

#include <odb/nullable.hxx>

#pragma db view object(object)
struct view
{

  odb::nullable<int> n; // If 'n' is NULL, then, logically, so is 'o'.
  unique_ptr<object> o;
};
  

10.3 Table Views

A table view is similar to an object view except that it is based on one or more database tables instead of persistent objects. Table views are primarily useful when dealing with ad-hoc tables that are not mapped to persistent classes.

To associate one or more tables with a view we use the db table pragma (Section 14.2.2, "table"). To associate the second and subsequent tables we repeat the db table pragma for each additional table. For example, the following view is based on the employee_extra legacy table we have defined at the beginning of the chapter.

#pragma db view table("employee_extra")
struct employee_vacation
{
  #pragma db column("employee_id") type("INTEGER")
  unsigned long employee_id;

  #pragma db column("vacation_days") type("INTEGER")
  unsigned short vacation_days;
};
  

Besides the table name in the db table pragma we also have to specify the column name for each view data member. Note that unlike for object views, the ODB compiler does not try to automatically come up with column names for table views. Furthermore, we cannot use references to object members either, since there are no associated objects in table views. Instead, the actual column name or column expression must be specified as a string literal. The column name can also be qualified with a table name either in the "table.column" form or, if either a table or a column name contains a period, in the "table"."column" form. The following example illustrates the use of a column expression:

#pragma db view table("employee_extra")
struct employee_max_vacation
{
  #pragma db column("max(vacation_days)") type("INTEGER")
  unsigned short max_vacation_days;
};
  

Both the asociated table names and the column names can be qualified with a database schema, for example:

#pragma db view table("hr.employee_extra")
struct employee_max_vacation
{
  #pragma db column("hr.employee_extra.vacation_days") type("INTEGER")
  unsigned short vacation_days;
};
  

For more information on database schemas and the format of the qualified names, refer to Section 14.1.8, "schema".

Note also that in the above examples we specified the SQL type for each of the columns to make sure that the ODB compiler has knowledge of the actual types as specified in the database schema. This is required to obtain correct and optimal generated code.

The complete syntax of the db table pragma is similar to the db object pragma and is shown below:

table("name" [= "alias"] [join-type] [: join-condition])

The name part is a database table name. The optional alias part gives this table an alias. If provided, the alias must be used instead of the table whenever a reference to a table is used. Contexts where such a reference may be needed include the join condition (discussed below), column names, and query expressions. The optional join-type part specifies the way this table is associated. It can be left, right, full, inner, and cross with left being the default. Finally, the optional join-condition part provides the criteria which should be used to associate this table with any of the previously associated tables or, as we will see in Section 10.4, "Mixed Views", objects. Note that while the first associated table can have an alias, it cannot have a join type or condition.

Similar to object views, for each subsequent associated table the ODB compiler needs a join condition. However, unlike for object views, for table views the ODB compiler does not try to come up with one automatically. Furthermore, we cannot use references to object members corresponding to object relationships either, since there are no associated objects in table views. Instead, for each subsequent associated table, a join condition must be specified as a custom query expression. While the syntax of the query expression is the same as in the query facility used to query the database for objects (Chapter 4, "Querying the Database"), a join condition for a table is normally specified as a single string literal containing a native SQL query expression.

As an example of a multi-table view, consider the employee_health table that we define in addition to employee_extra:

CREATE TABLE employee_health(
  employee_id INTEGER NOT NULL,
  sick_leave_days INTEGER NOT NULL)
  

Given these two tables we can now define a view that returns both the vacation and sick leave information for each employee:

#pragma db view table("employee_extra" = "extra") \
  table("employee_health" = "health": \
        "extra.employee_id = health.employee_id")
struct employee_leave
{
  #pragma db column("extra.employee_id") type("INTEGER")
  unsigned long employee_id;

  #pragma db column("vacation_days") type("INTEGER")
  unsigned short vacation_days;

  #pragma db column("sick_leave_days") type("INTEGER")
  unsigned short sick_leave_days;
};
  

Querying the database for a table view is the same as for an object view except that we can only use native query expressions. For example:

typedef odb::query<employee_leave> query;
typedef odb::result<employee_leave> result;

transaction t (db.begin ());

unsigned short v_min = ...
unsigned short l_min = ...

result r (db.query<employee_leave> (
  "vacation_days > " + query::_val(v_min) + "AND"
  "sick_leave_days > " + query::_val(l_min)));

t.commit ();
  

10.4 Mixed Views

A mixed view has both associated objects and tables. As a first example of a mixed view, let us improve employee_vacation from the previous section to return the employee's first and last names instead of the employee id. To achieve this we have to associate both the employee object and the employee_extra table with the view:

#pragma db view object(employee) \
  table("employee_extra" = "extra": "extra.employee_id = " + employee::id_)
struct employee_vacation
{
  std::string first;
  std::string last;

  #pragma db column("extra.vacation_days") type("INTEGER")
  unsigned short vacation_days;
};
  

When querying the database for a mixed view, we can use query members for the parts of the query expression that involves object members but have to fall back to using the native syntax for the parts that involve table columns. For example:

typedef odb::query<employee_vacation> query;
typedef odb::result<employee_vacation> result;

transaction t (db.begin ());

result r (db.query<employee_vacation> (
  (query::last == "Doe") + "AND extra.vacation_days <> 0"));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
  cout << i->first << " " << i->last << " " << i->vacation_days << endl;

t.commit ();
  

As another example, consider a more advanced view that associates two objects via a legacy table. This view allows us to find the previous employer name for each employee:

#pragma db view object(employee) \
  table("employee_extra" = "extra": "extra.employee_id = " + employee::id_) \
  object(employer: "extra.previous_employer_id = " + employer::id_)
struct employee_prev_employer
{
  std::string first;
  std::string last;

  // If previous_employer_id is NULL, then the name will be NULL as well.
  // We use the odb::nullable wrapper to handle this.
  //
  #pragma db column(employer::name_)
  odb::nullable<std::string> prev_employer_name;
};
  

10.5 View Query Conditions

Object, table, and mixed views can also specify an optional query condition that should be used whenever the database is queried for this view. To specify a query condition we use the db query pragma (Section 14.2.3, "query").

As an example, consider a view that returns some information about all the employees that are over a predefined retirement age. One way to implement this would be to define a standard object view as we have done in the previous sections and then use a query like this:

result r (db.query<employee_retirement> (query::age > 50));
  

The problem with the above approach is that we have to keep repeating the query::age > 50 expression every time we execute the query, even though this expression always stays the same. View query conditions allow us to solve this problem. For example:

#pragma db view object(employee) query(employee::age > 50)
struct employee_retirement
{
  std::string first;
  std::string last;
  unsigned short age;
};
  

With this improvement we can rewrite our query like this:

result r (db.query<employee_retirement> ());
  

But what if we may also need to restrict the result set based on some varying criteria, such as the employee's last name? Or, in other words, we may need to combine a constant query expression specified in the db query pragma with the varying expression specified at the query execution time. To allow this, the db query pragma syntax supports the use of the special (?) placeholder that indicates the position in the constant query expression where the runtime expression should be inserted. For example:

#pragma db view object(employee) query(employee::age > 50 && (?))
struct employee_retirement
{
  std::string first;
  std::string last;
  unsigned short name;
};
  

With this change we can now use additional query criteria in our view:

result r (db.query<employee_retirement> (query::last == "Doe"));
  

The syntax of the expression in a query condition is the same as in the query facility used to query the database for objects (Chapter 4, "Querying the Database") except for two differences. Firstly, for query members, instead of using odb::query<object>::member names, we refer directly to object members, using the object alias instead of the object name if an alias was assigned. Secondly, query conditions support the special (?) placeholder which can be used both in the C++-integrated query expressions as was shown above and in native SQL expressions specified as string literals. The following view is an example of the latter case:

#pragma db view table("employee_extra") \
  query("vacation_days <> 0 AND (?)")
struct employee_vacation
{
  ...
};
  

Another common use case for query conditions are views with the ORDER BY or GROUP BY clause. Such clauses are normally present in the same form in every query involving such views. As an example, consider an aggregate view which calculate the minimum and maximum ages of employees for each employer:

#pragma db view object(employee) object(employer) \
  query((?) + "GROUP BY" + employer::name_)
struct employer_age
{
  #pragma db column(employer::name_)
  std::string employer_name;

  #pragma db column("min(" + employee::age_ + ")")
  unsigned short min_age;

  #pragma db column("max(" + employee::age_ + ")")
  unsigned short max_age;
};
  

The query condition can be optionally followed (or replaced, if no constant query expression is needed) by one or more result modifiers. Currently supported result modifiers are distinct (which is translated to SELECT DISTINCT) and for_update (which is translated to FOR UPDATE or equivalent for database systems that support it). As an example, consider a view that allows us to get some information about employers ordered by the object id and without any duplicates:

#pragma db view object(employer) object(employee) \
  query((?) + "ORDER BY" + employer::name_, distinct)
struct employer_info
{
  ...
};
  

If we don't require ordering, then this view can be re-implemented like this:

#pragma db view object(employer) object(employee) query(distinct)
struct employer_info
{
  ...
};
  

10.6 Native Views

The last kind of view supported by ODB is a native view. Native views are a low-level mechanism for capturing results of native SQL queries, stored procedure calls, etc. Native views don't have associated tables or objects. Instead, we use the db query pragma to specify the native SQL query, which should normally include the select-list and, if applicable, the from-list. For example, here is how we can re-implement the employee_vacation table view from Section 10.3 above as a native view:

#pragma db view query("SELECT employee_id, vacation_days " \
                      "FROM employee_extra")
struct employee_vacation
{
  #pragma db type("INTEGER")
  unsigned long employee_id;

  #pragma db type("INTEGER")
  unsigned short vacation_days;
};
  

In native views the columns in the query select-list are associated with the view data members in the order specified. That is, the first column is stored in the first member, the second column — in the second member, and so on. The ODB compiler does not perform any error checking in this association. As a result you must make sure that the number and order of columns in the query select-list match the number and order of data members in the view. This is also the reason why we are not required to provide the column name for each data member in native views, as is the case for object and table views.

Note also that while it is always possible to implement a table view as a native view, the table views must be preferred since they are safer. In a native view, if you add, remove, or rearrange data members without updating the column list in the query, or vice versa, at best, this will result in a runtime error. In contrast, in a table view such changes will result in the query being automatically updated.

Similar to object and table views, the query specified for a native view can contain the special (?) placeholder which is replaced with the query expression specified at the query execution time. If the native query does not contain a placeholder, as in the example above, then any query expression specified at the query execution time is appended to the query text along with the WHERE keyword, if required. The following example shows the usage of the placeholder:

#pragma db view query("SELECT employee_id, vacation_days " \
                      "FROM employee_extra " \
                      "WHERE vacation_days <> 0 AND (?)")
struct employee_vacation
{
  ...
};
  

As another example, consider a view that returns the next value of a database sequence:

#pragma db view query("SELECT nextval('my_seq')")
struct sequence_value
{
  unsigned long long value;
};
  

While this implementation can be acceptable in some cases, it has a number of drawbacks. Firstly, the name of the sequence is fixed in the view, which means if we have a second sequence, we will have to define another, almost identical view. Similarly, the operation that we perform on the sequence is also fixed. In some situations, instead of returning the next value, we may need the last value.

Note that we cannot use the placeholder mechanism to resolve these problems since placeholders can only be used in the WHERE, GROUP BY, and similar clauses. In other words, the following won't work:

#pragma db view query("SELECT nextval('(?)')")
struct sequence_value
{
  unsigned long long value;
};

result r (db.query<sequence_value> ("my_seq"));
  

To support these kinds of use cases, ODB allows us to specify the complete query for a native view at runtime rather than at the view definition. To indicate that a native view has a runtime query, we can either specify the empty db query pragma or omit the pragma altogether. For example:

#pragma db view
struct sequence_value
{
  unsigned long long value;
};
  

Given this view, we can perform the following queries:

typedef odb::query<sequence_value> query;
typedef odb::result<sequence_value> result;

string seq_name = ...

result l (db.query<sequence_value> (
  "SELECT lastval('" + seq_name + "')"));

result n (db.query<sequence_value> (
  "SELECT nextval('" + seq_name + "')"));
  

Native views can also be used to call and handle results of stored procedures. The semantics and limitations of stored procedures vary greatly between database systems while some do not support this functionality at all. As a result, support for calling stored procedures using native views is described for each database system in Part II, "Database Systems".

10.7 Other View Features and Limitations

Views cannot be derived from other views. However, you can derive a view from a transient C++ class. View data members cannot be object pointers. If you need to access data from a pointed-to object, then you will need to associate such an object with the view. Similarly, view data members cannot be containers. These two limitations also apply to composite value types that contain object pointers or containers. Such composite values cannot be used as view data members.

On the other hand, composite values that do not contain object pointers or containers can be used in views. As an example, consider a modified version of the employee persistent class that stores a person's name as a composite value:

#pragma db value
class person_name
{
  std::string first_;
  std::string last_;
};

#pragma db object
class employee
{
  ...

  person_name name_;

  ...
};
  

Given this change, we can re-implement the employee_name view like this:

#pragma db view object(employee)
struct employee_name
{
  person_name name;
};
  

It is also possible to extract some or all of the nested members of a composite value into individual view data members. Here is how we could have defined the employee_name view if we wanted to keep its original structure:

#pragma db view object(employee)
struct employee_name
{
  #pragma db column(employee::name.first_)
  std::string first;

  #pragma db column(employee::name.last_)
  std::string last;
};
  

11 Session

A session is an application's unit of work that may encompass several database transactions. In this version of ODB a session is just an object cache. In future versions it may provide additional functionality, such as delayed database operations and automatic object state change tracking. As discussed later in Section 11.2, "Custom Sessions", it is also possible to provide a custom session implementation that provides these or other features.

Session support is optional and can be enabled or disabled on the per object basis using the db session pragma, for example:

#pragma db object session
class person
{
  ...
};
  

We can also enable or disable session support for a group of objects at the namespace level:

#pragma db namespace session
namespace accounting
{
  #pragma db object                // Session support is enabled.
  class employee
  {
    ...
  };

  #pragma db object session(false) // Session support is disabled.
  class employer
  {
    ...
  };
}
  

Finally, we can pass the --generate-session ODB compiler option to enable session support by default. With this option session support will be enabled for all the persistent classes except those for which it was explicitly disabled using the db session. An alternative to this method with the same effect is to enable session support for the global namespace:

#pragma db namespace() session
  

Each thread of execution in an application can have only one active session at a time. A session is started by creating an instance of the odb::session class and is automatically terminated when this instance is destroyed. You will need to include the <odb/session.hxx> header file to make this class available in your application. For example:

#include <odb/database.hxx>
#include <odb/session.hxx>
#include <odb/transaction.hxx>

using namespace odb::core;

{
  session s;

  // First transaction.
  //
  {
    transaction t (db.begin ());
    ...
    t.commit ();
  }

  // Second transaction.
  //
  {
    transaction t (db.begin ());
    ...
    t.commit ();
  }

  // Session 's' is terminated here.
}
  

The session class has the following interface:

namespace odb
{
  class session
  {
  public:
    session (bool make_current = true);
    ~session ();

    // Copying or assignment of sessions is not supported.
    //
  private:
    session (const session&);
    session& operator= (const session&);

    // Current session interface.
    //
  public:
    static session&
    current ();

    static bool
    has_current ();

    static void
    current (session&);

    static void
    reset_current ();

    static session*
    current_pointer ();

    static void
    current_pointer (session*);

    // Object cache interface.
    //
  public:
    template <typename T>
    struct cache_position {...};

    template <typename T>
    cache_position<T>
    cache_insert (database&,
                  const object_traits<T>::id_type&,
                  const object_traits<T>::pointer_type&);

    template <typename T>
    object_traits<T>::pointer_type
    cache_find (database&, const object_traits<T>::id_type&) const;

    template <typename T>
    void
    cache_erase (const cache_position<T>&);

    template <typename T>
    void
    cache_erase (database&, const object_traits<T>::id_type&);
  };
}
  

The session constructor creates a new session and, if the make_current argument is true, sets it as a current session for this thread. If we try to make a session current while there is already another session in effect for this thread, then the constructor throws the odb::already_in_session exception. The destructor clears the current session for this thread if this session is the current one.

The static current() accessor returns the currently active session for this thread. If there is no active session, this function throws the odb::not_in_session exception. We can check whether there is a session in effect in this thread using the has_current() static function.

The static current() modifier allows us to set the current session for this thread. The reset_current() static function clears the current session. These two functions allow for more advanced use cases, such as multiplexing two or more sessions on the same thread.

The static current_pointer() overloaded functions provided the same functionality but using pointers. Specifically, the current_pointer() accessor can be used to test whether there is a current session and get a pointer to it all with a single call.

We normally don't use the object cache interface directly. However, it could be useful in some cases, for example, to find out whether an object has already been loaded. Note that when calling cache_insert(), cache_find(), or the second version of cache_erase(), you need to specify the template argument (object type) explicitly. It is also possible to access the underlying cache data structures directly. This can be useful if, for example, you want to iterate over the objects store in the cache. Refer to the ODB runtime header files for more details on this direct access.

11.1 Object Cache

A session is an object cache. Every time a session-enabled object is made persistent by calling the database::persist() function (Section 3.8, "Making Objects Persistent"), loaded by calling the database::load() or database::find() function (Section 3.9, "Loading Persistent Objects"), or loaded by iterating over a query result (Section 4.4, "Query Result"), the pointer to the persistent object, in the form of the canonical object pointer (Section 3.3, "Object and View Pointers"), is stored in the session. For as long as the session is in effect, any subsequent calls to load the same object will return the cached instance. When an object's state is deleted from the database with the database::erase() function (Section 3.11, "Deleting Persistent Objects"), the cached object pointer is removed from the session. For example:

shared_ptr<person> p (new person ("John", "Doe"));

session s;
transaction t (db.begin ());

unsigned long id (db.persist (p));            // p is cached in s.
shared_ptr<person> p1 (db.load<person> (id)); // p1 same as p.

t.commit ();
  

The per-object caching policies depend on the object pointer kind (Section 6.5, "Using Custom Smart Pointers"). Objects with a unique pointer, such as std::auto_ptr or std::unique_ptr, as an object pointer are never cached since it is not possible to have two such pointers pointing to the same object. When an object is persisted via a pointer or loaded as a dynamically allocated instance, objects with both raw and shared pointers as object pointers are cached. If an object is persisted as a reference or loaded into a pre-allocated instance, the object is only cached if its object pointer is a raw pointer.

Also note that when we persist an object as a constant reference or constant pointer, the session caches such an object as unrestricted (non-const). This can lead to undefined behavior if the object being persisted was actually created as const and is later found in the session cache and used as non-const. As a result, when using sessions, it is recommended that all persistent objects be created as non-const instances. The following code fragment illustrates this point:

void save (database& db, shared_ptr<const person> p)
{
  transaction t (db.begin ());
  db.persist (p); // Persisted as const pointer.
  t.commit ();
}

session s;

shared_ptr<const person> p1 (new const person ("John", "Doe"));
unsigned long id1 (save (db, p1)); // p1 is cached in s as non-const.

{
  transaction t (db.begin ());
  shared_ptr<person> p (db.load<person> (id1)); // p == p1
  p->age (30); // Undefined behavior since p1 was created const.
  t.commit ();
}

shared_ptr<const person> p2 (new person ("Jane", "Doe"));
unsigned long id2 (save (db, p2)); // p2 is cached in s as non-const.

{
  transaction t (db.begin ());
  shared_ptr<person> p (db.load<person> (id2)); // p == p2
  p->age (30); // Ok, since p2 was not created const.
  t.commit ();
}
  

11.2 Custom Sessions

ODB can use a custom session implementation instead of the default odb::session. There could be multiple reasons for an application to provide its own session. For example, the application may already include a notion of an object cache or registry which ODB can re-use. A custom session can also provide additional functionality, such as automatic change tracking, delayed database operations, or object eviction. Finally, the session-per-thread approach used by odb::session may not be suitable for all applications. For instance, some may need a thread-safe session that can be shared among multiple threads. For an example of a custom session that implements automatic change tracking by keeping original copies of the objects, refer to the common/session/custom test in the odb-tests package.

To use a custom session we need to specify its type with the --session-type ODB compiler command line option. We also need to include its definition into the generated header file. This can be achieved with the --hxx-prologue option. For example, if our custom session is called app::session and is defined in the app/session.hxx header file, then the corresponding ODB compiler options would look like this:

odb --hxx-prologue "#include \"app/session.hxx\"" \
--session-type ::app::session ...
  

A custom session should provide the following interface:

class custom_session
{
public:
  template <typename T>
  struct cache_position
  {
    ...
  };

  // Cache management functions.
  //
  template <typename T>
  static cache_position<T>
  _cache_insert (odb::database&,
                 const typename odb::object_traits<T>::id_type&,
                 const typename odb::object_traits<T>::pointer_type&);

  template <typename T>
  static typename odb::object_traits<T>::pointer_type
  _cache_find (odb::database&,
               const typename odb::object_traits<T>::id_type&);

  template <typename T>
  static void
  _cache_erase (const cache_position<T>&);

  // Notification functions.
  //
  template <typename T>
  static void
  _cache_persist (const cache_position<T>&);

  template <typename T>
  static void
  _cache_load (const cache_position<T>&);

  template <typename T>
  static void
  _cache_update (odb::database&, const T& obj);

  template <typename T>
  static void
  _cache_erase (odb::database&,
                const typename odb::object_traits<T>::id_type&);
};
  

The cache_position class template represents a position in the cache of the inserted object. It should be default and copy-constructible as well as copy-assignable. The default constructor shall create a special empty/NULL position. A call of any of the cache management or notification functions with such an empty/NULL position shall be ignored.

The _cache_insert() function shall add the object into the object cache and return its position. The _cache_find() function looks an object up in the object cache given its id. It returns a NULL pointer if the object is not found. The _cache_erase() cache management function shall remove the object from the cache. It is called if the database operation that caused the object to be inserted (for example, load) failed. Note also that after insertion the object state is undefined. You can only access the object state (for example, make a copy or clear a flag) from one of the notification functions discussed below.

The notification functions are called after an object has been persisted, loaded, updated, or erased, respectively. If your session implementation does not need some of the notifications, you still have to provide their functions, however, you can leave their implementations empty.

Notice also that all the cache management and notification functions are static. This is done in order to allow for a custom notion of a current session. Normally, the first step a non-empty implementation will perform is lookup the current session.


12 Optimistic Concurrency

The ODB transaction model (Section 3.5, "Transactions") guarantees consistency as long as we perform all the database operations corresponding to a specific application transaction in a single database transaction. That is, if we load an object within a database transaction and update it in the same transaction, then we are guaranteed that the object state that we are updating in the database is exactly the same as the state we have loaded. In other words, it is impossible for another process or thread to modify the object state in the database between these load and update operations.

In this chapter we use the term application transaction to refer to a set of operations on persistent objects that an application needs to perform in order to implement some application-specific functionality. The term database transaction refers to the set of database operations performed between the ODB begin() and commit() calls. Up until now we have treated application transactions and database transactions as essentially the same thing.

While this model is easy to understand and straightforward to use, it may not be suitable for applications that have long application transactions. The canonical example of such a situation is an application transaction that requires user input between loading an object and updating it. Such an operation may take an arbitrary long time to complete and performing it within a single database transaction will consume database resources as well as prevent other processes/threads from updating the object for too long.

The solution to this problem is to break up the long-lived application transaction into several short-lived database transactions. In our example that would mean loading the object in one database transaction, waiting for user input, and then updating the object in another database transaction. For example:

unsigned long id = ...;
person p;

{
  transaction t (db.begin ());
  db.load (id, p);
  t.commit ();
}

cerr << "enter age for " << p.first () << " " << p.last () << endl;
unsigned short age;
cin >> age;
p.age (age);

{
  transaction t (db.begin ());
  db.update (p);
  t.commit ();
}
  

This approach works well if we only have one process/thread that can ever update the object. However, if we have multiple processes/threads modifying the same object, then this approach does not guarantee consistency anymore. Consider what happens in the above example if another process updates the person's last name while we are waiting for the user input. Since we loaded the object before this change occured, our version of the person's data will still have the old name. Once we receive the input from the user, we go ahead and update the object, overwriting both the old age with the new one (correct) and the new name with the old one (incorrect).

While there is no way to restore the consistency guarantee in an application transaction that consists of multiple database transactions, ODB provides a mechanism, called optimistic concurrency, that allows applications to detect and potentially recover from such inconsistencies.

In essence, the optimistic concurrency model detects mismatches between the current object state in the database and the state when it was loaded into the application memory. Such a mismatch would mean that the object was changed by another process or thread. There are several ways to implement such state mismatch detection. Currently, ODB uses object versioning while other methods, such as timestamps, may be supported in the future.

To declare a persistent class with the optimistic concurrency model we use the optimistic pragma (Section 14.1.5, "optimistic"). We also use the version pragma (Section 14.4.16, "version") to specify which data member will store the object version. For example:

#pragma db object optimistic
class person
{
  ...

  #pragma db version
  unsigned long version_;
};
  

The version data member is managed by ODB. It is initialized to 1 when the object is made persistent and incremented by 1 with each update. The 0 version value is not used by ODB and the application can use it as a special value, for example, to indicate that the object is transient. Note that for optimistic concurrency to function properly, the application should not modify the version member after making the object persistent or loading it from the database and until deleting the state of this object from the database. To avoid any accidental modifications to the version member, we can declare it const, for example:

#pragma db object optimistic
class person
{
  ...

  #pragma db version
  const unsigned long version_;
};
  

When we call the database::update() function (Section 3.10, "Updating Persistent Objects") and pass an object that has an outdated state, the odb::object_changed exception is thrown. At this point the application has two recovery options: it can abort and potentially restart the application transaction or it can reload the new object state from the database, re-apply or merge the changes, and call update() again. Note that aborting an application transaction that performs updates in multiple database transactions may require reverting changes that have already been committed to the database. As a result, this strategy works best if all the updates are performed in the last database transaction of the application transaction. This way the changes can be reverted by simply rolling back this last database transaction.

The following example shows how we can reimplement the above transaction using the second recovery option:

unsigned long id = ...;
person p;

{
  transaction t (db.begin ());
  db.load (id, p);
  t.commit ();
}

cerr << "enter age for " << p.first () << " " << p.last () << endl;
unsigned short age;
cin >> age;
p.age (age);

{
  transaction t (db.begin ());

  try
  {
    db.update (p);
  }
  catch (const object_changed&)
  {
    db.reload (p);
    p.age (age);
    db.update (p);
  }

  t.commit ();
}
  

An important point to note in the above code fragment is that the second update() call cannot throw the object_changed exception because we are reloading the state of the object and updating it within the same database transaction.

Depending on the recovery strategy employed by the application, an application transaction with a failed update can be significantly more expensive than a successful one. As a result, optimistic concurrency works best for situations with low to medium contention levels where the majority of the application transactions complete without update conflicts. This is also the reason why this concurrency model is called optimistic.

In addition to updates, ODB also performs state mismatch detection when we are deleting an object from the database (Section 3.11, "Deleting Persistent Objects"). To understand why this can be important, consider the following application transaction:

unsigned long id = ...;
person p;

{
  transaction t (db.begin ());
  db.load (id, p);
  t.commit ();
}

string answer;
cerr << "age is " << p.age () << ", delete?" << endl;
getline (cin, answer);

if (answer == "yes")
{
  transaction t (db.begin ());
  db.erase (p);
  t.commit ();
}
  

Consider again what happens if another process or thread updates the object by changing the person's age while we are waiting for the user input. In this case, the user makes the decision based on a certain age while we may delete (or not delete) an object that has a completely different age. Here is how we can fix this problem using optimistic concurrency:

unsigned long id = ...;
person p;

{
  transaction t (db.begin ());
  db.load (id, p);
  t.commit ();
}

string answer;
for (bool done (false); !done; )
{
  if (answer.empty ())
    cerr << "age is " << p.age () << ", delete?" << endl;
  else
    cerr << "age changed to " << p.age () << ", still delete?" << endl;

  getline (cin, answer);

  if (answer == "yes")
  {
    transaction t (db.begin ());

    try
    {
      db.erase (p);
      done = true;
    }
    catch (const object_changed&)
    {
      db.reload (p);
    }

    t.commit ();
  }
  else
    done = true;
}
  

Note that state mismatch detection is performed only if we delete an object by passing the object instance to the erase() function. If we want to delete an object with the optimistic concurrency model regardless of its state, then we need to use the erase() function that deletes an object given its id, for example:

{
  transaction t (db.begin ());
  db.erase (p.id ());
  t.commit ();
}
  

Finally, note that for persistent classes with the optimistic concurrency model both the update() function as well as the erase() function that accepts an object instance as its argument no longer throw the object_not_persistent exception if there is no such object in the database. Instead, this condition is treated as a change of object state and the object_changed exception is thrown instead.

For complete sample code that shows how to use optimistic concurrency, refer to the optimistic example in the odb-examples package.


13 Database Schema Evolution

When we add new persistent classes or change the existing ones, for example, by adding or deleting data members, the database schema necessary to store the new object model changes as well. At the same time, we may have existing databases that contain existing data. If new versions of your application don't need to handle old databases, then the schema creating functionality is all that you need. However, most applications will need to work with data stored by older versions of the same application.

We will call database schema evolution the overall task of updating the database to match the changes in the object model. Schema evolution usually consists of two sub-tasks: schema migration and data migration. Schema migration modifies the database schema to correspond to the current object model. In a relational database, this, for example, could require adding or dropping tables and columns. The data migration task involves converting the data stored in the existing database from the old format to the new one.

If performed manually, database schema evolution is a tedious and error-prone task. As a result, ODB provides comprehensive support for automated or, more precisely, semi-automated schema evolution. Specifically, ODB does fully-automatic schema migration and provides facilities to help you with data migration.

The topic of schema evolution is a complex and sensitive issue since normally there would be valuable, production data at stake. As a result, the approach taken by ODB is to provide simple and bullet-proof elementary building blocks (or migration steps) that we can understand and trust. Using these elementary blocks we can then implement more complex migration scenarios. In particular, ODB does not try to handle data migration automatically since in most cases this requires understanding of application-specific semantics. In other words, there is no magic.

There are two general approaches to working with older data: the application can either convert it to correspond to the new format or it can be made capable of working with multiple versions of this format. There is also a hybrid approach where the application may convert the data to the new format gradually as part of its normal functionality. ODB is capable of handling all these scenarios. That is, there is support for working with older models without performing any migration (schema or data). Alternatively, we can migrate the schema after which we have the choice of either also immediately migrating the data (immediate data migration) or doing it gradually (gradual data migration).

Schema evolution is already a complex task and we should not unnecessarily use a more complex approach where a simpler one would be sufficient. From the above, the simplest approach is the immediate schema migration that does not require any data migration. An example of such a change would be adding a new data member with the default value (Section 14.3.4, "default"). This case ODB can handle completely automatically.

If we do require data migration, then the next simplest approach is the immediate schema and data migration. Here we have to write custom migration code. However, it is separate from the rest of the core application logic and is executed at a well defined point (database migration). In other words, the core application logic need not be aware of older model versions. The potential drawback of this approach is performance. It may take a lot of resources and/or time to convert all the data upfront.

If the immediate migration is not possible, then the next option is the immediate schema migration followed by the gradual data migration. With this approach, both old and new data must co-exist in the new database. We also have to change the application logic to both account for different sources of the same data (for example, when either an old or new version of the object is loaded) as well as migrate the data when appropriate (for example, when the old version of the object is updated). At some point, usually when the majority of the data has been converted, gradual migrations are terminated with an immediate migration.

The most complex approach is working with multiple versions of the database without performing any migrations, schema or data. ODB does provide support for implementing this approach (Section 13.4, "Soft Object Model Changes"), however we will not cover it any further in this chapter. Generally, this will require embedding knowledge about each version into the core application logic which makes it hard to maintain for any non-trivial object model.

Note also that when it comes to data migration, we can use the immediate variant for some changes and gradual for others. We will discuss various migration scenarios in greater detail in section Section 13.3, "Data Migration".

13.1 Object Model Version and Changelog

To enable schema evolution support in ODB we need to specify the object model version, or, more precisely, two versions. The first is the base model version. It is the lowest version from which we will be able to migrate. The second version is the current model version. In ODB we can migrate from multiple previous versions by successively migrating from one to the next until we reach the current version. We use the db model version pragma to specify both the base and current versions.

When we enable schema evolution for the first time, our base and current versions will be the same, for example:

#pragma db model version(1, 1)
  

Once we release our application, its users may create databases with the schema corresponding to this version of the object model. This means that if we make any modifications to our object model that also change the schema, then we will need to be able to migrate the old databases to this new schema. As a result, before making any new changes after a release, we increment the current version, for example:

#pragma db model version(1, 2)
  

To put this another way, we can stay on the same version during development and keep adding new changes to it. But once we release it, any new changes to the object model will have to be done in a new version.

It is easy to forget to increment the version before making new changes to the object model. To help solve this problem, the db model version pragma accepts a third optional argument that specify whether the current version is open or closed for changes. For example:

#pragma db model version(1, 2, open)   // Can add new changes to
                                       // version 2.
  
#pragma db model version(1, 2, closed) // Can no longer add new
                                       // changes to version 2.
  

If the current version is closed, ODB will refuse to accept any new schema changes. In this situation you would normally increment the current version and mark it as open or you could re-open the existing version if, for example, you need to fix something. Note, however, that re-opening versions that have been released will most likely result in migration malfunctions. By default the version is open.

Normally, an application will have a range of older database versions from which it is able to migrate. When we change this range by removing support for older versions, we also need to adjust the base model version. This will make sure that ODB does not keep unnecessary information around.

A model version (both base and current) is a 64-bit unsigned integer (unsigned long long). 0 is reserved to signify special situations, such as the lack of schema in the database. Other than that, we can use any values as versions as long as they are monotonically increasing. In particular, we don't have to start with version 1 and can increase the versions by any increment.

One versioning approach is to use an independent object model version by starting from version 1 and also incrementing by 1. The alternative is to make the model version correspond to the application version. For example, if our application is using the X.Y.Z version format, then we could encode it as a hexadecimal number and use that as our model version, for example:

#pragma db model version(0x020000, 0x020306) // 2.0.0-2.3.6
  

Most real-world object models will be spread over multiple header files and it will be burdensome to repeat the db model version pragma in each of them. The recommended way to handle this situation is to place the version pragma into a separate header file and include it into the object model files. If your project already has a header file that defines the application version, then it is natural to place this pragma there. For example:

// version.hxx
//
// Define the application version.
//

#define MYAPP_VERSION 0x020306 // 2.3.6

#ifdef ODB_COMPILER
#pragma db model version(1, 7)
#endif
  

Note that we can also use macros in the version pragma which allows us to specify all the versions in a single place. For example:

#define MYAPP_VERSION      0x020306 // 2.3.6
#define MYAPP_BASE_VERSION 0x020000 // 2.0.0

#ifdef ODB_COMPILER
#pragma db model version(MYAPP_BASE_VERSION, MYAPP_VERSION)
#endif
  

It is also possible to have multiple object models within the same application that have different versions. Such models must be independent, that is, no headers from one model shall include a header from another. You will also need to assign different schema names to each model with the --schema-name ODB compiler option.

Once we specify the object model version, the ODB compiler starts tracking database schema changes in a changelog file. Changelog has an XML-based, line-oriented format. It uses XML in order to provide human readability while also facilitating, if desired, processing and analysis with custom tools. The line orientation makes it easy to review with tools like diff.

The changelog is maintained by the ODB compiler. Specifically, you do not need to make any manual changes to this file. You will, however, need to keep it around from one invocation of the ODB compiler to the next. In other words, the changelog file is both the input and the output of the ODB compiler. This, for example, means that if your project's source code is stored in a version control repository, then you will most likely want to store the changelog there as well. If you delete the changelog, then any ability to do schema migration will be lost.

The only operation that you may want to perform with the changelog is to review the database schema changes that resulted from the C++ object model changes. For this you can use a tool like diff or, better yet, the change review facilities offered by your revision control system. For this purpose the contents of a changelog will be self-explanatory.

As an example, consider the following initial object model:

// person.hxx
//

#include <string>

#pragma db model version(1, 1)

#pragma db object
class person
{
  ...

  #pragma db id auto
  unsigned long id_;

  std::string first_;
  std::string last_;
};
  

We then compile this header file with the ODB compiler (using the PostgreSQL database as an example):

odb --database pgsql --generate-schema person.hxx
  

If we now look at the list of generated files, then in addition to the now familiar person-odb.?xx and person.sql, we will also see person.xml — the changelog file. Just for illustration, below are the contents of this changelog.

<changelog database="pgsql">
  <model version="1">
    <table name="person" kind="object">
      <column name="id" type="BIGINT" null="false"/>
      <column name="first" type="TEXT" null="false"/>
      <column name="last" type="TEXT" null="false"/>
      <primary-key auto="true">
        <column name="id"/>
      </primary-key>
    </table>
  </model>
</changelog>
  

Let's say we now would like to add another data member to the person class — the middle name. We increment the version and make the change:

#pragma db model version(1, 2)

#pragma db object
class person
{
  ...

  #pragma db id auto
  unsigned long id_;

  std::string first_;
  std::string middle_;
  std::string last_;
};
  

We use exactly the same command line to re-compile our file:

odb --database pgsql --generate-schema person.hxx
  

This time the ODB compiler will read the old changelog, update it, and write out the new version. Again, for illustration only, below are the updated changelog contents:

<changelog database="pgsql">
  <changeset version="2">
    <alter-table name="person">
      <add-column name="middle" type="TEXT" null="false"/>
    </alter-table>
  </changeset>

  <model version="1">
    <table name="person" kind="object">
      <column name="id" type="BIGINT" null="false"/>
      <column name="first" type="TEXT" null="false"/>
      <column name="last" type="TEXT" null="false"/>
      <primary-key auto="true">
        <column name="id"/>
      </primary-key>
    </table>
  </model>
</changelog>
  

Just to reiterate, while the changelog may look like it could be written by hand, it is maintained completely automatically by the ODB compiler and the only reason you may want to look at its contents is to review the database schema changes. For example, if we compare the above two changelogs with diff, we will get the following summary of the database schema changes:

--- person.xml.orig
+++ person.xml
@@ -1,4 +1,10 @@
<changelog database="pgsql">
+  <changeset version="2">
+    <alter-table name="person">
+      <add-column name="middle" type="TEXT" null="false"/>
+    </alter-table>
+  </changeset>
+
  <model version="1">
    <table name="person" kind="object">
      <column name="id" type="BIGINT" null="false"/>
  

The changelog is only written when we generate the database schema, that is, the --generate-schema option is specified. Invocations of the ODB compiler that only produce the database support code (C++) do not read or update the changelog. To put it another way, the changelog tracks changes in the resulting database schema, not the C++ object model.

ODB ignores column order when comparing database schemas. This means that we can re-order data members in a class without causing any schema changes. Member renames, however, will result in schema changes since the column name changes as well (unless we specified the column name explicitly). From ODB's perspective such a rename looks like the deletion of one data member and the addition of another. If we don't want this to be treated as a schema change, then we will need to keep the old column name by explicitly specifying it with the db column pragma. For example, here is how we can rename middle_ to middle_name_ without causing any schema changes:

#pragma db model version(1, 2)

#pragma db object
class person
{
  ...

  #pragma db column("middle") // Keep the original column name.
  std::string middle_name_;

  ...
};
  

If your object model consists of a large number of header files and you generate the database schema for each of them individually, then a changelog will be created for each of your header files. This may be what you want, however, the large number of changelogs can quickly become unwieldy. In fact, if you are generating the database schema as standalone SQL files, then you may have already experienced a similar problem caused by a large number of .sql files, one for each header.

The solution to both of these problems is to generate a combined database schema file and a single changelog. For example, assume we have three header files in our object model: person.hxx, employee.hxx, and employer.hxx. To generate the database support code we compile them as usual but without specifying the --generate-schema option. In this case no changelog is created or updated:

odb --database pgsql person.hxx
odb --database pgsql employee.hxx
odb --database pgsql employer.hxx
  

To generate the database schema, we perform a separate invocation of the ODB compiler. This time, however, we instruct it to only generate the schema (--generate-schema-only) and produce it combined (--at-once) for all the files in our object model:

odb --database pgsql --generate-schema-only --at-once \
--input-name company person.hxx employee.hxx employer.hxx
  

The result of the above command is a single company.sql file (the name is derived from the --input-name value) that contains the database schema for our entire object model. There is also a single corresponding changelog file — company.xml.

The same can be achieved for the embedded schema by instructing the ODB compiler to generate the database creation code into a separate C++ file (--schema-format separate):

odb --database pgsql --generate-schema-only --schema-format separate \
--at-once --input-name company person.hxx employee.hxx employer.hxx
  

The result of this command is a single company-schema.cxx file and, again, company.xml.

Note also that by default the changelog file is not placed into the directory specified with the --output-dir option. This is due to the changelog being both an input and an output file at the same time. As a result, by default, the ODB compiler will place it in the directory of the input header file.

There is, however, a number of command line options (including --changelog-dir) that allow us to fine-tune the name and location of the changelog file. For example, you can instruct the ODB compiler to read the changelog from one file while writing it to another. This, for example, can be useful if you want to review the changes before discarding the old file. For more information on these options, refer to the ODB Compiler Command Line Manual and search for "changelog".

When we were discussing version increments above, we used the terms development and release. Specifically, we talked about keeping the same object model versions during development periods and incrementing them after releases. What is a development period and a release in this context? These definitions can vary from project to project. Generally, during a development period we work on one or more changes to the object model that result in the changes to the database schema. A release is a point where we make our changes available to someone else who may have an older database to migrate from. In the traditional sense, a release is a point where you make a new version of your application available to its users. However, for schema evolution purposes, a release could also mean simply making your schema-altering changes available to other developers on your team. Let us consider two common scenarios to illustrate how all this fits together.

One way to setup a project would be to re-use the application development period and application release for schema evolution. That is, during a new application version development we keep a single object model version and when we release the application, we increment the model version. In this case it makes sense to also reuse the application version as a model version for consistency. Here is a step-by-step guide for this setup:

  1. During development, keep the current object model version open.
  2. Before the release (for example, when entering a "feature freeze") close the version.
  3. After the release, update the version and open it.
  4. For each new feature, review the changeset at the top of the changelog, for example, with diff or your version control facilities. If you are using a version control, then this is best done just before committing your changes to the repository.

An alternative way to setup schema versioning in a project would be to define the development period as working on a single feature and the release as making this feature available to other people (developers, testers, etc.) on your team, for example, by committing the changes to a public version control repository. In this case, the object model version will be independent of the application version and can simply be a sequence that starts with 1 and is incremented by 1. Here is a step-by-step guide for this setup:

  1. Keep the current model version closed. Once a change is made that affects the database schema, the ODB compiler will refuse to update the changelog.
  2. If the change is legitimate, open a new version, that is, increment the current version and make it open.
  3. Once the feature is implemented and tested, review the final set of database changes (with diff or your version control facilities), close the version, and commit the changes to the version control repository (if using).

If you are using a version control repository that supports pre-commit checks, then you may want to consider adding such a check to make sure the committed version is always closed.

If we are just starting schema evolution in our project, which approach should we choose? The two approaches will work better in different situations since they have a different set of advantages and disadvantages. The first approach, which we can call version per application release, is best suited for simpler projects with smaller releases since otherwise a single migration will bundle a large number of unrelated actions corresponding to different features. This can become difficult to review and, if things go wrong, debug.

The second approach, which we can call version per feature, is much more modular and provides a number of additional benefits. We can perform migrations for each feature as a discreet step which makes it easier to debug. We can also place each such migration step into a separate transaction further improving reliability. It also scales much better in larger teams where multiple developers can work concurrently on features that affect the database schema. For example, if you find yourself in a situation where another developer on your team used the same version as you and managed to commit his changes before you (that is, you have a merge conflict), then you can simply change the version to the next available one, regenerate the changelog, and continue with your commit.

Overall, unless you have strong reasons to prefer the version per application release approach, rather choose version per feature even though it may seem more complex at the beginning. Also, if you do select the first approach, consider provisioning for switching to the second method by reserving a sub-version number. For example, for an application version in the form 2.3.4 you can make the object model version to be in the form 0x0203040000, reserving the last two bytes for a sub-version. Later on you can use it to switch to the version per feature approach.

13.2 Schema Migration

Once we enable schema evolution by specifying the object model version, in addition to the schema creation statements, the ODB compiler starts generating schema migration statements for each version all the way from the base to the current. As with schema creation, schema migration can be generated either as a set of SQL files or embedded into the generated C++ code (--schema-format option).

For each migration step, that is from one version to the next, ODB generates two sets of statements: pre-migration and post-migration. The pre-migration statements "relax" the database schema so that both old and new data can co-exist. At this stage new columns and tables are added while old constraints are dropped. The post-migration statements "tighten" the database schema back so that only data conforming to the new format can remain. At this stage old columns and tables are dropped and new constraints are added. Now you can probably guess where the data migration fits into this — between the pre and post schema migrations where we can both access the old data and create the new one.

If the schema is being generated as standalone SQL files, then we end up with a pair of files for each step: the pre-migration file and the post-migration file. For the person example we started in the previous section we will have the person-002-pre.sql and person-002-post.sql files. Here 002 is the version to which we are migrating while the pre and post suffixes specify the migration stage. So if we wanted to migrate a person database from version 1 to 2, then we would first execute person-002-pre.sql, then migrate the data, if any (discussed in more detail in the next section), and finally execute person-002-post.sql. If our database is several versions behind, for example the database has version 1 while the current version is 5, then we simply perform this set of steps for each version until we reach the current version.

If we look at the contents of the person-002-pre.sql file, we will see the following (or equivalent, depending on the database used) statement:

ALTER TABLE "person"
  ADD COLUMN "middle" TEXT NULL;
  

As we would expect, this statement adds a new column corresponding to the new data member. An observant reader would notice, however, that the column is added as NULL even though we never requested this semantics in our object model. Why is the column added as NULL? If during migration the person table already contains rows (that is, existing objects), then an attempt to add a non-NULL column that doesn't have a default value will fail. As a result, ODB will initially add a new column that doesn't have a default value as NULL but then clean this up at the post-migration stage. This way your data migration code is given a chance to assign some meaningful values for the new data member for all the existing objects. Here are the contents of the person-002-post.sql file:

ALTER TABLE "person"
  ALTER COLUMN "middle" SET NOT NULL;
  

Currently ODB directly supports the following elementary database schema changes:

More complex changes can normally be implemented in terms of these building blocks. For example, to change a type of a data member (which leads to a change of a column type), we can add a new data member with the desired type (add column), migrate the data, and then delete the old data member (drop column). ODB will issue diagnostics for cases that are currently not supported directly. Note also that some database systems (notably SQLite) have a number of limitations in their support for schema changes. For more information on these database-specific limitations, refer to the "Limitations" sections in Part II, "Database Systems".

How do we know what the current database version is? That is, the version from which we need to migrate? We need to know this, for example, in order to determine the set of migrations we have to perform. By default, when schema evolution is enabled, ODB maintains this information in a special table called schema_version that has the following (or equivalent, depending on the database used) definition:

CREATE TABLE "schema_version" (
  "name" TEXT NOT NULL PRIMARY KEY,
  "version" BIGINT NOT NULL,
  "migration" BOOLEAN NOT NULL);
  

The name column is the schema name as specified with the --schema-name option. It is empty for the default schema. The version column contains the current database version. And, finally, the migration flag indicates whether we are in the process of migrating the database, that is, between the pre and post-migration stages.

The schema creation statements (person.sql in our case) create this table and populate it with the initial model version. For example, if we executed person.sql corresponding to version 1 of our object model, then name would have been empty (which signifies the default schema since we didn't specify --schema-name), version will be 1 and migration will be FALSE.

The pre-migration statements update the version and set the migration flag to TRUE. Continuing with our example, after executing person-002-pre.sql, version will become 2 and migration will be set to TRUE. The post-migration statements simply clear the migration flag. In our case, after running person-002-post.sql, version will remain 2 while migration will be reset to FALSE.

Note also that above we mentioned that the schema creation statements (person.sql) create the schema_version table. This means that if we enable schema evolution support in the middle of a project, then we could already have existing databases that don't include this table. As a result, ODB will not be able to handle migrations for such databases unless we manually add the schema_version table and populate it with the correct version information. For this reason, it is highly recommended that you consider whether to use schema evolution and, if so, enable it from the beginning of your project.

The odb::database class provides an API for accessing and modifying the current database version:

namespace odb
{
  typedef unsigned long long schema_version;

  struct LIBODB_EXPORT schema_version_migration
  {
    schema_version_migration (schema_version = 0,
                              bool migration = false);

    schema_version version;
    bool migration;

    // This class also provides the ==, !=, <, >, <=, and >= operators.
    // Version ordering is as follows: {1,f} < {2,t} < {2,f} < {3,t}.
  };

  class database
  {
  public:
    ...

    schema_version
    schema_version (const std::string& name = "") const;

    bool
    schema_migration (const std::string& name = "") const;

    const schema_version_migration&
    schema_version_migration (const std::string& name = "") const;

    // Set schema version and migration state manually.
    //
    void
    schema_version_migration (schema_version,
                              bool migration,
                              const std::string& name = "");

    void
    schema_version_migration (const schema_version_migration&,
                              const std::string& name = "");

    // Set default schema version table for all schemas.
    //
    void
    schema_version_table (const std::string& table_name);

    // Set schema version table for a specific schema.
    //
    void
    schema_version_table (const std::string& table_name,
                          const std::string& name);
  };
}
  

The schema_version() and schema_migration() accessors return the current database version and migration flag, respectively. The optional name argument is the schema name. If the database schema hasn't been created (that is, there is no corresponding entry in the schema_version table or this table does not exist), then schema_version() returns 0. The schema_version_migration() accessor returns both version and migration flag together in the schema_version_migration struct.

You may already have a version table in your database or you (or your database administrator) may prefer to keep track of versions your own way. You can instruct ODB not to create the schema_version table with the --suppress-schema-version option. However, ODB still needs to know the current database version in order for certain schema evolution mechanisms to function properly. As a result, in this case, you will need to set the schema version on the database instance manually using the schema_version_migration() modifier. Note that the modifier API is not thread-safe. That is, you should not modify the schema version while other threads may be accessing or modifying the same information.

Note also that the accessors we discussed above will only query the schema_version table once and, if the version could be determined, cache the result. If, however, the version could not be determined (that is, schema_version() returned 0), then a subsequent call will re-query the table. While it is probably a bad idea to modify the database schema while the application is running (other than via the schema_catalog API, as discussed below), if for some reason you need ODB to re-query the version, then you can manually set it to 0 using the schema_version_migration() modifier.

It is also possible to change the name of the table that stores the schema version using the --schema-version-table option. You will also need to specify this alternative name on the database instance using the schema_version_table() modifier. The first version specifies the default table that is used for all the schema names. The second version specifies the table for a specific schema. The table name should be database-quoted, if necessary.

If we are generating our schema migrations as standalone SQL files, then the migration workflow could look like this:

  1. The database administrator determines the current database version. If migration is required, then for each migration step (that is, from one version to the next), he performs the following:
  2. Execute the pre-migration file.
  3. Execute our application (or a separate migration program) to perform data migration (discussed later). Our application can determine that is is being executed in the "migration mode" by calling schema_migration() and then which migration code to run by calling schema_version().
  4. Execute the post-migration file.

These steps become more integrated and automatic if we embed the schema creation and migration code into the generated C++ code. Now we can perform schema creation, schema migration, and data migration as well as determine when each step is necessary programmatically from within the application.

Schema evolution support adds the following extra functions to the odb::schema_catalog class, which we first discussed in Section 3.4, "Database".

namespace odb
{
  class schema_catalog
  {
  public:
    ...


    // Schema migration.
    //
    static void
    migrate_schema_pre (database&,
                        schema_version,
                        const std::string& name = "");

    static void
    migrate_schema_post (database&,
                         schema_version,
                         const std::string& name = "");

    static void
    migrate_schema (database&,
                    schema_version,
                    const std::string& name = "");

    // Data migration.
    //
    // Discussed in the next section.


    // Combined schema and data migration.
    //
    static void
    migrate (database&,
             schema_version = 0,
             const std::string& name = "");

    // Schema version information.
    //
    static schema_version
    base_version (const database&,
                  const std::string& name = "");

    static schema_version
    base_version (database_id,
                  const std::string& name = "");

    static schema_version
    current_version (const database&,
                     const std::string& name = "");

    static schema_version
    current_version (database_id,
                     const std::string& name = "");

    static schema_version
    next_version (const database&,
                  schema_version = 0,
                  const std::string& name = "");

    static schema_version
    next_version (database_id,
                  schema_version,
                  const std::string& name = "");
  };
}
  

The migrate_schema_pre() and migrate_schema_post() static functions perform a single stage (that is, pre or post) of a single migration step (that is, from one version to the next). The version argument specifies the version we are migrating to. For instance, in our person example, if we know that the database version is 1 and the next version is 2, then we can execute code like this:

transaction t (db.begin ());

schema_catalog::migrate_schema_pre (db, 2);

// Data migration goes here.

schema_catalog::migrate_schema_post (db, 2);

t.commit ();
  

If you don't have any data migration code to run, then you can perform both stages with a single call using the migrate_schema() static function.

The migrate() static function perform both schema and data migration (we discuss data migration in the next section). It can also perform several migration steps at once. If we don't specify its target version, then it will migrate (if necessary) all the way to the current model version. As an extra convenience, migrate() will also create the database schema if none exists. As a result, if we don't have any data migration code or we have registered it with schema_catalog (as discussed later), then the database schema creation and migration, whichever is necessary, if at all, can be performed with a single function call:

transaction t (db.begin ());
schema_catalog::migrate (db);
t.commit ();
  

Note also that schema_catalog is integrated with the odb::database schema version API. In particular, schema_catalog functions will query and synchronize the schema version on the database instance if and when required.

The schema_catalog class also allows you to iterate over known versions (remember, there could be "gaps" in version numbers) with the base_version(), current_version() and next_version() static functions. The base_version() and current_version() functions return the base and current object model versions, respectively. That is, the lowest version from which we can migrate and the version that we ultimately want to migrate to. The next_version() function returns the next known version. If the passed version is greater or equal to the current version, then this function will return the current version plus one (that is, one past current). If we don't specify the version, then next_version() will use the current database version as the starting point. Note also that the schema version information provided by these functions is only available if we embed the schema migration code into the generated C++ code. For standalone SQL file migrations this information is normally not needed since the migration process is directed by an external entity, such as a database administrator or a script.

Most schema_catalog functions presented above also accept the optional schema name argument. If the passed schema name is not found, then the odb::unknown_schema exception is thrown. Similarly, functions that accept the schema version argument will throw the odb::unknown_schema_version exception if the passed version is invalid. Refer to Section 3.14, "ODB Exceptions" for more information on these exceptions.

To illustrate how all these parts fit together, consider the following more realistic database schema management example. Here we want to handle the schema creation in a special way and perform each migration step in its own transaction.

schema_version v (db.schema_version ());
schema_version bv (schema_catalog::base_version (db));
schema_version cv (schema_catalog::current_version (db));

if (v == 0)
{
  // No schema in the database. Create the schema and
  // initialize the database.
  //
  transaction t (db.begin ());
  schema_catalog::create_schema (db);

  // Populate the database with initial data, if any.

  t.commit ();
}
else if (v < cv)
{
  // Old schema (and data) in the database, migrate them.
  //

  if (v < bv)
  {
    // Error: migration from this version is no longer supported.
  }

  for (v = schema_catalog::next_version (db, v);
       v <= cv;
       v = schema_catalog::next_version (db, v))
  {
    transaction t (db.begin ());
    schema_catalog::migrate_schema_pre (db, v);

    // Data migration goes here.

    schema_catalog::migrate_schema_post (db, v);
    t.commit ();
  }
}
else if (v > cv)
{
  // Error: old application trying to access new database.
}
  

13.3 Data Migration

In quite a few cases specifying the default value for new data members will be all that's required to handle the existing objects. For example, the natural default value for the new middle name that we have added is an empty string. And we can handle this case with the db default pragma and without any extra C++ code:

#pragma db model version(1, 2)

#pragma db object
class person
{
  ...


  #pragma db default("")
  std::string middle_;
};
  

However, there will be situations where we would need to perform more elaborate data migrations, that is, convert old data to the new format. As an example, suppose we want to add gender to our person class. And, instead of leaving it unassigned for all the existing objects, we will try to guess it from the first name. This is not particularly accurate but it could be sufficient for our hypothetical application:

#pragma db model version(1, 3)

enum gender {male, female};

#pragma db object
class person
{
  ...

  gender gender_;
};
  

As we have discussed earlier, there are two ways to perform data migration: immediate and gradual. To recap, with immediate migration we migrate all the existing objects at once, normally after the schema pre-migration statements but before the post-migration statements. With gradual migration, we make sure the new object model can accommodate both old and new data and gradually migrate existing objects as the application runs and the opportunities to do so arise, for example, an object is updated.

There is also another option for data migration that is not discussed further in this section. Instead of using our C++ object model we could execute ad-hoc SQL statements that perform the necessary conversions and migrations directly on the database server. While in certain cases this can be a better option from the performance point of view, this approach is often limited in terms of the migration logic that we can handle.

13.3.1 Immediate Data Migration

Let's first see how we can implement an immediate migration for the new gender_ data member we have added above. If we are using standalone SQL files for migration, then we could add code along these lines somewhere early in main(), before the main application logic:

int
main ()
{
  ...

  odb::database& db = ...

  // Migrate data if necessary.
  //
  if (db.schema_migration ())
  {
    switch (db.schema_version ())
    {
    case 3:
      {
        // Assign gender to all the existing objects.
        //
        transaction t (db.begin ());

        for (person& p: db.query<person> ())
        {
          p.gender (guess_gender (p.first ()));
          db.update (p);
        }

        t.commit ();
        break;
      }
    }
  }

  ...
}
  

If you have a large number of objects to migrate, it may also be a good idea, from the performance point of view, to break one big transaction that we now have into multiple smaller transactions (Section 3.5, "Transactions"). For example:

case 3:
  {
    transaction t (db.begin ());

    size_t n (0);
    for (person& p: db.query<person> ())
    {
      p.gender (guess_gender (p.first ()));
      db.update (p);

      // Commit the current transaction and start a new one after
      // every 100 updates.
      //
      if (n++ % 100 == 0)
      {
        t.commit ();
        t.reset (db.begin ());
      }
    }

    t.commit ();
    break;
  }
  

While it looks straightforward enough, as we add more migration snippets, this approach can quickly become unmaintainable. Instead of having all the migrations in a single function and determining when to run each piece ourselves, we can package each migration into a separate function, register it with the schema_catalog class, and let ODB figure out when to run which migration functions. To support this functionality, schema_catalog provides the following data migration API:

namespace odb
{
  class schema_catalog
  {
  public:
    ...

    // Data migration.
    //
    static std::size_t
    migrate_data (database&,
                  schema_version = 0,
                  const std::string& name = "");

    typedef void data_migration_function_type (database&);

    // Common (for all the databases) data migration, C++98/03 version:
    //
    template <schema_version v, schema_version base>
    static void
    data_migration_function (data_migration_function_type*,
                             const std::string& name = "");

    // Common (for all the databases) data migration, C++11 version:
    //
    template <schema_version v, schema_version base>
    static void
    data_migration_function (std::function<data_migration_function_type>,
                             const std::string& name = "");

    // Database-specific data migration, C++98/03 version:
    //
    template <schema_version v, schema_version base>
    static void
    data_migration_function (database&,
                             data_migration_function_type*,
                             const std::string& name = "");

    template <schema_version v, schema_version base>
    static void
    data_migration_function (database_id,
                             data_migration_function_type*,
                             const std::string& name = "");

    // Database-specific data migration, C++11 version:
    //
    template <schema_version v, schema_version base>
    static void
    data_migration_function (database&,
                             std::function<data_migration_function_type>,
                             const std::string& name = "");

    template <schema_version v, schema_version base>
    static void
    data_migration_function (database_id,
                             std::function<data_migration_function_type>,
                             const std::string& name = "");
  };

  // Static data migration function registration, C++98/03 version:
  //
  template <schema_version v, schema_version base>
  struct data_migration_entry
  {
    data_migration_entry (data_migration_function_type*,
                          const std::string& name = "");

    data_migration_entry (database_id,
                          data_migration_function_type*,
                          const std::string& name = "");
  };

  // Static data migration function registration, C++11 version:
  //
  template <schema_version v, schema_version base>
  struct data_migration_entry
  {
    data_migration_entry (std::function<data_migration_function_type>,
                          const std::string& name = "");

    data_migration_entry (database_id,
                          std::function<data_migration_function_type>,
                          const std::string& name = "");
  };
}
  

The migrate_data() static function performs data migration for the specified version. If no version is specified, then it will use the current database version and also check whether the database is in migration, that is, database::schema_migration() returns true. As a result, all we need to do in our main() is call this function. It will check if migration is required and if so, call all the migration functions registered for this version. For example:

int
main ()
{
  ...

  database& db = ...

  // Check if we need to migrate any data and do so
  // if that's the case.
  //
  schema_catalog::migrate_data (db);

  ...
}
  

The migrate_data() function returns the number of migration functions called. You can use this value for debugging or logging.

The only other step that we need to perform is register our data migration functions with schema_catalog. At the lower level we can call the data_migration_function() static function for every migration function we have, for example, at the beginning of main(). For each version, data migration functions are called in the order of registration.

A more convenient approach, however, is to use the data_migration_entry helper class template to register the migration functions during static initialization. This way we can keep the migration function and its registration code next to each other. Here is how we can reimplement our gender migration code to use this mechanism:

static void
migrate_gender (odb::database& db)
{
  transaction t (db.begin ());

  for (person& p: db.query<person> ())
  {
    p.gender (guess_gender (p.first ()));
    db.update (p);
  }

  t.commit ();
}

static const odb::data_migration_entry<3, MYAPP_BASE_VERSION>
migrate_gender_entry (&migrate_gender);
  

The first template argument to the data_migration_entry class template is the version we want this data migration function to be called for. The second template argument is the base model version. This second argument is necessary to detect the situation where we no longer need this data migration function. Remember that when we move the base model version forward, migrations from any version below the new base are no longer possible. We, however, may still have migration functions registered for those lower versions. Since these functions will never be called, they are effectively dead code and it would be useful to identify and remove them. To assist with this, data_migration_entry (and lower lever data_migration_function()) will check at compile time (that is, static_assert) that the registration version is greater than the base model version.

In the above example we use the MYAPP_BASE_VERSION macro that is presumably defined in a central place, for example, version.hxx. This is the recommended approach since we can update the base version in a single place and have the C++ compiler automatically identify all the data migration functions that can be removed.

In C++11 we can also create a template alias so that we don't have to repeat the base model macro in every registration, for example:

template <schema_version v>
using migration_entry = odb::data_migration_entry<v, MYAPP_BASE_VERSION>;

static const migration_entry<3>
migrate_gender_entry (&migrate_gender);
  

For cases where you need to by-pass the base version check, for example, to implement your own registration helper, ODB also provides "unsafe" versions of the data_migration_function() functions that take the version as a function argument rather than as a template parameter.

In C++11 we can also use lambdas as migration functions, which makes the migration code more concise:

static const migration_entry<3>
migrate_gender_entry (
  [] (odb::database& db)
  {
    transaction t (db.begin ());

    for (person& p: db.query<person> ())
    {
      p.gender (guess_gender (p.first ()));
      db.update (p);
    }

    t.commit ();
  });
  

If we are using embedded schema migrations, then both schema and data migration is integrated and can be performed with a single call to the schema_catalog::migrate() function that we discussed earlier. For example:

int
main ()
{
  ...

  database& db = ...

  // Check if we need to migrate the database and do so
  // if that's the case.
  //
  {
    transaction t (db.begin ());
    schema_catalog::migrate (db);
    t.commit ();
  }

  ...
}
  

Note, however, that in this case we call migrate() within a transaction (for the schema migration part) which means that our migration functions will also be called within this transaction. As a result, we will need to adjust our migration functions not to start their own transaction:

static void
migrate_gender (odb::database& db)
{
  // Assume we are already in a transaction.
  //
  for (person& p: db.query<person> ())
  {
    p.gender (guess_gender (p.first ()));
    db.update (p);
  }
}
  

If, however, we want more granular transactions, then we can use the lower-level schema_catalog functions to gain more control, as we have seen at the end of the previous section. Here is the relevant part of that example with an added data migration call:

  // Old schema (and data) in the database, migrate them.
  //
  for (v = schema_catalog::next_version (db, v);
       v <= cv;
       v = schema_catalog::next_version (db, v))
  {
    transaction t (db.begin ());
    schema_catalog::migrate_schema_pre (db, v);
    schema_catalog::migrate_data (db, v);
    schema_catalog::migrate_schema_post (db, v);
    t.commit ();
  }
  

13.3.2 Gradual Data Migration

If the number of existing objects that require migration is large, then an all-at-once, immediate migration, while simple, may not be practical from a performance point of view. In this case, we can perform a gradual migration as the application does its normal functions.

With gradual migrations, the object model must be capable of representing data that conforms to both old and new formats at the same time since, in general, the database will contain a mixture of old and new objects. For example, in case of our gender data member, we need a special value that represents the "no gender assigned yet" case (an old object). We also need to assign this special value to all the existing objects during the schema pre-migration stage. One way to do this would be add a special value to our gender enum and then make it the default value with the db default pragma. A cleaner and easier approach, however, is to use NULL as a special value. We can add support for the NULL value semantics to any existing type by wrapping it with odb::nullable, boost::optional or similar (Section 7.3, "Pointers and NULL Value Semantics"). We also don't need to specify the default value explicitly since NULL is used automatically. Here is how we can use this approach in our gender example:

#include <odb/nullable.hxx>

#pragma db object
class person
{
  ...

  odb::nullable<gender> gender_;
};
  

A variety of strategies can be employed to implement gradual migrations. For example, we can migrate the data when the object is updated as part of the normal application logic. While there is no migration cost associated with this approach (the object is updated anyway), depending on how often objects are typically updated, this strategy can take a long time to complete. An alternative strategy would be to perform an update whenever an old object is loaded. Yet another strategy is to have a separate thread that slowly migrates all the old objects as the application runs.

As an example, let us implement the first approach for our gender migration. While we could have added the necessary code throughout the application, from the maintenance point of view, it is best to try and localize the gradual migration logic to the persistent classes that it affects. And for this database operation callbacks (Section 14.1.7, "callback") are a very useful mechanism. In our case, all we have to do is handle the post_load event where we guess the gender if it is NULL:

#include <odb/core.hxx>     // odb::database
#include <odb/callback.hxx> // odb::callback_event
#include <odb/nullable.hxx>

#pragma db object callback(migrate)
class person
{
  ...

  void
  migrate (odb::callback_event e, odb::database&)
  {
    if (e == odb::callback_event::post_load)
    {
      // Guess gender if not assigned.
      //
      if (gender_.null ())
        gender_ = guess_gender (first_);
    }
  }

  odb::nullable<gender> gender_;
};
  

In particular, we don't have to touch any of the accessors or modifiers or the application logic — all of them can assume that the value can never be NULL. And when the object is next updated, the new gender value will be stored automatically.

All gradual migrations normally end up with a terminating immediate migration some number of versions down the line, when the bulk of the objects has presumably been converted. This way we don't have to keep the gradual migration code around forever. Here is how we could implement a terminating migration for our example:

// person.hxx
//
#pragma db model version(1, 4)

#pragma db object
class person
{
  ...

  gender gender_;
};

// person.cxx
//
static void
migrate_gender (odb::database& db)
{
  typedef odb::query<person> query;

  for (person& p: db.query<person> (query::gender.is_null ()))
  {
    p.gender (guess_gender (p.first ()));
    db.update (p);
  }
}

static const odb::data_migration_entry<4, MYAPP_BASE_VERSION>
migrate_gender_entry (&migrate_gender);
  

A couple of points to note about this code. Firstly, we removed all the gradual migration logic (the callback) from the class and replaced it with the immediate migration function. We also removed the odb::nullable wrapper (and therefore disallowed the NULL values) since after this migration all the objects will have been converted. Finally, in the migration function, we only query the database for objects that need migration, that is, have NULL gender.

13.4 Soft Object Model Changes

Let us consider another common kind of object model change: we delete an old member, add a new one, and need to copy the data from the old to the new, perhaps applying some conversion. For example, we may realize that in our application it is a better idea to store a person's name as a single string rather than split it into three fields. So what we would like to do is add a new data member, let's call it name_, convert all the existing split names, and then delete the first_, middle_, and last_ data members.

While this sounds straightforward, there is a problem. If we delete (that is, physically remove from the source code) the old data members, then we won't be able to access the old data. The data will still be available in the database between the schema pre and post-migrations, it is just we will no longer be able to access it through our object model. And if we keep the old data members around, then the old data will remain stored in the database even after the schema post-migration.

There is also a more subtle problem that has to do with existing migrations for the previous versions. Remember, in version 3 of our person example we added the gender_ data member. We also have a data migration function which guesses the gender based on the first name. Deleting the first_ data member from our class will obviously break this code. But even adding the new name_ data member will cause problems because when we try to update the object in order to store the new gender, ODB will try to update name_ as well. But there is no corresponding column in the database yet. When we run this migration function, we are still several versions away from the point where the name column will be added.

This is a very subtle but also very important implication to understand. Unlike the main application logic, which only needs to deal with the current model version, data migration code works on databases that can be multiple versions behind the current version.

How can we resolve this problem? It appears what we need is the ability to add or delete data members starting from a specific version. In ODB this mechanism is called soft member additions and deletions. A soft-added member is only treated as persistent starting from the addition version. A soft-deleted member is persistent until the deletion version (but including the migration stage). In its essence, soft model changes allow us to maintain multiple versions of our object model all with a single set of persistent classes. Let us now see how this functionality can help implement our changes:

#pragma db model version(1, 4)

#pragma db object
class person
{
  ...

  #pragma db id auto
  unsigned long id_;

  #pragma db deleted(4)
  std::string first_;

  #pragma db deleted(4)
  std::string middle_;

  #pragma db deleted(4)
  std::string last_;

  #pragma db added(4)
  std::string name_;

  gender gender_;
};
  

The migration function for this change could then look like this:

static void
migrate_name (odb::database& db)
{
  for (person& p: db.query<person> ())
  {
    p.name (p.first () + " " +
            p.middle () + (p.middle ().empty () ? "" : " ") +
            p.last ());
    db.update (p);
  }
}

static const odb::data_migration_entry<4, MYAPP_BASE_VERSION>
migrate_name_entry (&migrate_name);
  

Note also that no changes are required to the gender migration function.

As you may have noticed, in the code above we assumed that the person class still provides public accessors for the now deleted data members. This might not be ideal since now they should not be used by the application logic. The only code that may still need to access them is the migration functions. The recommended way to resolve this is to remove the accessors/modifiers corresponding to the deleted data member, make migration functions static functions of the class being migrated, and then access the deleted data members directly. For example:

#pragma db model version(1, 4)

#pragma db object
class person
{
  ...

private:
  friend class odb::access;

  #pragma db id auto
  unsigned long id_;

  #pragma db deleted(4)
  std::string first_;

  #pragma db deleted(4)
  std::string middle_;

  #pragma db deleted(4)
  std::string last_;

  #pragma db added(4)
  std::string name_;

  gender gender_;

private:
  static void
  migrate_gender (odb::database&);

  static void
  migrate_name (odb::database&);
};

void person::
migrate_gender (odb::database& db)
{
  for (person& p: db.query<person> ())
  {
    p.gender_ = guess_gender (p.first_);
    db.update (p);
  }
}

static const odb::data_migration_entry<3, MYAPP_BASE_VERSION>
migrate_name_entry (&migrate_gender);

void person::
migrate_name (odb::database& db)
{
  for (person& p: db.query<person> ())
  {
    p.name_ = p.first_ + " " +
              p.middle_ + (p.middle_.empty () ? "" : " ") +
              p.last_;
    db.update (p);
  }
}

static const odb::data_migration_entry<4, MYAPP_BASE_VERSION>
migrate_name_entry (&migrate_name);
  

Another potential issue with the soft-deletion is the requirement to keep the delete data members in the class. While they will not be initialized in the normal operation of the application (that is, not a migration), this can still be a problem if we need to minimize the memory footprint of our classes. For example, we may cache a large number of objects in memory and having three std::string data members can be a significant overhead.

The recommended way to resolve this issue is to place all the deleted data members into a dynamically allocated composite value type. For example:

#pragma db model version(1, 4)

#pragma db object
class person
{
  ...

  #pragma db id auto
  unsigned long id_;

  #pragma db added(4)
  std::string name_;

  gender gender_;

  #pragma db value
  struct deleted_data
  {
    #pragma db deleted(4)
    std::string first_;

    #pragma db deleted(4)
    std::string middle_;

    #pragma db deleted(4)
    std::string last_;
  };

  #pragma db column("")
  std::unique_ptr<deleted_data> dd_;

  ...
};
  

ODB will then automatically allocate the deleted value type if any of the deleted data members are being loaded. During the normal operation, however, the pointer will stay NULL and therefore reduce the common case overhead to a single pointer per class. Note that we make the composite value column prefix empty (the db column("") pragma) in order to keep the same column names for the deleted data members.

Soft-added and deleted data members can be used in objects, composite values, views, and container value types. We can also soft-add and delete data members of simple, composite, pointer to object, and container types. Only special data members, such as the object id and the optimistic concurrency version, cannot be soft-added or deleted.

It is also possible to soft-delete a persistent class. We can still work with the existing objects of such a class, however, no table is created in new databases for soft-deleted classes. To put it another way, a soft-delete class is like an abstract class (no table) but which can still be loaded, updated, etc. Soft-added persistent classes do not make much sense and are therefore not supported.

As an example of a soft-deleted class, suppose we want to replace our person class with the new employee object and migrate the data. Here is how we could do this:

#pragma db model version(1, 5)

#pragma db object deleted(5)
class person
{
  ...
};

#pragma db object
class employee
{
  ...

  #pragma db id auto
  unsigned long id_;

  std::string name_;
  gender gender_;

  static void
  migrate_person (odb::database&);
};

void employee::
migrate_person (odb::database& db)
{
  for (person& p: db.query<person> ())
  {
    employee e (p.name (), p.gender ());
    db.persist (e);
  }
}

static const odb::data_migration_entry<5, MYAPP_BASE_VERSION>
migrate_person_entry (&migrate_person);
  

As we have seen above, hard member additions and deletions can (and most likely will) break existing data migration code. Why, then, not treat all the changes, or at least additions, as soft? ODB requires you to explicitly request this semantics because support for soft-added and deleted data members incurs runtime overhead. And there can be plenty of cases where there is no existing data migration and therefore hard additions and deletions are sufficient.

In some cases a hard addition or deletion will result in a compile-time error. For example, one of the data migration functions may reference the data member we just deleted. In many cases, however, such errors can only be detected at runtime, and, worse yet, only when the migration function is executed. For example, we may hard-add a new data member that an existing migration function will try to indirectly store in the database as part of an object update. As a result, it is highly recommended that you always test your application with the database that starts at the base version so that every data migration function is called and therefore ensured to still work correctly.

To help with this problem you can also instruct ODB to warn you about any hard additions or deletions with the --warn-hard-add, --warn-hard-delete, and --warn-hard command line options. ODB will only warn you about hard changes in the current version and only for as long as it is open, which makes this mechanism fairly usable.

You may also be wondering why we have to specify the addition and deletion versions explicitly. It may seem like the ODB compiler should be able to figure this out automatically. While it is theoretically possible, to achieve this, ODB would have to also maintain a separate changelog of the C++ object model in addition to the database schema changelog it already maintains. While being a lot more complex, such an additional changelog would also complicate the workflow significantly. In this light, maintaining this change information as part of the original source files appears to be a cleaner and simpler approach.

As we discussed before, when we move the base model version forward we essentially drop support for migrations from versions before the new base. As a result, it is no longer necessary to maintain the soft semantics of additions and deletions up to and including the new base version. ODB will issue diagnostics for all such members and classes. For soft deletions we can simply remove the data member or class entirely. For soft additions we only need to remove the db added pragma.

13.4.1 Reuse Inheritance Changes

Besides adding and deleting data members, another way to alter the object's table is using reuse-style inheritance. If we add a new reuse base, then, from the database schema point of view, this is equivalent to adding all its columns to the derived object's table. Similarly, deleting reuse inheritance results in all the base's columns being deleted from the derived's table.

In the future ODB may provide direct support for soft addition and deletion of inheritance. Currently, however, this semantics can be emulated with soft-added and deleted data members. The following table describes the most common scenarios depending on where columns are added or deleted, that is, base table, derived table, or both.

DELETE HARD SOFT
In both (delete inheritance and base) Delete inheritance and base. Move object id to derived. Soft-delete base. Mark all data members (except id) in base as soft-deleted.
In base only (delete base) Option 1: mark base as abstract.

Option 2: move all the base member to derived, delete base.
Soft-delete base.
In derived only (delete inheritance) Delete inheritance, add object id to derived. Option 1: copy base to a new soft-deleted base, inherit from it instead. Mark all the data members (expect id) in this new base as soft-deleted. Note: we add the new base as soft-deleted to get notified when we can remove it.

Option 2: Copy all the data members from base to derived and mark them as soft-deleted in derived.
ADD HARD SOFT
In both (add new base and inheritance) Add new base and inheritance. Potentially move object id member from derived to base. Add new base and mark all its data members as soft-added. Add inheritance. Move object id from derived to base.
In base only (refactor existing data to new base) Add new base and move data members from derived to base. Note: in most cases the new base will be made abstract which make this scenario non-schema changing. The same as HARD.
In derived only (add inheritance to existing base) Add inheritance, delete object id in derived. Copy existing base to a new abstract base and inherit from it. Mark all the database members in the new base as soft-added (except object id). When notified by the ODB compiler that the soft addition of the data members is no longer necessary, delete the copy and inherit from the original base.

13.4.2 Polymorphism Inheritance Changes

Unlike reuse inheritance, adding or deleting a polymorphic base does not result in the base's data members being added or deleted from the derived object's table because each class in a polymorphic hierarchy is stored in a separate table. There are, however, other complications due to the presence of special columns (discriminator in the root table and object id links in derived tables) which makes altering the hierarchy structure difficult to handle automatically. Adding or deleting (including soft-deleting) of leaf classes (or leaf sub-hierarchies) in a polymorphic hierarchy is fully supported. Any more complex changes, such as adding or deleting the root or an intermediate base or getting an existing class into or out of a polymorphic hierarchy can be handled by creating a new leaf class (or leaf sub-hierarchy), soft-deleting the old class, and migrating the data.


14 ODB Pragma Language

As we have already seen in previous chapters, ODB uses a pragma-based language to capture database-specific information about C++ types. This chapter describes the ODB pragma language in more detail. It can be read together with other chapters in the manual to get a sense of what kind of configurations and mapping fine-tuning are possible. You can also use this chapter as a reference at a later stage.

An ODB pragma has the following syntax:

#pragma db qualifier [specifier specifier ...]

The qualifier tells the ODB compiler what kind of C++ construct this pragma describes. Valid qualifiers are object, view, value, member, namespace, model, index, and map. A pragma with the object qualifier describes a persistent object type. It tells the ODB compiler that the C++ class it describes is a persistent class. Similarly, pragmas with the view qualifier describe view types, the value qualifier describes value types and the member qualifier is used to describe data members of persistent object, view, and value types. The namespace qualifier is used to describe common properties of objects, views, and value types that belong to a C++ namespace while the model qualifier describes the whole C++ object model. The index qualifier defines a database index. And, finally, the map qualifier describes a mapping between additional database types and types for which ODB provides built-in support.

The specifier informs the ODB compiler about a particular database-related property of the C++ declaration. For example, the id member specifier tells the ODB compiler that this member contains this object's identifier. Below is the declaration of the person class that shows how we can use ODB pragmas:

#pragma db object
class person
{
  ...
private:
  #pragma db member id
  unsigned long id_;
  ...
};
  

In the above example we don't explicitly specify which C++ class or data member the pragma belongs to. Rather, the pragma applies to a C++ declaration that immediately follows the pragma. Such pragmas are called positioned pragmas. In positioned pragmas that apply to data members, the member qualifier can be omitted for brevity, for example:

  #pragma db id
  unsigned long id_;
  

Note also that if the C++ declaration immediately following a position pragma is incompatible with the pragma qualifier, an error will be issued. For example:

  #pragma db object  // Error: expected class instead of data member.
  unsigned long id_;
  

While keeping the C++ declarations and database declarations close together eases maintenance and increases readability, we can also place them in different parts of the same header file or even factor them to a separate file. To achieve this we use the so called named pragmas. Unlike positioned pragmas, named pragmas explicitly specify the C++ declaration to which they apply by adding the declaration name after the pragma qualifier. For example:

class person
{
  ...
private:
  unsigned long id_;
  ...
};

#pragma db object(person)
#pragma db member(person::id_) id
  

Note that in the named pragmas for data members the member qualifier is no longer optional. The C++ declaration name in the named pragmas is resolved using the standard C++ name resolution rules, for example:

namespace db
{
  class person
  {
    ...
  private:
    unsigned long id_;
    ...
  };
}

namespace db
{
  #pragma db object(person)  // Resolves db::person.
}

#pragma db member(db::person::id_) id
  

As another example, the following code fragment shows how to use the named value type pragma to map a C++ type to a native database type:

#pragma db value(bool) type("INT")

#pragma db object
class person
{
  ...
private:
  bool married_; // Mapped to INT NOT NULL database type.
  ...
};
  

If we would like to factor the ODB pragmas into a separate file, we can include this file into the original header file (the one that defines the persistent types) using the #include directive, for example:

// person.hxx

class person
{
  ...
};

#ifdef ODB_COMPILER
#  include "person-pragmas.hxx"
#endif
  

Alternatively, instead of using the #include directive, we can use the --odb-epilogue option to make the pragmas known to the ODB compiler when compiling the original header file, for example:

--odb-epilogue  '#include "person-pragmas.hxx"'
  

The following sections cover the specifiers applicable to all the qualifiers mentioned above.

The C++ header file that defines our persistent classes and normally contains one or more ODB pragmas is compiled by both the ODB compiler to generate the database support code and the C++ compiler to build the application. Some C++ compilers issue warnings about pragmas that they do not recognize. There are several ways to deal with this problem which are covered at the end of this chapter in Section 14.9, "C++ Compiler Warnings".

14.1 Object Type Pragmas

A pragma with the object qualifier declares a C++ class as a persistent object type. The qualifier can be optionally followed, in any order, by one or more specifiers summarized in the table below:

Specifier Summary Section
table table name for a persistent class 14.1.1
pointer pointer type for a persistent class 14.1.2
abstract persistent class is abstract 14.1.3
readonly persistent class is read-only 14.1.4
optimistic persistent class with the optimistic concurrency model 14.1.5
no_id persistent class has no object id 14.1.6
callback database operations callback 14.1.7
schema database schema for a persistent class 14.1.8
polymorphic persistent class is polymorphic 14.1.9
session enable/disable session support for a persistent class 14.1.10
definition definition location for a persistent class 14.1.11
transient all non-virtual data members in a persistent class are transient 14.1.12
sectionable support addition of new sections in derived classes 14.1.13
deleted persistent class is soft-deleted 14.1.14
bulk enable bulk operations for a persistent class 14.1.15

14.1.1 table

The table specifier specifies the table name that should be used to store objects of the persistent class in a relational database. For example:

#pragma db object table("people")
class person
{
  ...
};
  

If the table name is not specified, the class name is used as the table name. The table name can be qualified with a database schema, for example:

#pragma db object table("census.people")
class person
{
  ...
};
  

For more information on database schemas and the format of the qualified names, refer to Section 14.1.8, "schema".

14.1.2 pointer

The pointer specifier specifies the object pointer type for the persistent class. The object pointer type is used to return, pass, and cache dynamically allocated instances of a persistent class. For example:

#pragma db object pointer(std::tr1::shared_ptr<person>)
class person
{
  ...
};
  

There are several ways to specify an object pointer with the pointer specifier. We can use a complete pointer type as shown in the example above. Alternatively, we can specify only the template name of a smart pointer in which case the ODB compiler will automatically append the class name as a template argument. The following example is therefore equivalent to the one above:

#pragma db object pointer(std::tr1::shared_ptr)
class person
{
  ...
};
  

If you would like to use the raw pointer as an object pointer, you can use * as a shortcut:

#pragma db object pointer(*) // Same as pointer(person*)
class person
{
  ...
};
  

If a pointer type is not explicitly specified, the default pointer, specified at the namespace level (Section 14.5.1, "pointer") or with the --default-pointer ODB compiler option, is used. If neither of these two mechanisms is used to specify the pointer, then the raw pointer is used by default.

For a more detailed discussion of object pointers, refer to Section 3.3, "Object and View Pointers".

14.1.3 abstract

The abstract specifier specifies that the persistent class is abstract. An instance of an abstract class cannot be stored in the database and is normally used as a base for other persistent classes. For example:

#pragma db object abstract
class person
{
  ...
};

#pragma db object
class employee: public person
{
  ...
};

#pragma db object
class contractor: public person
{
  ...
};
  

Persistent classes with pure virtual functions are automatically treated as abstract by the ODB compiler. For a more detailed discussion of persistent class inheritance, refer to Chapter 8, "Inheritance".

14.1.4 readonly

The readonly specifier specifies that the persistent class is read-only. The database state of read-only objects cannot be updated. In particular, this means that you cannot call the database::update() function (Section 3.10, "Updating Persistent Objects") for such objects. For example:

#pragma db object readonly
class person
{
  ...
};
  

Read-only and read-write objects can derive from each other without any restrictions. When a read-only object derives from a read-write object, the resulting whole object is read-only, including the part corresponding to the read-write base. On the other hand, when a read-write object derives from a read-only object, all the data members that correspond to the read-only base are treated as read-only while the rest is treated as read-write.

Note that it is also possible to declare individual data members (Section 14.4.12, "readonly") as well as composite value types (Section 14.3.6, "readonly") as read-only.

14.1.5 optimistic

The optimistic specifier specifies that the persistent class has the optimistic concurrency model. A class with the optimistic concurrency model must also specify the data member that is used to store the object version using the version pragma (Section 14.4.16, "version"). For example:

#pragma db object optimistic
class person
{
  ...

  #pragma db version
  unsigned long version_;
};
  

If a base class has the optimistic concurrency model, then all its derived classes will automatically have the optimistic concurrency model. The current implementation also requires that in any given inheritance hierarchy the object id and the version data members reside in the same class.

For a more detailed discussion of optimistic concurrency, refer to Chapter 12, "Optimistic Concurrency".

14.1.6 no_id

The no_id specifier specifies that the persistent class has no object id. For example:

#pragma db object no_id
class person
{
  ...
};
  

A persistent class without an object id has limited functionality. Such a class cannot be loaded with the database::load() or database::find() functions (Section 3.9, "Loading Persistent Objects"), updated with the database::update() function (Section 3.10, "Updating Persistent Objects"), or deleted with the database::erase() function (Section 3.11, "Deleting Persistent Objects"). To load and delete objects without ids you can use the database::query() (Chapter 4, "Querying the Database") and database::erase_query() (Section 3.11, "Deleting Persistent Objects") functions, respectively. There is no way to update such objects except by using native SQL statements (Section 3.12, "Executing Native SQL Statements").

Furthermore, persistent classes without object ids cannot have container data members nor can they be used in object relationships. Such objects are not entered into the session object cache (Section 11.1, "Object Cache") either.

To declare a persistent class with an object id, use the data member id specifier (Section 14.4.1, "id").

14.1.7 callback

The callback specifier specifies the persist class member function that should be called before and after a database operation is performed on an object of this class. For example:

#include <odb/callback.hxx>

#pragma db object callback(init)
class person
{
  ...

  void
  init (odb::callback_event, odb::database&);
};
 

The callback function has the following signature and can be overloaded for constant objects:

void
name (odb::callback_event, odb::database&);

void
name (odb::callback_event, odb::database&) const;
  

The first argument to the callback function is the event that triggered this call. The odb::callback_event enum-like type is defined in the <odb/callback.hxx> header file and has the following interface:

namespace odb
{
  struct callback_event
  {
    enum value
    {
      pre_persist,
      post_persist,
      pre_load,
      post_load,
      pre_update,
      post_update,
      pre_erase,
      post_erase
    };

    callback_event (value v);
    operator value () const;
  };
}
  

The second argument to the callback function is the database on which the operation is about to be performed or has just been performed. A callback function can be inline or virtual.

The callback function for the *_persist, *_update, and *_erase events is always called on the constant object reference while for the *_load events — always on the unrestricted reference.

If only the non-const version of the callback function is provided, then only the *_load events will be delivered. If only the const version is provided, then all the events will be delivered to this function. Finally, if both versions are provided, then the *_load events will be delivered to the non-const version while all others — to the const version. If you need to modify the object in one of the "const" events, then you can safely cast away const-ness using the const_cast operator if you know that none of the objects will be created const. Alternatively, if you cannot make this assumption, then you can declare the data members you wish to modify as mutable.

A database operations callback can be used to implement object-specific pre and post initializations, registrations, and cleanups. As an example, the following code fragment outlines an implementation of a person class that maintains the transient age data member in addition to the persistent date of birth. A callback is used to calculate the value of the former from the latter every time a person object is loaded from the database.

#include <odb/core.hxx>
#include <odb/callback.hxx>

#pragma db object callback(init)
class person
{
  ...

private:
  friend class odb::access;

  date born_;

  #pragma db transient
  unsigned short age_;

  void
  init (odb::callback_event e, odb::database&)
  {
    switch (e)
    {
    case odb::callback_event::post_load:
    {
      // Calculate age from the date of birth.
      ...
      break;
    }
    default:
      break;
    }
  }
};
 

14.1.8 schema

The schema specifier specifies a database schema that should be used for the persistent class.

In relational databases the term schema can refer to two related but ultimately different concepts. Normally it means a collection of tables, indexes, sequences, etc., that are created in the database or the actual DDL statements that create these database objects. Some database implementations support what would be more accurately called a database namespace but is also called a schema. In this sense, a schema is a separate namespace in which tables, indexes, sequences, etc., can be created. For example, two tables that have the same name can coexist in the same database if they belong to different schemas. In this section when we talk about a schema, we refer to the database namespace meaning of this term.

When schemas are in use, a database object name is qualified with a schema. For example:

CREATE TABLE accounting.employee (...)

SELECT ... FROM accounting.employee WHERE ...
  

In the above example accounting is the schema and the employee table belongs to this schema.

Not all database implementations support schemas. Some implementation that don't support schemas (for example, MySQL, SQLite) allow the use of the above syntax to specify the database name. Yet others may support several levels of qualification. For example, Microsoft SQL Server has three levels starting with the linked database server, followed by the database, and then followed by the schema: server1.company1.accounting.employee. While the actual meaning of the qualifier in a qualified name vary from one database implementation to another, here we refer to all of them collectively as a schema.

In ODB, a schema for a table of a persistent class can be specified at the class level, C++ namespace level, or the file level. To assign a schema to a specific persistent class we can use the schema specifier, for example:

#pragma db object schema("accounting")
class employee
{
  ...
};
  

If we are also assigning a table name, then we can use a shorter notation by specifying both the schema and the table name in the table specifier:

#pragma db object table("accounting.employee")
class employee
{
  ...
};
  

If we want to assign a schema to all the persistent classes in a C++ namespace, then, instead of specifying the schema for each class, we can specify it once at the C++ namespace level. For example:

#pragma db namespace schema("accounting")
namespace accounting
{
  #pragma db object
  class employee
  {
    ...
  };

  #pragma db object
  class employer
  {
    ...
  };
}
  

If we want to assign a schema to all the persistent classes in a file, then we can use the --schema ODB compiler option. For example:

odb ... --schema accounting ...
  

An alternative to this approach with the same effect is to assign a schema to the global namespace:

#pragma db namespace() schema("accounting")
  

By default schema qualifications are accumulated starting from the persistent class, continuing with the namespace hierarchy to which this class belongs, and finishing with the schema specified with the --schema option. For example:

#pragma db namespace schema("audit_db")
namespace audit
{
  #pragma db namespace schema("accounting")
  namespace accounting
  {
    #pragma db object
    class employee
    {
      ...
    };
  }
}
  

If we compile the above code fragment with the --schema server1 option, then the employee table will have the server1.audit_db.accounting.employee qualified name.

In some situations we may want to prevent such accumulation of the qualifications. To accomplish this we can use the so-called fully-qualified names, which have the empty leading name component. This is analogous to the C++ fully-qualified names in the ::accounting::employee form. For example:

#pragma db namespace schema("accounting")
namespace accounting
{
  #pragma db object schema(".hr")
  class employee
  {
    ...
  };

  #pragma db object
  class employer
  {
    ...
  };
}
  

In the above code fragment, the employee table will have the hr.employee qualified name while the employeraccounting.employer. Note also that the empty leading name component is a special ODB syntax and is not propagated to the actual database names (using a name like .hr.employee to refer to a table will most likely result in an error).

Auxiliary database objects for a persistent class, such as indexes, sequences, triggers, etc., are all created in the same schema as the class table. By default, this is also true for the container tables. However, if you need to store a container table in a different schema, then you can provide a qualified name using the table specifier, for example:

#pragma db object table("accounting.employee")
class employee
{
  ...

  #pragma db object table("operations.projects")
  std::vector<std::string> projects_;
};
  

The standard syntax for qualified names used in the schema and table specifiers as well as the view column specifier (Section 14.4.10, "column (view)") has the "name.name..." form where, as discussed above, the leading name component can be empty to denote a fully qualified name. This form, however, doesn't work if one of the name components contains periods. To support such cases the alternative form is available: "name"."name"... For example:

#pragma db object table("accounting_1.2"."employee")
class employee
{
  ...
};
  

Finally, to specify an unqualified name that contains periods we can use the following special syntax:

#pragma db object schema(."accounting_1.2") table("employee")
class employee
{
  ...
};
  

Table prefixes (Section 14.5.2, "table") can be used as an alternative to database schemas if the target database system does not support schemas.

14.1.9 polymorphic

The polymorphic specifier specifies that the persistent class is polymorphic. For more information on polymorphism support, refer to Chapter 8, "Inheritance".

14.1.10 session

The session specifier specifies whether to enable session support for the persistent class. For example:

#pragma db object session        // Enable.
class person
{
  ...
};

#pragma db object session(true)  // Enable.
class employee
{
  ...
};

#pragma db object session(false) // Disable.
class employer
{
  ...
};
  

Session support is disabled by default unless the --generate-session ODB compiler option is specified or session support is enabled at the namespace level (Section 14.5.4, "session"). For more information on sessions, refer to Chapter 11, "Session".

14.1.11 definition

The definition specifier specifies an alternative definition location for the persistent class. By default, the ODB compiler generates the database support code for a persistent class when we compile the header file that defines this class. However, if the definition specifier is used, then the ODB compiler will instead generate the database support code when we compile the header file containing this pragma.

For more information on this functionality, refer to Section 14.3.7, "definition".

14.1.12 transient

The transient specifier instructs the ODB compiler to treat all non-virtual data members in the persistent class as transient (Section 14.4.1, "transient"). This specifier is primarily useful when declaring virtual data members, as discussed in Section 14.4.13, "virtual".

14.1.13 sectionable

The sectionable specifier instructs the ODB compiler to generate support for the addition of new object sections in derived classes in a hierarchy with the optimistic concurrency model. For more information on this functionality, refer to Section 9.2, "Sections and Optimistic Concurrency".

14.1.14 deleted

The deleted specifier marks the persistent class as soft-deleted. The single required argument to this specifier is the deletion version. For more information on this functionality, refer to Section 13.4, "Soft Object Model Changes".

14.1.15 bulk

The bulk specifier enables bulk operation support for the persistent class. The single required argument to this specifier is the batch size. For more information on this functionality, refer to Section 15.3, "Bulk Database Operations".

14.2 View Type Pragmas

A pragma with the view qualifier declares a C++ class as a view type. The qualifier can be optionally followed, in any order, by one or more specifiers summarized in the table below:

Specifier Summary Section
object object associated with a view 14.2.1
table table associated with a view 14.2.2
query view query condition 14.2.3
pointer pointer type for a view 14.2.4
callback database operations callback 14.2.5
definition definition location for a view 14.2.6
transient all non-virtual data members in a view are transient 14.2.7

For more information on view types refer to Chapter 10, "Views".

14.2.1 object

The object specifier specifies a persistent class that should be associated with the view. For more information on object associations refer to Section 10.1, "Object Views".

14.2.2 table

The table specifier specifies a database table that should be associated with the view. For more information on table associations refer to Section 10.3, "Table Views".

14.2.3 query

The query specifier specifies a query condition and, optionally, result modifiers for an object or table view or a native SQL query for a native view. An empty query specifier indicates that a native SQL query is provided at runtime. For more information on query conditions refer to Section 10.5, "View Query Conditions". For more information on native SQL queries, refer to Section 10.6, "Native Views".

14.2.4 pointer

The pointer specifier specifies the view pointer type for the view class. Similar to objects, the view pointer type is used to return dynamically allocated instances of a view class. The semantics of the pointer specifier for a view are the same as those of the pointer specifier for an object (Section 14.1.2, "pointer").

14.2.5 callback

The callback specifier specifies the view class member function that should be called before and after an instance of this view class is created as part of the query result iteration. The semantics of the callback specifier for a view are similar to those of the callback specifier for an object (Section 14.1.7, "callback") except that the only events that can trigger a callback call in the case of a view are pre_load and post_load.

14.2.6 definition

The definition specifier specifies an alternative definition location for the view class. By default, the ODB compiler generates the database support code for a view class when we compile the header file that defines this class. However, if the definition specifier is used, then the ODB compiler will instead generate the database support code when we compile the header file containing this pragma.

For more information on this functionality, refer to Section 14.3.7, "definition".

14.2.7 transient

The transient specifier instructs the ODB compiler to treat all non-virtual data members in the view class as transient (Section 14.4.1, "transient"). This specifier is primarily useful when declaring virtual data members, as discussed in Section 14.4.13, "virtual".

14.3 Value Type Pragmas

A pragma with the value qualifier describes a value type. It can be optionally followed, in any order, by one or more specifiers summarized in the table below:

Specifier Summary Section
type database type for a value type 14.3.1
id_type database type for a value type when used as an object id 14.3.2
null/not_null type can/cannot be NULL 14.3.3
default default value for a value type 14.3.4
options database options for a value type 14.3.5
readonly composite value type is read-only 14.3.6
definition definition location for a composite value type 14.3.7
transient all non-virtual data members in a composite value are transient 14.3.8
unordered ordered container should be stored unordered 14.3.9
index_type database type for a container's index type 14.3.10
key_type database type for a container's key type 14.3.11
value_type database type for a container's value type 14.3.12
value_null/value_not_null container's value can/cannot be NULL 14.3.13
id_options database options for a container's id column 14.3.14
index_options database options for a container's index column 14.3.15
key_options database options for a container's key column 14.3.16
value_options database options for a container's value column 14.3.17
id_column column name for a container's object id 14.3.18
index_column column name for a container's index 14.3.19
key_column column name for a container's key 14.3.20
value_column column name for a container's value 14.3.21

Many of the value type specifiers have corresponding member type specifiers with the same names (Section 14.4, "Data Member Pragmas"). The behavior of such specifiers for members is similar to that for value types. The only difference is the scope. A particular value type specifier applies to all the members of this value type that don't have a pre-member version of the specifier, while the member specifier always applies only to a single member. Also, with a few exceptions, member specifiers take precedence over and override parameters specified with value specifiers.

14.3.1 type

The type specifier specifies the native database type that should be used for data members of this type. For example:

#pragma db value(bool) type("INT")

#pragma db object
class person
{
  ...

  bool married_; // Mapped to INT NOT NULL database type.
};
  

The ODB compiler provides the default mapping between common C++ types, such as bool, int, and std::string and the database types for each supported database system. For more information on the default mapping, refer to Part II, "Database Systems". The null and not_null (Section 14.3.3, "null/not_null") specifiers can be used to control the NULL semantics of a type.

In the above example we changed the mapping for the bool type which is now mapped to the INT database type. In this case, the value pragma is all that is necessary since the ODB compiler will be able to figure out how to store a boolean value as an integer in the database. However, there could be situations where the ODB compiler will not know how to handle the conversion between the C++ and database representations of a value. Consider, as an example, a situation where the boolean value is stored in the database as a string:

#pragma db value(bool) type("VARCHAR(5)")
  

The possible database values for the C++ true value could be "true", or "TRUE", or "True". Or, maybe, all of the above could be valid. The ODB compiler has no way of knowing how your application wants to convert bool to a string and back. To support such custom value type mappings, ODB allows you to provide your own database conversion functions by specializing the value_traits class template. The mapping example in the odb-examples package shows how to do this for all the supported database systems.

14.3.2 id_type

The id_type specifier specifies the native database type that should be used for data members of this type that are designated as object identifiers (Section 14.4.1, "id"). In combination with the type specifier (Section 14.3.1, "type") id_type allows you to map a C++ type differently depending on whether it is used in an ordinary member or an object id. For example:

#pragma db value(std::string) type("TEXT") id_type("VARCHAR(128)")

#pragma db object
class person
{
  ...

  #pragma db id
  std::string email_; // Mapped to VARCHAR(128) NOT NULL.

  std::string name_;  // Mapped to TEXT NOT NULL.
};
  

Note that there is no corresponding member type specifier for id_type since the desired result can be achieved with just the type specifier, for example:

#pragma db object
class person
{
  ...

  #pragma db id type("VARCHAR(128)")
  std::string email_;
};
  

14.3.3 null/not_null

The null and not_null specifiers specify that a value type or object pointer can or cannot be NULL, respectively. By default, value types are assumed not to allow NULL values while object pointers are assumed to allow NULL values. Data members of types that allow NULL values are mapped in a relational database to columns that allow NULL values. For example:

using std::tr1::shared_ptr;

typedef shared_ptr<std::string> string_ptr;
#pragma db value(string_ptr) type("TEXT") null

#pragma db object
class person
{
  ...

  string_ptr name_; // Mapped to TEXT NULL.
};

typedef shared_ptr<person> person_ptr;
#pragma db value(person_ptr) not_null
  

The NULL semantics can also be specified on the per-member basis (Section 14.4.6, "null/not_null"). If both a type and a member have null/not_null specifiers, then the member specifier takes precedence. If a member specifier relaxes the NULL semantics (that is, if a member has the null specifier and the type has the explicit not_null specifier), then a warning is issued.

It is also possible to override a previously specified null/not_null specifier. This is primarily useful if a third-party type, for example, one provided by a profile library (Part III, "Profiles"), allows NULL values but in your object model data members of this type should never be NULL. In this case you can use the not_null specifier to disable NULL values for this type for the entire translation unit. For example:

// By default, null_string allows NULL values.
//
#include <null-string.hxx>

// Disable NULL values for all the null_string data members.
//
#pragma db value(null_string) not_null
  

For a more detailed discussion of the NULL semantics for values, refer to Section 7.3, "Pointers and NULL Value Semantics". For a more detailed discussion of the NULL semantics for object pointers, refer to Chapter 6, "Relationships".

14.3.4 default

The default specifier specifies the database default value that should be used for data members of this type. For example:

#pragma db value(std::string) default("")

#pragma db object
class person
{
  ...

  std::string name_; // Mapped to TEXT NOT NULL DEFAULT ''.
};
  

The semantics of the default specifier for a value type are similar to those of the default specifier for a data member (Section 14.4.7, "default").

14.3.5 options

The options specifier specifies additional column definition options that should be used for data members of this type. For example:

#pragma db value(std::string) options("COLLATE binary")

#pragma db object
class person
{
  ...

  std::string name_; // Mapped to TEXT NOT NULL COLLATE binary.
};
  

The semantics of the options specifier for a value type are similar to those of the options specifier for a data member (Section 14.4.8, "options").

14.3.6 readonly

The readonly specifier specifies that the composite value type is read-only. Changes to data members of a read-only composite value type are ignored when updating the database state of an object (Section 3.10, "Updating Persistent Objects") containing such a value type. Note that this specifier is only valid for composite value types. For example:

#pragma db value readonly
class person_name
{
  ...
};
  

Read-only and read-write composite values can derive from each other without any restrictions. When a read-only value derives from a read-write value, the resulting whole value is read-only, including the part corresponding to the read-write base. On the other hand, when a read-write value derives from a read-only value, all the data members that correspond to the read-only base are treated as read-only while the rest is treated as read-write.

Note that it is also possible to declare individual data members (Section 14.4.12, "readonly") as well as whole objects (Section 14.1.4, "readonly") as read-only.

14.3.7 definition

The definition specifier specifies an alternative definition location for the composite value type. By default, the ODB compiler generates the database support code for a composite value type when we compile the header file that defines this value type. However, if the definition specifier is used, then the ODB compiler will instead generate the database support code when we compile the header file containing this pragma.

This mechanism is primarily useful for converting third-party types to ODB composite value types. In such cases we normally cannot modify the header files to add the necessary pragmas. It is also often inconvenient to compile these header files with the ODB compiler. With the definition specifier we can create a wrapper header that contains the necessary pragmas and instructs the ODB compiler to generate the database support code for a third-party type when we compile the wrapper header. As an example, consider struct timeval that is defined in the <sys/time.h> system header. This type has the following (or similar) definition:

struct timeval
{
  long tv_sec;
  long tv_usec;
};
  

If we would like to make this type an ODB composite value type, then we can create a wrapper header, for example time-mapping.hxx, with the following content:

#ifndef TIME_MAPPING_HXX
#define TIME_MAPPING_HXX

#include <sys/time.h>

#pragma db value(timeval) definition
#pragma db member(timeval::tv_sec) column("sec")
#pragma db member(timeval::tv_usec) column("usec")

#endif // TIME_MAPPING_HXX
  

If we now compile this header with the ODB compiler, the resulting time-mapping-odb.?xx files will contain the database support code for struct timeval. To use timeval in our persistent classes, we simply include the time-mapping.hxx header:

#include <sys/time.h>
#include "time-mapping.hxx"

#pragma db object
class object
{
  timeval timestamp;
};
  

14.3.8 transient

The transient specifier instructs the ODB compiler to treat all non-virtual data members in the composite value type as transient (Section 14.4.1, "transient"). This specifier is primarily useful when declaring virtual data members, as discussed in Section 14.4.13, "virtual".

14.3.9 unordered

The unordered specifier specifies that the ordered container should be stored unordered in the database. The database table for such a container will not contain the index column and the order in which elements are retrieved from the database may not be the same as the order in which they were stored. For example:

typedef std::vector<std::string> names;
#pragma db value(names) unordered
  

For a more detailed discussion of ordered containers and their storage in the database, refer to Section 5.1, "Ordered Containers".

14.3.10 index_type

The index_type specifier specifies the native database type that should be used for the ordered container's index column. The semantics of index_type are similar to those of the type specifier (Section 14.3.1, "type"). The native database type is expected to be an integer type. For example:

typedef std::vector<std::string> names;
#pragma db value(names) index_type("SMALLINT UNSIGNED")
  

14.3.11 key_type

The key_type specifier specifies the native database type that should be used for the map container's key column. The semantics of key_type are similar to those of the type specifier (Section 14.3.1, "type"). For example:

typedef std::map<unsigned short, float> age_weight_map;
#pragma db value(age_weight_map) key_type("INT UNSIGNED")
  

14.3.12 value_type

The value_type specifier specifies the native database type that should be used for the container's value column. The semantics of value_type are similar to those of the type specifier (Section 14.3.1, "type"). For example:

typedef std::vector<std::string> names;
#pragma db value(names) value_type("VARCHAR(255)")
  

The value_null and value_not_null (Section 14.3.13, "value_null/value_not_null") specifiers can be used to control the NULL semantics of a value column.

14.3.13 value_null/value_not_null

The value_null and value_not_null specifiers specify that the container type's element value can or cannot be NULL, respectively. The semantics of value_null and value_not_null are similar to those of the null and not_null specifiers (Section 14.3.3, "null/not_null"). For example:

using std::tr1::shared_ptr;

#pragma db object
class account
{
  ...
};

typedef std::vector<shared_ptr<account> > accounts;
#pragma db value(accounts) value_not_null
  

For set and multiset containers (Section 5.2, "Set and Multiset Containers") the element value is automatically treated as not allowing a NULL value.

14.3.14 id_options

The id_options specifier specifies additional column definition options that should be used for the container's id column. For example:

typedef std::vector<std::string> nicknames;
#pragma db value(nicknames) id_options("COLLATE binary")
  

The semantics of the id_options specifier for a container type are similar to those of the id_options specifier for a container data member (Section 14.4.29, "id_options").

14.3.15 index_options

The index_options specifier specifies additional column definition options that should be used for the container's index column. For example:

typedef std::vector<std::string> nicknames;
#pragma db value(nicknames) index_options("ZEROFILL")
  

The semantics of the index_options specifier for a container type are similar to those of the index_options specifier for a container data member (Section 14.4.30, "index_options").

14.3.16 key_options

The key_options specifier specifies additional column definition options that should be used for the container's key column. For example:

typedef std::map<std::string, std::string> properties;
#pragma db value(properties) key_options("COLLATE binary")
  

The semantics of the key_options specifier for a container type are similar to those of the key_options specifier for a container data member (Section 14.4.31, "key_options").

14.3.17 value_options

The value_options specifier specifies additional column definition options that should be used for the container's value column. For example:

typedef std::set<std::string> nicknames;
#pragma db value(nicknames) value_options("COLLATE binary")
  

The semantics of the value_options specifier for a container type are similar to those of the value_options specifier for a container data member (Section 14.4.32, "value_options").

14.3.18 id_column

The id_column specifier specifies the column name that should be used to store the object id in the container's table. For example:

typedef std::vector<std::string> names;
#pragma db value(names) id_column("id")
  

If the column name is not specified, then object_id is used by default.

14.3.19 index_column

The index_column specifier specifies the column name that should be used to store the element index in the ordered container's table. For example:

typedef std::vector<std::string> names;
#pragma db value(names) index_column("name_number")
  

If the column name is not specified, then index is used by default.

14.3.20 key_column

The key_column specifier specifies the column name that should be used to store the key in the map container's table. For example:

typedef std::map<unsigned short, float> age_weight_map;
#pragma db value(age_weight_map) key_column("age")
  

If the column name is not specified, then key is used by default.

14.3.21 value_column

The value_column specifier specifies the column name that should be used to store the element value in the container's table. For example:

typedef std::map<unsigned short, float> age_weight_map;
#pragma db value(age_weight_map) value_column("weight")
  

If the column name is not specified, then value is used by default.

14.4 Data Member Pragmas

A pragma with the member qualifier or a positioned pragma without a qualifier describes a data member. It can be optionally followed, in any order, by one or more specifiers summarized in the table below:

Specifier Summary Section
id member is an object id 14.4.1
auto id is assigned by the database 14.4.2
type database type for a member 14.4.3
id_type database type for a member when used as an object id 14.4.4
get/set/access member accessor/modifier expressions 14.4.5
null/not_null member can/cannot be NULL 14.4.6
default default value for a member 14.4.7
options database options for a member 14.4.8
column column name for a member of an object or composite value 14.4.9
column column name for a member of a view 14.4.10
transient member is not stored in the database 14.4.11
readonly member is read-only 14.4.12
virtual declare a virtual data member 14.4.13
inverse member is an inverse side of a bidirectional relationship 14.4.14
on_delete ON DELETE clause for object pointer member 14.4.15
version member stores object version 14.4.16
index define database index for a member 14.4.17
unique define unique database index for a member 14.4.18
unordered ordered container should be stored unordered 14.4.19
table table name for a container 14.4.20
load/update loading/updating behavior for a section 14.4.21
section member belongs to a section 14.4.22
added member is soft-added 14.4.23
deleted member is soft-deleted 14.4.24
index_type database type for a container's index type 14.4.25
key_type database type for a container's key type 14.4.26
value_type database type for a container's value type 14.4.27
value_null/value_not_null container's value can/cannot be NULL 14.4.28
id_options database options for a container's id column 14.4.29
index_options database options for a container's index column 14.4.30
key_options database options for a container's key column 14.4.31
value_options database options for a container's value column 14.4.32
id_column column name for a container's object id 14.4.33
index_column column name for a container's index 14.4.34
key_column column name for a container's key 14.4.35
value_column column name for a container's value 14.4.36

Many of the member specifiers have corresponding value type specifiers with the same names (Section 14.3, "Value Type Pragmas"). The behavior of such specifiers for members is similar to that for value types. The only difference is the scope. A particular value type specifier applies to all the members of this value type that don't have a pre-member version of the specifier, while the member specifier always applies only to a single member. Also, with a few exceptions, member specifiers take precedence over and override parameters specified with value specifiers.

14.4.1 id

The id specifier specifies that the data member contains the object id. In a relational database, an identifier member is mapped to a primary key. For example:

#pragma db object
class person
{
  ...

  #pragma db id
  std::string email_;
};
  

Normally, every persistent class has a data member designated as an object's identifier. However, it is possible to declare a persistent class without an id using the object no_id specifier (Section 14.1.6, "no_id").

Note also that the id specifier cannot be used for data members of composite value types or views.

14.4.2 auto

The auto specifier specifies that the object's identifier is automatically assigned by the database. Only a member that was designated as an object id can have this specifier. For example:

#pragma db object
class person
{
  ...

  #pragma db id auto
  unsigned long id_;
};
  

Note that automatically-assigned object ids are not reused. If you have a high object turnover (that is, objects are routinely made persistent and then erased), then care must be taken not to run out of object ids. In such situations, using unsigned long long as the identifier type is a safe choice.

For additional information on the automatic identifier assignment, refer to Section 3.8, "Making Objects Persistent".

Note also that the auto specifier cannot be specified for data members of composite value types or views.

14.4.3 type

The type specifier specifies the native database type that should be used for the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db type("INT")
  bool married_;
};
  

The null and not_null (Section 14.4.6, "null/not_null") specifiers can be used to control the NULL semantics of a data member. It is also possible to specify the database type on the per-type instead of the per-member basis using the value type specifier (Section 14.3.1, "type").

14.4.4 id_type

The id_type specifier specifies the native database type that should be used for the data member when it is part of an object identifier. This specifier only makes sense when applied to a member of a composite value type that is used for both id and non-id members. For example:

#pragma db value
class name
{
  ...

  #pragma db type("VARCHAR(256)") id_type("VARCHAR(64)")
  std::string first_;

  #pragma db type("VARCHAR(256)") id_type("VARCHAR(64)")
  std::string last_;
};

#pragma db object
class person
{
  ...

  #pragma db id
  name name_;  // name_.first_, name_.last_ mapped to VARCHAR(64)

  name alias_; // alias_.first_, alias_.last_ mapped to VARCHAR(256)
};
  

14.4.5 get/set/access

The get and set specifiers specify the data member accessor and modifier expressions, respectively. If provided, the generated database support code will use these expressions to access and modify the data member when performing database operations. The access specifier can be used as a shortcut to specify both the accessor and modifier if they happen to be the same.

In its simplest form the accessor or modifier expression can be just a name. Such a name should resolve either to another data member of the same type or to a suitable accessor or modifier member function. For example:

#pragma db object
class person
{
  ...

public:
  const std::string& name () const;
  void name (const std::string&);
private:
  #pragma db access(name)
  std::string name_;
};
  

A suitable accessor function is a const member function that takes no arguments and whose return value can be implicitly converted to the const reference to the member type (const std::string& in the example above). An accessor function that returns a const reference to the data member is called by-reference accessor. Otherwise, it is called by-value accessor.

A suitable modifier function can be of two forms. It can be the so called by-reference modifier which is a member function that takes no arguments and returns a non-const reference to the data member (std::string& in the example above). Alternatively, it can be the so called by-value modifier which is a member function taking a single argument — the new value — that can be implicitly initialized from a variable of the member type (std::string in the example above). The return value of a by-value modifier, if any, is ignored. If both by-reference and by-value modifiers are available, then ODB prefers the by-reference version since it is more efficient. For example:

#pragma db object
class person
{
  ...

public:
  std::string get_name () const;      // By-value accessor.
  std::string& set_name ();           // By-reference modifier.
  void set_name (std::string const&); // By-value modifier.
private:
  #pragma db get(get_name) \ // Uses by-value accessor.
             set(set_name)   // Uses by-reference modifier.
  std::string name_;
};
  

Note that in many cases it is not necessary to specify accessor and modifier functions explicitly since the ODB compiler will try to discover them automatically in case the data member will be inaccessible to the generated code. In particular, in both of the above examples the ODB compiler would have successfully discovered the necessary functions. For more information on this functionality, refer to Section 3.2, "Declaring Persistent Objects and Values".

Note also that by-value accessors and by-value modifiers cannot be used for certain data members in certain situations. These limitations are discussed in more detail later in this section.

Accessor and modifier expressions can be more elaborate than simple names. An accessor expression is any C++ expression that can be used to initialize a const reference to the member type. Similar to accessor functions, which are just a special case of accessor expressions, an accessor expression that evaluates to a const reference to the data member is called by-reference accessor expression. Otherwise, it is called by-value accessor expression.

Modifier expressions can also be of two forms: by-reference modifier expression and by-value modifier expression (again, modifier functions are just a special case of modifier expressions). A by-reference modifier expression is any C++ expression that evaluates to the non-const reference to the member type. A by-value modifier expression can be a single or multiple (separated by semicolon) C++ statements with the effect of setting the new member value.

There are two special placeholders that are recognized by the ODB compiler in accessor and modifier expressions. The first is the this keyword which denotes a reference (note: not a pointer) to the persistent object. In accessor expressions this reference is const while in modifier expressions it is non-const. If an expression does not contain the this placeholder, then the ODB compiler automatically prefixes it with this. sequence.

The second placeholder, the (?) sequence, is used to denote the new value in by-value modifier expressions. The ODB compiler replaces the question mark with the variable name, keeping the surrounding parenthesis. The following example shows a few more interesting accessor and modifier expressions:

#pragma db value
struct point
{
  point (int, int);

  int x;
  int y;
};

#pragma db object
class person
{
  ...

  public:
    const char* name () const;
    void name (const char*);
  private:
    #pragma db get(std::string (this.name ())) \
               set(name ((?).c_str ())) // The same as this.name (...).
    std::string name_;

  public:
    const std::unique_ptr<account>& acc () const;
    void acc (std::unique_ptr<account>);
  private:
    #pragma db set(acc (std::move (?)))
    std::unique_ptr<account> acc_;

  public:
    int loc_x () const
    int loc_y () const
    void loc_x (int);
    void loc_y (int);
  private:
    #pragma db get(point (this.loc_x (), this.loc_y ()))    \
               set(this.loc_x ((?).x); this.loc_y ((?).y))
    point loc_;
};
  

When the data member is of an array type, then the terms "reference" and "member type" in the above discussion should be replaced with "pointer" and "array element type", respectively. That is, the accessor expression for an array member is any C++ expression that can be used to initialize a const pointer to the array element type, and so on. The following example shows common accessor and modifier signatures for array members:

#pragma db object
class person
{
  ...

  public:
    const char* id () const; // By-reference accessor.
    void id (const char*);   // By-value modifier.
  private:
    char id_[16];

  public:
    const char* pub_key () const; // By-reference accessor.
    char* pub_key ();             // By-reference modifier.
  private:
    char pub_key_[2048];
};
  

Accessor and modifier expressions can be used with data members of simple value, composite value, container, and object pointer types. They can be used for data members in persistent classes, composite value types, and views. There is also a mechanism related to accessors and modifiers called virtual data members and which is discussed in Section 14.4.13, "virtual".

There are, however, certain limitations when it comes to using by-value accessor and modifier expressions. First of all, if a by-value modifier is used, then the data member type should be default-constructible. Furthermore, a composite value type that has a container member cannot be modified with a by-value modifier. Only a by-reference modifier expression can be used. The ODB compiler will detect such cases and issue diagnostics. For example:

#pragma db value
struct name
{
  std::string first_;
  std::string last_;
  std::vector<std::string> aliases_;
};

#pragma db object
class person
{
  ...

public:
  const name& name () const;
  void name (const name&);
private:
  #pragma db access(name) // Error: by-value modifier.
  name name_;
};
  

In certain database systems it is also not possible to use by-value accessor and modifier expression with certain database types. The ODB compiler is only able to detect such cases and issue diagnostics if you specified accessor/modifier function names as opposed to custom expressions. For more information on these database and type-specific limitations, refer to the "Limitations" sections in Part II, "Database Systems".

14.4.6 null/not_null

The null and not_null specifiers specify that the data member can or cannot be NULL, respectively. By default, data members of basic value types for which database mapping is provided by the ODB compiler do not allow NULL values while data members of object pointers allow NULL values. Other value types, such as those provided by the profile libraries (Part III, "Profiles"), may or may not allow NULL values, depending on the semantics of each value type. Consult the relevant documentation to find out more about the NULL semantics for such value types. A data member containing the object id (Section 14.4.1, "id") is automatically treated as not allowing a NULL value. Data members that allow NULL values are mapped in a relational database to columns that allow NULL values. For example:

using std::tr1::shared_ptr;

#pragma db object
class person
{
  ...

  #pragma db null
  std::string name_;
};

#pragma db object
class account
{
  ...

  #pragma db not_null
  shared_ptr<person> holder_;
};
  

The NULL semantics can also be specified on the per-type basis (Section 14.3.3, "null/not_null"). If both a type and a member have null/not_null specifiers, then the member specifier takes precedence. If a member specifier relaxes the NULL semantics (that is, if a member has the null specifier and the type has the explicit not_null specifier), then a warning is issued.

For a more detailed discussion of the NULL semantics for values, refer to Section 7.3, "Pointers and NULL Value Semantics". For a more detailed discussion of the NULL semantics for object pointers, refer to Chapter 6, "Relationships".

14.4.7 default

The default specifier specifies the database default value that should be used for the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db default(-1)
  int age_;              // Mapped to INT NOT NULL DEFAULT -1.
};
  

A default value can be the special null keyword, a bool literal (true or false), an integer literal, a floating point literal, a string literal, or an enumerator name. If you need to specify a default value that is an expression, for example an SQL function call, then you can use the options specifier (Section 14.4.8, "options") instead. For example:

enum gender {male, female, undisclosed};

#pragma db object
class person
{
  ...

  #pragma db default(null)
  odb::nullable<std::string> middle_; // DEFAULT NULL

  #pragma db default(false)
  bool married_;                      // DEFAULT 0/FALSE

  #pragma db default(0.0)
  float weight_;                      // DEFAULT 0.0

  #pragma db default("Mr")
  string title_;                      // DEFAULT 'Mr'

  #pragma db default(undisclosed)
  gender gender_;                     // DEFAULT 2/'undisclosed'

  #pragma db options("DEFAULT CURRENT_TIMESTAMP()")
  date timestamp_;                    // DEFAULT CURRENT_TIMESTAMP()
};
  

Default values specified as enumerators are only supported for members that are mapped to an ENUM or an integer type in the database, which is the case for the automatic mapping of C++ enums and enum classes to suitable database types as performed by the ODB compiler. If you have mapped a C++ enum or enum class to another database type, then you should use a literal corresponding to that type to specify the default value. For example:

enum gender {male, female, undisclosed};
#pragma db value(gender) type("VARCHAR(11)")

#pragma db object
class person
{
  ...

  #pragma db default("undisclosed")
  gender gender_;                   // DEFAULT 'undisclosed'
};
  

A default value can also be specified on the per-type basis (Section 14.3.4, "default"). An empty default specifier can be used to reset a default value that was previously specified on the per-type basis. For example:

#pragma db value(std::string) default("")

#pragma db object
class person
{
  ...

  #pragma db default()
  std::string name_;   // No default value.
};
  

A data member containing the object id (Section 14.4.1, "id" ) is automatically treated as not having a default value even if its type specifies a default value.

Note also that default values do not affect the generated C++ code in any way. In particular, no automatic initialization of data members with their default values is performed at any point. If you need such an initialization, you will need to implement it yourself, for example, in your persistent class constructors. The default values only affect the generated database schemas and, in the context of ODB, are primarily useful for schema evolution.

Additionally, the default specifier cannot be specified for view data members.

14.4.8 options

The options specifier specifies additional column definition options that should be used for the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db options("CHECK(email != '')")
  std::string email_; // Mapped to TEXT NOT NULL CHECK(email != '').
};
  

Options can also be specified on the per-type basis (Section 14.3.5, "options"). By default, options are accumulating. That is, the ODB compiler first adds all the options specified for a value type followed by all the options specified for a data member. To clear the accumulated options at any point in this sequence you can use an empty options specifier. For example:

#pragma db value(std::string) options("COLLATE binary")

#pragma db object
class person
{
  ...

  std::string first_; // TEXT NOT NULL COLLATE binary

  #pragma db options("CHECK(email != '')")
  std::string last_;  // TEXT NOT NULL COLLATE binary CHECK(email != '')

  #pragma db options()
  std::string title_; // TEXT NOT NULL

  #pragma db options() options("CHECK(email != '')")
  std::string email_; // TEXT NOT NULL CHECK(email != '')
};
  

ODB provides dedicated specifiers for specifying column types (Section 14.4.3, "type"), NULL constraints (Section 14.4.6, "null/not_null"), and default values (Section 14.4.7, "default"). For ODB to function correctly these specifiers should always be used instead of the opaque options specifier for these components of a column definition.

Note also that the options specifier cannot be specified for view data members.

14.4.9 column (object, composite value)

The column specifier specifies the column name that should be used to store the data member of a persistent class or composite value type in a relational database. For example:

#pragma db object
class person
{
  ...

  #pragma db id column("person_id")
  unsigned long id_;
};
  

For a member of a composite value type, the column specifier specifies the column name prefix. Refer to Section 7.2.2, "Composite Value Column and Table Names" for details.

If the column name is not specified, it is derived from the member's so-called public name. A public member name is obtained by removing the common data member name decorations, such as leading and trailing underscores, the m_ prefix, etc.

14.4.10 column (view)

The column specifier can be used to specify the associated object data member, the potentially qualified column name, or the column expression for the data member of a view class. For more information, refer to Section 10.1, "Object Views" and Section 10.3, "Table Views".

14.4.11 transient

The transient specifier instructs the ODB compiler not to store the data member in the database. For example:

#pragma db object
class person
{
  ...

  date born_;

  #pragma db transient
  unsigned short age_; // Computed from born_.
};
  

This pragma is usually used on computed members, pointers and references that are only meaningful in the application's memory, as well as utility members such as mutexes, etc.

14.4.12 readonly

The readonly specifier specifies that the data member of an object or composite value type is read-only. Changes to a read-only data member are ignored when updating the database state of an object (Section 3.10, "Updating Persistent Objects") containing such a member. Since views are read-only, it is not necessary to use this specifier for view data members. Object id (Section 14.4.1, "id") and inverse (Section 14.4.14, "inverse") data members are automatically treated as read-only and must not be explicitly declared as such. For example:

#pragma db object
class person
{
  ...

  #pragma db readonly
  date born_;
};
  

Besides simple value members, object pointer, container, and composite value members can also be declared read-only. A change of a pointed-to object is ignored when updating the state of a read-only object pointer. Similarly, any changes to the number or order of elements or to the element values themselves are ignored when updating the state of a read-only container. Finally, any changes to the members of a read-only composite value type are also ignored when updating the state of such a composite value.

ODB automatically treats const data members as read-only. For example, the following person object is equivalent to the above declaration for the database persistence purposes:

#pragma db object
class person
{
  ...

  const date born_; // Automatically read-only.
};
  

When declaring an object pointer const, make sure to declare the pointer as const rather than (or in addition to) the object itself. For example:

#pragma db object
class person
{
  ...

  const person* father_; // Read-write pointer to a read-only object.
  person* const mother_; // Read-only pointer to a read-write object.
};
  

Note that in case of a wrapper type (Section 7.3, "Pointers and NULL Value Semantics"), both the wrapper and the wrapped type must be const in order for the ODB compiler to automatically treat the data member as read-only. For example:

#pragma db object
class person
{
  ...

  const std::auto_ptr<const date> born_;
};
  

Read-only members are useful when dealing with asynchronous changes to the state of a data member in the database which should not be overwritten. In other cases, where the state of a data member never changes, declaring such a member read-only allows ODB to perform more efficient object updates. In such cases, however, it is conceptually more correct to declare such a data member as const rather than as read-only.

Note that it is also possible to declare composite value types (Section 14.3.6, "readonly") as well as whole objects (Section 14.1.4, "readonly") as read-only.

14.4.13 virtual

The virtual specifier is used to declare a virtual data member in an object, view, or composite value type. A virtual data member is an imaginary data member that is only used for the purpose of database persistence. A virtual data member does not actually exist (that is, occupy space) in the C++ class. Note also that virtual data members have nothing to do with C++ virtual functions or virtual inheritance. Specifically, no virtual function call overhead is incurred when using virtual data members.

To declare a virtual data member we must specify the data member name using the member specifier. We must also specify the data member type with the virtual specifier. Finally, the virtual data member declaration must also specify the accessor and modifier expressions, unless suitable accessor and modifier functions can automatically be found by the ODB compiler (Section 14.4.5, "get/set/access"). For example:

#pragma db object
class person
{
  ...

  // Transient real data member that actually stores the data.
  //
  #pragma db transient
  std::string name_;

  // Virtual data member.
  //
  #pragma db member(name) virtual(std::string) access(name_)
};
  

From the pragma language point of view, a virtual data member behaves exactly like a normal data member. Specifically, we can reference the virtual data member after it has been declared and use positioned pragmas before its declaration. For example:

#pragma db object
class person
{
  ...

  #pragma db transient
  std::string name_;

  #pragma db access(name_)
  #pragma db member(name) virtual(std::string)
};

#pragma db member(person::name) column("person_name")
#pragma db index member(person::name)
  

We can also declare a virtual data member outside the class scope:

#pragma db object
class person
{
  ...

  std::string name_;
};

#pragma db member(person::name_) transient
#pragma db member(person::name) virtual(std::string) access(name_)
  

While in the above examples using virtual data members doesn't seem to yield any benefits, this mechanism can be useful in a number of situations. As one example, consider the need to aggregate or dis-aggregate a data member:

#pragma db object
class person
{
  ...

  #pragma db transient
  std::pair<std::string, std::string> name_;

  #pragma db member(first) virtual(std::string) access(name_.first)
  #pragma db member(last) virtual(std::string) access(name_.second)
};
  

We can also use virtual data members to implement composite object ids that are spread over multiple data members:

#pragma db value
struct name
{
  name () {}
  name (std::string const& f, std::string const& l)
    : first (f), last(l) {}

  std::string first;
  std::string last;
};

#pragma db object
class person
{
  ...

  #pragma db transient
  std::string first_;

  #pragma db transient
  std::string last_;

  #pragma db member(name) virtual(name) id                       \
             get(::name (this.first_, this.last_))               \
             set(this.first_ = (?).first; this.last_ = (?).last)
};
  

Another common situation that calls for virtual data members is a class that uses the pimpl idiom. While the following code fragment outlines the idea, for details refer to the pimpl example in the odb-examples package.

#pragma db object
class person
{
public:
  std::string const& name () const;
  void name (std::string const&);

  unsigned short age () const;
  void age (unsigned short);

  ...

private:
  class impl;

  #pragma db transient
  impl* pimpl_;

  #pragma db member(name) virtual(std::string)   // Uses name().
  #pragma db member(age) virtual(unsigned short) // Uses age().
};
  

The above example also shows that names used for virtual data members (name and age in our case) can be the same as the names of accessor/modifier functions. The only names that virtual data members cannot clash with are those of other data members, virtual or real.

A common pattern in the above examples is the need to declare the real data member that actually stores the data as transient. If all the real data members in a class are treated as transient, then we can use the class-level transient specifier (Section 14.1.12, "transient (object)", Section 14.3.8, "transient (composite value)", Section 14.2.7, "transient (view)") instead of doing it for each individual member. For example:

#pragma db object transient
class person
{
  ...

  std::string first_; // Transient.
  std::string last_;  // Transient.

  #pragma db member(name) virtual(name) ...
};
  

The ability to treat all the real data members as transient becomes more important if we don't know the names of these data members. This is often the case when we are working with third-party types that document the accessor and modifier functions but not the names of their private data members. As an example, consider the point class defined in a third-party <point> header file:

class point
{
public:
  point ();
  point (int x, int y);

  int x () const;
  int y () const;

  void x (int);
  void y (int);

private:
  ...
};
  

To convert this class to an ODB composite value type we could create the point-mapping.hxx file with the following content:

#include <point>

#pragma db value(point) transient definition
#pragma db member(point::x) virtual(int)
#pragma db member(point::y) virtual(int)
  

Virtual data members can be of simple value, composite value, container, or object pointer types. They can be used in persistent classes, composite value types, and views.

14.4.14 inverse

The inverse specifier specifies that the data member of an object pointer or a container of object pointers type is an inverse side of a bidirectional object relationship. The single required argument to this specifier is the corresponding data member name in the referenced object. For example:

using std::tr1::shared_ptr;
using std::tr1::weak_ptr;

class person;

#pragma db object pointer(shared_ptr)
class employer
{
  ...

  std::vector<shared_ptr<person> > employees_;
};

#pragma db object pointer(shared_ptr)
class person
{
  ...

  #pragma db inverse(employee_)
  weak_ptr<employer> employer_;
};
  

An inverse member does not have a corresponding column or, in case of a container, table in the resulting database schema. Instead, the column or table from the referenced object is used to retrieve the relationship information. Only ordered and set containers can be used for inverse members. If an inverse member is of an ordered container type, it is automatically marked as unordered (Section 14.4.19, "unordered").

For a more detailed discussion of inverse members, refer to Section 6.2, "Bidirectional Relationships".

14.4.15 on_delete

The on_delete specifier specifies the on-delete semantics for a data member of an object pointer or a container of object pointers type. The single required argument to this specifier must be either cascade or set_null.

The on_delete specifier is translated directly to the corresponding ON DELETE SQL clause. That is, if cascade is specified, then when a pointed-to object is erased from the database, the database state of the pointing object is automatically erased as well. If set_null is specified, then when a pointed-to object is erased from the database, the database state of the pointing object is automatically updated to set the pointer column to NULL. For example:

#pragma db object
class employer
{
  ...

  #pragma db id auto
  unsigned long id_;
};

#pragma db object
class person
{
  ...

  #pragma db on_delete(cascade)
  employer* employer_;
};

unsigned long id;

{
  employer e;
  person p;
  p.employer_ = &e;

  transaction t (db.begin ());

  id = db.persist (e);
  db.persist (p);

  t.commit ();
}

{
  transaction t (db.begin ());

  // Database state of the person object is erased as well.
  //
  db.erase<employer> (id);

  t.commit ();
}
  

Note that this is a database-level functionality and care must be taken in order not to end up with inconsistent object states in the application's memory and database. The following example illustrates the kind of problems one may encounter:

#pragma db object
class employer
{
  ...
};

#pragma db object
class person
{
  ...

  #pragma db on_delete(set_null)
  employer* employer_;
};

employer e;
person p;
p.employer_ = &e;

{
  transaction t (db.begin ());
  db.persist (e);
  db.persist (p);
  t.commit ();
}

{
  transaction t (db.begin ());

  // The employer column is set to NULL in the database but
  // not the p.employer_ data member in the application.
  //
  db.erase (e);
  t.commit ();
}

{
  transaction t (db.begin ());

  // Override the employer column with an invalid pointer.
  //
  db.update (p);

  t.commit ();
}
  

Note that even optimistic concurrency will not resolve such issues unless you are using database-level support for optimistic concurrency as well (for example, ROWVERSION in SQL Server).

The on_delete specifier is only valid for non-inverse object pointer data members. If the set_null semantics is used, then the pointer must allow the NULL value.

14.4.16 version

The version specifier specifies that the data member stores the object version used to support optimistic concurrency. If a class has a version data member, then it must also be declared as having the optimistic concurrency model using the optimistic pragma (Section 14.1.5, "optimistic"). For example:

#pragma db object optimistic
class person
{
  ...

  #pragma db version
  unsigned long version_;
};
  

A version member must be of an integral C++ type and must map to an integer or similar database type. Note also that object versions are not reused. If you have a high update frequency, then care must be taken not to run out of versions. In such situations, using unsigned long long as the version type is a safe choice.

For a more detailed discussion of optimistic concurrency, refer to Chapter 12, "Optimistic Concurrency".

14.4.17 index

The index specifier instructs the ODB compiler to define a database index for the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db index
  std::string name_;
};
  

For more information on defining database indexes, refer to Section 14.7, "Index Definition Pragmas".

14.4.18 unique

The index specifier instructs the ODB compiler to define a unique database index for the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db unique
  std::string name_;
};
  

For more information on defining database indexes, refer to Section 14.7, "Index Definition Pragmas".

14.4.19 unordered

The unordered specifier specifies that the member of an ordered container type should be stored unordered in the database. The database table for such a member will not contain the index column and the order in which elements are retrieved from the database may not be the same as the order in which they were stored. For example:

#pragma db object
class person
{
  ...

  #pragma db unordered
  std::vector<std::string> nicknames_;
};
  

For a more detailed discussion of ordered containers and their storage in the database, refer to Section 5.1, "Ordered Containers".

14.4.20 table

The table specifier specifies the table name that should be used to store the contents of the container member. For example:

#pragma db object
class person
{
  ...

  #pragma db table("nicknames")
  std::vector<std::string> nicknames_;
};
  

If the table name is not specified, then the container table name is constructed by concatenating the object's table name, underscore, and the public member name. The public member name is obtained by removing the common member name decorations, such as leading and trailing underscores, the m_ prefix, etc. In the example above, without the table specifier, the container's table name would have been person_nicknames.

The table specifier can also be used for members of composite value types. In this case it specifies the table name prefix for container members inside the composite value type. Refer to Section 7.2.2, "Composite Value Column and Table Names" for details.

The container table name can be qualified with a database schema, for example:

#pragma db object
class person
{
  ...

  #pragma db table("extras.nicknames")
  std::vector<std::string> nicknames_;
};
  

For more information on database schemas and the format of the qualified names, refer to Section 14.1.8, "schema".

14.4.21 load/update

The load and update specifiers specify the loading and updating behavior for an object section, respectively. Valid values for the load specifier are eager (default) and lazy. Valid values for the update specifier are always (default), change, and manual. For more information on object sections, refer to Chapter 9, "Sections".

14.4.22 section

The section specifier indicates that a data member of a persistent class belongs to an object section. The single required argument to this specifier is the name of the section data member. This specifier can only be used on direct data members of a persistent class. For more information on object sections, refer to Chapter 9, "Sections".

14.4.23 added

The added specifier marks the data member as soft-added. The single required argument to this specifier is the addition version. For more information on this functionality, refer to Section 13.4, "Soft Object Model Changes".

14.4.24 deleted

The deleted specifier marks the data member as soft-deleted. The single required argument to this specifier is the deletion version. For more information on this functionality, refer to Section 13.4, "Soft Object Model Changes".

14.4.25 index_type

The index_type specifier specifies the native database type that should be used for an ordered container's index column of the data member. The semantics of index_type are similar to those of the type specifier (Section 14.4.3, "type"). The native database type is expected to be an integer type. For example:

#pragma db object
class person
{
  ...

  #pragma db index_type("SMALLINT UNSIGNED")
  std::vector<std::string> nicknames_;
};
  

14.4.26 key_type

The key_type specifier specifies the native database type that should be used for a map container's key column of the data member. The semantics of key_type are similar to those of the type specifier (Section 14.4.3, "type"). For example:

#pragma db object
class person
{
  ...

  #pragma db key_type("INT UNSIGNED")
  std::map<unsigned short, float> age_weight_map_;
};
  

14.4.27 value_type

The value_type specifier specifies the native database type that should be used for a container's value column of the data member. The semantics of value_type are similar to those of the type specifier (Section 14.4.3, "type"). For example:

#pragma db object
class person
{
  ...

  #pragma db value_type("VARCHAR(255)")
  std::vector<std::string> nicknames_;
};
  

The value_null and value_not_null (Section 14.4.28, "value_null/value_not_null") specifiers can be used to control the NULL semantics of a value column.

14.4.28 value_null/value_not_null

The value_null and value_not_null specifiers specify that a container's element value for the data member can or cannot be NULL, respectively. The semantics of value_null and value_not_null are similar to those of the null and not_null specifiers (Section 14.4.6, "null/not_null"). For example:

using std::tr1::shared_ptr;

#pragma db object
class person
{
  ...
};

#pragma db object
class account
{
  ...

  #pragma db value_not_null
  std::vector<shared_ptr<person> > holders_;
};
  

For set and multiset containers (Section 5.2, "Set and Multiset Containers") the element value is automatically treated as not allowing a NULL value.

14.4.29 id_options

The id_options specifier specifies additional column definition options that should be used for a container's id column of the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db id options("COLLATE binary")
  std::string name_;

  #pragma db id_options("COLLATE binary")
  std::vector<std::string> nicknames_;
};
  

The semantics of id_options are similar to those of the options specifier (Section 14.4.8, "options").

14.4.30 index_options

The index_options specifier specifies additional column definition options that should be used for a container's index column of the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db index_options("ZEROFILL")
  std::vector<std::string> nicknames_;
};
  

The semantics of index_options are similar to those of the options specifier (Section 14.4.8, "options").

14.4.31 key_options

The key_options specifier specifies additional column definition options that should be used for a container's key column of the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db key_options("COLLATE binary")
  std::map<std::string, std::string> properties_;
};
  

The semantics of key_options are similar to those of the options specifier (Section 14.4.8, "options").

14.4.32 value_options

The value_options specifier specifies additional column definition options that should be used for a container's value column of the data member. For example:

#pragma db object
class person
{
  ...

  #pragma db value_options("COLLATE binary")
  std::set<std::string> nicknames_;
};
  

The semantics of value_options are similar to those of the options specifier (Section 14.4.8, "options").

14.4.33 id_column

The id_column specifier specifies the column name that should be used to store the object id in a container's table for the data member. The semantics of id_column are similar to those of the column specifier (Section 14.4.9, "column"). For example:

#pragma db object
class person
{
  ...

  #pragma db id_column("person_id")
  std::vector<std::string> nicknames_;
};
  

If the column name is not specified, then object_id is used by default.

14.4.34 index_column

The index_column specifier specifies the column name that should be used to store the element index in an ordered container's table for the data member. The semantics of index_column are similar to those of the column specifier (Section 14.4.9, "column"). For example:

#pragma db object
class person
{
  ...

  #pragma db index_column("nickname_number")
  std::vector<std::string> nicknames_;
};
  

If the column name is not specified, then index is used by default.

14.4.35 key_column

The key_column specifier specifies the column name that should be used to store the key in a map container's table for the data member. The semantics of key_column are similar to those of the column specifier (Section 14.4.9, "column"). For example:

#pragma db object
class person
{
  ...

  #pragma db key_column("age")
  std::map<unsigned short, float> age_weight_map_;
};
  

If the column name is not specified, then key is used by default.

14.4.36 value_column

The value_column specifier specifies the column name that should be used to store the element value in a container's table for the data member. The semantics of value_column are similar to those of the column specifier (Section 14.4.9, "column"). For example:

#pragma db object
class person
{
  ...

  #pragma db value_column("weight")
  std::map<unsigned short, float> age_weight_map_;
};
  

If the column name is not specified, then value is used by default.

14.5 Namespace Pragmas

A pragma with the namespace qualifier describes a C++ namespace. Similar to other qualifiers, namespace can also refer to a named C++ namespace, for example:

namespace test
{
  ...
}

#pragma db namespace(test) ...
  

To refer to the global namespace in the namespace qualifier the following special syntax is used:

#pragma db namespace() ....
  

The namespace qualifier can be optionally followed, in any order, by one or more specifiers summarized in the table below:

Specifier Summary Section
pointer pointer type for persistent classes and views inside a namespace 14.5.1
table table name prefix for persistent classes inside a namespace 14.5.2
schema database schema for persistent classes inside a namespace 14.5.3
session enable/disable session support for persistent classes inside a namespace 14.5.4

14.5.1 pointer

The pointer specifier specifies the default pointer type for persistent classes and views inside the namespace. For example:

#pragma db namespace pointer(std::tr1::shared_ptr)
namespace accounting
{
  #pragma db object
  class employee
  {
    ...
  };

  #pragma db object
  class employer
  {
    ...
  };
}
  

There are only two valid ways to specify a pointer with the pointer specifier at the namespace level. We can specify the template name of a smart pointer in which case the ODB compiler will automatically append the class name as a template argument. Or we can use * to denote a raw pointer.

Note also that we can always override the default pointer specified at the namespace level for any persistent class or view inside this namespace. For example:

#pragma db namespace pointer(std::unique_ptr)
namespace accounting
{
  #pragma db object pointer(std::shared_ptr)
  class employee
  {
    ...
  };

  #pragma db object
  class employer
  {
    ...
  };
}
  

For a more detailed discussion of object and view pointers, refer to Section 3.3, "Object and View Pointers".

14.5.2 table

The table specifier specifies a table prefix that should be added to table names of persistent classes inside the namespace. For example:

#pragma db namespace table("acc_")
namespace accounting
{
  #pragma db object table("employees")
  class employee
  {
    ...
  };

  #pragma db object table("employers")
  class employer
  {
    ...
  };
}
  

In the above example the resulting table names will be acc_employees and acc_employers.

The table name prefix can also be specified with the --table-prefix ODB compiler option. Note that table prefixes specified at the namespace level as well as with the command line option are accumulated. For example:

#pragma db namespace() table("audit_")

#pragma db namespace table("hr_")
namespace hr
{
  #pragma db object table("employees")
  class employee
  {
    ...
  };
}

#pragma db object table("employers")
class employer
{
  ...
};
  

If we compile the above example with the --table-prefix test_ option, then the employee class table will be called test_audit_hr_employees and employertest_audit_employers.

Table prefixes can be used as an alternative to database schemas (Section 14.1.8, "schema") if the target database system does not support schemas.

14.5.3 schema

The schema specifier specifies a database schema that should be used for persistent classes inside the namespace. For more information on specifying a database schema refer to Section 14.1.8, "schema".

14.5.4 session

The session specifier specifies whether to enable session support for persistent classes inside the namespace. For example:

#pragma db namespace session
namespace hr
{
  #pragma db object                // Enabled.
  class employee
  {
    ...
  };

  #pragma db object session(false) // Disabled.
  class employer
  {
    ...
  };
}
  

Session support is disabled by default unless the --generate-session ODB compiler option is specified. Session support specified at the namespace level can be overridden on the per object basis (Section 14.1.10, "session"). For more information on sessions, refer to Chapter 11, "Session".

14.6 Object Model Pragmas

A pragma with the model qualifier describes the whole C++ object model. For example:

#pragma db model ...
  

The model qualifier can be followed, in any order, by one or more specifiers summarized in the table below:

Specifier Summary Section
version object model version 14.6.1

14.6.1 version

The version specifier specifies the object model version when schema evolution support is used. The first two required arguments to this specifier are the base and current model versions, respectively. The third optional argument specifies whether the current version is open for changes. Valid values for this argument are open (the default) and closed. For more information on this functionality, refer to Chapter 13, "Database Schema Evolution".

14.7 Index Definition Pragmas

While it is possible to manually add indexes to the generated database schema, it is more convenient to do this as part of the persistent class definitions. A pragma with the index qualifier describes a database index. It has the following general format:

#pragma db index[("<name>")]                       \
           [unique|type("<type>")]                 \
           [method("<method>")]                    \
           [options("<index-options>")]            \
           member(<name>[, "<column-options>"])... \
           members(<name>[,<name>...])...

The index qualifier can optionally specify the index name. If the index name is not specified, then one is automatically derived by appending the _i suffix to the column name of the index member. If the name is not specified and the index contains multiple members, then the index definition is invalid.

The optional type, method, and options clauses specify the index type, for example UNIQUE, index method, for example BTREE, and index options, respectively. The unique clause is a shortcut for type("UNIQUE"). Note that not all database systems support specifying an index method or options. For more information on the database system-specific index types, methods, and options, refer to Part II, "Database Systems".

To specify index members we can use the member or members clauses, or a mix of the two. The member clause allows us to specify a single index member with optional column options, for example, "ASC". If we need to create a composite index that contains multiple members, then we can repeat the member clause several times or, if the members don't have any column options, we can use a single members clause instead. Similar to the index type, method, and options, the format of column options is database system-specific. For more details, refer to Part II, "Database Systems".

The following code fragment shows some typical examples of index definitions:

#pragma db object
class object
{
  ...

  int x;
  int y;
  int z1;
  int z2;

  // An index for member x with automatically-assigned name x_i.
  //
  #pragma db index member(x)

  // A unique index named y_index for member y which is sorted in
  // the descending order. The index is using the BTREE method.
  //
  #pragma db index("y_index") unique method("BTREE") member(y, "DESC")

  // A composite BITMAP index named z_i for members z1 and z2.
  //
  #pragma db index("z_i") type("BITMAP") members(z1, z2)
};

ODB also offers a shortcut for defining an index with the default method and options for a single data member. Such an index can be defined using the index (Section 14.4.17, "index") or unique (Section 14.4.18, "unique") member specifier. For example:

#pragma db object
class object
{
  ...

  #pragma db index
  int x;

  #pragma db type("INT") unique
  int y;
};

The above example is semantically equivalent to the following more verbose version:

#pragma db object
class object
{
  ...

  int x;

  #pragma db type("INT")
  int y;

  #pragma db index member(x)
  #pragma db index unique member(y)
};

While it is convenient to define an index inside a persistent class, it is also possible to do that out of the class body. In this case, the index name, if specified, must be prefixed with the potentially-qualified class name. For example:

namespace n
{
  #pragma db object
  class object
  {
    ...

    int x;
    int y;
  };

  // An index for member x in persistent class object with automatically-
  // assigned name x_i.
  //
  #pragma db index(object) member(x)
}

// An index named y_index for member y in persistent class n::object.
//
#pragma db index(n::object::"y_index") member(y)

It is possible to define an index on a member that is of a composite value type or on some of its nested members. For example:

#pragma db value
struct point
{
  int x;
  int y;
  int z;
};

#pragma db object
class object
{
  // An index that includes all of the p1's nested members.
  //
  #pragma db index
  point p1;

  point p2;

  // An index that includes only p2.x and p2.y.
  //
  #pragma db index("p2_xy_i") members(p2.x, p2.y)
};

When generating a schema for a container member (Chapter 5, "Containers"), ODB automatically defines two indexes on the container table. One is for the object id that references the object table and the other is for the index column in case the container is ordered (Section 5.1, "Ordered Containers"). By default these indexes use the default index name, type, method, and options. The index pragma allows us to customize these two indexes by recognizing the special id and index nested member names when specified after a container member. For example:

#pragma db object
class object
{
  std::vector<int> v;

  // Change the container id index name.
  //
  #pragma db index("id_index") member(v.id)

  // Change the container index index method.
  //
  #pragma db index method("BTREE") member(v.index)
};

14.8 Database Type Mapping Pragmas

A pragma with the map qualifier describes a mapping between two database types. For each database system ODB provides built-in support for a core set of database types, such as integers, strings, binary, etc. However, many database systems provide extended types such as geospatial types, user-defined types, and collections (arrays, table types, key-value stores, etc). In order to support such extended types, ODB allows us to map them to one of the built-in types, normally a string or a binary. Given the text or binary representation of the data we can then extract it into our chosen C++ data type and thus establish a mapping between an extended database type and its C++ equivalent.

The map pragma has the following format:

#pragma db map type("regex") as("subst") [to("subst")] [from("subst")]

The type clause specifies the name of the database type that we are mapping. We will refer to it as the mapped type from now on. The name of the mapped type is a Perl-like regular expression pattern that is matched in the case-insensitive mode.

The as clause specifies the name of the database type that we are mapping the mapped type to. We will refer to it as the interface type from now on. The name of the interface type is a regular expression substitution and should expand to a name of a database type for which ODB provides built-in support.

The optional to and from clauses specify the database conversion expressions between the mapped type and the interface type. The to expression converts from the interface type to the mapped type and from converts in the other direction. If no explicit conversion is required for either direction, then the corresponding clause can be omitted. The conversion expressions are regular expression substitutions. They must contain the special (?) placeholder which will be replaced with the actual value to be converted. Turning on SQL statement tracing (Section 3.13, "Tracing SQL Statement Execution") can be useful for debugging conversion expressions. This allows you to see the substituted expressions as used in the actual statements.

As an example, the following map pragma maps the PostgreSQL array of INTEGER's to TEXT:

#pragma db map type("INTEGER *\\[(\\d*)\\]") \
               as("TEXT")                    \
               to("(?)::INTEGER[$1]")        \
               from("(?)::TEXT")

With the above mapping we can now have a persistent class that has a member of the PostgreSQL array type:

#pragma db object
class object
{
  ...

  #pragma db type("INTEGER[]")
  std::string array_;
};

In PostgreSQL the array literal has the {n1,n2,...} form. As a result, we need to make sure that we pass the correct text representation in the array_ member, for example:

object o;
o.array_ = "{1,2,3}";
db.persist (o);

Of course, std::string is not the most natural representation of an array of integers in C++. Instead, std::vector<int> would have been much more appropriate. To add support for mapping std::vector<int> to PostgreSQL INTEGER[] we need to provide a value_traits specialization that implements conversion between the PostgreSQL text representation of an array and std::vector<int>. Below is a sample implementation:

namespace odb
{
  namespace pgsql
  {
    template <>
    class value_traits<std::vector<int>, id_string>
    {
    public:
      typedef std::vector<int> value_type;
      typedef value_type query_type;
      typedef details::buffer image_type;

      static void
      set_value (value_type& v,
                 const details::buffer& b,
                 std::size_t n,
                 bool is_null)
      {
        v.clear ();

        if (!is_null)
        {
          char c;
          std::istringstream is (std::string (b.data (), n));

          is >> c; // '{'

          for (c = static_cast<char> (is.peek ()); c != '}'; is >> c)
          {
            v.push_back (int ());
            is >> v.back ();
          }
        }
      }

      static void
      set_image (details::buffer& b,
                 std::size_t& n,
                 bool& is_null,
                 const value_type& v)
      {
        is_null = false;
        std::ostringstream os;

        os << '{';

        for (value_type::const_iterator i (v.begin ()), e (v.end ());
             i != e;)
        {
          os << *i;

          if (++i != e)
            os << ',';
        }

        os << '}';

        const std::string& s (os.str ());
        n = s.size ();

        if (n > b.capacity ())
          b.capacity (n);

        std::memcpy (b.data (), s.c_str (), n);
      }
    };
  }
}

Once this specialization is included in the generated code (see the mapping example in the odb-examples package for details), we can use std::vector<int> instead of std::string in our persistent class:

#pragma db object
class object
{
  ...

  #pragma db type("INTEGER[]")
  std::vector<int> array_;
};

If we wanted to always map std::vector<int> to PostgreSQL INTEGER[], then we could instead write:

typedef std::vector<int> int_vector;
#pragma db value(int_vector) type("INTEGER[]")

#pragma db object
class object
{
  ...

  std::vector<int> array_; // Mapped to INTEGER[].
};

While the above example only shows how to handle PostgreSQL arrays, other types in PostgreSQL and in other databases can be supported in a similar way. The odb-tests package contains a set of tests in the <database>/custom directories that, for each database, shows how to provide custom mapping for some of the extended types.

14.9 C++ Compiler Warnings

When a C++ header file defining persistent classes and containing ODB pragmas is used to build the application, the C++ compiler may issue warnings about pragmas that it doesn't recognize. There are several ways to deal with this problem. The easiest is to disable such warnings using one of the compiler-specific command line options or warning control pragmas. This method is described in the following sub-section for popular C++ compilers.

There are also several C++ compiler-independent methods that we can employ. The first is to use the PRAGMA_DB macro, defined in <odb/core.hxx>, instead of using #pragma db directly. This macro expands to the ODB pragma when compiled with the ODB compiler and to an empty declaration when compiled with other compilers. The following example shows how we can use this macro:

#include <odb/core.hxx>

PRAGMA_DB(object)
class person
{
  ...

  PRAGMA_DB(id)
  unsigned long id_;
};
  

An alternative to using the PRAGMA_DB macro is to group the #pragma db directives in blocks that are conditionally included into compilation only when compiled with the ODB compiler. For example:

class person
{
  ...

  unsigned long id_;
};

#ifdef ODB_COMPILER
#  pragma db object(person)
#  pragma db member(person::id_) id
#endif
  

The disadvantage of this approach is that it can quickly become overly verbose when positioned pragmas are used.

14.9.1 GNU C++

GNU g++ does not issue warnings about unknown pragmas unless requested with the -Wall command line option. To disable only the unknown pragma warning, we can add the -Wno-unknown-pragmas option after -Wall, for example:

g++ -Wall -Wno-unknown-pragmas ...
  

14.9.2 Visual C++

Microsoft Visual C++ issues an unknown pragma warning (C4068) at warning level 1 or higher. This means that unless we have disabled the warnings altogether (level 0), we will see this warning.

To disable this warning via the compiler command line, we can add the /wd4068 C++ compiler option in Visual Studio 2008 and earlier. In Visual Studio 2010 and later there is now a special GUI field where we can enter warning numbers that should be disabled. Simply enter 4068 into this field.

We can also disable this warning for only a specific header or a fragment of a header using the warning control pragma. For example:

#include <odb/core.hxx>

#pragma warning (push)
#pragma warning (disable:4068)

#pragma db object
class person
{
  ...

  #pragma db id
  unsigned long id_;
};

#pragma warning (pop)
  

14.9.3 Sun C++

The Sun C++ compiler does not issue warnings about unknown pragmas unless the +w or +w2 option is specified. To disable only the unknown pragma warning we can add the -erroff=unknownpragma option anywhere on the command line, for example:

CC +w -erroff=unknownpragma ...
  

14.9.4 IBM XL C++

IBM XL C++ issues an unknown pragma warning (1540-1401) by default. To disable this warning we can add the -qsuppress=1540-1401 command line option, for example:

xlC -qsuppress=1540-1401 ...
  

14.9.5 HP aC++

HP aC++ (aCC) issues an unknown pragma warning (2161) by default. To disable this warning we can add the +W2161 command line option, for example:

aCC +W2161 ...
  

14.9.6 Clang

Clang does not issue warnings about unknown pragmas unless requested with the -Wall command line option. To disable only the unknown pragma warning, we can add the -Wno-unknown-pragmas option after -Wall, for example:

clang++ -Wall -Wno-unknown-pragmas ...
  

We can also disable this warning for only a specific header or a fragment of a header using the warning control pragma. For example:

#include <odb/core.hxx>

#pragma clang diagnostic push
#pragma clang diagnostic ignored "-Wunknown-pragmas"

#pragma db object
class person
{
  ...

  #pragma db id
  unsigned long id_;
};

#pragma clang diagnostic pop
  

15 Advanced Techniques and Mechanisms

This chapter covers more advanced techniques and mechanisms provided by ODB that may be useful in certain situations.

15.1 Transaction Callbacks

The ODB transaction class (odb::transaction) allows an application to register a callback that will be called after the transaction is finalized, that is, committed or rolled back. This mechanism can be used, for example, to restore values that were updated during the transaction execution to their original states if the transaction is rolled back.

The callback management interface of the transaction class is shown below.

namespace odb
{
  class transaction
  {
    ...

  public:
    static const unsigned short event_commit = 0x01;
    static const unsigned short event_rollback = 0x02;
    static const unsigned short event_all = event_commit | event_rollback;

    typedef void (*callback_type) (
      unsigned short event, void* key, unsigned long long data);

    void
    callback_register (callback_type callback,
                       void* key,
                       unsigned short event = event_all,
                       unsigned long long data = 0,
                       transaction** state = 0);


    void
    callback_unregister (void* key);

    void
    callback_update (void* key,
                     unsigned short event,
                     unsigned long long data = 0,
                     transaction** state = 0);
  }
}
  

The callback_register() function registers a post-commit/rollback callback. The callback argument is the function that should be called. The key argument is used by the transaction to identify this callback. It is also normally used to pass an address of the data object on which the callback function will work. The event argument is the bitwise-or of the events that should trigger the callback.

The optional data argument can be used to store any POD user data that doesn't exceed 8 bytes in size and doesn't require alignment greater than unsigned long long. For example, we could store an old value of a flag or a counter that needs to be restored in case of a roll back.

The optional state argument can be used to indicate that the callback has been unregistered because the transaction was finalized. In this case the transaction automatically resets the passed pointer to 0. This is primarily useful if we are interested in only one of the events (commit or rollback).

The callback_unregister() function unregisters a previously registered callback. If the number of registered callbacks is large, then this can be a slow operation. Generally, the callback mechanism is optimized for cases where the callbacks stay registered until the transaction is finalized.

Note also that you don't need to unregister a callback that has been called or auto-reset using the state argument passed to callback_register(). This function does nothing if the key is not found.

The callback_update() function can be used to update the event, data, and state values of a previously registered callback. Similar to callback_unregister(), this is a potentially slow operation.

When the callback is called, it is passed the event that triggered it, as well as the key and data values that were passed to the callback_register() function. Note also that the order in which the callbacks are called is unspecified. The rollback event can be triggered by an exception. In this case, if the callback throws, the program will be terminated.

The following example shows how we can use transaction callbacks together with database operation callbacks (Section 14.1.7, "callback") to manage the object's "dirty" flag.

#include <odb/callback.hxx>
#include <odb/transaction.hxx>

#pragma db object callback(update)
class object
{
  ...

  #pragma db transient
  mutable bool dirty_;

  // Non-NULL value indicates that we are registered
  // with this transaction.
  //
  #pragma db transient
  mutable odb::transaction* tran_;

  void
  update (odb::callback_event e, odb::database&) const
  {
    using namespace odb::core;

    if (e == callback_event::post_update)
      return;

    // Mark the object as clean again but register a
    // transaction callback in case the update is rolled
    // back.
    //
    tran_ = &transaction::current ();
    tran_->callback_register (&rollback,
                              const_cast<object*> (this),
                              transaction::event_rollback,
                              0,
                              &tran_);
    dirty_ = false;
  }

  static void
  rollback (unsigned short, void* key, unsigned long long)
  {
    // Restore the dirty flag since the changes have been
    // rolled back.
    //
    object& o (*static_cast<object*> (key));
    o.dirty_ = true;
  }

  ~object ()
  {
    // Unregister the callback if we are going away before
    // the transaction.
    //
    if (tran_ != 0)
      tran_->callback_unregister (this);
  }
};
  

15.2 Persistent Class Template Instantiations

Similar to composite value types (Section 7.2, "Composite Value Types"), a persistent object can be defined as an instantiation of a C++ class template, for example:

template <typename T>
class person
{
  ...

  T first_;
  T last_;
};

typedef person<std::string> std_person;

#pragma db object(std_person)
#pragma db member(std_person::last_) id
  

Note that the database support code for such a persistent object is generated when compiling the header containing the db object pragma and not the header containing the template definition or the typedef name. This allows us to use templates defined in other files, for example:

#include <utility> // std::pair

typedef std::pair<unsigned int, std::string> person;
#pragma db object(person)
#pragma db member(person::first) id auto column("id")
#pragma db member(person::second) column("name")
  

You may also have to explicitly specify the object type in calls to certain database class functions due to the inability do distinguish, at the API level, between smart pointers and persistent objects defined as class template instantiations. For example:

person p;

db.update (p); // Error.
db.reload (p); // Error.
db.erase (p);  // Error.

db.update<person> (p); // Ok.
db.reload<person> (p); // Ok.
db.erase<person> (p);  // Ok.
  

It also makes sense to factor persistent data members that do not depend on template arguments into a common, non-template base class. The following more realistic example illustrates this approach:

#pragma db object abstract
class base_common
{
  ...

  #pragma db id auto
  unsigned long id;
};

template <typename T>
class base: public base_common
{
  ...

  T value;
};

typedef base<std::string> string_base;
#pragma db object(string_base) abstract

#pragma db object
class derived: public string_base
{
  ...
};
  

15.3 Bulk Database Operations

Some database systems supported by ODB provide a mechanism, often called bulk or batch statement execution, that allows us to execute the same SQL statement on multiple sets of data at once and with a single database API call. This often results in significantly better performance if we need to execute the same statement for a large number of data sets (thousands to millions).

ODB translates this mechanism to bulk operations which allow us to persist, update, or erase a range of objects in the database. Currently, from all the database systems supported by ODB, only Oracle and Microsoft SQL Server are capable of bulk operations. There is also currently no emulation of the bulk API for other databases nor dynamic multi-database support. As a result, if you are using dynamic multi-database support, you will need to "drop down" to static support in order to access the bulk API. Refer to Chapter 16, "Multi-Database Support" for details.

As we will discuss later in this section, bulk operations have complex failure semantics that is dictated by the underlying database API. As a result, support for bulk persist, update, and erase is limited to persistent classes for which these operations can be performed with a single database statement execution. In particular, bulk operations are not available for polymorphic objects (Section 8.2, "Polymorphism Inheritance") or objects that have containers (inverse containers of object pointers are an exception). Furthermore, for objects that have sections (Chapter 9, "Sections") the bulk update operation will only be available if all the sections are manually-updated. On the other hand, bulk operations are supported for objects that use optimistic concurrency (Chapter 12, "Optimistic Concurrency") or have no object id (Section 14.1.6, "no_id").

To enable the generation of bulk operation support for a persistent class we use the bulk pragma. For example:

#pragma db object bulk(5000)
class person
{
  ...

  #pragma db id auto
  unsigned long id;
};
  

The single argument to the bulk pragma is the batch size. The batch size specifies the maximum number of data sets that should be handled with a single underlying statement execution. If the range that we want to perform the bulk operation on contains more objects than the batch size, then ODB will split this operation into multiple underlying statement executions (batches). To illustrate this point with an example, suppose we want to persist 53,000 objects and the batch size is 5,000. ODB will then execute the statement 11 times, the first 10 times with 5,000 data sets each, and the last time with the remaining 3,000 data sets.

The commonly used batch sizes are in the 2,000-5,000 range, though smaller or larger batches could provide better performance, depending on the situation. As a result, it is recommended to experiment with different batch sizes to determine the optimum number for a particular object and its use-cases. Note also that you may achieve better performance by also splitting a large bulk operation into multiple transactions (Section 3.5, "Transactions").

For database systems that do not support bulk operations the bulk pragma is ignored. It is also possible to specify different batch sizes for different database systems by using the database prefix, for example:

#pragma db object mssql:bulk(3000) oracle:bulk(4000)
class person
{
  ...
};
  

Note that while specifying the batch size at compile time might seem inflexible, this approach allows ODB to place internal arrays of the fixed batch size on the stack rather than allocating them in the dynamic memory. However, specifying the batch size at runtime may be supported in the future.

Once the bulk support is enabled for a particular object, we can use the following database functions to perform bulk operations:

template <typename I>
void
persist (I begin, I end, bool continue_failed = true);

template <typename I>
void
update (I begin, I end, bool continue_failed = true);

template <typename I>
void
erase (I obj_begin, I obj_end, bool continue_failed = true);

template <typename T, typename I>
void
erase (I id_begin, I id_end, bool continue_failed = true);
  

Every bulk API function expects a range of elements, passed in the canonical C++ form as a pair of input iterators. In case of persist(), update(), and the first erase() overload, we pass a range of objects, either as references or as pointers, raw or smart. The following example illustrates the most common scenarios using the persist() call:

// C array of objects.
//
person a[2] {{"John", "Doe"}, {"Jane", "Doe"}};

db.persist (a, a + sizeof(a) / sizeof(a[0]));


// Vector of objects.
//
std::vector<person> v {{"John", "Doe"}, {"Jane", "Doe"}};

db.persist (v.begin (), v.end ());


// C array of raw pointers to objects.
//
person p1 ("John", "Doe");
person p2 ("Jane", "Doe");
person* pa[2] {&p1, &p2};

db.persist (pa, pa + sizeof(pa) / sizeof(pa[0]));


// Vector of raw pointers to objects.
//
std::vector<person*> pv {&p1, &p2};

db.persist (pv.begin (), pv.end ());


// Vector of smart (shared) pointers to objects.
//
std::vector<std::shared_ptr<person>> sv {
  std::make_shared<person> ("John", "Doe"),
  std::make_shared<person> ("Jane", "Doe")};

db.persist (sv.begin (), sv.end ());
  

The ability to perform a bulk operation on a range of raw pointers to objects can be especially useful when the application stores objects in a way that does not easily conform to the pair of iterators interface. In such cases we can create a temporary container of shallow pointers to objects and use that to perform the bulk operation, for example:

struct person_entry
{
  person obj;

  // Some additional data.
  ...
};

typedef std::vector<person_entry> people;

void
persist (odb::database& db, people& p)
{
  std::vector<person*> tmp;
  tmp.reserve (p.size ());
  std::for_each (p.begin (),
                 p.end (),
                 [&tmp] (person_entry& pe)
                 {
                   tmp.push_back (&pe.obj);
                 });


  db.persist (tmp.begin (), tmp.end ());
}
  

The second overload of the bulk erase() function allows us to pass a range of object ids rather than objects themselves. As with the corresponding non-bulk version, we have to specify the object type explicitly, for example:

std::vector<unsigned long> ids {1, 2};

db.erase<person> (ids.begin (), ids.end ());
  

Conceptually, a bulk operation is equivalent to performing the corresponding non-bulk version in a loop, except when it comes to the failure semantics. Both databases that currently are capable of bulk operations (Oracle and SQL Server) do not stop when a data set in a batch fails (for example, because of a unique constraint violation). Instead, they continue executing subsequent data sets until every element in the batch has been attempted. The continue_failed argument in the bulk functions listed above specifies whether ODB should extend this behavior and continue with subsequent batches if the one it has tried to execute has failed elements. The default behavior is to continue.

The consequence of this failure semantics is that we may have multiple elements in the range failed for different reasons. For example, if we tried to persist a number of objects, some of them might have failed because they are already persistent while others — because of a unique constraint violation. As a result, ODB uses the special odb::multiple_exceptions class to report failures in the bulk API functions. This exception is thrown if one or more elements in the range have failed and it contains the error information in the form of other ODB exception for each failed position. The multiple_exceptions class has the following interface:

struct multiple_exceptions: odb::exception
{
  // Element type.
  //
  struct value_type
  {
    std::size_t
    position () const;

    const odb::exception&
    exception () const;

    bool
    maybe () const;
  };

  // Iteration.
  //
  typedef std::set<value_type> set_type;

  typedef set_type::const_iterator iterator;
  typedef set_type::const_iterator const_iterator;

  iterator
  begin () const;

  iterator
  end () const;

  // Lookup.
  //
  const value_type*
  operator[] (std::size_t) const;

  // Severity, failed and attempted counts.
  //
  std::size_t
  attempted () const;

  std::size_t
  failed () const;

  bool
  fatal () const;

  void
  fatal (bool);

  // Direct data access.
  //
  const set_type&
  set () const;

  // odb::exception interface.
  //
  virtual const char*
  what () const throw ();
};
  

The multiple_exceptions class has a map-like interface with the key being the position in the range and the value being the exception plus the maybe flag (discussed below). As a result, we can either iterate over the failed positions or we can check whether a specific position in the range has failed. The following example shows what a catch-handler for this exception might look like:

std::vector<person> objs {{"John", "Doe"}, {"Jane", "Doe"}};

try
{
  db.persist (objs.begin (), objs.end ());
}
catch (const odb::multiple_exceptions& me)
{
  for (const auto& v: me)
  {
    size_t p (v.position ());

    try
    {
      throw v.exception ();
    }
    catch (const odb::object_already_persistent&)
    {
      cerr << p << ": duplicate id: " << objs[p].id () << endl;
    }
    catch (const odb::exception& e)
    {
      cerr << p << ": " << e.what () << endl;
    }
  }
}
  

If, however, all we want is to show the diagnostics to the user, then the string returned by the what() function will contain the error information for each failed position. Here is what it might look like (using Oracle as an example):

multiple exceptions, 4 elements attempted, 2 failed:
[0] object already persistent
[3] 1: ORA-00001: unique constraint (ODB_TEST.person_last_i) violated
  

Both databases that currently are capable of bulk operations return a total count of affected rows rather than individual counts for each data set. This limitation prevents ODB from being able to always determine which elements in the batch haven't affected any rows and, for the update and erase operations, translate this to the object_not_persistent exceptions. As a result, if some elements in the batch haven't affected any rows and ODB is unable to determine exactly which ones, it will mark all the elements in this batch as "maybe not persistent". That is, it will insert the object_not_persistent exception and set the maybe flag for every position in the batch. The diagnostics string returned by what() will also reflect this situation, for example (assuming batch size of 3):

multiple exceptions, 4 elements attempted, 4 failed:
[0-2] (some) object not persistent
[3] object not persistent
  

The way to handle and recover from such "maybe failures" will have to be application-specific. For example, for some applications the fact that some objects no longer exist in the database when performing bulk erase might be an ignorable error. If, however, the application needs to determine exactly which elements in the batch have failed, then a load() call will be required for each element in the batch (or a query using a view to avoid loading all the data members; Chapter 10, "Views"). This is also something to keep in mind when selecting the batch size since for larger sizes it will be more expensive (more loads to perform) to handle such "maybe failures". If the failures are not uncommon, as is the case, for example, when using optimistic concurrency, then it may make sense to use a smaller batch.

The lookup operator (operator[]) returns NULL if the element at this position has no exception. Note also that the returned value is value_type* and not odb::exception* in order to provide access to the maybe flag discussed above.

The multiple_exceptions class also provides access to the number of positions attempted (the attempted() accessor) and failed (the failed() accessor). Note that the failed count includes the "maybe failed" positions.

The multiple_exceptions exception can also be fatal. If the fatal() accessor returns true, then (some of) the exceptions were fatal. In this case, even for positions that did not fail, no attempts were made to complete the operation and the transaction must be aborted.

If fatal() returns false, then the operation on the elements that don't have an exception has succeeded. The application can ignore the errors or try to correct the errors and re-attempt the operation on the elements that did fail. In either case, the transaction can be committed.

An example of a fatal exception would be the situation where the execution of the underlying statement failed summarily, without attempting any data sets, for instance, because of an error in the statement itself.

The fatal() modifier allows you to "upgrade" an exception to fatal, for example, for specific database error codes.


PART II   DATABASE SYSTEMS

Part II covers topics specific to the database system implementations and their support in ODB. The first chapter in Part II discusses how to use multiple database systems in the same application. The subsequent chapters describe the system-specific database classes as well as the default mapping between basic C++ value types and native database types. Part II consists of the following chapters.

16Multi-Database Support
17MySQL Database
18SQLite Database
19PostgreSQL Database
20Oracle Database
21Microsoft SQL Server Database

16 Multi-Database Support

Some applications may need to access multiple database systems, either simultaneously or one at a time. For example, an application may utilize an embedded database such as SQLite as a local cache and use a client-server database such as PostgreSQL for more permanent but slower to access remote storage. Or an application may need to be able to store its data in any database selected at runtime by the user. Yet another scenario is the data migration from one database system to another. In this case, multi-database support is only required for a short period. It is also plausible that an application implements all three of these scenarios, that is, it uses SQLite as a local cache, allows the user to select the remote database system, and supports data migration from one remote database system to another.

ODB provides two types of multi-database support: static and dynamic. With static support we use the database system-specific interfaces to perform database operations. That is, instead of using odb::database, odb::transaction, or odb::query, we would use, for example, odb::sqlite::database, odb::sqlite::transaction, or odb::sqlite::query to access an SQLite database.

In contrast, with dynamic multi-database support we can use the common interface to access any database without having to know which one it is. At runtime, ODB will automatically dispatch a call on the common interface to the specific database implementation based on the actual database instance being used. In fact, this mechanism is very similar to C++ virtual functions.

Both static and dynamic multi-database support have a different set of advantages and disadvantages which makes them more or less suitable for different use cases. Static support has zero overhead compared to single-database support and allows us to use database system-specific features, extensions, etc. At the same time, the code that we write will be tied to the specific database system. As a result, this type of multi-database support is more suitable for situations where different parts of an application access different but specific database systems. For example, using SQLite as a local cache most likely falls into this category since we are using a specific database system (SQLite) and the code that will check the cache will most likely (but not necessarily) be separate from the code that interact with the remote database. Another example where static multi-database support might be more suitable is a once-off data migration from one database system to another. In this case both the source and target are specific database systems. In contrast, if data migration from one database system to another is a general feature in an application, then dynamic multi-database support might be more suitable.

The main advantage of dynamic multi-database support is the database system-independence of the code that we write. The same application code will work with any database system supported by ODB and the generated database support code can be packaged into separate libraries and loaded dynamically by the application. The disadvantages of dynamic support are slight overhead and certain limitations in functionality compared to static support (see Section 16.2, "Dynamic Multi-Database Support" for details). As a result, dynamic multi-database support is most suitable to situations where we need the same code to work with a range of database systems. For example, if your application must be able to store its data in any database selected by the user, then dynamic support is probably the best option.

Note also that it is possible to mix and match static and dynamic support in the same application. In fact, dynamic support is built on top of static support so it is possible to use the same database system both "statically" and "dynamically". In particular, the ability to "drop down" from dynamic to static support can be used to overcome the functionality limitations mentioned above. Finally, single-database support is just a special case of static multi-database support with a single database system.

By default ODB assumes single-database support. To enable multi-database support we use the --multi-database (or -m) ODB compiler option. This option is also used to specify the support type: static or dynamic. For example:

odb -m static ... person.hxx
  

With multi-database support enabled, we can now generate the database support code for several database systems. This can be accomplished either with a single ODB compiler invocation by specifying multiple --database (or -d) options or with multiple ODB compiler invocations. Both approaches produce the same result, for example:

odb -m static -d common -d sqlite -d pgsql person.hxx
  

Is equivalent to:

odb -m static -d common person.hxx
odb -m static -d sqlite person.hxx
odb -m static -d pgsql person.hxx
  

Notice that the first -d option has common as its value. This is not a real database system. Rather, it instructs the ODB compiler to generate code that is common to all the database systems and, in case of dynamic support, is also the common interfaces.

If you look at the result of the above commands, you will also notice changes in the output file names. In the single-database mode the ODB compiler produces a single set of the person-odb.?xx files which contain both the common as well as the database specific generated code (since there is only one database system in use, there is no reason to split the two). In contrast, in the multi-database mode, the person-odb.?xx set of files contains the common code while the database system-specific code is written to files in the form person-odb-<db>.?xx. That is, person-odb-sqlite.?xx for SQLite, person-odb-pgsql.?xx for PostgreSQL, etc.

If we need dynamic support for some databases and static for others, then the common code must be generated in the dynamic mode. For example, if we need static support for SQLite and dynamic support for PostgreSQL and Oracle, then the ODB compiler invocations could look like this:

odb -m dynamic -d common person.hxx
odb -m static -d sqlite person.hxx
odb -m dynamic -d pgsql person.hxx
odb -m dynamic -d oracle person.hxx
  

With multi-database support enabled, it is possible to restrict ODB pragmas to apply only to a specific database system (unrestricted pragmas apply to all the databases). For example:

#pragma db object
class person
{
  ...

  #pragma db pgsql:type("VARCHAR(128)") sqlite:type("TEXT")
  std::string name_;

  unsigned short age_;

  #pragma db pgsql index member(age_)
};
  

Above, the pragma for the name_ data member shows the use of a database prefix (for example, pgsql:) that only applies to the specifier that follows. The pragma that defines an index on the age_ data member shows the use of a database prefix that applies to the whole pragma. In this case the database name must immediately follow the db keyword.

Similar to pragmas, ODB compiler options that determine the kind (for example, --schema-format), names (for example, --odb-file-suffix), or content (for example, prologue and epilogue options) of the output files can be prefixed with the database name. For example:

odb --odb-file-suffix common:-odb-common ...
  

Dynamic multi-database support requires consistent mapping across all the databases. That is, the same classes and data members should be mapped to objects, simple/composite values, etc., for all the databases. In contrast, static multi-database support does not have this restriction. Specifically, with static support, some data members can be transient for some database systems. Similarly, the same class (for example, point) can be mapped to a simple value in one database (for example, to the POINT PostgreSQL type) and to a composite value in another (for example, in SQLite, which does not have a built-in point type).

The following sections discuss static and dynamic multi-database support in more detail.

16.1 Static Multi-Database Support

With static multi-database support, instead of including person-odb.hxx, application source code has to include person-odb-<db>.hxx header files corresponding to the database systems that will be used.

The application code has to also use database system-specific interfaces when performing database operations. As an example, consider the following transaction in a single-database application. It uses the common interfaces, that is, classes from the odb namespace.

#include "person-odb.hxx"

odb::database& db = ...

typedef odb::query<person> query;
typedef odb::result<person> result;

odb::transaction t (db.begin ());
result r (db.query<person> (query::age < 30));
...
t.commit ();
  

In an application that employs static multi-database support the same transaction for SQLite would be rewritten like this:

#include "person-odb-sqlite.hxx"

odb::sqlite::database& db = ...

typedef odb::sqlite::query<person> query;
typedef odb::result<person> result;      // odb:: not odb::sqlite::

odb::sqlite::transaction t (db.begin ());
result r (db.query<person> (query::age < 30));
...
t.commit ();
  

That is, the database, transaction, and query classes now come from the odb::sqlite namespace instead of odb. Other classes that have database system-specific interfaces are connection, statement, and tracer. Note that all of them derive from the corresponding common versions. It is also possible to use common transaction, connection, and statement classes with static support, if desired.

Notice that we didn't use the odb::sqlite namespace for the result class template. This is because result is database system-independent. All other classes defined in namespace odb, except those specifically mentioned above, are database system-independent. In particular, result, prepared_query, session, schema_catalog, and all the exceptions are database system-independent.

Writing odb::sqlite:: before every name can quickly become burdensome. As we have seen before, in single-database applications that use the common interface we can add the using namespace directive to avoid qualifying each name. For example:

#include "person-odb.hxx"

odb::database& db = ...

{
  using namespace odb::core;

  typedef query<person> person_query;
  typedef result<person> person_result;

  transaction t (db.begin ());
  person_result r (db.query<person> (person_query::age < 30));
  ...
  t.commit ();
}
  

A similar mechanism is available in multi-database support. Each database runtime defines the odb::<db>::core namespace that contains all the database system-independent names as well as the database system-specific ones for this database. Here is how we can rewire the above transaction using this approach:

#include "person-odb-sqlite.hxx"

odb::sqlite::database& db = ...

{
  using namespace odb::sqlite::core;

  typedef query<person> person_query;
  typedef result<person> person_result;

  transaction t (db.begin ());
  person_result r (db.query<person> (person_query::age < 30));
  ...
  t.commit ();
}
  

If the using namespace directive cannot be used, for example, because the same code fragment accesses several databases, then we can still make the namespace qualifications more concise by assigning shorter aliases to database namespaces. For example:

#include "person-odb-pgsql.hxx"
#include "person-odb-sqlite.hxx"

namespace pg = odb::pgsql;
namespace sl = odb::sqlite;

pg::database& pg_db = ...
sl::database& sl_db = ...

typedef pg::query<person> pg_query;
typedef sl::query<person> sl_query;
typedef odb::result<person> result;

// First check the local cache.
//
odb::transaction t (sl_db.begin ()); // Note: using common transaction.
result r (sl_db.query<person> (sl_query::age < 30));

// If no hits, try the remote database.
//
if (r.empty ())
{
  t.commit ();              // End the SQLite transaction.
  t.reset (pg_db.begin ()); // Start the PostgreSQL transaction.

  r = pg_db.query<person> (pg_query::age < 30);
}

// Handle the result.
//
...

t.commit ();
  

With static multi-database support we can make one of the databases the default database with the --default-database option. The default database can be accessed via the common interface, just like with single-database support. For example:

odb -m static -d common -d pgsql -d sqlite --default-database pgsql ...
  

The default database mechanism can be useful when one of the databases is primary or when retrofitting multi-database support into an existing single-database application. For example, if we are adding SQLite as a local cache into an existing application that uses PostgreSQL as its only database, then by making PostgreSQL the default database we avoid having to change all the existing code. Note that if dynamic multi-database support is enabled, then the common (dynamic) interface is always made the default database.

16.2 Dynamic Multi-Database Support

With dynamic multi-database support, application source code only needs to include the person-odb.hxx header file, just like with single-database support. In particular, we don't need to include any of the person-odb-<db>.hxx files unless we would also like to use certain database systems in the static multi-database mode.

When performing database operations, the application code uses the common interfaces from the odb namespace, just like with single-database support. As an example, consider a function that can be used to load an object either from a local SQLite cache or a remote PostgreSQL database (in reality, this function can be used with any database system support by ODB provided we generated the database support code for this database and linked it into our application):

#include "person-odb.hxx"

std::unique_ptr<person>
load (odb::database& db, const std::string& name)
{
  odb::transaction t (db.begin ());
  std::unique_ptr<person> p (db.find (name));
  t.commit ();
  return p;
}

odb::pgsql::database& pg_db = ...
odb::sqlite::database& sl_db = ...

// First try the local cache.
//
std::unique_ptr<person> p (load (sl_db, "John Doe"));

// If not found, try the remote database.
//
if (p == 0)
  p = load (pg_db, "John Doe");

...
  

As you can see, we can use dynamic multi-database support just like single-database support except that now our code can work with different database systems. Note, however, one difference: with single-database support we could perform database operations using either the common odb::database or a database system-specific (for example, odb::sqlite::database) interface with the same effect. In contrast, with dynamic multi-database support, the use of the database system-specific interface results in the switch to the static mode (for which, as was mentioned earlier, we would need to include the corresponding person-odb-<db>.hxx header file). As we will discuss shortly, switching from dynamic to static mode can be used to overcome limitations imposed by dynamic multi-database support.

Dynamic multi-database support has certain overheads and limitations compared to static support. For database operations, the generated code maintains function tables that are used to dispatch calls to the database system-specific implementations. In single-database and static multi-database support, the query type implements a thin wrapper around the underlying database system's SELECT statement. With dynamic multi-database support, because the underlying database system is only known at query execution (or preparation) time, the query type stores a database system-independent representation of the query that is then translated to the database system-specific form. Because of this database system-independent representation, dynamic support queries have a number of limitations. Specifically, dynamic queries do not support parameter binding in native query fragments. They also make copies of by-value parameterd (by-reference parameters can be used to remove this overhead). Finally, parameters of array types (for example, char[256]) can only be bound by-reference.

As we mentioned earlier, switching from dynamic to static mode can be an effective way to overcome these limitations. As an example, consider a function that prints the list of people of a certain age. The caller also specified the limit on the number of entries to print. Some database systems, for example, PostgreSQL, allow us to propagate this limit to the database server with the LIMIT clause. To add this clause we would need to construct a native query fragment and, as we discussed above, we won't be able to bind a parameter (the limit) while in the dynamic mode. The following implementation shows how we can overcome this by switching to the static mode and using the PostgreSQL-specific interface:

#include "person-odb.hxx"
#include "person-odb-pgsql.hxx" // Needed for static mode.

void
print (odb::database& db, unsigned short age, unsigned long limit)
{
  typedef odb::query<person> query;
  typedef odb::result<person> result;

  odb::transaction t (db.begin ());

  query q (query::age == age);
  result r;

  if (db.id () == odb::id_pgsql)
  {
    // We are using PostgreSQL. Drop down to the static mode and
    // add the LIMIT clause to the query.
    //
    namespace pg = odb::pgsql;
    typedef pg::query<person> pg_query;

    pg::database& pg_db (static_cast<pg::database&> (db));
    pg_query pg_q (pg_query (q) + "LIMIT" + pg_query::_val (limit));
    r = pg_db.query<person> (pg_q);
  }
  else
    r = db.query<person> (q);

  // Handle the result up to the limit elements.
  //
  ...

  t.commit ();
}

odb::pgsql::database& pg_db = ...
odb::sqlite::database& sl_db = ...

print (sl_db, 30, 100);
print (sl_db, 30, 100);
  

A few things to note about this example. First, we use the database::id() function to determine the actual database system we use. This function has the following signature:

namespace odb
{
  enum database_id
  {
    id_mysql,
    id_sqlite,
    id_pgsql,
    id_oracle,
    id_mssql,
    id_common
  };

  class database
  {
  public:
    ...

    database_id
    id () const;
  }
}
  

Note that database::id() can never return the id_common value.

The other thing to note is how we translate the dynamic query to the database system-specific one (the pg_query (q) expression). Every odb::<db>::query class provides such a translation constructor.

16.2.2 Dynamic Loading of Database Support Code

With dynamic multi-database support, the generated database support code automatically registers itself with the function tables that we mentioned earlier. This makes it possible to package the generated code for each database into a separate dynamic-link library (Windows DLL) or dynamic shared object (Unix DSO; collectively referred to as DLLs from now on) and load/unload them from the application dynamically using APIs such as Win32 LoadLibrary() or POSIX dlopen(). This allows the application address space to contain code only for database systems that are actually needed in any particular moment. Another advantage of this approach is the ability to distribute individual database system support separately.

This section provides an overview of how to package the generated database support code into DLLs for both Windows and Unix using GNU/Linux as an example. Note also that if static multi-database support is used for a particular database system, then the dynamic loading cannot be used for this database. It is, however, still possible to package the generated code into a DLL but this DLL will have to be linked to the executable at link-time rather than at runtime. If dynamic loading is desirable in this situation, then another alternative would be to package the functionality that requires static support together with the database support code into the DLL and import this functionality dynamically using the GetProcAddress() (Win32) or dlsym() (Unix) function.

The first step in packaging the generated code into DLLs is to set up the symbol exporting. This step is required for Windows DLLs but is optional for Unix DSOs. Most modern Unix systems (such as GNU/Linux) provide control over symbol visibility, which is a mechanism similar to Windows symbol exporting. Notable advantages of using this mechanism to explicitly specify which symbols are visible include smaller Unix DSOs and faster load times. If, however, you are not planning to control symbol visibility on Unix, then you can skip directly to the second step below.

An important point to understand is that we only need to export the common interface, that is, the classes defined in the person-odb.hxx header. In particular, we don't need to export the database system-specific classes defined in the person-odb-<db>.hxx, unless we are also using this database in the static mode (in which case, the procedure described below will need to be repeated for that database as well).

The ODB compiler provides two command line options, --export-symbol and --extern-symbol, which can be used to insert the export and extern macros in all the necessary places in the generated header file. You are probably familiar with the concept of export macro which expands to an export directive if we are building the DLL and to an import directive if we are building client code. The extern macro is a supplementary mechanism which is necessary to export explicit template instantiations used by the generated code when query support is enabled. As we will see shortly, the extern macro must expand into the extern C++ keyword in certain situations and must be left undefined in others. To manage all these macro definitions, it is customary to create the so called export header. Based on a single macro that is normally defined in the project file or on the command line and which indicates whether we are building the DLL or client code, the export header file sets the export and extern macros to their appropriate values. Continuing with our person example, on Windows the export header, which we will call person-export.hxx, could look like this:

// person-export.hxx
//
// Define PERSON_BUILD_DLL if we are building the DLL. Leave it
// undefined in client code.
//
#ifndef PERSON_EXPORT_HXX
#define PERSON_EXPORT_HXX

#ifdef PERSON_BUILD_DLL
#  define PERSON_EXPORT __declspec(dllexport)
#else
#  define PERSON_EXPORT __declspec(dllimport)
#  define PERSON_EXTERN extern
#endif

#endif // PERSON_EXPORT_HXX
  

The equivalent export header for GCC on GNU/Linux is shown below. Note also that on GNU/Linux, by default, all symbols are visible and we need to add the GCC -fvisibility=hidden option to make them hidden by default.

// person-export.hxx
//
#ifndef PERSON_EXPORT_HXX
#define PERSON_EXPORT_HXX

#define PERSON_EXPORT __attribute__ ((visibility ("default")))
#define PERSON_EXTERN extern

#endif // PERSON_EXPORT_HXX
  

Next we need to export the person persistent class using the export macro and re-compile our person.hxx file with the --export-symbol and --extern-symbol options. We will also need to include person-export.hxx into the generated person-odb.hxx file. For that we use the --hxx-prologue option. Here is how we can do this with multiple invocations of the ODB compiler:

odb -m dynamic -d common --hxx-prologue "#include \"person-export.hxx\"" \
--export-symbol PERSON_EXPORT --extern-symbol PERSON_EXTERN person.hxx

odb -m dynamic -d sqlite person.hxx
odb -m dynamic -d pgsql person.hxx
  

It is also possible to achieve the same with a single invocation. Here we need to restrict some option values to apply only to the common database:

odb -m dynamic -d common -d sqlite -d pgsql \
--hxx-prologue "common:#include \"person-export.hxx\"" \
--export-symbol common:PERSON_EXPORT --extern-symbol common:PERSON_EXTERN \
person.hxx
  

The second step in packaging the generated code into DLLs is to decide where to place the generated common interface code. One option is to place it into a DLL of its own so that we will end up with (replace *.dll with lib*.so for Unix): person.dll plus person-sqlite.dll and person-pgsql.dll, which both link to person.dll, as well as person.exe, which links to person.dll and dynamically loads person-sqlite.dll and/or person-pgsql.dll. If this is the organization that you prefer, then the next step is to build all the DLLs as you normally would any other DLL, placing person-odb.cxx and person.cxx into person.dll, person-odb-sqlite.cxx into person-sqlite.dll, etc. Note that in the pure dynamic multi-database support, person-sqlite.dll and person-pgsql.dll do not export any symbols.

We can improve on the above organization by getting rid of person.dll, which is not really necessary unless we have multiple executables sharing the same database support. To achieve this, we will place person-odb.cxx into person.exe and export its symbols from the executable instead of a DLL. Exporting symbols from an executable is a seldom used functionality, especially on Windows, however, it is well supported on both Windows and most Unix platforms. Note also that this approach won't work if we also use one of the databases in the static mode.

On Windows all we have to do is place person-odb.cxx into the executable and compile it as we would in a DLL (that is, with the PERSON_BUILD_DLL macro defined). If Windows linker detects that an executable exports any symbols, then it will automatically create the corresponding import library (person.lib in our case). We then use this import library to build person-sqlite.dll and person-pgsql.dll as before.

To export symbols from an executable on GNU/Linux all we need to do is add the -rdynamic option when linking our executable.


17 MySQL Database

To generate support code for the MySQL database you will need to pass the "--database mysql" (or "-d mysql") option to the ODB compiler. Your application will also need to link to the MySQL ODB runtime library (libodb-mysql). All MySQL-specific ODB classes are defined in the odb::mysql namespace.

17.1 MySQL Type Mapping

The following table summarizes the default mapping between basic C++ value types and MySQL database types. This mapping can be customized on the per-type and per-member basis using the ODB Pragma Language (Chapter 14, "ODB Pragma Language").

C++ Type MySQL Type Default NULL Semantics
bool TINYINT(1) NOT NULL
char CHAR(1) NOT NULL
signed char TINYINT NOT NULL
unsigned char TINYINT UNSIGNED NOT NULL
short SMALLINT NOT NULL
unsigned short SMALLINT UNSIGNED NOT NULL
int INT NOT NULL
unsigned int INT UNSIGNED NOT NULL
long BIGINT NOT NULL
unsigned long BIGINT UNSIGNED NOT NULL
long long BIGINT NOT NULL
unsigned long long BIGINT UNSIGNED NOT NULL
float FLOAT NOT NULL
double DOUBLE NOT NULL
std::string TEXT/VARCHAR(255) NOT NULL
char[N] VARCHAR(N-1) NOT NULL

It is possible to map the char C++ type to an integer database type (for example, TINYINT) using the db type pragma (Section 14.4.3, "type").

Note that the std::string type is mapped differently depending on whether a member of this type is an object id or not. If the member is an object id, then for this member std::string is mapped to the VARCHAR(255) MySQL type. Otherwise, it is mapped to TEXT.

Additionally, by default, C++ enums and C++11 enum classes are automatically mapped to suitable MySQL types. Contiguous enumerations with the zero first enumerator are mapped to the MySQL ENUM type. All other enumerations are mapped to the MySQL types corresponding to their underlying integral types (see table above). In both cases the default NULL semantics is NOT NULL. For example:

enum color {red, green, blue};
enum class taste: unsigned char
{
  bitter = 1, // Non-zero first enumerator.
  sweet,
  sour = 4,   // Non-contiguous.
  salty
};

#pragma db object
class object
{
  ...

  color color_; // Mapped to ENUM ('red', 'green', 'blue') NOT NULL.
  taste taste_; // Mapped to TINYNT UNSIGNED NOT NULL.
};
  

It is also possible to add support for additional MySQL types, such as geospatial types. For more information, refer to Section 14.8, "Database Type Mapping Pragmas".

17.1.1 String Type Mapping

The MySQL ODB runtime library provides support for mapping the std::string, char[N], and std::array<char, N> types to the MySQL CHAR, VARCHAR, TEXT, NCHAR, and NVARCHAR types. However, these mappings are not enabled by default (in particular, by default, std::array will be treated as a container). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("CHAR(2)")
  char state_[2];

  #pragma db type("VARCHAR(128)")
  std::string name_;
};
  

Alternatively, this can be done on the per-type basis, for example:

#pragma db value(std::string) type("VARCHAR(128)")

#pragma db object
class object
{
  ...

  std::string name_; // Mapped to VARCHAR(128).
};
  

The char[N] and std::array<char, N> values may or may not be zero-terminated. When extracting such values from the database, ODB will append the zero terminator if there is enough space.

17.1.2 Binary Type Mapping

The MySQL ODB runtime library provides support for mapping the std::vector<char>, std::vector<unsigned char>, char[N], unsigned char[N], std::array<char, N>, and std::array<unsigned char, N> types to the MySQL BINARY, VARBINARY, and BLOB types. However, these mappings are not enabled by default (in particular, by default, std::vector and std::array will be treated as containers). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("BLOB")
  std::vector<char> buf_;

  #pragma db type("BINARY(16)")
  unsigned char uuid_[16];
};
  

Alternatively, this can be done on the per-type basis, for example:

typedef std::vector<char> buffer;
#pragma db value(buffer) type("BLOB")

#pragma db object
class object
{
  ...

  buffer buf_; // Mapped to BLOB.
};
  

Note also that in native queries (Chapter 4, "Querying the Database") char[N] and std::array<char, N> parameters are by default passed as a string rather than a binary. To pass such parameters as a binary, we need to specify the database type explicitly in the _val()/_ref() calls. Note also that we don't need to do this for the integrated queries, for example:

char u[16] = {...};

db.query<object> ("uuid = " + query::_val<odb::mysql::id_blob> (u));
db.query<object> (query::uuid == query::_ref (u));
  

17.2 MySQL Database Class

The MySQL database class has the following interface:

namespace odb
{
  namespace mysql
  {
    class database: public odb::database
    {
    public:
      database (const char* user,
                const char* passwd,
                const char* db,
                const char* host = 0,
                unsigned int port = 0,
                const char* socket = 0,
                const char* charset = 0,
                unsigned long client_flags = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string& passwd,
                const std::string& db,
                const std::string& host = "",
                unsigned int port = 0,
                const std::string* socket = 0,
                const std::string& charset = "",
                unsigned long client_flags = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string* passwd,
                const std::string& db,
                const std::string& host = "",
                unsigned int port = 0,
                const std::string* socket = 0,
                const std::string& charset = "",
                unsigned long client_flags = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string& passwd,
                const std::string& db,
                const std::string& host,
                unsigned int port,
                const std::string& socket,
                const std::string& charset = "",
                unsigned long client_flags = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string* passwd,
                const std::string& db,
                const std::string& host,
                unsigned int port,
                const std::string& socket,
                const std::string& charset = "",
                unsigned long client_flags = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (int& argc,
                char* argv[],
                bool erase = false,
                const std::string& charset = "",
                unsigned long client_flags = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      static void
      print_usage (std::ostream&);

    public:
      const char*
      user () const;

      const char*
      password () const;

      const char*
      db () const;

      const char*
      host () const;

      unsigned int
      port () const;

      const char*
      socket () const;

      const char*
      charset () const;

      unsigned long
      client_flags () const;

    public:
      connection_ptr
      connection ();
    };
  }
}
  

You will need to include the <odb/mysql/database.hxx> header file to make this class available in your application.

The overloaded database constructors allow us to specify MySQL database parameters that should be used when connecting to the database. In MySQL NULL and an empty string are treated as the same values for all the string parameters except password and socket.

The charset argument allows us to specify the client character set, that is, the character set in which the application will encode its text data. Note that this can be different from the MySQL server character set. If this argument is not specified or is empty, then the default MySQL client character set is used, normally latin1. Commonly used values for this argument are latin1 (equivalent to Windows cp1252 and similar to ISO-8859-1) and utf8. For other possible values as well as more information on character set support in MySQL, refer to the MySQL documentation.

The client_flags argument allows us to specify various MySQL client library flags. For more information on the possible values, refer to the MySQL C API documentation. The CLIENT_FOUND_ROWS flag is always set by the MySQL ODB runtime regardless of whether it was passed in the client_flags argument.

The last constructor extracts the database parameters from the command line. The following options are recognized:

  --user <login>
  --password <password>
  --database <name>
  --host <host>
  --port <integer>
  --socket <socket>
  --options-file <file>
  

The --options-file option allows us to specify some or all of the database options in a file with each option appearing on a separate line followed by a space and an option value.

If the erase argument to this constructor is true, then the above options are removed from the argv array and the argc count is updated accordingly. This is primarily useful if your application accepts other options or arguments and you would like to get the MySQL options out of the argv array.

This constructor throws the odb::mysql::cli_exception exception if the MySQL option values are missing or invalid. See section Section 17.4, "MySQL Exceptions" for more information on this exception.

The static print_usage() function prints the list of options with short descriptions that are recognized by this constructor.

The last argument to all of the constructors is a pointer to the connection factory. In C++98/03, it is std::auto_ptr while in C++11 std::unique_ptr is used instead. If we pass a non-NULL value, the database instance assumes ownership of the factory instance. The connection factory interface as well as the available implementations are described in the next section.

The set of accessor functions following the constructors allows us to query the parameters of the database instance.

The connection() function returns a pointer to the MySQL database connection encapsulated by the odb::mysql::connection class. For more information on mysql::connection, refer to Section 17.3, "MySQL Connection and Connection Factory".

17.3 MySQL Connection and Connection Factory

The mysql::connection class has the following interface:

namespace odb
{
  namespace mysql
  {
    class connection: public odb::connection
    {
    public:
      connection (database&);
      connection (database&, MYSQL*);

      MYSQL*
      handle ();
    };

    typedef details::shared_ptr<connection> connection_ptr;
  }
}
  

For more information on the odb::connection interface, refer to Section 3.6, "Connections". The first overloaded mysql::connection constructor establishes a new MySQL connection. The second constructor allows us to create a connection instance by providing an already connected native MySQL handle. Note that the connection instance assumes ownership of this handle. The handle() accessor returns the MySQL handle corresponding to the connection.

The mysql::connection_factory abstract class has the following interface:

namespace odb
{
  namespace mysql
  {
    class connection_factory
    {
    public:
      virtual void
      database (database&) = 0;

      virtual connection_ptr
      connect () = 0;
    };
  }
}
  

The database() function is called when a connection factory is associated with a database instance. This happens in the odb::mysql::database class constructors. The connect() function is called whenever a database connection is requested.

The two implementations of the connection_factory interface provided by the MySQL ODB runtime are new_connection_factory and connection_pool_factory. You will need to include the <odb/mysql/connection-factory.hxx> header file to make the connection_factory interface and these implementation classes available in your application.

The new_connection_factory class creates a new connection whenever one is requested. When a connection is no longer needed, it is released and closed. The new_connection_factory class has the following interface:

namespace odb
{
  namespace mysql
  {
    class new_connection_factory: public connection_factory
    {
    public:
      new_connection_factory ();
    };
};
  

The connection_pool_factory class implements a connection pool. It has the following interface:

namespace odb
{
  namespace mysql
  {
    class connection_pool_factory: public connection_factory
    {
    public:
      connection_pool_factory (std::size_t max_connections = 0,
                               std::size_t min_connections = 0,
                               bool ping = true);

    protected:
      class pooled_connection: public connection
      {
      public:
        pooled_connection (database_type&);
        pooled_connection (database_type&, MYSQL*);
      };

      typedef details::shared_ptr<pooled_connection> pooled_connection_ptr;

      virtual pooled_connection_ptr
      create ();
    };
};
  

The max_connections argument in the connection_pool_factory constructor specifies the maximum number of concurrent connections that this pool factory will maintain. Similarly, the min_connections argument specifies the minimum number of available connections that should be kept open. The ping argument specifies whether the factory should validate the connection before returning it to the caller.

Whenever a connection is requested, the pool factory first checks if there is an unused connection that can be returned. If there is none, the pool factory checks the max_connections value to see if a new connection can be created. If the total number of connections maintained by the pool is less than this value, then a new connection is created and returned. Otherwise, the caller is blocked until a connection becomes available.

When a connection is released, the pool factory first checks if there are blocked callers waiting for a connection. If so, then one of them is unblocked and is given the connection. Otherwise, the pool factory checks whether the total number of connections maintained by the pool is greater than the min_connections value. If that's the case, the connection is closed. Otherwise, the connection is added to the pool of available connections to be returned on the next request. In other words, if the number of connections maintained by the pool exceeds min_connections and there are no callers waiting for a new connection, then the pool will close the excess connections.

If the max_connections value is 0, then the pool will create a new connection whenever all of the existing connections are in use. If the min_connections value is 0, then the pool will never close a connection and instead maintain all the connections that were ever created.

Connection validation (the ping argument) is useful if your application may experience long periods of inactivity. In such cases the MySQL server may close network connections that have been inactive for too long. If during connection validation the pool factory detects that the connection has been terminated, it silently closes it and tries to find or create another connection instead.

The create() virtual function is called whenever the pool needs to create a new connection. By deriving from the connection_pool_factory class and overriding this function we can implement custom connection establishment and configuration.

If you pass NULL as the connection factory to one of the database constructors, then the connection_pool_factory instance will be created by default with the min and max connections values set to 0 and connection validation enabled. The following code fragment shows how we can pass our own connection factory instance:

#include <odb/database.hxx>

#include <odb/mysql/database.hxx>
#include <odb/mysql/connection-factory.hxx>

int
main (int argc, char* argv[])
{
  auto_ptr<odb::mysql::connection_factory> f (
    new odb::mysql::connection_pool_factory (20));

  auto_ptr<odb::database> db (
    new mysql::database (argc, argv, false, 0, f));
}
  

17.4 MySQL Exceptions

The MySQL ODB runtime library defines the following MySQL-specific exceptions:

namespace odb
{
  namespace mysql
  {
    class database_exception: odb::database_exception
    {
    public:
      unsigned int
      error () const;

      const std::string&
      sqlstate () const;

      const std::string&
      message () const;

      virtual const char*
      what () const throw ();
    };

    class cli_exception: odb::exception
    {
    public:
      virtual const char*
      what () const throw ();
    };
  }
}
  

You will need to include the <odb/mysql/exceptions.hxx> header file to make these exceptions available in your application.

The odb::mysql::database_exception is thrown if a MySQL database operation fails. The MySQL-specific error information is accessible via the error(), sqlstate(), and message() functions. All this information is also combined and returned in a human-readable form by the what() function.

The odb::mysql::cli_exception is thrown by the command line parsing constructor of the odb::mysql::database class if the MySQL option values are missing or invalid. The what() function returns a human-readable description of an error.

17.5 MySQL Limitations

The following sections describe MySQL-specific limitations imposed by the current MySQL and ODB runtime versions.

17.5.1 Foreign Key Constraints

ODB relies on standard SQL behavior which requires that foreign key constraints checking is deferred until the transaction is committed. The only behaviors supported by MySQL are to either check such constraints immediately (InnoDB engine) or to ignore foreign key constraints altogether (all other engines). As a result, by default, schemas generated by the ODB compiler for MySQL have foreign key definitions commented out. They are retained only for documentation.

You can override the default behavior and instruct the ODB compiler to generate non-deferrable foreign keys by specifying the --fkeys-deferrable-mode not_deferrable ODB compiler option. Note, however, that in this case the order in which you persist, update, and erase objects within a transaction becomes important.

17.6 MySQL Index Definitions

When the index pragma (Section 14.7, "Index Definition Pragmas") is used to define a MySQL index, the type clause specifies the index type (for example, UNIQUE, FULLTEXT, SPATIAL), the method clause specifies the index method (for example, BTREE, HASH), and the options clause is not used. The column options can be used to specify column length limits and the sort order. For example:

#pragma db object
class object
{
  ...

  std::string name_;

  #pragma db index method("HASH") member(name_, "(100) DESC")
};
  

17.7 MySQL Stored Procedures

ODB native views (Section 10.6, "Native Views") can be used to call MySQL stored procedures. For example, assuming we are using the person class from Chapter 2, "Hello World Example" (and the corresponding person table), we can create a stored procedure that given the min and max ages returns some information about all the people in that range:

CREATE PROCEDURE person_range (
  IN min_age SMALLINT,
  IN max_age SMALLINT)
BEGIN
  SELECT age, first, last FROM person
    WHERE age >= min_age AND age <= max_age;
END
  

Given the above stored procedure we can then define an ODB view that can be used to call it and retrieve its result:

#pragma db view query("CALL person_range((?))")
struct person_range
{
  unsigned short age;
  std::string first;
  std::string last;
};
  

The following example shows how we can use the above view to print the list of people in a specific age range:

typedef odb::query<person_range> query;
typedef odb::result<person_range> result;

transaction t (db.begin ());

result r (
  db.query<person_range> (
    query::_val (1) + "," + query::_val (18)));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
  cerr << i->first << " " << i->last << " " << i->age << endl;

t.commit ();
  

Note that as with all native views, the order and types of data members must match those of columns in the SELECT list inside the stored procedure.

There are also a number of limitations when it comes to support for MySQL stored procedures in ODB views. First of all, you have to use MySQL server and client libraries version 5.5.3 or later since this is the version in which support for calling stored procedures with prepared statements was first added (the mysql_stmt_next_result() function).

In MySQL, a stored procedure can produce multiple results. For example, if a stored procedure executes several SELECT statements, then the result of calling such a procedure consists of two row sets, one for each SELECT statement. Additionally, if the procedure has any OUT or INOUT parameters, then their values are returned as an additional special row set containing only a single row. Because such multiple row sets can contain varying number and type of columns, they cannot be all extracted into a single view. As a result, an ODB view will only extract the data from the first row set and ignore all the subsequent ones.

In particular, this means that we can use an ODB view to extract the values of the OUT and INOUT parameters provided that the stored procedure does not generate any other row sets. For example:

CREATE PROCEDURE person_min_max_age (
  OUT min_age SMALLINT,
  OUT max_age SMALLINT)
BEGIN
  SELECT MIN(age), MAX(age) INTO min_age, max_age FROM person;
END
  
#pragma db view query("CALL person_min_max_age((?))")
struct person_min_max_age
{
  unsigned short min_age;
  unsigned short max_age;
};
  
typedef odb::query<person_min_max_age> query;

transaction t (db.begin ());

// We know this query always returns a single row, so use query_value().
// We have to pass dummy values for OUT parameters.
//
person_min_max_age mma (
  db.query_value<person_min_max_age> (
    query::_val (0) + "," + query::_val (0)));

cerr << mma.min_age << " " << mma.max_age << endl;

t.commit ();
  

Another limitation that stems from having multiple results is the inability to cache the result of a stored procedure call. In other words, a MySQL stored procedure call always produces an uncached query result (Section 4.4, "Query Result").


18 SQLite Database

To generate support code for the SQLite database you will need to pass the "--database sqlite" (or "-d sqlite") option to the ODB compiler. Your application will also need to link to the SQLite ODB runtime library (libodb-sqlite). All SQLite-specific ODB classes are defined in the odb::sqlite namespace.

18.1 SQLite Type Mapping

The following table summarizes the default mapping between basic C++ value types and SQLite database types. This mapping can be customized on the per-type and per-member basis using the ODB Pragma Language (Chapter 14, "ODB Pragma Language").

C++ Type SQLite Type Default NULL Semantics
bool INTEGER NOT NULL
char TEXT NOT NULL
signed char INTEGER NOT NULL
unsigned char INTEGER NOT NULL
short INTEGER NOT NULL
unsigned short INTEGER NOT NULL
int INTEGER NOT NULL
unsigned int INTEGER NOT NULL
long INTEGER NOT NULL
unsigned long INTEGER NOT NULL
long long INTEGER NOT NULL
unsigned long long INTEGER NOT NULL
float REAL NULL
double REAL NULL
std::string TEXT NOT NULL
char[N] TEXT NOT NULL
std::wstring (Windows only) TEXT NOT NULL
wchar_t[N] (Windows only) TEXT NOT NULL

It is possible to map the char C++ type to the INTEGER SQLite type using the db type pragma (Section 14.4.3, "type").

SQLite represents the NaN FLOAT value as a NULL value. As a result, columns of the float and double types are by default declared as NULL. However, you can override this by explicitly declaring them as NOT NULL with the db not_null pragma (Section 14.4.6, "null/not_null").

Additionally, by default, C++ enums and C++11 enum classes are automatically mapped to the SQLite INTEGER type with the default NULL semantics being NOT NULL. For example:

enum color {red, green, blue};
enum class taste: unsigned char
{
  bitter = 1,
  sweet,
  sour = 4,
  salty
};

#pragma db object
class object
{
  ...

  color color_; // Automatically mapped to INTEGER.
  taste taste_; // Automatically mapped to INTEGER.
};
  

Note also that SQLite only operates with signed integers and the largest value that an SQLite database can store is a signed 64-bit integer. As a result, greater unsigned long and unsigned long long values will be represented in the database as negative values.

It is also possible to add support for additional SQLite types, such as NUMERIC. For more information, refer to Section 14.8, "Database Type Mapping Pragmas".

18.1.1 String Type Mapping

The SQLite ODB runtime library provides support for mapping the std::array<char, N> and, on Windows, std::array<wchar_t, N> types to the SQLite TEXT type. However, this mapping is not enabled by default (in particular, by default, std::array will be treated as a container). To enable the alternative mapping for this type we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("TEXT")
  std::array<char, 128> name_;
};
  

Alternatively, this can be done on the per-type basis, for example:

typedef std::array<char, 128> name_type;
#pragma db value(name_type) type("TEXT")

#pragma db object
class object
{
  ...

  name_type name_; // Mapped to TEXT.
};
  

The char[N], std::array<char, N>, wchar_t[N], and std::array<wchar_t, N> values may or may not be zero-terminated. When extracting such values from the database, ODB will append the zero terminator if there is enough space.

18.1.2 Binary Type Mapping

The SQLite ODB runtime library provides support for mapping the std::vector<char>, std::vector<unsigned char>, char[N], unsigned char[N], std::array<char, N>, and std::array<unsigned char, N> types to the SQLite BLOB type. However, these mappings are not enabled by default (in particular, by default, std::vector and std::array will be treated as containers). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("BLOB")
  std::vector<char> buf_;

  #pragma db type("BLOB")
  unsigned char uuid_[16];
};
  

Alternatively, this can be done on the per-type basis, for example:

typedef std::vector<char> buffer;
#pragma db value(buffer) type("BLOB")

#pragma db object
class object
{
  ...

  buffer buf_; // Mapped to BLOB.
};
  

Note also that in native queries (Chapter 4, "Querying the Database") char[N] and std::array<char, N> parameters are by default passed as a string rather than a binary. To pass such parameters as a binary, we need to specify the database type explicitly in the _val()/_ref() calls. Note also that we don't need to do this for the integrated queries, for example:

char u[16] = {...};

db.query<object> ("uuid = " + query::_val<odb::sqlite::id_blob> (u));
db.query<object> (query::uuid == query::_ref (u));
  

18.2 SQLite Database Class

The SQLite database class has the following interface:

namespace odb
{
  namespace sqlite
  {
    class database: public odb::database
    {
    public:
      database (const std::string& name,
                int flags = SQLITE_OPEN_READWRITE,
                bool foreign_keys = true,
                const std::string& vfs = "",
                std::[auto|unique]_ptr<connection_factory> = 0);

#ifdef _WIN32
      database (const std::wstring& name,
                int flags = SQLITE_OPEN_READWRITE,
                bool foreign_keys = true,
                const std::string& vfs = "",
                std::[auto|unique]_ptr<connection_factory> = 0);
#endif

      database (int& argc,
                char* argv[],
                bool erase = false,
                int flags = SQLITE_OPEN_READWRITE,
                bool foreign_keys = true,
                const std::string& vfs = "",
                std::[auto|unique]_ptr<connection_factory> = 0);

      static void
      print_usage (std::ostream&);

    public:
      const std::string&
      name () const;

      int
      flags () const;

    public:
      transaction
      begin_immediate ();

      transaction
      begin_exclusive ();

    public:
      connection_ptr
      connection ();
    };
  }
}
  

You will need to include the <odb/sqlite/database.hxx> header file to make this class available in your application.

The first constructor opens the specified SQLite database. The name argument is the database file name to open in the UTF-8 encoding. If this argument is empty, then a temporary, on-disk database is created. If this argument is the :memory: special value, then a temporary, in-memory database is created. The flags argument allows us to specify SQLite opening flags. For more information on the possible values, refer to the sqlite3_open_v2() function description in the SQLite C API documentation. The foreign_keys argument specifies whether foreign key constraints checking should be enabled. See Section 18.5.3, "Foreign Key Constraints" for more information on foreign keys. The vfs argument specifies the SQLite virtual file system module that should be used to access the database. If this argument is empty, then the default vfs module is used. Again, refer to the sqlite3_open_v2() function documentation for detail.

The following example shows how we can open the test.db database in the read-write mode and create it if it does not exist:

auto_ptr<odb::database> db (
  new odb::sqlite::database (
    "test.db",
    SQLITE_OPEN_READWRITE | SQLITE_OPEN_CREATE));
  

The second constructor is the same as the first except that the database name is passes as std::wstring in the UTF-16 encoding. This constructor is only available when compiling for Windows.

The third constructor extracts the database parameters from the command line. The following options are recognized:

  --database <name>
  --create
  --read-only
  --options-file <file>
  

By default, this constructor opens the database in the read-write mode (SQLITE_OPEN_READWRITE flag). If the --create flag is specified, then the database file is created if it does not already exist (SQLITE_OPEN_CREATE flag). If the --read-only flag is specified, then the database is opened in the read-only mode (SQLITE_OPEN_READONLY flag instead of SQLITE_OPEN_READWRITE). The --options-file option allows us to specify some or all of the database options in a file with each option appearing on a separate line followed by a space and an option value.

If the erase argument to this constructor is true, then the above options are removed from the argv array and the argc count is updated accordingly. This is primarily useful if your application accepts other options or arguments and you would like to get the SQLite options out of the argv array.

The flags argument has the same semantics as in the first constructor. Flags from the command line always override the corresponding values specified with this argument.

The third constructor throws the odb::sqlite::cli_exception exception if the SQLite option values are missing or invalid. See Section 18.4, "SQLite Exceptions" for more information on this exception.

The static print_usage() function prints the list of options with short descriptions that are recognized by the third constructor.

The last argument to all of the constructors is a pointer to the connection factory. In C++98/03, it is std::auto_ptr while in C++11 std::unique_ptr is used instead. If we pass a non-NULL value, the database instance assumes ownership of the factory instance. The connection factory interface as well as the available implementations are described in the next section.

The set of accessor functions following the constructors allows us to query the parameters of the database instance.

The begin_immediate() and begin_exclusive() functions are the SQLite-specific extensions to the standard odb::database::begin() function (see Section 3.5, "Transactions"). They allow us to start an immediate (BEGIN IMMEDIATE) and an exclusive (BEGIN EXCLUSIVE) SQLite transaction, respectively. For more information on the semantics of the immediate and exclusive transactions, refer to the BEGIN statement description in the SQLite documentation.

The connection() function returns a pointer to the SQLite database connection encapsulated by the odb::sqlite::connection class. For more information on sqlite::connection, refer to Section 18.3, "SQLite Connection and Connection Factory".

18.3 SQLite Connection and Connection Factory

The sqlite::connection class has the following interface:

namespace odb
{
  namespace sqlite
  {
    class connection: public odb::connection
    {
    public:
      connection (database&, int extra_flags = 0);
      connection (database&, sqlite3*);

      transaction
      begin_immediate ();

      transaction
      begin_exclusive ();

      sqlite3*
      handle ();
    };

    typedef details::shared_ptr<connection> connection_ptr;
  }
}
  

For more information on the odb::connection interface, refer to Section 3.6, "Connections". The first overloaded sqlite::connection constructor opens a new SQLite connection. The extra_flags argument can be used to specify extra sqlite3_open_v2() flags that are combined with the flags specified in the sqlite::database constructor. The second constructor allows us to create a connection instance by providing an already open native SQLite handle. Note that the connection instance assumes ownership of this handle.

The begin_immediate() and begin_exclusive() functions allow us to start an immediate and an exclusive SQLite transaction on the connection, respectively. Their semantics are equivalent to the corresponding functions defined in the sqlite::database class (Section 18.2, "SQLite Database Class"). The handle() accessor returns the SQLite handle corresponding to the connection.

The sqlite::connection_factory abstract class has the following interface:

namespace odb
{
  namespace sqlite
  {
    class connection_factory
    {
    public:
      virtual void
      database (database&) = 0;

      virtual connection_ptr
      connect () = 0;
    };
  }
}
  

The database() function is called when a connection factory is associated with a database instance. This happens in the odb::sqlite::database class constructors. The connect() function is called whenever a database connection is requested.

The three implementations of the connection_factory interface provided by the SQLite ODB runtime library are single_connection_factory, new_connection_factory, and connection_pool_factory. You will need to include the <odb/sqlite/connection-factory.hxx> header file to make the connection_factory interface and these implementation classes available in your application.

The single_connection_factory class creates a single connection that is shared between all the threads in an application. If the connection is currently not in use, then it is returned to the caller. Otherwise, the caller is blocked until the connection becomes available. The single_connection_factory class has the following interface:

namespace odb
{
  namespace sqlite
  {
    class single_connection_factory: public connection_factory
    {
    public:
      single_connection_factory ();

    protected:
      class single_connection: public connection
      {
      public:
        single_connection (database_type&);
        single_connection (database_type&, MYSQL*);
      };

      typedef details::shared_ptr<single_connection> single_connection_ptr;

      virtual single_connection_ptr
      create ();
    };
};
  

The create() virtual function is called when the factory needs to create the connection. By deriving from the single_connection_factory class and overriding this function we can implement custom connection establishment and configuration.

The new_connection_factory class creates a new connection whenever one is requested. When a connection is no longer needed, it is released and closed. The new_connection_factory class has the following interface:

namespace odb
{
  namespace sqlite
  {
    class new_connection_factory: public connection_factory
    {
    public:
      new_connection_factory ();
    };
};
  

The connection_pool_factory class implements a connection pool. It has the following interface:

namespace odb
{
  namespace sqlite
  {
    class connection_pool_factory: public connection_factory
    {
    public:
      connection_pool_factory (std::size_t max_connections = 0,
                               std::size_t min_connections = 0);

    protected:
      class pooled_connection: public connection
      {
      public:
        pooled_connection (database_type&, int extra_flags = 0);
        pooled_connection (database_type&, sqlite3*);
      };

      typedef details::shared_ptr<pooled_connection> pooled_connection_ptr;

      virtual pooled_connection_ptr
      create ();
    };
};
  

The max_connections argument in the connection_pool_factory constructor specifies the maximum number of concurrent connections that this pool factory will maintain. Similarly, the min_connections argument specifies the minimum number of available connections that should be kept open.

Whenever a connection is requested, the pool factory first checks if there is an unused connection that can be returned. If there is none, the pool factory checks the max_connections value to see if a new connection can be created. If the total number of connections maintained by the pool is less than this value, then a new connection is created and returned. Otherwise, the caller is blocked until a connection becomes available.

When a connection is released, the pool factory first checks if there are blocked callers waiting for a connection. If so, then one of them is unblocked and is given the connection. Otherwise, the pool factory checks whether the total number of connections maintained by the pool is greater than the min_connections value. If that's the case, the connection is closed. Otherwise, the connection is added to the pool of available connections to be returned on the next request. In other words, if the number of connections maintained by the pool exceeds min_connections and there are no callers waiting for a new connection, then the pool will close the excess connections.

If the max_connections value is 0, then the pool will create a new connection whenever all of the existing connections are in use. If the min_connections value is 0, then the pool will never close a connection and instead maintain all the connections that were ever created.

The create() virtual function is called whenever the pool needs to create a new connection. By deriving from the connection_pool_factory class and overriding this function we can implement custom connection establishment and configuration.

By default, connections created by new_connection_factory and connection_pool_factory enable the SQLite shared cache mode and use the unlock notify functionality to aid concurrency. To disable the shared cache mode you can pass the SQLITE_OPEN_PRIVATECACHE flag when creating the database instance. For more information on the shared cache mode refer to the SQLite documentation.

If you pass NULL as the connection factory to one of the database constructors, then the connection_pool_factory instance will be created by default with the min and max connections values set to 0. The following code fragment shows how we can pass our own connection factory instance:

#include <odb/database.hxx>

#include <odb/sqlite/database.hxx>
#include <odb/sqlite/connection-factory.hxx>

int
main (int argc, char* argv[])
{
  auto_ptr<odb::sqlite::connection_factory> f (
    new odb::sqlite::connection_pool_factory (20));

  auto_ptr<odb::database> db (
    new sqlite::database (argc, argv, false, SQLITE_OPEN_READWRITE, f));
}
  

18.4 SQLite Exceptions

The SQLite ODB runtime library defines the following SQLite-specific exceptions:

namespace odb
{
  namespace sqlite
  {
    class forced_rollback: odb::recoverable
    {
    public:
      virtual const char*
      what () const throw ();
    };

    class database_exception: odb::database_exception
    {
    public:
      int
      error () const

      int
      extended_error () const;

      const std::string&
      message () const;

      virtual const char*
      what () const throw ();
    };

    class cli_exception: odb::exception
    {
    public:
      virtual const char*
      what () const throw ();
    };
  }
}
  

You will need to include the <odb/sqlite/exceptions.hxx> header file to make these exceptions available in your application.

The odb::sqlite::forced_rollback exception is thrown if SQLite is forcing the current transaction to roll back. For more information on this behavior refer to Section 18.5.6, "Forced Rollback".

The odb::sqlite::database_exception is thrown if an SQLite database operation fails. The SQLite-specific error information is accessible via the error(), extended_error(), and message() functions. All this information is also combined and returned in a human-readable form by the what() function.

The odb::sqlite::cli_exception is thrown by the command line parsing constructor of the odb::sqlite::database class if the SQLite option values are missing or invalid. The what() function returns a human-readable description of an error.

18.5 SQLite Limitations

The following sections describe SQLite-specific limitations imposed by the current SQLite and ODB runtime versions.

18.5.1 Query Result Caching

SQLite ODB runtime implementation does not perform query result caching (Section 4.4, "Query Result") even when explicitly requested. The SQLite API supports interleaving execution of multiple prepared statements on a single connection. As a result, with SQLite, it is possible to have multiple uncached results and calls to other database functions do not invalidate them. The only limitation of the uncached SQLite results is the unavailability of the result::size() function. If you call this function on an SQLite query result, then the odb::result_not_cached exception (Section 3.14, "ODB Exceptions") is always thrown. Future versions of the SQLite ODB runtime library may add support for result caching.

18.5.2 Automatic Assignment of Object Ids

Due to SQLite API limitations, every automatically assigned object id (Section 14.4.2, "auto") should have the INTEGER SQLite type. While SQLite will treat other integer type names (such as INT, BIGINT, etc.) as INTEGER, automatic id assignment will not work. By default, ODB maps all C++ integral types to INTEGER. This means that the only situation that requires consideration is the assignment of a custom database type using the db type pragma (Section 14.4.3, "type"). For example:

#pragma db object
class person
{
  ...

  //#pragma db id auto type("INT")     // Will not work.
  //#pragma db id auto type("INTEGER") // Ok.
  #pragma db id auto                   // Ok, Mapped to INTEGER.
  unsigned int id_;
};
  

18.5.3 Foreign Key Constraints

By default the SQLite ODB runtime enables foreign key constraints checking (PRAGMA foreign_keys=ON). You can disable foreign keys by passing false as the foreign_keys argument to one of the odb::sqlite::database constructors. Foreign keys will also be disabled if the SQLite library is built without support for foreign keys (SQLITE_OMIT_FOREIGN_KEY and SQLITE_OMIT_TRIGGER macros) or if you are using an SQLite version prior to 3.6.19, which does not support foreign key constraints checking.

If foreign key constraints checking is disabled or not available, then inconsistencies in object relationships will not be detected. Furthermore, using the erase_query() function (Section 3.11, "Deleting Persistent Objects") to delete persistent objects that contain containers will not work correctly. Container data for such objects will not be deleted.

When foreign key constraints checking is enabled, then you may get the "foreign key constraint failed" error while re-creating the database schema. This error is due to bugs in the SQLite DDL foreign keys support. The recommended work-around for this problem is to temporarily disable foreign key constraints checking while re-creating the schema. The following code fragment shows how this can be done:

#include <odb/connection.hxx>
#include <odb/transaction.hxx>
#include <odb/schema-catalog.hxx>

odb::database& db = ...

{
  odb::connection_ptr c (db.connection ());

  c->execute ("PRAGMA foreign_keys=OFF");

  odb::transaction t (c->begin ());
  odb::schema_catalog::create_schema (db);
  t.commit ();

  c->execute ("PRAGMA foreign_keys=ON");
}
  

Finally, ODB assumes the standard SQL behavior which requires that foreign key constraints checking is deferred until the transaction is committed. Default SQLite behavior is to check such constraints immediately. As a result, when used with ODB, a custom database schema that defines foreign key constraints may need to declare such constraints as DEFERRABLE INITIALLY DEFERRED, as shown in the following example. By default, schemas generated by the ODB compiler meet this requirement automatically.

CREATE TABLE Employee (
  ...
  employer INTEGER REFERENCES Employer(id)
           DEFERRABLE INITIALLY DEFERRED);
  

You can override the default behavior and instruct the ODB compiler to generate non-deferrable foreign keys by specifying the --fkeys-deferrable-mode not_deferrable ODB compiler option. Note, however, that in this case the order in which you persist, update, and erase objects within a transaction becomes important.

18.5.4 Constraint Violations

Due to the granularity of the SQLite error codes, it is impossible to distinguish between the duplicate primary key and other constraint violations. As a result, when making an object persistent, the SQLite ODB runtime will translate all constraint violation errors to the object_already_persistent exception (Section 3.14, "ODB Exceptions").

18.5.5 Sharing of Queries

As discussed in Section 4.3, "Executing a Query", a query instance that does not have any by-reference parameters is immutable and can be shared between multiple threads without synchronization. Currently, the SQLite ODB runtime does not support this functionality. Future versions of the library will remove this limitation.

18.5.6 Forced Rollback

In SQLite 3.7.11 or later, if one of the connections participating in the shared cache rolls back a transaction, then ongoing transactions on other connections in the shared cache may also be forced to roll back. An example of such behavior would be a read-only transaction that is forced to roll back while iterating over the query result because another transaction on another connection was rolled back.

If a transaction is being forced to roll back by SQLite, then ODB throws odb::sqlite::forced_rollback (Section 18.4, "SQLite Exceptions") which is a recoverable exception (3.7 Error Handling and Recovery). As a result, the recommended way to handle this exception is to re-execute the affected transaction.

18.5.7 Database Schema Evolution

From the list of schema migration changes supported by ODB (Section 13.2, "Schema Migration"), the following are not supported by SQLite:

The biggest problem is the lack of support for dropping columns. This means that it would be impossible to delete a data member in a persistent class. To work around this limitation ODB implements logical delete for columns that allow NULL values. In this case, instead of dropping the column (in the post-migration stage), the schema migration statements will automatically reset this column in all the existing rows to NULL. Any new rows that are inserted later will also automatically have this column set to NULL (unless the column specifies a default value).

Since it is also impossible to change the column's NULL/NOT NULL attribute after it has been added, to make schema evolution support usable in SQLite, all the columns should be added as NULL even if semantically they should not allow NULL values. We should also normally refrain from assigning default values to columns (Section 14.4.7, default), unless the space overhead of a default value is not a concern. Explicitly making all the data members NULL would be burdensome and ODB provides the --sqlite-override-null command line option that forces all the columns, even those that were explicitly marked NOT NULL, to be NULL in SQLite.

SQLite only supports adding foreign keys as part of the column addition. As a result, we can only add a new data member of an object pointer type if it points to an object with a simple (single-column) object id.

SQLite also doesn't support dropping foreign keys. Leaving a foreign key around works well with logical delete unless we also want to delete the pointed-to object. In this case we will have to leave an empty table corresponding to the pointed-to object around. An alternative would be to make a copy of the pointing object without the object pointer, migrate the data, and then delete both the old pointing and the pointed-to objects. Since this will result in dropping the pointing table, the foreign key will be dropped as well. Yet another, more radical, solution to this problem is to disable foreign keys checking altogether (see the foreign_keys SQLite pragma).

To summarize, to make schema evolution support usable in SQLite we should pass the --sqlite-override-null option when compiling our persistent classes and also refrain from assigning default values to data members. Note also that this has to be done from the start so that every column is added as NULL and therefore can be logically deleted later. In particular, you cannot add the --sqlite-override-null option when you realize you need to delete a data member. At this point it is too late since the column has already been added as NOT NULL in existing databases. We should also avoid composite object ids if we are planning to use object relationships.

18.6 SQLite Index Definitions

When the index pragma (Section 14.7, "Index Definition Pragmas") is used to define an SQLite index, the type clause specifies the index type (for example, UNIQUE) while the method and options clauses are not used. The column options can be used to specify collations and the sort order. For example:

#pragma db object
class object
{
  ...

  std::string name_;

  #pragma db index member(name_, "COLLATE binary DESC")
};
  

Index names in SQLite are database-global. To avoid name clashes, ODB automatically prefixes each index name with the table name on which it is defined.


19 PostgreSQL Database

To generate support code for the PostgreSQL database you will need to pass the "--database pgsql" (or "-d pgsql") option to the ODB compiler. Your application will also need to link to the PostgreSQL ODB runtime library (libodb-pgsql). All PostgreSQL-specific ODB classes are defined in the odb::pgsql namespace.

ODB utilizes prepared statements extensively. Support for prepared statements was added in PostgreSQL version 7.4 with the introduction of the messaging protocol version 3.0. For this reason, ODB supports only PostgreSQL version 7.4 and later.

19.1 PostgreSQL Type Mapping

The following table summarizes the default mapping between basic C++ value types and PostgreSQL database types. This mapping can be customized on the per-type and per-member basis using the ODB Pragma Language (Chapter 14, "ODB Pragma Language").

C++ Type PostgreSQL Type Default NULL Semantics
bool BOOLEAN NOT NULL
char CHAR(1) NOT NULL
signed char SMALLINT NOT NULL
unsigned char SMALLINT NOT NULL
short SMALLINT NULL NOT NULL
unsigned short SMALLINT NOT NULL
int INTEGER NOT NULL
unsigned int INTEGER NOT NULL
long BIGINT NOT NULL
unsigned long BIGINT NOT NULL
long long BIGINT NOT NULL
unsigned long long BIGINT NOT NULL
float REAL NOT NULL
double DOUBLE PRECISION NOT NULL
std::string TEXT NOT NULL
char[N] VARCHAR(N-1) NOT NULL

It is possible to map the char C++ type to an integer database type (for example, SMALLINT) using the db type pragma (Section 14.4.3, "type").

Additionally, by default, C++ enums and C++11 enum classes are automatically mapped to the PostgreSQL types corresponding to their underlying integral types (see table above). The default NULL semantics is NOT NULL. For example:

enum color {red, green, blue};
enum class taste: unsigned char
{
  bitter = 1,
  sweet,
  sour = 4,
  salty
};

#pragma db object
class object
{
  ...

  color color_; // Automatically mapped to INTEGER.
  taste taste_; // Automatically mapped to SMALLINT.
};
  

Note also that because PostgreSQL does not support unsigned integers, the unsigned short, unsigned int, and unsigned long/unsigned long long C++ types are by default mapped to the SMALLINT, INTEGER, and BIGINT PostgreSQL types, respectively. The sign bit of the value stored by the database for these types will contain the most significant bit of the actual unsigned value being persisted.

It is also possible to add support for additional PostgreSQL types, such as NUMERIC, geometry types, XML, JSON, enumeration types, composite types, arrays, geospatial types, and the key-value store (HSTORE). For more information, refer to Section 14.8, "Database Type Mapping Pragmas".

19.1.1 String Type Mapping

The PostgreSQL ODB runtime library provides support for mapping the std::string, char[N], and std::array<char, N> types to the PostgreSQL CHAR, VARCHAR, and TEXT types. However, these mappings are not enabled by default (in particular, by default, std::array will be treated as a container). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("CHAR(2)")
  char state_[2];

  #pragma db type("VARCHAR(128)")
  std::string name_;
};
  

Alternatively, this can be done on the per-type basis, for example:

#pragma db value(std::string) type("VARCHAR(128)")

#pragma db object
class object
{
  ...

  std::string name_; // Mapped to VARCHAR(128).
};
  

The char[N] and std::array<char, N> values may or may not be zero-terminated. When extracting such values from the database, ODB will append the zero terminator if there is enough space.

19.1.2 Binary Type and UUID Mapping

The PostgreSQL ODB runtime library provides support for mapping the std::vector<char>, std::vector<unsigned char>, char[N], unsigned char[N], std::array<char, N>, and std::array<unsigned char, N> types to the PostgreSQL BYTEA type. There is also support for mapping the char[16] array to the PostgreSQL UUID type. However, these mappings are not enabled by default (in particular, by default, std::vector and std::array will be treated as containers). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("UUID")
  char uuid_[16];

  #pragma db type("BYTEA")
  std::vector<char> buf_;

  #pragma db type("BYTEA")
  unsigned char data_[256];
};
  

Alternatively, this can be done on the per-type basis, for example:

typedef std::vector<char> buffer;
#pragma db value(buffer) type("BYTEA")

#pragma db object
class object
{
  ...

  buffer buf_; // Mapped to BYTEA.
};
  

Note also that in native queries (Chapter 4, "Querying the Database") char[N] and std::array<char, N> parameters are by default passed as a string rather than a binary. To pass such parameters as a binary, we need to specify the database type explicitly in the _val()/_ref() calls. Note also that we don't need to do this for the integrated queries, for example:

char u[16] = {...};

db.query<object> ("uuid = " + query::_val<odb::pgsql::id_uuid> (u));
db.query<object> ("buf = " + query::_val<odb::pgsql::id_bytea> (u));
db.query<object> (query::uuid == query::_ref (u));
  

19.2 PostgreSQL Database Class

The PostgreSQL database class has the following interface:

namespace odb
{
  namespace pgsql
  {
    class database: public odb::database
    {
    public:
      database (const std::string& user,
                const std::string& password,
                const std::string& db,
                const std::string& host = "",
                unsigned int port = 0,
                const std::string& extra_conninfo = "",
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string& password,
                const std::string& db,
                const std::string& host,
                const std::string& socket_ext,
                const std::string& extra_conninfo = "",
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& conninfo,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (int& argc,
                char* argv[],
                bool erase = false,
                const std::string& extra_conninfo = "",
                std::[auto|unique]_ptr<connection_factory> = 0);

      static void
      print_usage (std::ostream&);

    public:
      const std::string&
      user () const;

      const std::string&
      password () const;

      const std::string&
      db () const;

      const std::string&
      host () const;

      unsigned int
      port () const;

      const std::string&
      socket_ext () const;

      const std::string&
      extra_conninfo () const;

      const std::string&
      conninfo () const;

    public:
      connection_ptr
      connection ();
    };
  }
}
  

You will need to include the <odb/pgsql/database.hxx> header file to make this class available in your application.

The overloaded database constructors allow us to specify the PostgreSQL database parameters that should be used when connecting to the database. The port argument in the first constructor is an integer value specifying the TCP/IP port number to connect to. A zero port number indicates that the default port should be used. The socket_ext argument in the second constructor is a string value specifying the UNIX-domain socket file name extension.

The third constructor allows us to specify all the database parameters as a single conninfo string. All other constructors accept additional database connection parameters as the extra_conninfo argument. For more information about the format of the conninfo string, refer to the PQconnectdb() function description in the PostgreSQL documentation. In the case of extra_conninfo, all the database parameters provided in this string will take precedence over those explicitly specified with other constructor arguments.

The last constructor extracts the database parameters from the command line. The following options are recognized:

  --user <login> | --username <login>
  --password <password>
  --database <name> | --dbname <name>
  --host <host>
  --port <integer>
  --options-file <file>
  

The --options-file option allows us to specify some or all of the database options in a file with each option appearing on a separate line followed by a space and an option value.

If the erase argument to this constructor is true, then the above options are removed from the argv array and the argc count is updated accordingly. This is primarily useful if your application accepts other options or arguments and you would like to get the PostgreSQL options out of the argv array.

This constructor throws the odb::pgsql::cli_exception exception if the PostgreSQL option values are missing or invalid. See section Section 19.4, "PostgreSQL Exceptions" for more information on this exception.

The static print_usage() function prints the list of options with short descriptions that are recognized by this constructor.

The last argument to all of the constructors is a pointer to the connection factory. In C++98/03, it is std::auto_ptr while in C++11 std::unique_ptr is used instead. If we pass a non-NULL value, the database instance assumes ownership of the factory instance. The connection factory interface as well as the available implementations are described in the next section.

The set of accessor functions following the constructors allows us to query the parameters of the database instance. Note that the conninfo() accessor returns a complete conninfo string which includes parameters that were explicitly specified with the various constructor arguments, as well as the extra parameters passed in the extra_conninfo argument. The extra_conninfo() accessor will return the conninfo string as passed in the extra_conninfo argument.

The connection() function returns a pointer to the PostgreSQL database connection encapsulated by the odb::pgsql::connection class. For more information on pgsql::connection, refer to Section 19.3, "PostgreSQL Connection and Connection Factory".

19.3 PostgreSQL Connection and Connection Factory

The pgsql::connection class has the following interface:

namespace odb
{
  namespace pgsql
  {
    class connection: public odb::connection
    {
    public:
      connection (database&);
      connection (database&, PGconn*);

      PGconn*
      handle ();
    };

    typedef details::shared_ptr<connection> connection_ptr;
  }
}
  

For more information on the odb::connection interface, refer to Section 3.6, "Connections". The first overloaded pgsql::connection constructor establishes a new PostgreSQL connection. The second constructor allows us to create a connection instance by providing an already connected native PostgreSQL handle. Note that the connection instance assumes ownership of this handle. The handle() accessor returns the PostgreSQL handle corresponding to the connection.

The pgsql::connection_factory abstract class has the following interface:

namespace odb
{
  namespace pgsql
  {
    class connection_factory
    {
    public:
      virtual void
      database (database&) = 0;

      virtual connection_ptr
      connect () = 0;
    };
  }
}
  

The database() function is called when a connection factory is associated with a database instance. This happens in the odb::pgsql::database class constructors. The connect() function is called whenever a database connection is requested.

The two implementations of the connection_factory interface provided by the PostgreSQL ODB runtime are new_connection_factory and connection_pool_factory. You will need to include the <odb/pgsql/connection-factory.hxx> header file to make the connection_factory interface and these implementation classes available in your application.

The new_connection_factory class creates a new connection whenever one is requested. When a connection is no longer needed, it is released and closed. The new_connection_factory class has the following interface:

namespace odb
{
  namespace pgsql
  {
    class new_connection_factory: public connection_factory
    {
    public:
      new_connection_factory ();
    };
};
  

The connection_pool_factory class implements a connection pool. It has the following interface:

namespace odb
{
  namespace pgsql
  {
    class connection_pool_factory: public connection_factory
    {
    public:
      connection_pool_factory (std::size_t max_connections = 0,
                               std::size_t min_connections = 0);

    protected:
      class pooled_connection: public connection
      {
      public:
        pooled_connection (database_type&);
        pooled_connection (database_type&, PGconn*);
      };

      typedef details::shared_ptr<pooled_connection> pooled_connection_ptr;

      virtual pooled_connection_ptr
      create ();
    };
};
  

The max_connections argument in the connection_pool_factory constructor specifies the maximum number of concurrent connections that this pool factory will maintain. Similarly, the min_connections argument specifies the minimum number of available connections that should be kept open.

Whenever a connection is requested, the pool factory first checks if there is an unused connection that can be returned. If there is none, the pool factory checks the max_connections value to see if a new connection can be created. If the total number of connections maintained by the pool is less than this value, then a new connection is created and returned. Otherwise, the caller is blocked until a connection becomes available.

When a connection is released, the pool factory first checks if there are blocked callers waiting for a connection. If so, then one of them is unblocked and is given the connection. Otherwise, the pool factory checks whether the total number of connections maintained by the pool is greater than the min_connections value. If that's the case, the connection is closed. Otherwise, the connection is added to the pool of available connections to be returned on the next request. In other words, if the number of connections maintained by the pool exceeds min_connections and there are no callers waiting for a new connection, the pool will close the excess connections.

If the max_connections value is 0, then the pool will create a new connection whenever all of the existing connections are in use. If the min_connections value is 0, then the pool will never close a connection and instead maintain all the connections that were ever created.

The create() virtual function is called whenever the pool needs to create a new connection. By deriving from the connection_pool_factory class and overriding this function we can implement custom connection establishment and configuration.

If you pass NULL as the connection factory to one of the database constructors, then the connection_pool_factory instance will be created by default with the min and max connections values set to 0. The following code fragment shows how we can pass our own connection factory instance:

#include <odb/database.hxx>

#include <odb/pgsql/database.hxx>
#include <odb/pgsql/connection-factory.hxx>

int
main (int argc, char* argv[])
{
  auto_ptr<odb::pgsql::connection_factory> f (
    new odb::pgsql::connection_pool_factory (20));

  auto_ptr<odb::database> db (
    new pgsql::database (argc, argv, false, "", f));
}
  

19.4 PostgreSQL Exceptions

The PostgreSQL ODB runtime library defines the following PostgreSQL-specific exceptions:

namespace odb
{
  namespace pgsql
  {
    class database_exception: odb::database_exception
    {
    public:
      const std::string&
      message () const;

      const std::string&
      sqlstate () const;

      virtual const char*
      what () const throw ();
    };

    class cli_exception: odb::exception
    {
    public:
      virtual const char*
      what () const throw ();
    };
  }
}
  

You will need to include the <odb/pgsql/exceptions.hxx> header file to make these exceptions available in your application.

The odb::pgsql::database_exception is thrown if a PostgreSQL database operation fails. The PostgreSQL-specific error information is accessible via the message() and sqlstate() functions. All this information is also combined and returned in a human-readable form by the what() function.

The odb::pgsql::cli_exception is thrown by the command line parsing constructor of the odb::pgsql::database class if the PostgreSQL option values are missing or invalid. The what() function returns a human-readable description of an error.

19.5 PostgreSQL Limitations

The following sections describe PostgreSQL-specific limitations imposed by the current PostgreSQL and ODB runtime versions.

19.5.1 Query Result Caching

The PostgreSQL ODB runtime implementation will always return a cached query result (Section 4.4, "Query Result") even when explicitly requested not to. This is a limitation of the PostgreSQL client library (libpq) which does not support uncached (streaming) query results.

19.5.2 Foreign Key Constraints

ODB assumes the standard SQL behavior which requires that foreign key constraints checking is deferred until the transaction is committed. Default PostgreSQL behavior is to check such constraints immediately. As a result, when used with ODB, a custom database schema that defines foreign key constraints may need to declare such constraints as INITIALLY DEFERRED, as shown in the following example. By default, schemas generated by the ODB compiler meet this requirement automatically.

CREATE TABLE Employee (
  ...
  employer BIGINT REFERENCES Employer(id) INITIALLY DEFERRED);
  

You can override the default behavior and instruct the ODB compiler to generate non-deferrable foreign keys by specifying the --fkeys-deferrable-mode not_deferrable ODB compiler option. Note, however, that in this case the order in which you persist, update, and erase objects within a transaction becomes important.

19.5.3 Unique Constraint Violations

Due to the granularity of the PostgreSQL error codes, it is impossible to distinguish between the duplicate primary key and other unique constraint violations. As a result, when making an object persistent, the PostgreSQL ODB runtime will translate all unique constraint violation errors to the object_already_persistent exception (Section 3.14, "ODB Exceptions").

19.5.4 Date-Time Format

ODB expects the PostgreSQL server to use integers as a binary format for the date-time types, which is the default for most PostgreSQL configurations. When creating a connection, ODB examines the integer_datetimes PostgreSQL server parameter and if it is false, odb::pgsql::database_exception is thrown. You may check the value of this parameter for your server by executing the following SQL query:

SHOW integer_datetimes
  

19.5.5 Timezones

ODB does not currently natively support the PostgreSQL date-time types with timezone information. However, these types can be accessed by mapping them to one of the natively supported types, as discussed in Section 14.8, "Database Type Mapping Pragmas".

19.5.6 NUMERIC Type Support

Support for the PostgreSQL NUMERIC type is limited to providing a binary buffer containing the binary representation of the value. For more information on the binary format used to store NUMERIC values refer to the PostgreSQL documentation. An alternative approach to accessing NUMERIC values is to map this type to one of the natively supported ones, as discussed in Section 14.8, "Database Type Mapping Pragmas".

19.6 PostgreSQL Index Definitions

When the index pragma (Section 14.7, "Index Definition Pragmas") is used to define a PostgreSQL index, the type clause specifies the index type (for example, UNIQUE), the method clause specifies the index method (for example, BTREE, HASH, GIN, etc.), and the options clause specifies additional index options, such as storage parameters, table spaces, and the WHERE predicate. To support the definition of concurrent indexes, the type clause can end with the word CONCURRENTLY (upper and lower cases are recognized). The column options can be used to specify collations, operator classes, and the sort order. For example:

#pragma db object
class object
{
  ...

  std::string name_;

  #pragma db index                            \
             type("UNIQUE CONCURRENTLY")      \
             method("HASH")                   \
             member(name_, "DESC")            \
             options("WITH(FILLFACTOR = 80)")
};
  

Index names in PostgreSQL are schema-global. To avoid name clashes, ODB automatically prefixes each index name with the table name on which it is defined.


20 Oracle Database

To generate support code for the Oracle database you will need to pass the "--database oracle" (or "-d oracle") option to the ODB compiler. Your application will also need to link to the Oracle ODB runtime library (libodb-oracle). All Oracle-specific ODB classes are defined in the odb::oracle namespace.

20.1 Oracle Type Mapping

The following table summarizes the default mapping between basic C++ value types and Oracle database types. This mapping can be customized on the per-type and per-member basis using the ODB Pragma Language (Chapter 14, "ODB Pragma Language").

C++ Type Oracle Type Default NULL Semantics
bool NUMBER(1) NOT NULL
char CHAR(1) NOT NULL
signed char NUMBER(3) NOT NULL
unsigned char NUMBER(3) NOT NULL
short NUMBER(5) NOT NULL
unsigned short NUMBER(5) NOT NULL
int NUMBER(10) NOT NULL
unsigned int NUMBER(10) NOT NULL
long NUMBER(19) NOT NULL
unsigned long NUMBER(20) NOT NULL
long long NUMBER(19) NOT NULL
unsigned long long NUMBER(20) NOT NULL
float BINARY_FLOAT NOT NULL
double BINARY_DOUBLE NOT NULL
std::string VARCHAR2(512) NULL
char[N] VARCHAR2(N-1) NULL

It is possible to map the char C++ type to an integer database type (for example, NUMBER(3)) using the db type pragma (Section 14.4.3, "type").

In Oracle empty VARCHAR2 and NVARCHAR2 strings are represented as a NULL value. As a result, columns of the std::string and char[N] types are by default declared as NULL except for primary key columns. However, you can override this by explicitly declaring such columns as NOT NULL with the db not_null pragma (Section 14.4.6, "null/not_null"). This also means that for object ids that are mapped to these Oracle types, an empty string is an invalid value.

Additionally, by default, C++ enums and C++11 enum classes are automatically mapped to the Oracle types corresponding to their underlying integral types (see table above). The default NULL semantics is NOT NULL. For example:

enum color {red, green, blue};
enum class taste: unsigned char
{
  bitter = 1,
  sweet,
  sour = 4,
  salty
};

#pragma db object
class object
{
  ...

  color color_; // Automatically mapped to NUMBER(10).
  taste taste_; // Automatically mapped to NUMBER(3).
};
  

It is also possible to add support for additional Oracle types, such as XML, geospatial types, user-defined types, and collections (arrays, table types). For more information, refer to Section 14.8, "Database Type Mapping Pragmas".

20.1.1 String Type Mapping

The Oracle ODB runtime library provides support for mapping the std::string, char[N], and std::array<char, N> types to the Oracle CHAR, VARCHAR2, CLOB, NCHAR, NVARCHAR2, and NCLOB types. However, these mappings are not enabled by default (in particular, by default, std::array will be treated as a container). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type ("CHAR(2)")
  char state_[2];

  #pragma db type ("VARCHAR(128)") null
  std::string name_;

  #pragma db type ("CLOB")
  std::string text_;
};
  

Alternatively, this can be done on the per-type basis, for example:

#pragma db value(std::string) type("VARCHAR(128)") null

#pragma db object
class object
{
  ...

  std::string name_; // Mapped to VARCHAR(128).

  #pragma db type ("CLOB")
  std::string text_; // Mapped to CLOB.
};
  

The char[N] and std::array<char, N> values may or may not be zero-terminated. When extracting such values from the database, ODB will append the zero terminator if there is enough space.

20.1.2 Binary Type Mapping

The Oracle ODB runtime library provides support for mapping the std::vector<char>, std::vector<unsigned char>, char[N], unsigned char[N], std::array<char, N>, and std::array<unsigned char, N> types to the Oracle BLOB and RAW types. However, these mappings are not enabled by default (in particular, by default, std::vector and std::array will be treated as containers). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("BLOB")
  std::vector<char> buf_;

  #pragma db type("RAW(16)")
  unsigned char uuid_[16];
};
  

Alternatively, this can be done on the per-type basis, for example:

typedef std::vector<char> buffer;
#pragma db value(buffer) type("BLOB")

#pragma db object
class object
{
  ...

  buffer buf_; // Mapped to BLOB.
};
  

Note also that in native queries (Chapter 4, "Querying the Database") char[N] and std::array<char, N> parameters are by default passed as a string rather than a binary. To pass such parameters as a binary, we need to specify the database type explicitly in the _val()/_ref() calls. Note also that we don't need to do this for the integrated queries, for example:

char u[16] = {...};

db.query<object> ("uuid = " + query::_val<odb::oracle::id_raw> (u));
db.query<object> (query::uuid == query::_ref (u));
  

20.2 Oracle Database Class

The Oracle database class encapsulates the OCI environment handle as well as the database connection string and user credentials that are used to establish connections to the database. It has the following interface:

namespace odb
{
  namespace oracle
  {
    class database: public odb::database
    {
    public:
      database (const std::string& user,
                const std::string& password,
                const std::string& db,
                ub2 charset = 0,
                ub2 ncharset = 0,
                OCIEnv* environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string& password,
                const std::string& service,
                const std::string& host,
                unsigned int port = 0,
                ub2 charset = 0,
                ub2 ncharset = 0,
                OCIEnv* environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (int& argc,
                char* argv[],
                bool erase = false,
                ub2 charset = 0,
                ub2 ncharset = 0,
                OCIEnv* environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      static void
      print_usage (std::ostream&);

    public:
      const std::string&
      user () const;

      const std::string&
      password () const;

      const std::string&
      db () const;

      const std::string&
      service () const;

      const std::string&
      host () const;

      unsigned int
      port () const;

      ub2
      charset () const;

      ub2
      ncharset () const;

      OCIEnv*
      environment ();

    public:
      connection_ptr
      connection ();
    };
  }
}
  

You will need to include the <odb/oracle/database.hxx> header file to make this class available in your application.

The overloaded database constructors allow us to specify the Oracle database parameters that should be used when connecting to the database. The db argument in the first constructor is a connection identifier that specifies the database to connect to. For more information on the format of the connection identifier, refer to the Oracle documentation.

The second constructor allows us to specify the individual components of a connection identifier as the service, host, and port arguments. If the host argument is empty, then localhost is used by default. Similarly, if the port argument is zero, then the default port is used.

The last constructor extracts the database parameters from the command line. The following options are recognized:

  --user <login>
  --password <password>
  --database <connect-id>
  --service <name>
  --host <host>
  --port <integer>
  --options-file <file>
  

The --options-file option allows us to specify some or all of the database options in a file with each option appearing on a separate line followed by a space and an option value. Note that it is invalid to specify the --database option together with --service, --host, or --port options.

If the erase argument to this constructor is true, then the above options are removed from the argv array and the argc count is updated accordingly. This is primarily useful if your application accepts other options or arguments and you would like to get the Oracle options out of the argv array.

This constructor throws the odb::oracle::cli_exception exception if the Oracle option values are missing or invalid. See section Section 20.4, "Oracle Exceptions" for more information on this exception.

The static print_usage() function prints the list of options with short descriptions that are recognized by this constructor.

Additionally, all the constructors have the charset, ncharset, and environment arguments. The charset argument specifies the client-side database character encoding. Character data corresponding to the CHAR, VARCHAR2, and CLOB types will be delivered to and received from the application in this encoding. Similarly, the ncharset argument specifies the client-side national character encoding. Character data corresponding to the NCHAR, NVARCHAR2, and NCLOB types will be delivered to and received from the application in this encoding. For the complete list of available character encoding values, refer to the Oracle documentation. Commonly used encoding values are 873 (UTF-8), 31 (ISO-8859-1), and 1000 (UTF-16). If the database character encoding is not specified, then the NLS_LANG environment/registry variable is used. Similarly, if the national character encoding is not specified, then the NLS_NCHAR environment/registry variable is used. For more information on character encodings, refer to the OCIEnvNlsCreate() function in the Oracle Call Interface (OCI) documentation.

The environment argument allows us to provide a custom OCI environment handle. If this argument is not NULL, then the passed handle is used in all the OCI function calls made by this database class instance. Note also that the database instance does not assume ownership of the passed environment handle and this handle should be valid for the lifetime of the database instance. If a custom environment handle is used, then the charset and ncharset arguments have no effect.

The last argument to all of the constructors is a pointer to the connection factory. In C++98/03, it is std::auto_ptr while in C++11 std::unique_ptr is used instead. If we pass a non-NULL value, the database instance assumes ownership of the factory instance. The connection factory interface as well as the available implementations are described in the next section.

The set of accessor functions following the constructors allows us to query the parameters of the database instance.

The connection() function returns a pointer to the Oracle database connection encapsulated by the odb::oracle::connection class. For more information on oracle::connection, refer to Section 20.3, "Oracle Connection and Connection Factory".

20.3 Oracle Connection and Connection Factory

The oracle::connection class has the following interface:

namespace odb
{
  namespace oracle
  {
    class connection: public odb::connection
    {
    public:
      connection (database&);
      connection (database&, OCISvcCtx*);

      OCISvcCtx*
      handle ();

      OCIError*
      error_handle ();

      details::buffer&
      lob_buffer ();
    };

    typedef details::shared_ptr<connection> connection_ptr;
  }
}
  

For more information on the odb::connection interface, refer to Section 3.6, "Connections". The first overloaded oracle::connection constructor creates a new OCI service context. The OCI statement caching is enabled for the underlying session while the OCI connection pooling and session pooling are not used. The second constructor allows us to create a connection instance by providing an already connected Oracle service context. Note that the connection instance assumes ownership of this handle. The handle() accessor returns the OCI service context handle associated with the connection instance.

An OCI error handle is allocated for each connection instance and is available via the error_handle() accessor function.

Additionally, each connection instance maintains a large object (LOB) buffer. This buffer is used by the Oracle ODB runtime as an intermediate storage for piecewise handling of LOB data. By default, the LOB buffer has zero initial capacity and is expanded to 4096 bytes when the first LOB operation is performed. If your application requires a bigger or smaller LOB buffer, you can specify a custom capacity using the lob_buffer() accessor.

The oracle::connection_factory abstract class has the following interface:

namespace odb
{
  namespace oracle
  {
    class connection_factory
    {
    public:
      virtual void
      database (database&) = 0;

      virtual connection_ptr
      connect () = 0;
    };
  }
}
  

The database() function is called when a connection factory is associated with a database instance. This happens in the odb::oracle::database class constructors. The connect() function is called whenever a database connection is requested.

The two implementations of the connection_factory interface provided by the Oracle ODB runtime are new_connection_factory and connection_pool_factory. You will need to include the <odb/oracle/connection-factory.hxx> header file to make the connection_factory interface and these implementation classes available in your application.

The new_connection_factory class creates a new connection whenever one is requested. When a connection is no longer needed, it is released and closed. The new_connection_factory class has the following interface:

namespace odb
{
  namespace oracle
  {
    class new_connection_factory: public connection_factory
    {
    public:
      new_connection_factory ();
    };
};
  

The connection_pool_factory class implements a connection pool. It has the following interface:

namespace odb
{
  namespace oracle
  {
    class connection_pool_factory: public connection_factory
    {
    public:
      connection_pool_factory (std::size_t max_connections = 0,
                               std::size_t min_connections = 0);

    protected:
      class pooled_connection: public connection
      {
      public:
        pooled_connection (database_type&);
        pooled_connection (database_type&, OCISvcCtx*);
      };

      typedef details::shared_ptr<pooled_connection> pooled_connection_ptr;

      virtual pooled_connection_ptr
      create ();
    };
};
  

The max_connections argument in the connection_pool_factory constructor specifies the maximum number of concurrent connections that this pool factory will maintain. Similarly, the min_connections argument specifies the minimum number of available connections that should be kept open.

Whenever a connection is requested, the pool factory first checks if there is an unused connection that can be returned. If there is none, the pool factory checks the max_connections value to see if a new connection can be created. If the total number of connections maintained by the pool is less than this value, then a new connection is created and returned. Otherwise, the caller is blocked until a connection becomes available.

When a connection is released, the pool factory first checks if there are blocked callers waiting for a connection. If so, then one of them is unblocked and is given the connection. Otherwise, the pool factory checks whether the total number of connections maintained by the pool is greater than the min_connections value. If that's the case, the connection is closed. Otherwise, the connection is added to the pool of available connections to be returned on the next request. In other words, if the number of connections maintained by the pool exceeds min_connections and there are no callers waiting for a new connection, the pool will close the excess connections.

If the max_connections value is 0, then the pool will create a new connection whenever all of the existing connections are in use. If the min_connections value is 0, then the pool will never close a connection and instead maintain all the connections that were ever created.

The create() virtual function is called whenever the pool needs to create a new connection. By deriving from the connection_pool_factory class and overriding this function we can implement custom connection establishment and configuration.

If you pass NULL as the connection factory to one of the database constructors, then the connection_pool_factory instance will be created by default with the min and max connections values set to 0. The following code fragment shows how we can pass our own connection factory instance:

#include <odb/database.hxx>

#include <odb/oracle/database.hxx>
#include <odb/oracle/connection-factory.hxx>

int
main (int argc, char* argv[])
{
  auto_ptr<odb::oracle::connection_factory> f (
    new odb::oracle::connection_pool_factory (20));

  auto_ptr<odb::database> db (
    new oracle::database (argc, argv, false, 0, 0, 0, f));
}
  

20.4 Oracle Exceptions

The Oracle ODB runtime library defines the following Oracle-specific exceptions:

namespace odb
{
  namespace oracle
  {
    class database_exception: odb::database_exception
    {
    public:
      class record
      {
      public:
        sb4
        error () const;

        const std::string&
        message () const;
      };

      typedef std::vector<record> records;

      typedef records::size_type size_type;
      typedef records::const_iterator iterator;

      iterator
      begin () const;

      iterator
      end () const;

      size_type
      size () const;

      virtual const char*
      what () const throw ();
    };

    class cli_exception: odb::exception
    {
    public:
      virtual const char*
      what () const throw ();
    };

    class invalid_oci_handle: odb::exception
    {
    public:
      virtual const char*
      what () const throw ();
    };
  }
}
  

You will need to include the <odb/oracle/exceptions.hxx> header file to make these exceptions available in your application.

The odb::oracle::database_exception is thrown if an Oracle database operation fails. The Oracle-specific error information is stored as a series of records, each containing the error code as a signed 4-byte integer and the message string. All this information is also combined and returned in a human-readable form by the what() function.

The odb::oracle::cli_exception is thrown by the command line parsing constructor of the odb::oracle::database class if the Oracle option values are missing or invalid. The what() function returns a human-readable description of an error.

The odb::oracle::invalid_oci_handle is thrown if an invalid handle is passed to an OCI function or if an OCI function was unable to allocate a handle. The former normally indicates a programming error while the latter indicates an out of memory condition. The what() function returns a human-readable description of an error.

20.5 Oracle Limitations

The following sections describe Oracle-specific limitations imposed by the current Oracle and ODB runtime versions.

20.5.1 Identifier Truncation

Oracle limits the length of database identifiers (table, column, etc., names) to 30 characters. The ODB compiler automatically truncates any identifier that is longer than 30 characters. This, however, can lead to duplicate names. A common symptom of this problem are errors during the database schema creation indicating that a database object with the same name already exists. To resolve this problem we can assign custom, shorter identifiers using the db table and db column pragmas (Chapter 14, "ODB Pragma Language"). For example:

#pragma db object
class long_class_name
{
  ...

  std::vector<int> long_container_x_;
  std::vector<int> long_container_y_;
};
  

In the above example, the names of the two container tables will be long_class_name_long_container_x_ and long_class_name_long_container_y_. However, when truncated to 30 characters, they both become long_class_name_long_container. To resolve this collision we can assign a custom table name for each container:

#pragma db object
class long_class_name
{
  ...

  #pragma db table("long_class_name_cont_x")
  std::vector<int> long_container_x_;

  #pragma db table("long_class_name_cont_y")
  std::vector<int> long_container_y_;
};
  

20.5.2 Query Result Caching

Oracle ODB runtime implementation does not perform query result caching (Section 4.4, "Query Result") even when explicitly requested. The OCI API supports interleaving execution of multiple prepared statements on a single connection. As a result, with OCI, it is possible to have multiple uncached results and calls to other database functions do not invalidate them. The only limitation of the uncached Oracle results is the unavailability of the result::size() function. If you call this function on an Oracle query result, then the odb::result_not_cached exception (Section 3.14, "ODB Exceptions") is always thrown. Future versions of the Oracle ODB runtime library may add support for result caching.

20.5.3 Foreign Key Constraints

ODB assumes the standard SQL behavior which requires that foreign key constraints checking is deferred until the transaction is committed. Default Oracle behavior is to check such constraints immediately. As a result, when used with ODB, a custom database schema that defines foreign key constraints may need to declare such constraints as INITIALLY DEFERRED, as shown in the following example. By default, schemas generated by the ODB compiler meet this requirement automatically.

CREATE TABLE Employee (
  ...
  employer NUMBER(20) REFERENCES Employer(id)
           DEFERRABLE INITIALLY DEFERRED);
  

You can override the default behavior and instruct the ODB compiler to generate non-deferrable foreign keys by specifying the --fkeys-deferrable-mode not_deferrable ODB compiler option. Note, however, that in this case the order in which you persist, update, and erase objects within a transaction becomes important.

20.5.4 Unique Constraint Violations

Due to the granularity of the Oracle error codes, it is impossible to distinguish between the duplicate primary key and other unique constraint violations. As a result, when making an object persistent, the Oracle ODB runtime will translate all unique constraint violation errors to the object_already_persistent exception (Section 3.14, "ODB Exceptions").

20.5.5 Large FLOAT and NUMBER Types

The Oracle FLOAT type with a binary precision greater than 53 and fixed-point NUMBER type with a decimal precision greater than 15 cannot be automatically extracted into the C++ float and double types. Instead, the Oracle ODB runtime uses a 21-byte buffer containing the binary representation of a value as an image type for such FLOAT and NUMBER types. In order to convert them into an application-specific large number representation, you will need to provide a suitable value_traits template specialization. For more information on the binary format used to store the FLOAT and NUMBER values, refer to the Oracle Call Interface (OCI) documentation.

An alternative approach to accessing large FLOAT and NUMBER values is to map these type to one of the natively supported ones, as discussed in Section 14.8, "Database Type Mapping Pragmas".

Note that a NUMBER type that is used to represent a floating point number (declared by specifying NUMBER without any range and scale) can be extracted into the C++ float and double types.

20.5.6 Timezones

ODB does not currently support the Oracle date-time types with timezone information. However, these types can be accessed by mapping them to one of the natively supported types, as discussed in Section 14.8, "Database Type Mapping Pragmas".

20.5.7 LONG Types

ODB does not support the deprecated Oracle LONG and LONG RAW data types. However, these types can be accessed by mapping them to one of the natively supported types, as discussed in Section 14.8, "Database Type Mapping Pragmas".

20.5.8 LOB Types and By-Value Accessors/Modifiers

As discussed in Section 14.4.5, "get/set/access", by-value accessor and modifier expressions cannot be used with data members of Oracle large object (LOB) data types: BLOB, CLOB, and NCLOB. The Oracle ODB runtime uses streaming for reading/writing LOB data directly from/to data members. As a result, by-reference accessors and modifiers should be used for these data types.

20.5.9 Database Schema Evolution

In Oracle, the type of the name column in the schema_version table is VARCHAR2(512). Because this column is a primary key and VARCHAR2 represents empty strings as NULL values, it is impossible to store an empty string in this column, which is what is used to represent the default schema name. As a result, in Oracle, the empty schema name is stored as a string containing a single space character. ODB performs all the necessary translations automatically and normally you do not need to worry about this implementation detail unless you are querying or modifying the schema_version table directly.

20.6 Oracle Index Definitions

When the index pragma (Section 14.7, "Index Definition Pragmas") is used to define an Oracle index, the type clause specifies the index type (for example, UNIQUE, BITMAP), the method clause is not used, and the options clause specifies additional index properties, such as partitioning, table spaces, etc. The column options can be used to specify the sort order. For example:

#pragma db object
class object
{
  ...

  std::string name_;

  #pragma db index                     \
             type("BITMAP")            \
             member(name_, "DESC")     \
             options("TABLESPACE TBS1")
};
  

Index names in Oracle are schema-global. To avoid name clashes, ODB automatically prefixes each index name with the table name on which it is defined.


21 Microsoft SQL Server Database

To generate support code for the SQL Server database you will need to pass the "--database mssql" (or "-d mssql") option to the ODB compiler. Your application will also need to link to the SQL Server ODB runtime library (libodb-mssql). All SQL Server-specific ODB classes are defined in the odb::mssql namespace.

21.1 SQL Server Type Mapping

The following table summarizes the default mapping between basic C++ value types and SQL Server database types. This mapping can be customized on the per-type and per-member basis using the ODB Pragma Language (Chapter 14, "ODB Pragma Language").

C++ Type SQL Server Type Default NULL Semantics
bool BIT NOT NULL
char CHAR(1) NOT NULL
signed char TINYINT NOT NULL
unsigned char TINYINT NOT NULL
short SMALLINT NOT NULL
unsigned short SMALLINT NOT NULL
int INT NOT NULL
unsigned int INT NOT NULL
long BIGINT NOT NULL
unsigned long BIGINT NOT NULL
long long BIGINT NOT NULL
unsigned long long BIGINT NOT NULL
float REAL NOT NULL
double FLOAT NOT NULL
std::string VARCHAR(512)/VARCHAR(256) NOT NULL
char[N] VARCHAR(N-1) NOT NULL
std::wstring NVARCHAR(512)/NVARCHAR(256) NOT NULL
wchar_t[N] NVARCHAR(N-1) NOT NULL
GUID UNIQUEIDENTIFIER NOT NULL

It is possible to map the char C++ type to an integer database type (for example, TINYINT) using the db type pragma (Section 14.4.3, "type").

Note that the std::string and std::wstring types are mapped differently depending on whether a member of one of these types is an object id or not. If the member is an object id, then for this member std::string is mapped to VARCHAR(256) and std::wstring — to NVARCHAR(256). Otherwise, std::string is mapped to VARCHAR(512) and std::wstring — to NVARCHAR(512). Note also that you can always change this mapping using the db type pragma (Section 14.4.3, "type").

Additionally, by default, C++ enums and C++11 enum classes are automatically mapped to the SQL Server types corresponding to their underlying integral types (see table above). The default NULL semantics is NOT NULL. For example:

enum color {red, green, blue};
enum class taste: unsigned char
{
  bitter = 1,
  sweet,
  sour = 4,
  salty
};

#pragma db object
class object
{
  ...

  color color_; // Automatically mapped to INT.
  taste taste_; // Automatically mapped to TINYINT.
};
  

Note also that because SQL Server does not support unsigned integers, the unsigned short, unsigned int, and unsigned long/unsigned long long C++ types are by default mapped to the SMALLINT, INT, and BIGINT SQL Server types, respectively. The sign bit of the value stored by the database for these types will contain the most significant bit of the actual unsigned value being persisted. Similarly, because there is no signed version of the TINYINT SQL Server type, by default, the signed char C++ type is mapped to TINYINT. As a result, the most significant bit of the value stored by the database for this type will contain the sign bit of the actual signed value being persisted.

It is also possible to add support for additional SQL Server types, such as geospatial types, XML, and user-defined types. For more information, refer to Section 14.8, "Database Type Mapping Pragmas".

21.1.1 String Type Mapping

The SQL Server ODB runtime library provides support for mapping the std::string, char[N], and std::array<char, N> types to the SQL Server CHAR, VARCHAR, and TEXT types as well as the std::wstring, wchar_t[N], and std::array<wchar_t, N> types to NCHAR, NVARCHAR, and NTEXT. However, these mappings are not enabled by default (in particular, by default, std::array will be treated as a container). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type ("CHAR(2)")
  char state_[2];

  #pragma db type ("NVARCHAR(max)")
  std::wstring text_;
};
  

Alternatively, this can be done on the per-type basis, for example:

#pragma db value(std::wstring) type("NVARCHAR(max)")

#pragma db object
class object
{
  ...

  std::wstring text_; // Mapped to NVARCHAR(max).
};
  

The char[N], std::array<char, N>, wchar_t[N], and std::array<wchar_t, N> values may or may not be zero-terminated. When extracting such values from the database, ODB will append the zero terminator if there is enough space.

See also Section 21.1.4, "Long String and Binary Types" for certain limitations of long string types.

21.1.2 Binary Type and UNIQUEIDENTIFIER Mapping

The SQL Server ODB runtime library also provides support for mapping the std::vector<char>, std::vector<unsigned char>, char[N], unsigned char[N], std::array<char, N>, and std::array<unsigned char, N> types to the SQL Server BINARY, VARBINARY, and IMAGE types. There is also support for mapping the char[16] array to the SQL Server UNIQUEIDENTIFIER type. However, these mappings are not enabled by default (in particular, by default, std::vector and std::array will be treated as containers). To enable the alternative mappings for these types we need to specify the database type explicitly using the db type pragma (Section 14.4.3, "type"), for example:

#pragma db object
class object
{
  ...

  #pragma db type("UNIQUEIDENTIFIER")
  char uuid_[16];

  #pragma db type("VARBINARY(max)")
  std::vector<char> buf_;

  #pragma db type("BINARY(256)")
  unsigned char data_[256];
};
  

Alternatively, this can be done on the per-type basis, for example:

typedef std::vector<char> buffer;
#pragma db value(buffer) type("VARBINARY(max)")

#pragma db object
class object
{
  ...

  buffer buf_; // Mapped to VARBINARY(max).
};
  

Note also that in native queries (Chapter 4, "Querying the Database") char[N] and std::array<char, N> parameters are by default passed as a string rather than a binary. To pass such parameters as a binary, we need to specify the database type explicitly in the _val()/_ref() calls. Note also that we don't need to do this for the integrated queries, for example:

char u[16] = {...};

db.query<object> ("uuid = " + query::_val<odb::mssql::id_binary> (u));
db.query<object> (
  "uuid = " + query::_val<odb::mssql::id_uniqueidentifier> (u));
db.query<object> (query::uuid == query::_ref (u));
  

See also Section 21.1.4, "Long String and Binary Types" for certain limitations of long binary types.

21.1.3 ROWVERSION Mapping

ROWVERSION is a special SQL Server data type that is automatically incremented by the database server whenever a row is inserted or updated. As such, it is normally used to implement optimistic concurrency and ODB provides support for using ROWVERSION instead of the more portable approach for optimistic concurrency (Chapter 12, "Optimistic Concurrency").

ROWVERSION is a 64-bit value which is mapped by ODB to unsigned long long. As a result, to use ROWVERSION for optimistic concurrency we need to make sure that the version column is of the unsigned long long type. We also need to explicitly specify that it should be mapped to the ROWVERSION data type. For example:

#pragma db object optimistic
class person
{
  ...

  #pragma db version type("ROWVERSION")
  unsigned long long version_;
};
  

21.1.4 Long String and Binary Types

For SQL Server, ODB handles character, national character, and binary data in two different ways depending on its maximum length. If the maximum length (in bytes) is less than or equal to the limit specified with the --mssql-short-limit ODB compiler option (1024 by default), then it is treated as short data, otherwise — long data. For short data ODB pre-allocates an intermediate buffer of the maximum size and binds it directly to a parameter or result column. This way the underlying database API (ODBC) can read/write directly from/to this buffer. In the case of long data, the data is read/written in chunks using the SQLGetData()/SQLPutData() ODBC functions. While the long data approach reduces the amount of memory used by the application, it may require greater CPU resources.

Long data has a number of limitations. In particular, when setting a custom short data limit, make sure that it is sufficiently large so that no object id in the application is treated as long data. It is also impossible to load an object or view with long data more than once as part of a query result iteration (Section 4.4, "Query Result"). Any such attempt will result in the odb::mssql::long_data_reload exception (Section 21.4, "SQL Server Exceptions"). For example:

#pragma db object
class object
{
  ...

  int num_;

  #pragma db type("VARCHAR(max)") // Long data.
  std::string str_;
};

typedef odb::query<object> query;
typedef odb::result<object> result;

transaction t (db.begin ());

result r (db.query<object> (query::num < 100));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
{
  if (!i->str_.empty ()) // First load.
  {
    object o;
    i.load (o); // Error: second load, long_data_reload is thrown.
  }
}

t.commit ();
  

Finally, if a native view (Section 10.6, "Native Views") contains one or more long data members, then such members should come last both in the select-list of the native SQL query and the list of data members in the C++ class.

21.2 SQL Server Database Class

The SQL Server database class encapsulates the ODBC environment handle as well as the server instance address and user credentials that are used to establish connections to the database. It has the following interface:

namespace odb
{
  namespace mssql
  {
    enum protocol
    {
      protocol_auto,
      protocol_tcp, // TCP/IP.
      protocol_lpc, // Shared memory (local procedure call).
      protocol_np   // Named pipes.
    };

    enum transaction_isolation
    {
      isolation_read_uncommitted,
      isolation_read_committed,   // SQL Server default.
      isolation_repeatable_read,
      isolation_snapshot,
      isolation_serializable
    };

    class database: public odb::database
    {
    public:
      typedef protocol protocol_type;
      typedef transaction_isolation transaction_isolation_type;

      database (const std::string& user,
                const std::string& password,
                const std::string& db,
                const std::string& server,
                const std::string& driver = "",
                const std::string& extra_connect_string = "",
                transaction_isolation_type = isolation_read_committed,
                SQLHENV environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string& password,
                const std::string& db,
                protocol_type protocol = protocol_auto,
                const std::string& host = "",
                const std::string& instance = "",
                const std::string& driver = "",
                const std::string& extra_connect_string = "",
                transaction_isolation_type = isolation_read_committed,
                SQLHENV environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& user,
                const std::string& password,
                const std::string& db,
                const std::string& host,
                unsigned int port,
                const std::string& driver = "",
                const std::string& extra_connect_string = "",
                transaction_isolation_type = isolation_read_committed,
                SQLHENV environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (const std::string& connect_string,
                transaction_isolation_type = isolation_read_committed,
                SQLHENV environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      database (int& argc,
                char* argv[],
                bool erase = false,
                const std::string& extra_connect_string = "",
                transaction_isolation_type = isolation_read_committed,
                SQLHENV environment = 0,
                std::[auto|unique]_ptr<connection_factory> = 0);

      static void
      print_usage (std::ostream&);

    public:
      const std::string&
      user () const;

      const std::string&
      password () const;

      const std::string&
      db () const;

      protocol_type
      protocol () const;

      const std::string&
      host () const;

      const std::string&
      instance () const;

      unsigned int
      port () const;

      const std::string&
      server () const;

      const std::string&
      driver () const;

      const std::string&
      extra_connect_string () const;

      transaction_isolation_type
      transaction_isolation () const;

      const std::string&
      connect_string () const;

      SQLHENV
      environment ();

    public:
      connection_ptr
      connection ();
    };
  }
}
  

You will need to include the <odb/mssql/database.hxx> header file to make this class available in your application.

The overloaded database constructors allow us to specify the SQL Server database parameters that should be used when connecting to the database. The user and password arguments specify the login name and password. If user is empty, then Windows authentication is used and the password argument is ignored. The db argument specifies the database name to open. If it is empty, then the default database for the user is used.

The server argument in the first constructor specifies the SQL Server instance address in the standard SQL Server address format:

[protocol:]host[\instance][,port]

Where protocol can be tcp (TCP/IP), lpc (shared memory), or np (named pipe). If protocol is not specified, then a suitable protocol is automatically selected based on the SQL Server Native Client configuration. The host component can be a host name or an IP address. If instance is not specified, then the default SQL Server instance is assumed. If port is not specified, then the default SQL Server port is used (1433). Note that you would normally specify either the instance name or the port, but not both. If both are specified, then the instance name is ignored by the SQL Server Native Client ODBC driver. For more information on the format of the SQL Server address, refer to the SQL Server Native Client ODBC driver documentation.

The second and third constructors allow us to specify all these address components (protocol, host, instance, and port) as separate arguments. The third constructor always connects using TCP/IP to the specified host and port.

The driver argument specifies the SQL Server Native Client ODBC driver that should be used to connect to the database. If not specified, then the latest available version is used. The following examples show common ways of connecting to the database using the first three constructors:

// Connect to the default SQL Server instance on the local machine
// using the default protocol. Login as 'test' with password 'secret'
// and open the 'example_db' database.
//
odb::mssql::database db1 ("test",
                          "secret",
                          "example_db");

// As above except use Windows authentication and open the default
// database for this user.
//
odb::mssql::database db2 ("",
                          "",
                          "");

// Connect to the default SQL Server instance on 'onega' using the
// default protocol. Login as 'test' with password 'secret' and open
// the 'example_db' database.
//
odb::mssql::database db3 ("test",
                          "secret",
                          "example_db"
                          "onega");

// As above but connect to the 'production' SQL Server instance.
//
odb::mssql::database db4 ("test",
                          "secret",
                          "example_db"
                          "onega\\production");

// Same as above but specify protocol, host, and instance as separate
// arguments.
//
odb::mssql::database db5 ("test",
                          "secret",
                          "example_db",
                          odb::mssql::protocol_auto,
                          "onega",
                          "production");

// As above, but use TCP/IP as the protocol.
//
odb::mssql::database db6 ("test",
                          "secret",
                          "example_db"
                          "tcp:onega\\production");

// Same as above but using separate arguments.
//
odb::mssql::database db7 ("test",
                          "secret",
                          "example_db",
                          odb::mssql::protocol_tcp,
                          "onega",
                          "production");

// As above, but use TCP/IP port instead of the instance name.
//
odb::mssql::database db8 ("test",
                          "secret",
                          "example_db"
                          "tcp:onega,1435");

// Same as above but using separate arguments. Note that here we
// don't need to specify protocol explicitly since it can only
// be TCP/IP.
//
odb::mssql::database db9 ("test",
                          "secret",
                          "example_db",
                          "onega",
                          1435);

// As above but use the specific SQL Server Native Client ODBC
// driver version.
//
odb::mssql::database dbA ("test",
                          "secret",
                          "example_db"
                          "tcp:onega,1435",
                          "SQL Server Native Client 10.0");
  

The fourth constructor allows us to pass a custom ODBC connection string that provides all the information necessary to connect to the database. Note also that all the other constructors have the extra_connect_string argument which can be used to specify additional ODBC connection attributes. For more information on the format of the ODBC connection string, refer to the SQL Server Native Client ODBC driver documentation.

The last constructor extracts the database parameters from the command line. The following options are recognized:

  --user | -U <login>
  --password | -P <password>
  --database | -d <name>
  --server | -S <address>
  --driver <name>
  --options-file <file>
  

The --options-file option allows us to specify some or all of the database options in a file with each option appearing on a separate line followed by a space and an option value.

If the erase argument to this constructor is true, then the above options are removed from the argv array and the argc count is updated accordingly. This is primarily useful if your application accepts other options or arguments and you would like to get the SQL Server options out of the argv array.

This constructor throws the odb::mssql::cli_exception exception if the SQL Server option values are missing or invalid. See section Section 21.4, "SQL Server Exceptions" for more information on this exception.

The static print_usage() function prints the list of options with short descriptions that are recognized by this constructor.

Additionally, all the constructors have the transaction_isolation and environment arguments. The transaction_isolation argument allows us to specify an alternative transaction isolation level that should be used by all the connections created by this database instance. The environment argument allows us to provide a custom ODBC environment handle. If this argument is not NULL, then the passed handle is used in all the ODBC function calls made by this database instance. Note also that the database instance does not assume ownership of the passed environment handle and this handle should be valid for the lifetime of the database instance.

The last argument to all of the constructors is a pointer to the connection factory. In C++98/03, it is std::auto_ptr while in C++11 std::unique_ptr is used instead. If we pass a non-NULL value, the database instance assumes ownership of the factory instance. The connection factory interface as well as the available implementations are described in the next section.

The set of accessor functions following the constructors allows us to query the parameters of the database instance.

The connection() function returns a pointer to the SQL Server database connection encapsulated by the odb::mssql::connection class. For more information on mssql::connection, refer to Section 21.3, "SQL Server Connection and Connection Factory".

21.3 SQL Server Connection and Connection Factory

The mssql::connection class has the following interface:

namespace odb
{
  namespace mssql
  {
    class connection: public odb::connection
    {
    public:
      connection (database&);
      connection (database&, SQLHDBC handle);

      SQLHDBC
      handle ();

      details::buffer&
      long_data_buffer ();
    };

    typedef details::shared_ptr<connection> connection_ptr;
  }
}
  

For more information on the odb::connection interface, refer to Section 3.6, "Connections". The first overloaded mssql::connection constructor creates a new ODBC connection. The created connection is configured to use the manual commit mode with multiple active result sets (MARS) enabled. The second constructor allows us to create a connection instance by providing an already established ODBC connection. Note that the connection instance assumes ownership of this handle. The handle() accessor returns the underlying ODBC connection handle associated with the connection instance.

Additionally, each connection instance maintains a long data buffer. This buffer is used by the SQL Server ODB runtime as an intermediate storage for piecewise handling of long data. By default, the long data buffer has zero initial capacity and is expanded to 4096 bytes when the first long data operation is performed. If your application requires a bigger or smaller long data buffer, you can specify a custom capacity using the long_data_buffer() accessor.

The mssql::connection_factory abstract class has the following interface:

namespace odb
{
  namespace mssql
  {
    class connection_factory
    {
    public:
      virtual void
      database (database&) = 0;

      virtual connection_ptr
      connect () = 0;
    };
  }
}
  

The database() function is called when a connection factory is associated with a database instance. This happens in the odb::mssql::database class constructors. The connect() function is called whenever a database connection is requested.

The two implementations of the connection_factory interface provided by the SQL Server ODB runtime are new_connection_factory and connection_pool_factory. You will need to include the <odb/mssql/connection-factory.hxx> header file to make the connection_factory interface and these implementation classes available in your application.

The new_connection_factory class creates a new connection whenever one is requested. When a connection is no longer needed, it is released and closed. The new_connection_factory class has the following interface:

namespace odb
{
  namespace mssql
  {
    class new_connection_factory: public connection_factory
    {
    public:
      new_connection_factory ();
    };
};
  

The connection_pool_factory class implements a connection pool. It has the following interface:

namespace odb
{
  namespace mssql
  {
    class connection_pool_factory: public connection_factory
    {
    public:
      connection_pool_factory (std::size_t max_connections = 0,
                               std::size_t min_connections = 0);

    protected:
      class pooled_connection: public connection
      {
      public:
        pooled_connection (database_type&);
        pooled_connection (database_type&, SQLHDBC handle);
      };

      typedef details::shared_ptr<pooled_connection> pooled_connection_ptr;

      virtual pooled_connection_ptr
      create ();
    };
};
  

The max_connections argument in the connection_pool_factory constructor specifies the maximum number of concurrent connections that this pool factory will maintain. Similarly, the min_connections argument specifies the minimum number of available connections that should be kept open.

Whenever a connection is requested, the pool factory first checks if there is an unused connection that can be returned. If there is none, the pool factory checks the max_connections value to see if a new connection can be created. If the total number of connections maintained by the pool is less than this value, then a new connection is created and returned. Otherwise, the caller is blocked until a connection becomes available.

When a connection is released, the pool factory first checks if there are blocked callers waiting for a connection. If so, then one of them is unblocked and is given the connection. Otherwise, the pool factory checks whether the total number of connections maintained by the pool is greater than the min_connections value. If that's the case, the connection is closed. Otherwise, the connection is added to the pool of available connections to be returned on the next request. In other words, if the number of connections maintained by the pool exceeds min_connections and there are no callers waiting for a new connection, the pool will close the excess connections.

If the max_connections value is 0, then the pool will create a new connection whenever all of the existing connections are in use. If the min_connections value is 0, then the pool will never close a connection and instead maintain all the connections that were ever created.

The create() virtual function is called whenever the pool needs to create a new connection. By deriving from the connection_pool_factory class and overriding this function we can implement custom connection establishment and configuration.

If you pass NULL as the connection factory to one of the database constructors, then the connection_pool_factory instance will be created by default with the min and max connections values set to 0. The following code fragment shows how we can pass our own connection factory instance:

#include <odb/database.hxx>

#include <odb/mssql/database.hxx>
#include <odb/mssql/connection-factory.hxx>

int
main (int argc, char* argv[])
{
  auto_ptr<odb::mssql::connection_factory> f (
    new odb::mssql::connection_pool_factory (20));

  auto_ptr<odb::database> db (
    new mssql::database (argc, argv, false, "", 0, f));
}
  

21.4 SQL Server Exceptions

The SQL Server ODB runtime library defines the following SQL Server-specific exceptions:

namespace odb
{
  namespace mssql
  {
    class database_exception: odb::database_exception
    {
    public:
      class record
      {
      public:
        SQLINTEGER
        error () const;

        const std::string&
        sqlstate () const;

        const std::string&
        message () const;
      };

      typedef std::vector<record> records;

      typedef records::size_type size_type;
      typedef records::const_iterator iterator;

      iterator
      begin () const;

      iterator
      end () const;

      size_type
      size () const;

      virtual const char*
      what () const throw ();
    };

    class cli_exception: odb::exception
    {
    public:
      virtual const char*
      what () const throw ();
    };

    class long_data_reload: odb::exception
    {
    public:
      virtual const char*
      what () const throw ();
    };
  }
}
  

You will need to include the <odb/mssql/exceptions.hxx> header file to make these exceptions available in your application.

The odb::mssql::database_exception is thrown if an SQL Server database operation fails. The SQL Server-specific error information is stored as a series of records, each containing the error code as a signed 4-byte integer, the SQLSTATE code, and the message string. All this information is also combined and returned in a human-readable form by the what() function.

The odb::mssql::cli_exception is thrown by the command line parsing constructor of the odb::mssql::database class if the SQL Server option values are missing or invalid. The what() function returns a human-readable description of an error.

The odb::mssql::long_data_reload is thrown if an attempt is made to re-load an object or view with long data as part of a query result iteration. For more information, refer to Section 21.1, "SQL Server Type Mapping".

21.5 SQL Server Limitations

The following sections describe SQL Server-specific limitations imposed by the current SQL Server and ODB runtime versions.

21.5.1 Query Result Caching

SQL Server ODB runtime implementation does not perform query result caching (Section 4.4, "Query Result") even when explicitly requested. The ODBC API and the SQL Server Native Client ODBC driver support interleaving execution of multiple prepared statements on a single connection. As a result, it is possible to have multiple uncached results and calls to other database functions do not invalidate them. The only limitation of the uncached SQL Server results is the unavailability of the result::size() function. If you call this function on an SQL Server query result, then the odb::result_not_cached exception (Section 3.14, "ODB Exceptions") is always thrown. Future versions of the SQL Server ODB runtime library may add support for result caching.

21.5.2 Foreign Key Constraints

ODB assumes the standard SQL behavior which requires that foreign key constraints checking is deferred until the transaction is committed. The only behavior supported by SQL Server is to check such constraints immediately. As a result, by default, schemas generated by the ODB compiler for SQL Server have foreign key definitions commented out. They are retained only for documentation.

You can override the default behavior and instruct the ODB compiler to generate non-deferrable foreign keys by specifying the --fkeys-deferrable-mode not_deferrable ODB compiler option. Note, however, that in this case the order in which you persist, update, and erase objects within a transaction becomes important.

21.5.3 Unique Constraint Violations

Due to the granularity of the ODBC error codes, it is impossible to distinguish between the duplicate primary key and other unique constraint violations. As a result, when making an object persistent, the SQL Server ODB runtime will translate all unique constraint violation errors to the object_already_persistent exception (Section 3.14, "ODB Exceptions").

21.5.4 Multi-threaded Windows Applications

Multi-threaded Windows applications must use the _beginthread()/_beginthreadex() and _endthread()/_endthreadex() CRT functions instead of the CreateThread() and EndThread() Win32 functions to start and terminate threads. This is a limitation of the ODBC implementation on Windows.

21.5.5 Affected Row Count and DDL Statements

SQL Server always returns zero as the number of affected rows for DDL statements. In particular, this means that the database::execute() (Section 3.12, "Executing Native SQL Statements") function will always return zero for such statements.

21.5.6 Long Data and Auto Object Ids, ROWVERSION

SQL Server 2005 has a bug that causes it to fail on an INSERT or UPDATE statement with the OUTPUT clause (used to return automatically assigned object ids as well as ROWVERSION values) if one of the inserted columns is long data. The symptom of this bug in ODB is an exception thrown by the database::persist() or database::update() function when used on an object that contains long data and has an automatically assigned object id or uses ROWVERSION-based optimistic concurrency (Section 21.1.1, "ROWVERSION Support"). The error message reads "This operation conflicts with another pending operation on this transaction. The operation failed."

For automatically assigned object ids ODB includes a workaround for this bug which uses a less efficient method to obtain id values for objects that contain long data. To enable this workaround you need to specify that the generated code will be used with SQL Server 2005 or later by passing the --mssql-server-version 9.0 ODB compiler option.

For ROWVERSION-based optimistic concurrency no workaround is currently provided. The ODB compiler will issue an error for objects that use ROWVERSION for optimistic concurrency and containing long data.

21.5.7 Long Data and By-Value Accessors/Modifiers

As discussed in Section 14.4.5, "get/set/access", by-value accessor and modifier expressions cannot be used with data members of long data types. The SQL Server ODB runtime uses streaming for reading/writing long data directly from/to data members. As a result, by-reference accessors and modifiers should be used for these data types.

21.5.8 Bulk Update and ROWVERSION

The bulk update operation (Section 15.3, "Bulk Database Operations") is not yet supported for persistent classes that use ROWVERSION-based optimistic concurrency. For such classes the bulk update() function is not available. The bulk persist and erase support is still provided.

21.6 SQL Server Index Definitions

When the index pragma (Section 14.7, "Index Definition Pragmas") is used to define an SQL Server index, the type clause specifies the index type (for example, UNIQUE, CLUSTERED), the method clause is not used, and the options clause specifies additional index properties. The column options can be used to specify the sort order. For example:

#pragma db object
class object
{
  ...

  std::string name_;

  #pragma db index                             \
             type("UNIQUE CLUSTERED")          \
             member(name_, "DESC")             \
             options("WITH(FILLFACTOR = 80)")
};
  

21.7 SQL Server Stored Procedures

ODB native views (Section 10.6, "Native Views") can be used to call SQL Server stored procedures. For example, assuming we are using the person class from Chapter 2, "Hello World Example" (and the corresponding person table), we can create a stored procedure that given the min and max ages returns some information about all the people in that range:

CREATE PROCEDURE dbo.person_range (
  @min_age SMALLINT,
  @max_age SMALLINT)
AS
  SELECT age, first, last FROM person
    WHERE age >= @min_age AND age <= @max_age;
  

Given the above stored procedure we can then define an ODB view that can be used to call it and retrieve its result:

#pragma db view query("EXEC person_range (?)")
struct person_range
{
  unsigned short age;
  std::string first;
  std::string last;
};
  

The following example shows how we can use the above view to print the list of people in a specific age range:

typedef odb::query<person_range> query;
typedef odb::result<person_range> result;

transaction t (db.begin ());

result r (
  db.query<person_range> (
    query::_val (1) + "," + query::_val (18)));

for (result::iterator i (r.begin ()); i != r.end (); ++i)
  cerr << i->first << " " << i->last << " " << i->age << endl;

t.commit ();
  

Note that as with all native views, the order and types of data members must match those of columns in the SELECT list inside the stored procedure.

There are also a number of limitations when it comes to calling SQL Server stored procedures with ODB views. There is currently no support for output parameters, however, this is planned for a future version. In the meantime, to call a stored procedure that has output parameters we have to use a wrapper procedure that converts such parameters to a SELECT result. For example, given the following procedure that calculates the age range of the people in our database:

CREATE PROCEDURE dbo.person_age_range (
  @min_age SMALLINT = NULL OUTPUT,
  @max_age SMALLINT = NULL OUTPUT)
AS
  SELECT @min_age = MIN(age), @max_age = MAX(max) FROM person;
  

We can create a wrapper procedure like this:

CREATE PROCEDURE dbo.person_age_range_odb
AS
  DECLARE @min_age SMALLINT, @max_age SMALLINT;
  EXEC person_age_range @min_age OUTPUT, @max_age OUTPUT;
  SELECT @min_age, @max_age;
   

And a view like this:

#pragma db view query("EXEC person_age_range_odb")
struct person_age_range
{
  unsigned short min_age;
  unsigned short max_age;
};
  

Which we can then use to call the stored procedure:

transaction t (db.begin ());

person_age_range ar (db.query_value<person_age_range> ());
cerr << ar.min_age << " " << ar.max_age << endl;

t.commit ();
  

In SQL Server, a stored procedure can produce multiple results. For example, if a stored procedure executes several SELECT statements, then the result of calling such a procedure consists of two row sets, one for each SELECT statement. Because such multiple row sets can contain varying number and type of columns, they cannot be all extracted into a single view. Consequently, these kind of stored procedures are currently not supported.

A stored procedure may also produce no row sets at all. For example, a stored procedure that only executes DML statements would exhibit this behavior. To call such a procedure we use an empty view, for example:

CREATE PROCEDURE dbo.insert_person (
  @first VARCHAR(512),
  @last VARCHAR(512),
  @age SMALLINT)
AS
  INSERT INTO person(first, last, age)
    VALUES(@first, @last, @age);
  
#pragma db view
struct no_result {};

transaction t (db.begin ());

db.query_one<no_result> (
  "EXEC insert_person" +
    query::_val ("John") + "," +
    query::_val ("Doe") + "," +
    query::_val (21));

t.commit ();
  

Finally, an SQL Server stored procedure can also return an integer status code. Similar to output parameters, this code can only be observed by an ODB view if it is converted to a SELECT result. For more information on how to do this and for other examples of stored procedure calls, refer to the mssql/stored-proc test in the odb-tests package.


PART III   PROFILES

Part III covers the integration of ODB with popular C++ frameworks and libraries. It consists of the following chapters.

22Profiles Introduction
23Boost Profile
24Qt Profile

22 Profiles Introduction

ODB profiles are a generic mechanism for integrating ODB with widely-used C++ frameworks and libraries. A profile provides glue code which allows you to seamlessly persist various components, such as smart pointers, containers, and value types found in these frameworks or libraries. The code necessary to implement a profile is packaged into the so called profile library. For example, the Boost profile implementation is provided by the libodb-boost profile library.

Besides linking the profile library to our application, it is also necessary to let the ODB compiler know which profiles we are using. This is accomplished with the --profile (or -p alias) option. For example:

odb --profile boost ...
  

Some profiles, especially those covering frameworks or libraries that consist of multiple sub-libraries, provide sub-profiles that allow you to pick and choose which components you would like to use in your application. For example, the boost profile contains the boost/data-time sub-profile. If we are only interested in the date_time types, then we can pass boost/data-time instead of boost to the --profile option, for example:

odb --profile boost/date-time ...
  

To summarize, you will need to perform the following steps in order to make use of a profile in your application:

  1. ODB compiler: if necessary, specify the path to the profile library headers (-I option).
  2. ODB compiler: specify the profile you would like to use with the --profile option.
  3. C++ compiler: if necessary, specify the path to the profile library headers (normally -I option).
  4. Linker: link the profile library to the application.

The remaining chapters in this part of the manual describe the standard profiles provided by ODB.


23 Boost Profile

The ODB profile implementation for Boost is provided by the libodb-boost library and consists of multiple sub-profiles corresponding to the individual Boost libraries. To enable all the available Boost sub-profiles, pass boost as the profile name to the --profile ODB compiler option. Alternatively, you can enable only specific sub-profiles by passing individual sub-profile names to --profile. The following sections in this chapter discuss each Boost sub-profile in detail. The boost example in the odb-examples package shows how to enable and use the Boost profile.

Some sub-profiles may throw exceptions to indicate error conditions, such as the inability to store a specific value in a particular database system. All such exceptions derive from the odb::boost::exception class which in turn derives from the root of the ODB exception hierarchy, class odb::exception (Section 3.14, "ODB Exceptions"). The odb::boost::exception class is defined in the <odb/boost/exception.hxx> header file and has the same interface as odb::exception. Concrete exceptions that can be thrown by the Boost sub-profiles are described in the following sections.

23.1 Smart Pointers Library

The smart-ptr sub-profile provides persistence support for a subset of smart pointers from the Boost smart_ptr library. To enable only this profile, pass boost/smart-ptr to the --profile ODB compiler option.

The currently supported smart pointers are boost::shared_ptr and boost::weak_ptr. For more information on using smart pointers as pointers to objects and views, refer to Section 3.3, "Object and View Pointers" and Chapter 6, "Relationships". For more information on using smart pointers as pointers to values, refer to Section 7.3, "Pointers and NULL Value Semantics". When used as a pointer to a value, only boost::shared_ptr is supported. For example:

#pragma db object
class person
{
  ...

  #pragma db null
  boost::shared_ptr<std::string> middle_name_;
};
  

To provide finer grained control over object relationship loading, the smart-ptr sub-profile also provides the lazy counterparts for the above pointers: odb::boost::lazy_shared_ptr and odb::boost::lazy_weak_ptr. You will need to include the <odb/boost/lazy-ptr.hxx> header file to make the lazy variants available in your application. For a description of the lazy pointer interface and semantics refer to Section 6.4, "Lazy Pointers". The following example shows how we can use these smart pointers to establish a relationship between persistent objects.

class employee;

#pragma db object
class position
{
  ...

  #pragma db inverse(position_)
  odb::boost::lazy_weak_ptr<employee> employee_;
};

#pragma db object
class employee
{
  ...

  #pragma db not_null
  boost::shared_ptr<position> position_;
};
  

Besides providing persistence support for the above smart pointers, the smart-ptr sub-profile also changes the default pointer (Section 3.3, "Object and View Pointers") to boost::shared_ptr. In particular, this means that database functions that return dynamically allocated objects and views will return them as boost::shared_ptr pointers. To override this behavior, add the --default-pointer option specifying the alternative pointer type after the --profile option.

23.2 Unordered Containers Library

The unordered sub-profile provides persistence support for the containers from the Boost unordered library. To enable only this profile, pass boost/unordered to the --profile ODB compiler option.

The supported containers are boost::unordered_set, boost::unordered_map, boost::unordered_multiset, and boost::unordered_multimap. For more information on using the set and multiset containers with ODB, refer to Section 5.2, "Set and Multiset Containers". For more information on using the map and multimap containers with ODB, refer to Section 5.3, "Map and Multimap Containers". The following example shows how the unordered_set container may be used within a persistent object.

#pragma db object
class person
{
  ...
  boost::unordered_set<std::string> emails_;
};
  

23.3 Multi-Index Container Library

The multi-index sub-profile provides persistence support for boost::multi_index_container from the Boost Multi-Index library. To enable only this profile, pass boost/multi-index to the --profile ODB compiler option. The following example shows how multi_index_container may be used within a persistent object.

namespace mi = boost::multi_index;

#pragma db object
class person
{
  ...

  typedef
  mi::multi_index_container<
    std::string,
    mi::indexed_by<
      mi::sequenced<>,
      mi::ordered_unique<mi::identity<std::string> >
    >
  > emails;

  emails emails_;
};
  

Note that a multi_index_container instantiation is stored differently in the database depending on whether it has any sequenced or random_access indexes. If it does, then it is treated as an ordered container (Section 5.1, "Ordered Containers") with the first such index establishing the order. Otherwise, it is treated as a set container (Section 5.2, "Set and Multiset Containers").

Note also that there is a terminology clash between ODB and Boost Multi-Index. The ODB term ordered container translates to Multi-Index terms sequenced index and random access index while the ODB term set container translates to Multi-Index terms ordered index and hashed index.

The emails container from the above example is stored as an ordered container. In contrast, the following aliases container is stored as a set.

namespace mi = boost::multi_index;

#pragma db value
struct name
{
  std::string first;
  std::string last;
};

bool operator< (const name&, const name&);

#pragma db object
class person
{
  ...

  typedef
  mi::multi_index_container<
    name,
    mi::indexed_by<
      mi::ordered_unique<mi::identity<name> >
      mi::ordered_non_unique<
        mi::member<name, std::string, &name::first>
      >,
      mi::ordered_non_unique<
        mi::member<name, std::string, &name::last>
      >
    >
  > aliases;

  aliases aliases_;
};
  

23.4 Optional Library

The optional sub-profile provides persistence support for the boost::optional container from the Boost optional library. To enable only this profile, pass boost/optional to the --profile ODB compiler option.

In a relational database boost::optional is mapped to a column that can have a NULL value. Similar to odb::nullable (Section 7.3, "Pointers and NULL Value Semantics"), it can be used to add the NULL semantics to existing C++ types. For example:

#include <boost/optional.hpp>

#pragma db object
class person
{
  ...

  std::string first_;                    // TEXT NOT NULL
  boost::optional<std::string> middle_;  // TEXT NULL
  std::string last_;                     // TEXT NOT NULL
};
  

Note also that similar to odb::nullable, when this profile is used, the NULL values are automatically enabled for data members of the boost::optional type.

23.5 Date Time Library

The date-time sub-profile provides persistence support for a subset of types from the Boost date_time library. It is further subdivided into two sub-profiles, gregorian and posix_time. The gregorian sub-profile provides support for types from the boost::gregorian namespace, while the posix-time sub-profile provides support for types from the boost::posix_time namespace. To enable the entire date-time sub-profile, pass boost/date-time to the --profile ODB compiler option. To enable only the gregorian sub-profile, pass boost/date-time/gregorian, and to enable only the posix-time sub-profile, pass boost/date-time/posix-time.

The only type that the gregorian sub-profile currently supports is gregorian::date. The types currently supported by the posix-time sub-profile are posix_time::ptime and posix_time::time_duration. The manner in which these types are persisted is database system dependent and is discussed in the sub-sections that follow. The example below shows how gregorian::date may be used within a persistent object.

#pragma db object
class person
{
  ...
  boost::gregorian::date date_of_birth_;
};
  

Concrete exceptions that can be thrown by the date-time sub-profile implementation are presented below.

namespace odb
{
  namespace boost
  {
    namespace date_time
    {
      struct special_value: odb::boost::exception
      {
        virtual const char*
        what () const throw ();
      };

      struct value_out_of_range: odb::boost::exception
      {
        virtual const char*
        what () const throw ();
      };
    }
  }
}
  

You will need to include the <odb/boost/date-time/exceptions.hxx> header file to make these exceptions available in your application.

The special_value exception is thrown if an attempt is made to store a Boost date-time special value that cannot be represented in the target database. The value_out_of_range exception is thrown if an attempt is made to store a date-time value that is out of the target database range. The specific conditions under which these exceptions are thrown are database system dependent and are discussed in more detail in the following sub-sections.

23.5.1 MySQL Database Type Mapping

The following table summarizes the default mapping between the currently supported Boost date_time types and the MySQL database types.

Boost date_time Type MySQL Type Default NULL Semantics
gregorian::date DATE NULL
posix_time::ptime DATETIME NULL
posix_time::time_duration TIME NULL

The Boost special value date_time::not_a_date_time is stored as a NULL value in a MySQL database.

The posix-time sub-profile implementation also provides support for mapping posix_time::ptime to the TIMESTAMP MySQL type. However, this mapping has to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class person
{
  ...
  #pragma db type("TIMESTAMP") not_null
  boost::posix_time::ptime updated_;
};
  

Starting with MySQL version 5.6.4 it is possible to store fractional seconds up to microsecond precision in TIME, DATETIME, and TIMESTAMP columns. However, to enable sub-second precision, the corresponding type with the desired precision has to be specified explicitly, as shown in the following example:

#pragma db object
class person
{
  ...
  #pragma db type("DATETIME(6)")     // Microsecond precision.
  boost::posix_time::ptime updated_;
};
  

Alternatively, you can enable sub-second precision on the per-type basis, for example:

#pragma db value(boost::posix_time::ptime) type("DATETIME(6)")

#pragma db object
class person
{
  ...
  boost::posix_time::ptime created_; // Microsecond precision.
  boost::posix_time::ptime updated_; // Microsecond precision.
};
  

Some valid Boost date-time values cannot be stored in a MySQL database. An attempt to persist any Boost date-time special value other than date_time::not_a_date_time will result in the special_value exception. An attempt to persist a Boost date-time value that is out of the MySQL type range will result in the out_of_range exception. Refer to the MySQL documentation for more information on the MySQL data type ranges.

23.5.2 SQLite Database Type Mapping

The following table summarizes the default mapping between the currently supported Boost date_time types and the SQLite database types.

Boost date_time Type SQLite Type Default NULL Semantics
gregorian::date TEXT NULL
posix_time::ptime TEXT NULL
posix_time::time_duration TEXT NULL

The Boost special value date_time::not_a_date_time is stored as a NULL value in an SQLite database.

The date-time sub-profile implementation also provides support for mapping gregorian::date and posix_time::ptime to the INTEGER SQLite type, with the integer value representing the UNIX time. Similarly, an alternative mapping for posix_time::time_duration to the INTEGER type represents the duration as a number of seconds. These mappings have to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class person
{
  ...
  #pragma db type("INTEGER")
  boost::gregorian::date born_;
};
  

Some valid Boost date-time values cannot be stored in an SQLite database. An attempt to persist any Boost date-time special value other than date_time::not_a_date_time will result in the special_value exception. An attempt to persist a negative posix_time::time_duration value as SQLite TEXT will result in the out_of_range exception.

23.5.3 PostgreSQL Database Type Mapping

The following table summarizes the default mapping between the currently supported Boost date_time types and the PostgreSQL database types.

Boost date_time Type PostgreSQL Type Default NULL Semantics
gregorian::date DATE NULL
posix_time::ptime TIMESTAMP NULL
posix_time::time_duration TIME NULL

The Boost special value date_time::not_a_date_time is stored as a NULL value in a PostgreSQL database. posix_time::ptime values representing the special values date_time::pos_infin and date_time::neg_infin are stored as the special PostgreSQL TIMESTAMP values infinity and -infinity, respectively.

Some valid Boost date-time values cannot be stored in a PostgreSQL database. The PostgreSQL TIME type represents a clock time, and can therefore only store positive durations with a total length of time less than 24 hours. An attempt to persist a posix_time::time_duration value outside of this range will result in the value_out_of_range exception. An attempt to persist a posix_time::time_duration value representing any special value other than date_time::not_a_date_time will result in the special_value exception.

23.5.4 Oracle Database Type Mapping

The following table summarizes the default mapping between the currently supported Boost date_time types and the Oracle database types.

Boost date_time Type Oracle Type Default NULL Semantics
gregorian::date DATE NULL
posix_time::ptime TIMESTAMP NULL
posix_time::time_duration INTERVAL DAY TO SECOND NULL

The Boost special value date_time::not_a_date_time is stored as a NULL value in an Oracle database.

The date-time sub-profile implementation also provides support for mapping posix_time::ptime to the DATE Oracle type with fractional seconds that may be stored in a ptime instance being ignored. This alternative mapping has to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class person
{
  ...
  #pragma db type("DATE")
  boost::posix_time::ptime updated_;
};
  

Some valid Boost date-time values cannot be stored in an Oracle database. An attempt to persist a gregorian::date, posix_time::ptime, or posix_time::time_duration value representing any special value other than date_time::not_a_date_time will result in the special_value exception.

23.5.5 SQL Server Database Type Mapping

The following table summarizes the default mapping between the currently supported Boost date_time types and the SQL Server database types.

Boost date_time Type SQL Server Type Default NULL Semantics
gregorian::date DATE NULL
posix_time::ptime DATETIME2 NULL
posix_time::time_duration TIME NULL

The Boost special value date_time::not_a_date_time is stored as a NULL value in an SQL Server database.

Note that the DATE, TIME, and DATETIME2 types are only available in SQL Server 2008 and later. SQL Server 2005 only supports the DATETIME and SMALLDATETIME date-time types. The new types are also unavailable when connecting to an SQL Server 2008 or later with the SQL Server 2005 Native Client ODBC driver.

The date-time sub-profile implementation provides support for mapping posix_time::ptime to the DATETIME and SMALLDATETIME types, however, this mapping has to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class person
{
  ...
  #pragma db type("DATETIME")
  boost::posix_time::ptime updated_;
};
  

Some valid Boost date-time values cannot be stored in an SQL Server database. An attempt to persist a gregorian::date, posix_time::ptime, or posix_time::time_duration value representing any special value other than date_time::not_a_date_time will result in the special_value exception. The range of the TIME type in SQL server is from 00:00:00.0000000 to 23:59:59.9999999. An attempt to persist a posix_time::time_duration value out of this range will result in the value_out_of_range exception.

23.6 Uuid Library

The uuid sub-profile provides persistence support for the uuid type from the Boost uuid library. To enable only this profile, pass boost/uuid to the --profile ODB compiler option.

The manner in which these types are persisted is database system dependent and is discussed in the sub-sections that follow. By default a data member of the uuid type is mapped to a database column with NULL enabled and nil uuid instances are stored as a NULL value. However, you can change this behavior by declaring the data member NOT NULL with the not_null pragma (Section 14.4.6, "null/not_null"). In this case, or if the data member is an object id, the implementation will store nil uuid instances as zero UUID values ({00000000-0000-0000-0000-000000000000}). For example:

#pragma db object
class object
{
  ...

  boost::uuids::uuid x_; // Nil values stored as NULL.

  #pragma db not_null
  boost::uuids::uuid y_; // Nil values stored as zero.
};
  

23.6.1 MySQL Database Type Mapping

The following table summarizes the default mapping between the Boost uuid type and the MySQL database type.

Boost Type MySQL Type Default NULL Semantics
boost::uuids::uuid BINARY(16) NULL

23.6.2 SQLite Database Type Mapping

The following table summarizes the default mapping between the Boost uuid type and the SQLite database type.

Boost Type SQLite Type Default NULL Semantics
boost::uuids::uuid BLOB NULL

23.6.3 PostgreSQL Database Type Mapping

The following table summarizes the default mapping between the Boost uuid type and the PostgreSQL database type.

Boost Type PostgreSQL Type Default NULL Semantics
boost::uuids::uuid UUID NULL

23.6.4 Oracle Database Type Mapping

The following table summarizes the default mapping between the Boost uuid type and the Oracle database type.

Boost Type Oracle Type Default NULL Semantics
boost::uuids::uuid RAW(16) NULL

23.6.5 SQL Server Database Type Mapping

The following table summarizes the default mapping between the Boost uuid type and the SQL Server database type.

Boost Type SQL Server Type Default NULL Semantics
boost::uuids::uuid UNIQUEIDENTIFIER NULL

24 Qt Profile

The ODB profile implementation for Qt is provided by the libodb-qt library. Both Qt4 and Qt5 as well as C++98/03 and C++11 are supported.

The Qt profile consists of multiple sub-profiles corresponding to the common type groups within Qt. Currently, only types from the QtCore module are supported. To enable all the available Qt sub-profiles, pass qt as the profile name to the --profile ODB compiler option. Alternatively, you can enable only specific sub-profiles by passing individual sub-profile names to --profile. The following sections in this chapter discuss each Qt sub-profile in detail. The qt example in the odb-examples package shows how to enable and use the Qt profile.

Some sub-profiles may throw exceptions to indicate error conditions, such as the inability to store a specific value in a particular database system. All such exceptions derive from the odb::qt::exception class which in turn derives from the root of the ODB exception hierarchy, class odb::exception (Section 3.14, "ODB Exceptions"). The odb::qt::exception class is defined in the <odb/qt/exception.hxx> header file and has the same interface as odb::exception. Concrete exceptions that can be thrown by the Qt sub-profiles are described in the following sections.

24.1 Basic Types Library

The basic sub-profile provides persistence support for basic types defined by Qt. To enable only this profile, pass qt/basic to the --profile ODB compiler option.

The currently supported basic types are QString, QByteArray, and QUuid. The manner in which these types are persisted is database system dependent and is discussed in the sub-sections that follow. The example below shows how QString may be used within a persistent object.

#pragma db object
class Person
{
  ...
  QString name_;
};
  

By default a data member of the QUuid type is mapped to a database column with NULL enabled and null QUuid instances are stored as a NULL value. However, you can change this behavior by declaring the data member NOT NULL with the not_null pragma (Section 14.4.6, "null/not_null"). In this case, or if the data member is an object id, the implementation will store null QUuid instances as zero UUID values ({00000000-0000-0000-0000-000000000000}). For example:

#pragma db object
class object
{
  ...

  QUuid x_; // Null values stored as NULL.

  #pragma db not_null
  QUuid y_; // Null values stored as zero.
};
  

24.1.1 MySQL Database Type Mapping

The following table summarizes the default mapping between the currently supported basic Qt types and the MySQL database types.

Qt Type MySQL Type Default NULL Semantics
QString TEXT/VARCHAR(255) NULL
QByteArray BLOB NULL
QUuid BINARY(16) NULL

Instances of the QString and QByteArray types are stored as a NULL value if their isNull() member function returns true.

Note also that the QString type is mapped differently depending on whether a member of this type is an object id or not. If the member is an object id, then for this member QString is mapped to the VARCHAR(255) MySQL type. Otherwise, it is mapped to TEXT.

The basic sub-profile also provides support for mapping QString to the CHAR, NCHAR, and NVARCHAR MySQL types. However, these alternative mappings have to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class Person
{
  ...

  #pragma db type("CHAR(2)") not_null
  QString licenseState_;
};
  

24.1.2 SQLite Database Type Mapping

The following table summarizes the default mapping between the currently supported basic Qt types and the SQLite database types.

Qt Type SQLite Type Default NULL Semantics
QString TEXT NULL
QByteArray BLOB NULL
QUuid BLOB NULL

Instances of the QString and QByteArray types are stored as a NULL value if their isNull() member function returns true.

24.1.3 PostgreSQL Database Type Mapping

The following table summarizes the default mapping between the currently supported basic Qt types and the PostgreSQL database types.

Qt Type PostgreSQL Type Default NULL Semantics
QString TEXT NULL
QByteArray BYTEA NULL
QUuid UUID NULL

Instances of the QString and QByteArray types are stored as a NULL value if their isNull() member function returns true.

The basic sub-profile also provides support for mapping QString to the CHAR and VARCHAR PostgreSQL types. However, these alternative mappings have to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class Person
{
  ...

  #pragma db type("CHAR(2)") not_null
  QString licenseState_;
};
  

24.1.4 Oracle Database Type Mapping

The following table summarizes the default mapping between the currently supported basic Qt types and the Oracle database types.

Qt Type Oracle Type Default NULL Semantics
QString VARCHAR2(512) NULL
QByteArray BLOB NULL
QUuid RAW(16) NULL

Instances of the QString and QByteArray types are stored as a NULL value if their isNull() member function returns true.

The basic sub-profile also provides support for mapping QString to the CHAR, NCHAR, NVARCHAR, CLOB, and NCLOB Oracle types, and for mapping QByteArray to the RAW Oracle type. However, these alternative mappings have to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class Person
{
  ...

  #pragma db type("CLOB") not_null
  QString firstName_;

  #pragma db type("RAW(16)") null
  QByteArray uuid_;
};
  

24.1.5 SQL Server Database Type Mapping

The following table summarizes the default mapping between the currently supported basic Qt types and the SQL Server database types.

Qt Type SQL Server Type Default NULL Semantics
QString VARCHAR(512)/VARCHAR(256) NULL
QByteArray VARBINARY(max) NULL
QUuid UNIQUEIDENTIFIER NULL

Instances of the QString and QByteArray types are stored as a NULL value if their isNull() member function returns true.

Note also that the QString type is mapped differently depending on whether a member of this type is an object id or not. If the member is an object id, then for this member QString is mapped to the VARCHAR(256) SQL Server type. Otherwise, it is mapped to VARCHAR(512).

The basic sub-profile also provides support for mapping QString to the CHAR, NCHAR, NVARCHAR, TEXT, and NTEXT SQL Server types, and for mapping QByteArray to the BINARY and IMAGE SQL Server types. However, these alternative mappings have to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class Person
{
  ...

  #pragma db type("NVARCHAR(256)") not_null
  QString firstName_;

  #pragma db type("BINARY(16)") null
  QByteArray uuid_;
};
  

24.2 Smart Pointers Library

The smart-ptr sub-profile provides persistence support the Qt smart pointers. To enable only this profile, pass qt/smart-ptr to the --profile ODB compiler option.

The currently supported smart pointers are QSharedPointer and QWeakPointer. For more information on using smart pointers as pointers to objects and views, refer to Section 3.3, "Object and View Pointers" and Chapter 6, "Relationships". For more information on using smart pointers as pointers to values, refer to Section 7.3, "Pointers and NULL Value Semantics". When used as a pointer to a value, only QSharedPointer is supported. For example:

#pragma db object
class person
{
  ...

  #pragma db null
  QSharedPointer<QString> middle_name_;
};
  

To provide finer grained control over object relationship loading, the smart-ptr sub-profile also provides the lazy counterparts for the above pointers: QLazySharedPointer and QLazyWeakPointer. You will need to include the <odb/qt/lazy-ptr.hxx> header file to make the lazy variants available in your application. For a description of the lazy pointer interface and semantics refer to Section 6.4, "Lazy Pointers". The following example shows how we can use these smart pointers to establish a relationship between persistent objects.

class Employee;

#pragma db object
class Position
{
  ...

  #pragma db inverse(position_)
  QLazyWeakPointer<Employee> employee_;
};

#pragma db object
class Employee
{
  ...

  #pragma db not_null
  QSharedPointer<Position> position_;
};
  

Besides providing persistence support for the above smart pointers, the smart-ptr sub-profile also changes the default pointer (Section 3.3, "Object and View Pointers") to QSharedPointer. In particular, this means that database functions that return dynamically allocated objects and views will return them as QSharedPointer pointers. To override this behavior, add the --default-pointer option specifying the alternative pointer type after the --profile option.

24.3 Containers Library

The containers sub-profile provides persistence support for Qt containers. To enable only this profile, pass qt/containers to the --profile ODB compiler option.

The currently supported ordered containers are QVector, QList, and QLinkedList. Supported map containers are QMap, QMultiMap, QHash, and QMultiHash. The supported set container is QSet. For more information on using containers with ODB, refer to Chapter 5, "Containers". The following example shows how the QSet container may be used within a persistent object.

#pragma db object
class Person
{
  ...
  QSet<QString> emails_;
};
  

The containers sub-profile also provide a change-tracking equivalent for QList (Section 24.3.1, "Change-Tracking QList") with support for other Qt container equivalents planned for future releases. For general information on change-tracking containers refer to Section 5.4, "Change-Tracking Containers".

24.3.1 Change-Tracking QList

Class template QOdbList, defined in <odb/qt/list.hxx>, is a change-tracking equivalent for QList. It is implemented in terms of QList and is implicit-convertible to and implicit-constructible from const QList&. In particular, this means that we can use QOdbList instance anywhere const QList& is expected. In addition, QOdbList constant iterator (const_iterator) is the same type as that of QList.

QOdbList incurs 2-bit per element overhead in order to store the change state. It cannot be stored unordered in the database (Section 14.4.19 "unordered") but can be used as an inverse side of a relationship (6.2 "Bidirectional Relationships"). In this case, no change tracking is performed since no state for such a container is stored in the database.

The number of database operations required to update the state of QOdbList corresponds well to the complexity of QList functions, except for prepend()/push_front(). In particular, adding or removing an element from the back of the list (for example, with append()/push_back() and removeLast()/pop_back()), requires only a single database statement execution. In contrast, inserting or erasing an element at the beginning or in the middle of the list will require a database statement for every element that follows it.

QOdbList replicates most of the QList interface as defined in both Qt4 and Qt5 and includes support for C++11. However, functions and operators that provide direct write access to the elements had to be altered or disabled in order to support change tracking. Additional functions used to interface with QList and to control the change tracking state were also added. The following listing summarizes the differences between the QOdbList and QList interfaces. Any QList function or operator not mentioned in this listing has exactly the same signature and semantics in QOdbList. Functions and operators that were disabled are shown as commented out and are followed by functions/operators that replace them.

template <typename T>
class QOdbList
{
  ...

  // Element access.
  //

  //T& operator[] (int);
    T& modify (int);

  //T& first();
    T& modifyFirst();

  //T& last();
    T& modifyLast();

  //T& front();
    T& modify_front();

  //T& back();
    T& modify_back();

  // Iterators.
  //
  typedef typename QList<T>::const_iterator const_iterator;

  class iterator
  {
    ...

    // Element Access.
    //

    //reference       operator* () const;
      const_reference operator* () const;
      reference       modify () const;

    //pointer       operator-> () const;
      const_pointer operator-> () const;

    //reference       operator[] (difference_type);
      const_reference operator[] (difference_type);
      reference       modify (difference_type) const;

    // Interfacing with QList::iterator.
    //
    typename QList<T>::iterator base () const;
  };

  // Return QList iterators. The begin() functions mark all
  // the elements as modified.
  //
  typename QList<T>::iterator mbegin ();
  typename QList<T>::iterator modifyBegin ();
  typename QList<T>::iterator mend ();
  typename QList<T>::iterator modifyEnd ();

  // Interfacing with QList.
  //
  QOdbList (const QList<T>&);
  QOdbList (QList<T>&&); // C++11 only.

  QOdbList& operator= (const QList<T>&);
  QOdbList& operator= (QList<T>&&);

  operator const QList<T>& () const;
  QList<T>& base ();
  const QList<T>& base () const;

  // Change tracking.
  //
  bool _tracking () const;
  void _start () const;
  void _stop () const;
  void _arm (transaction&) const;
};
  

The following example highlights some of the differences between the two interfaces. QList versions are commented out.

#include <QtCore/QList>
#include <odb/qt/list.hxx>

void f (const QList<int>&);

QOdbList<int> l ({1, 2, 3});

f (l); // Ok, implicit conversion.

if (l[1] == 2) // Ok, const access.
  //l[1]++;
  l.modify (1)++;

//l.last () = 4;
l.modifyLast () = 4;

for (auto i (l.begin ()); i != l.end (); ++i)
{
  if (*i != 0) // Ok, const access.
    //*i += 10;
    i.modify () += 10;
}

qSort (l.modifyBegin (), l.modifyEnd ());
  

Note also the subtle difference between copy/move construction and copy/move assignment of QOdbList instances. While copy/move constructor will copy/move both the elements as well as their change state, in contrast, assignment is tracked as any other change to the vector content.

The QListIterator and QMutableListIterator equivalents are also provided. These are QOdbListIterator and QMutableOdbListIterator and are defined in <odb/qt/list-iterator.hxx> and <odb/qt/mutable-list-iterator.hxx>, respectively.

QOdbListIterator has exactly the same interface and semantics as QListIterator. In fact, we can use QListIterator to iterate over a QOdbList instance.

QMutableOdbListIterator also has exactly the same interface as QMutableListIterator. Note, however, that any element that such an iterator passes over with the call to next() is marked as modified.

24.4 Date Time Library

The date-time sub-profile provides persistence support for the Qt date-time types. To enable only this profile, pass qt/date-time to the --profile ODB compiler option.

The currently supported date-time types are QDate, QTime, and QDateTime. The manner in which these types are persisted is database system dependent and is discussed in the sub-sections that follow. The example below shows how QDate may be used within a persistent object.

#pragma db object
class Person
{
  ...
  QDate dateOfBirth_;
};
  

The single concrete exception that can be thrown by the date-time sub-profile implementation is presented below.

namespace odb
{
  namespace qt
  {
    namespace date_time
    {
      struct value_out_of_range: odb::qt::exception
      {
        virtual const char*
        what () const throw ();
      };
    }
  }
}
  

You will need to include the <odb/qt/date-time/exceptions.hxx> header file to make this exception available in your application.

The value_out_of_range exception is thrown if an attempt is made to store a date-time value that is out of the target database range. The specific conditions under which it is thrown is database system dependent and is discussed in more detail in the following sub-sections.

24.4.1 MySQL Database Type Mapping

The following table summarizes the default mapping between the currently supported Qt date-time types and the MySQL database types.

Qt Date Time Type MySQL Type Default NULL Semantics
QDate DATE NULL
QTime TIME NULL
QDateTime DATETIME NULL

Instances of the QDate, QTime, and QDateTime types are stored as a NULL value if their isNull() member function returns true.

The date-time sub-profile implementation also provides support for mapping QDateTime to the TIMESTAMP MySQL type. However, this mapping has to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class Person
{
  ...
  #pragma db type("TIMESTAMP") not_null
  QDateTime updated_;
};
  

Starting with MySQL version 5.6.4 it is possible to store fractional seconds up to microsecond precision in TIME, DATETIME, and TIMESTAMP columns. However, to enable sub-second precision, the corresponding type with the desired precision has to be specified explicitly, as shown in the following example:

#pragma db object
class Person
{
  ...
  #pragma db type("DATETIME(3)") // Millisecond precision.
  QDateTime updated_;
};
  

Alternatively, you can enable sub-second precision on the per-type basis, for example:

#pragma db value(QDateTime) type("DATETIME(3)")

#pragma db object
class Person
{
  ...
  QDateTime created_; // Millisecond precision.
  QDateTime updated_; // Millisecond precision.
};
  

Some valid Qt date-time values cannot be stored in a MySQL database. An attempt to persist a Qt date-time value that is out of the MySQL type range will result in the out_of_range exception. Refer to the MySQL documentation for more information on the MySQL data type ranges.

24.4.2 SQLite Database Type Mapping

The following table summarizes the default mapping between the currently supported Qt date-time types and the SQLite database types.

Qt Date Time Type SQLite Type Default NULL Semantics
QDate TEXT NULL
QTime TEXT NULL
QDateTime TEXT NULL

Instances of the QDate, QTime, and QDateTime types are stored as a NULL value if their isNull() member function returns true.

The date-time sub-profile implementation also provides support for mapping QDate and QDateTime to the SQLite INTEGER type, with the integer value representing the UNIX time. Similarly, an alternative mapping for QTime to the INTEGER type represents a clock time as the number of seconds since midnight. These mappings have to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class Person
{
  ...
  #pragma db type("INTEGER")
  QDate born_;
};
  

Some valid Qt date-time values cannot be stored in an SQLite database. An attempt to persist any Qt date-time value representing a negative UNIX time (any point in time prior to the 1970-01-01 00:00:00 UNIX time epoch) as an SQLite INTEGER will result in the out_of_range exception.

24.4.3 PostgreSQL Database Type Mapping

The following table summarizes the default mapping between the currently supported Qt date-time types and the PostgreSQL database types.

Qt Date Time Type PostgreSQL Type Default NULL Semantics
QDate DATE NULL
QTime TIME NULL
QDateTime TIMESTAMP NULL

Instances of the QDate, QTime, and QDateTime types are stored as a NULL value if their isNull() member function returns true.

24.4.4 Oracle Database Type Mapping

The following table summarizes the default mapping between the currently supported Qt date-time types and the Oracle database types.

Qt Date Time Type Oracle Type Default NULL Semantics
QDate DATE NULL
QTime INTERVAL DAY(0) TO SECOND(3) NULL
QDateTime TIMESTAMP(3) NULL

Instances of the QDate, QTime, and QDateTime types are stored as a NULL value if their isNull() member function returns true.

The date-time sub-profile implementation also provides support for mapping QDateTime to the DATE Oracle type with fractional seconds that may be stored in a QDateTime instance being ignored. This alternative mapping has to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class person
{
  ...
  #pragma db type("DATE")
  QDateTime updated_;
};
  

24.4.5 SQL Server Database Type Mapping

The following table summarizes the default mapping between the currently supported Qt date-time types and the SQL Server database types.

Qt Date Time Type SQL Server Type Default NULL Semantics
QDate DATE NULL
QTime TIME(3) NULL
QDateTime DATETIME2(3) NULL

Instances of the QDate, QTime, and QDateTime types are stored as a NULL value if their isNull() member function returns true.

Note that the DATE, TIME, and DATETIME2 types are only available in SQL Server 2008 and later. SQL Server 2005 only supports the DATETIME and SMALLDATETIME date-time types. The new types are also unavailable when connecting to an SQL Server 2008 or later with the SQL Server 2005 Native Client ODBC driver.

The date-time sub-profile implementation provides support for mapping QDateTime to the DATETIME and SMALLDATETIME types, however, this mapping has to be explicitly requested using the db type pragma (Section 14.4.3, "type"), as shown in the following example:

#pragma db object
class person
{
  ...
  #pragma db type("DATETIME")
  QDateTime updated_;
};