Simple promise-based data access layer for PostgreSQL written on top of pg-io.
pg-dao is designed for scenarios when connection to the database is needed for a series of short and relatively simple requests. If you need a connection to execute long running queries (or queries that return large amounts of data) or require complex transaction logic, pg-dao is probably not for you.
pg-dao adheres to the following principles:
- Single transaction - only one transaction is allowed per connection session. A transaction can be started at any point during the session, but can be committed (or rolled back) only at the end of the session
- Low error tolerance - any error in query execution will terminate the session and release the connection back to the pool
The above would work well for many web-server scenarios when connection is needed to process a single user request. If an error is encountered during the request, all changes are rolled back, an error is returned to the user, and the connection is released to handle next user request.
In addition to the functionality of pg-io, pg-dao provides a flexible managed model mechanism to make syncing data with the database simpler.
pg-dao is written in TypeScript and compiled down to JavaScript (ES6). Aa such, it can be used in any JavaScript application which complies with runtime requirements of pg-dao. The most recent version of pg-dao requires Node.js 6.0 or later.
Since pg-dao is ES6-centric, many examples below utilize ES6 syntax. Some examples also use TypeScript syntax to communicate expected types of variables.
$ npm install --save pg-dao
import { Database } from 'pg-dao';
// create a database object
const db = new Database({ /* database options */ });
// connect to the database
db.connect().then((dao) => {
// create a query object
var query = {
text: 'SELECT * FROM users WHERE status = {{status}};',
params: {
status: 'active'
},
mask: 'list'
};
// start a transaction
return dao.startTransaction()
// execute the query
.then(() => dao.execute(query))
.then((results) => {
// do some work with results
// maybe run some other queries to update the database
})
// commit transaction and release the connection back to the pool
.then(() => dao.close('commit'));
});
import { Database, Dao, AbstractModel, PgIdGenerator, dbModel, dbField } from 'pg-dao';
// Define a simple model backed by the 'users' table in the database
@dbModel('users', new PgIdGenerator('users_id_seq'))
export class User extends AbstractModel {
// username field is expected to be a string
@dbField(String)
username: string;
// password field will be encrypted
@dbField(String, { secret: 'secret' })
password: string;
}
// create a database object
const db = new Database({ /* database options */ });
// define a function to update a model
async function updatePassowrd(userId: string, newPassword: string) {
let dao: Dao;
try {
// connect to the database and start a transaction
dao = await db.connect({ startTransaction: true });
// fetch a user model from the database, and lock it for edit
const user = await dao.fetchOne(User, { id: userId }, true);
// update the model
user.password = newPassword;
// commit transaction and release the connection back to the pool
await dao.close('commit');
}
catch(error) {
// log error somewhere
// then make sure connection is closed
if (dao && dao.isActive) {
await dao.close(dao.inTransaction ? 'rollback' : undefined);
}
// maybe re-throw the error
}
}
Complete public API definitions can be found in pg-dao.d.ts file. The below sections will explain in more detail how the API can be used for specific tasks:
- Obtaining Database Connection
- Managing Transactions
- Querying the Database
- Working with Models
- Errors
- Refactored most interfaces
- Query
mask
values have been redefined:object
value has been replaced withsingle
- Model timestamps (
createdOn
andupdatedOn
) have been changed from dates to numbers - Model constructor has been updated: the meaning of the second parameter has been changed
pg-dao exposes a Database
class which can be created like so:
const db = new Database(options, logger?);
where options
should have the following form:
{
name? : string; // defaults to 'database', used for logging
pool?: { // optional connection pool settings
maxSize? : number; // defaults to 20
idleTimeout? : number; // defaults to 30000 milliseconds
reapInterval? : number; // defaults to 1000 milliseconds
};
connection: { // required connection settings
host : string;
port? : number; // optional, default 5432
ssl? : boolean; // optional, default false
user : string;
password : string;
database : string;
};
session?: { // optional default session options
startTransaction? : boolean; // defaults to false
logQueryText? : boolean; // defaults to false
};
}
and, if provided, logger
must comply with the following interface:
interface Logger {
debug(message: string, source?: string);
info(message: string, source?: string);
warn(message: string, source?: string);
trace(source: string, command: string, time: number, success?: boolean);
}
Creation of the database object does not establish a database connection but rather allocates a pool to hold connections to the database specified by the options object.
Once a Database object is created, it can be used to acquire connection sessions from the pool like so:
database.connect(options?) : Promise<Dao>;
The method returns a promise for a Dao
object which represents a connection session. The optional options
object has the following form:
{
startTransaction? : boolean; // defaults to false
validateImmutability? : boolean; // default true
validateHandlerOutput? : boolean; // default true
manageUpdatedOn? : boolean; // default true
logQueryText? : boolean; // defaults to false
}
The startTransaction
option specifies whether a transaction should be started on the connection (more on this below). The meaning of other options will be explained in the following sections.
Additionally, Database object exposes a method for checking connection pool state:
database.getPoolState() : PoolState;
Where PoolState
has the following form:
{
size : number; // current size of the connection pool
available : number; // number of available connections in the pool
}
Database connections must always be released back to the pool after they are no longer needed by calling dao.close()
method (more on this below). If you do not release connections, connection pool will be exhausted and bad things will happen.
pg-dao supports a simple transactions mechanism. Queries executed when DAO is in transaction mode will all be a part of a single transaction. Only one transaction is allowed per connection session. A transaction can be started at any point during the connection session, and must be committed or rolled back when the session is closed.
Starting a transaction can be done via the following method:
dao.startTransaction(lazy?: boolean) : Promise<void>;
If an optional lazy
parameter is set to true (the default), the transaction will be started upon the first call to dao.execute()
method. If lazy
is set to false, the transaction will be started immediately.
It is also possible to start a transaction at the time of DAO creation by setting startTransaction
option to true for the database.connect()
method:
import { Database } from 'pg-io';
const db = new Database({ /* database settings */ });
db.connect({ stratTransaction: true }).then((dao) => {
// DAO is now in transaction and all queries executed through
// this DAO session will be executed in a single transaction
return dao.release('commit');
});
In the above example, the transaction is actually not started immediately but is delayed until the first call to dao.execute()
method (this is basically equivalent to starting a transaction in lazy
mode).
Do not start transactions manually by executing BEGIN
commands. Doing so will confuse the DAO object and bad things may happen.
DAO exits transaction mode when the transaction is either committed or rolled back. This can only be done at the end of the connection session using the following method:
dao.close(action?: 'commit' | 'rollback') : Promise<void>;
where action
can be one of the following values:
- 'commit' - if there is an active transaction it will be committed
- 'rollback' - if there is an active transaction it will be rolled back
- undefined - if no transactions were started on the connection,
dao.close()
method can be called withoutaction
parameter. However, if a transaction is in progress, and action parameter is omitted, an error will be thrown and the active transaction will be rolled back before the connection is released back to the pool
In the example below, query1 and query2 are executed in the context of the same transaction, then transaction is committed and DAO is released back to the pool.
dao.startTransaction()
.then(() => {
const query1 = { ... };
return dao.execute(query1);
})
.then((query1Result) => {
// do something with the results of the first query
const query2 = { ... };
return dao.execute(query);
})
.then((query2Result) => {
// do something with the results of the second query
})
.then(() => dao.close('commit'));
Once the connection is released back to the pool, DAO object will become inactive and trying to execute queries on it will throw errors.
Always call the session.close()
method after session object is no longer needed. This will release the connection for use by other requests. If you do not release the connection, the connection pool will become exhausted and bad things will happen.
To check whether DAO is active, the following property can be used:
dao.isActive : boolean;
DAO session is considered to be active from the point it is created, and until the point it is released (or until query execution error happens).
To check whether DAO is in transaction, the following property can be used:
dao.inTransaction : boolean;
DAO is considered to be in transaction from the point startTransaction()
method is called, and until it is closed (via dao.close()
method).
Once reference to DAO object is obtained it can be used to execute queries against the database using dao.execute()
method.
// executes a single query - and return a promise for the result
dao.execute(query) : Promise<any>;
// execute multiple queries and return a promise for a map of results
dao.execute([query1, query2]) : Promise<Map>;
A query object passed to the execute method should have the following form:
{
text : string;
mask? : 'list' | 'single';
mode? : 'object' | 'array';
name? : string;
params? : any;
handler?: ResultHandler;
}
The only required property for a query is text
, however, the behavior of the execute()
method is directly controlled by other query properties. The meaning of the properties is as follows:
Property | Type | Description |
---|---|---|
text | string | SQL code to be executed against the database |
mask | enum? | Optional result mask; can be one of the following values: [list , single ]. If mask is not provided, no results will be returned to the caller (even for SELECT statements).When mask=list , an array of rows retrieved from the database will be returned to the caller (or [] if no rows were returned).When mask=single , first row retrieved from the database will be returned to the caller (or undefined if no rows were returned). |
mode | enum? | Optional row mode; can be one of the following values: [object , array ]; default is object .When mode=object , each row will be returned as an object with property keys being field names.When mode=array each row will be returned as an array of values (without the field names). |
name | string? | Optional query name; used for logging. Also, when execute() is called with an array of queries, the returned map of results will be indexed by query name. For queries which don't have a name, the results will be held under the undefined key. If several executed queries have the same name, an array of results will be stored under the key for that name |
params | object? | Optional parameters to apply to to the query (see parameterized queries) |
handler | ResultHandler? | Optional result handler to apply custom parsing logic (see result parsing ) |
A few examples of executing different queries:
var query1 = {
text: `UPDATE users SET username = 'User1' WHERE id = 1;`
};
dao.execute(query1).then((result) => {
// query is executed, and the result object is undefined
});
var query2 = {
text: 'SELECT * FROM users;',
mask: 'list'
};
dao.execute(query2).then((result) => {
// result is an array of user objects
var user1 = result[0];
});
var query3 = {
text: 'SELECT * FROM users WHERE id = 1;',
mask: 'single'
};
dao.execute(query3).then((result) => {
// result is a single user object
var user1 = result;
});
dao.execute([query1, query2, query3]).then((result) => {
// result is a map layed out as follows:
// result.get(undefined)[0] contains results from query2
var user1 = result.get(undefined)[0][0];
// result.get(undefined)[1] contains results from query3
var user2 = result.get(undefined)[1];
// results from query1 are not in the map
});
var query4 = {
text: 'SELECT * FROM users;',
mask: 'list',
name: 'q1'
};
var query5 = {
text: 'SELECT * FROM users WHERE id = 1;',
mask: 'single',
name: 'q2'
};
dao.execute([query4, query5]).then((result) => {
// result is a map laid out as follows:
// result.get(query4.name) contains results from query4
var user1 = result.get(query4.name)[0];
// result.get(query5.name) contains results from query5
var user2 = result.get(query5.name);
});
Queries can be parametrized using named parameters. Parameters must be enclosed in {{}}
brackets and params
object should be provided with parameter values.
var query = {
text: 'UPDATE users SET username = {{username}} WHERE id = {{id}};',
params: {
username: 'joe',
id: 1
}
};
dao.execute(query).then(() => {
// the query is executed as
// UPDATE users SET username = 'joe' WHERE id = 1;
});
Safe parameters (e.g. booleans, numbers, safe strings) are inlined into the query text before the query is sent to the database. If one of the parameters is an unsafe string, the query is executed as a parametrized query on the database to avoid possibility of SQL-injection. In general, properties in the params
object are treated as follows:
- boolean - always inlined
- number - always inlined
- Date - converted to ISO string and always inlined
- string - if the string is safe, it is inlined, otherwise the query is executed as a parametrized query
- object - object parameters are treated as follows:
valueOf()
method is called on the object and if it returns a number, a boolean, a safe string, or a date, the value is inlined; if the returned value is an unsafe string, the query is executed as parametrized query- if
valueOf()
method returns an object, the parameter is converted to string usingJSON.stringify()
and if the resulting string is safe, inlined; otherwise the query is executed as parametrized query
- arrays - arrays are parametrized same as objects
- null or undefined - always inlined as 'null'
- functions - functions are parametrized as follwos:
valueOf()
mehtod is called on the function, and if it returns a primitive value, the value is inlined- otherwise QueryError will be thrown
It is also possible to parametrize arrays of primitives in a special way to make them useful for IN
clauses. This can be done by using [[]]
brackets. In this case, the parameterization logic is as follows:
- arrays of numbers are always inlined using commas as a separator
- arrays of strings are either inlined (if the strings are safe) or sent to the database as parametrized queries (if strings are unsafe)
- all other array types (and arrays of mixed numbers and strings) are not supported and will throw QueryError
Examples of array parametrization:
var query1 = {
text: 'SELECT * FROM users WHERE id IN ([[ids]]);',
params: {
ids: [1, 2]
}
};
// query1 will be executed as:
// SELECT * FROM users WHERE id IN (1,2);
// if {{}} was used instead, the query would have been:
// SELECT * FROM users WHERE id IN ('[1,2]');
var query2 = {
text: 'SELECT * FROM users WHERE type IN ([[types]]);',
params: {
types: ['personal', 'business']
}
};
// query2 will be executed as:
// SELECT * FROM users WHERE type IN ('personal','business');
// if {{}} was used instead, the query would have been:
// SELECT * FROM users WHERE type IN ('["personal","business"]');
var query3 = {
text: 'SELECT * FROM users WHERE name IN ([[names]]);',
params: {
names: [`Test`, `T'est`, `Test2` ]
}
};
// query3 will be executed as:
// SELECT * FROM users WHERE firstName IN ('Test',$1,'Test2');
It is possible to parse query results using custom logic by providing a ResultHandler
object for a query. The handler object must have a single parse()
method which takes a row as input and produces custom output. For example:
var query = {
text: 'SELECT * FORM users;',
handler: {
parse: (row) => row.id
}
};
dao.execute(query).then((result) => {
// the result will contain an array of user IDs
});
pg-dao provides a very flexible mechanism for defining managed models. Once models are defined, pg-dao takes care of synchronizing models with the database whenever changes are made.
Any object can be a model as long as the object has the following properties:
{
id : string, // unique identifier for the model
createdOn: number, // Unix timestamp of when the model was created
updatedOn: number // Unix timestamp of when the model was last updated
[handler]: ModelHandler // handler for the model (described below)
}
Please note that createdOn
and updatedOn
timestamps are in milliseconds, and therefore, must be stored in the database as 64-bit integers.
ModelHandler is an object which provides services needed by DAO to work with the model. Model handler must have the following form:
{
parse(row: any): Model;
build(id: string, attributes: any): Model;
clone(model: Model): Model;
compare(original: Model, current: Model): string[];
areEqual(model1: Model, model2: Model): boolean;
infuse(target: Model, source: Model);
getSyncQueries(original: Model, current: Model, changes?: string[]): Query[];
getFetchOneQuery(selector: any, forUpdate: boolean, name?: string): Query;
getFetchAllQuery(selector: any, forUpdate: boolean, name?: string): Query;
getIdGenerator(): idGenerator;
}
The meaning of the above methods is as follows:
- parse(row) - should take a single database row as input and return a model object
- build(id, attributes) - should take an ID and a set of attributes, and build a fully functional model
- clone(model) - should take a model as an input and produce a new object identical to the original model
- compare(original, current) - if two models are different, should return a list of dirty field names
- areEqual(model1, model2) - should return true if both models are identical
- infuse(target, source) - should change the properties of the
target
model to make it identical to thesource
model - getSyncQueries(original, current, changes?) - given the original and the current state of the model (as well as a list of dirty fields), should produce an array of queries which can be executed to synchronize the model with the database
- getFetchOneQuery(selector, forUpdate, name?) - given the selector, returns a query which can be executed to retrieve a single model
- getFetchAllQuery(selector, forUpdate, name?) - given the selector, returns a query which can be executed to retrieve a list of models
- getIdGenerator() - should return an IdGenerator class which can be used to generate unique IDs for the model
The above mechanism is extremely flexible and allows the user to define models of nearly arbitrary complexity (e.g. it is possible to implement models which span multiple tables and contain complex object hierarchies). However, it would be extremely tedious to manually define handlers for all models from scratch. To make this task simpler, pg-dao provides a base class called AbstractModel
which implements most of the boilerplate functionality for you.
If you are using TypeScript, the task of defining models can be made even simplear by using decorators.
Below is an example of a very simple User
model implemented by extending AbstractModel
base class. For this model, the data is stored in the users
table which has id
, username
, password
, created_on
, and updated_on
fields. There is also a sequence called users_id_seq
defined in the database.
import { AbstractModel, PgIdGenerator, dbModel, dbField } from 'pg-dao';
@dbModel('users', new PgIdGenerator('users_id_seq'))
export class User extends AbstractModel {
@dbField(String)
username: string;
@dbField(String, { secret: 'my secret' })
password: string;
}
The models can be further customized almost at will. For example, providing a specialized fetch query can be done as follows:
import { AbstractModel, PgIdGenerator, dbModel, dbField } from 'pg-dao';
@dbModel('users', new PgIdGenerator('users_id_seq'))
export class User extends AbstractModel {
@dbField(String)
username: string;
@dbField(String, { secret: 'my secret' })
password: string;
static getFetchAllQuery(selector: any, forUpdate = false, name?: string) {
if ('conversationId' in selector) {
// fetch all users participating in a specific conversation
return {
text: `SELECT ... FROM users WHERE id IN
(SELECT user_id FROM conversations WHERE id = ${selector.conversationId});`,
mask: 'list',
handler: this,
mutable: forUpdate,
name: name
};
}
else {
return super.getFetchAllQuery(selector, forUpdate, name);
}
}
}
Using AbstractModel
(as opposed to defining model handler from scratch) does impose a few limitations:
- The underlying table must have
id
,created_on
, andupdated_on
fields.id
must be a primary key and can be either varchar or bigint. The other two fields must be bigints. These fields (in camelCase) will be added to all abstract models automatically. - All model properties must be in camelCase while all database fields must be in snake_case.
AbstractModel
assumes this conventions and queries generated automatically will have syntax errors if this convention is not adhered to - If you decide to override model constructor, the constructor signature must be
constructor(seed: any, deepCopy?: boolean)
, and the first call inside the constructor must besuper(seed, deepCopy)
pg-dao provides two decorators which can be used to define a model: @dbModel
and @dbField
.
As the name implies, @dbModel
defines parameters for the entire model. The two parameters currently required for this decorator are:
- tableName - the name of the table backing the model
- idGenerator - the ID Generator class which can be used by the model to generate unique IDs
@dbField
decorator must be attached to each model field. Any property not decorated with @dbField
will not be synced with the database. The following paramters can be specified for the @dbField
decorator:
- fieldType - specifies the type of the field. This parameter is required. Currently allowed field types are:
Number
,Boolean
,String
,Timestamp
,Date
,Object
, andArray
- fieldOptions - optional paramter to specify additional options for the field. Currently, the following options are supported:
- readonly - a boolean flag which specifies if the field is read-only. Read-only fields are assumed to never change, and will not be synced with the database
- secret - if specified, the filed values are assumed to be encryped in the stabases with the specified
secret
. pg-dao will decrypt the values upon retrieval from the database, and will encrypt them back upon syncing with the database. Currently, encryption is supported only forString
,Object
, andArray
fields. The corresponding field in the database must be able to store string values - handler - an optional custom handler for the filed to be used to compare and clone field values. Providing custom handlers is only allowed for
Object
andArray
fields. Providing custom handlers can improve performace of comparison and cloning operations - especially for complex objects. When provided, a custom handler must haveclone(model) => Model
andareEqual(model1, model2) => boolean
methods.
As described above, pg-dao models require ID Generators to be provided. Such generators can be anything as long as they have the following form:
{
getNextId(dao?: Dao): Promise<string>;
}
The getnextId()
method will receive a reference to a DAO object whenever this method is called, but ID generator does not need to rely on this object to generate unique IDs. This approach makes it possible to generate unique IDs in a variety of ways (e.g. in-memory using timestamps, distributed ID generation using redis server etc.) making pg-dao models even more flexible.
Out of the box, pg-dao provides a PgIdGenerator
that takes a name of a database sequence and whenever a new ID is requested, makes a call to the database to get the next value from that sequence.
Defining models using TypeScript decorators is super convinient, but if you are not using TypeScript, you can still use pg-dao. Below is an example of a simple User model equivalent to the one defined above:
import { AbstractModel, PgIdGenerator } from 'pg-dao';
class User extends AbstractModel {
// nothing to do here, unless you need to override the constructor
// or other handler methods
}
// no need to set id, createdOn, updatedOn fields - they will be set automatically
User.setSchema('users', new PgIdGenerator('users_id_seq'), {
username: { type: String },
password: { type: String, secret: 'my secret' }
});
The above will create a fully functional User model.
Retrieving models from the database can be done via the regular dao.execute()
method or via specialized fetchOne()
and fetchAll()
methods.
The main difference from executing regular queries is that model queries should have ModelHandler
specified for the handler
property and can have an additional mutable
property to specify whether retrieved models can be updated.
If you are using AbstractModel
as a base for your models, the mode
of the model queries must be set to array
as AbastractModel
expects all models read from the database to be in array form.
For example, given the User model defined above, a query to retrieve a single user by ID would look like this:
var userId = '1';
var qFetchUserById = {
text: `SELECT id, username, created_on AS "createdOn", updated_on AS "updatedOn"
FROM users WHERE id = ${userId};`,
mask: 'single',
mode: 'array',
handler: User,
mutable: false
};
dao.execute(qFetchUserById).then((user) => {
// user is now User model retrieved from the database
});
A query to retrieve multiple users by ID could look like this:
var userIdList = ['1', '2', '3'];
var qFetchUsersByIdList = {
text: `SELECT id, username, created_on AS "createdOn", updated_on AS "updatedOn"
FROM users WHERE id IN (${userIdList.join(',')});`,
mask: 'list',
mode: 'array',
handler: User,
mutable: false
};
dao.execute(qFetchUsersByIdList).then((users) => {
// users is now an array of 3 User model retrieved from the database
});
Retrieving the same model multiple times does not create a new model object - but rather updates an existing model object with fresh data from the database (this effectively reloads the model in memory):
var userId = '1';
var qFetchUserById = {
text: `SELECT id, username, created_on AS "createdOn", updated_on AS "updatedOn"
FROM users WHERE id = ${userId};`,
mask: 'single',
mode: 'array',
handler: User,
mutable: false
};
dao.execute(qFetchUserById).then((user1) => {
return dao.execute(qFetchUserById).then((user2) => {
user1 === user2; // true
});
});
The mutable
property indicates whether the models retrieved by the query can be updated during the DAO session. Setting mutable
to true instructs DAO to monitor the model and detect if any changes take place. This has performance implications, so setting mutable
to true should be done only if you are planning to modify the models during the session.
In some scenarios it might make sense to mark models as mutable only if SELECT ... FOR UPDATE
statement was used to retrieve it from the database. For example, the following might be a query to select a user model for update:
var userId = '1';
var qFetchUserById = {
text: `SELECT id, username, created_on AS "createdOn", updated_on AS "updatedOn"
FROM users WHERE id = ${userId} FOR UPDATE;`,
mask: 'single',
mode: 'array',
handler: User,
mutable: true
};
dao.startTransaction.then(() => {
return dao.execute(qFetchUserById).then((user) => {
// user row with ID = '1' is now locked in the database - no other client
// can modify the row until this transaction is committed.
// so, we can update the user model safely knowing that we have the most
// recent version of the user data
});
}).then(() => dao.close('commit'));
Checking whether the model was retrieved as mutable can be done as follows:
dao.isMutable(model) : boolean;
If you only need to retrieve one type of model at a time, it might be easier to use fetchOne()
or fetchAll()
methods. The signatures of the methods are as follows:
dao.fetchOne(handler, selector, forEdit?): Model;
dao.fetchAll(handler, selector, forEdit?): Model[];
The meaning of the parameters is as follows:
handler
- model handler for which to retrieve modelsselector
- an object describing parameters based on which models should be selectedforEdit
- an optional parameter (default false) indicating whether the retrieved models are mutable
For the User model defiend above, the fetch methods can be used as follows:
dao.fetchOne(User, { id: '1' }).then((user) => {
// fetches User model with ID = '1' from the database
// the fetched model is immutable
});
dao.fetchOne(User, { id: '2' }, true).then((user) => {
// fetches User model with ID = '2' from the database
// the fetched model is mutable
});
dao.fetchAll(User, { id: ['2', '3'] }).then((users) => {
// fetches User models with IDs '2' and '3' from the database
// the fetched models are immutable
});
Internally, DAO fetch methods call model handler's get fetch query methods (getFetchOneQuery()
and getFetchAllQuery()
) and execute returned queries using dao.execute()
method. So, any custom fetch queries defined for the model will automatically work in these methods too.
When working with models derived from AbstractModel
, keep in mind that FOR UPDATE
statement will be added to all automatically generated fetch queries when forUpdate
parameter is set to true for getFetchOneQuery()
and getFetchAllQuery()
methods. Also, the mode
for all fetch queries used with AbstractModel
must be array
as AbstractModel
expects rows read from the database to be in array form.
For models which were retrieved from the database as mutable, DAO observes the changes and then writes changes out to the database on dao.sync()
call or on dao.close('commit')
call.
Updating existing models is done simply by modifying model properties. No additional works is needed:
dao.startTransaction.then(() => {
// retrieve user model from the database
return dao.fetchOne(User, { id: '1' }, true).then((user) => {
// update the model
user.username = 'test';
dao.isModified(user); // true
});
})
// sync changes with the database and release connection
.then(() => dao.close('commit'));
Deleting existing models can be done as follows:
dao.startTransaction.then(() => {
// retrieve user model from the database
return dao.fetchOne(User, { id: '1' }, true).then((user) => {
dao.destroy(user);
dao.isDestroyed(user); // true
});
})
// sync changes with the database and release connection
.then(() => dao.close('commit'));
Inserting new models into the database can be done by using a combination of dao.create()
and dao.insert()
method:
dao.startTransaction.then(() => {
// create a new user model
return dao.create(User, { username: 'Test' }).then((user) => {
dao.hasModel(user); // false - the model is created but not yet inserted into DAO
// insert the model into DAO
dao.insert(user);
dao.isNew(user); // true - the model is inserted but not yet saved to the database
});
})
// sync changes with the database and release connection
.then(() => dao.close('commit'));
It is possible to revert the changes made to a model by using the following method:
dao.clean(model);
The changes will be reverted to the point when last call to dao.sync()
was made. If this method is called on a new model, the model will be removed form DAO.
All pending model changes must be either committed or rolled-back upon DAO release. This can be done as follows:
dao.close('commit')
- this will write out all pending model changes to the database, commit any active transactions, and release connection back to the pooldao.close('rollback')
- this will discard any pending model changes, rollback any active transaction, and release connection back to the pool
If dao.close()
is called without any parameters and there are pending model changes, the changes will be discarded, any active transaction will be rolled back, and an error will be thrown.
It is also possible to sync model changes with the database without releasing DAO connection by using dao.sync()
method.
pg-dao does not actively enforce model immutability. This means that models retrieved as immutable can still be modified by the user. As pg-dao only observes mutable models, any changes to immutable models will be ignored. However, it is possible to force pg-dao to validate model immutability on syncing changes. This can be done via setting validateImmutability
property for the connection to true. In such a case, if any changes to immutable models are detected during model synchronization, an error will be thrown. There are performance implications to setting validateImmutability
property to true - so, it might be a good idea to use it in development environments only.
pg-dao will also automatically set updatedOn
property of any models that have been updated to the current date upon model synchronization. This behavior can be overridden by setting manageUpdatedOn
connection option to false.
It is possible to check the state of a specific model using the following methods:
dao.isModified(model) : boolean // true if the model has pending changes
dao.isNew(model) : boolean // true if the new model has not yet been saved to the database
dao.isDestroyed(model): boolean // true if the deleted model has not yet been removed from the databsae
This methods will throw an error if the model has not been registered with DAO. To check whether a model is registered with DAO the following method can be used:
dao.hasModel(model) : boolean
To check whether DAO has any pending changes (updates, inserts, or deletes), the following method can be used:
dao.isSynchronized() : boolean // returns false if DAO has any pending changes
pg-io provides several customized errors which extend the built-in Error object (via base PgError class). These errors are:
- ConnectionError, thrown when:
- establishing a database connection fails
- an attempt to use an already released connection is made
- an attempt to release an already released connection is made
- TransactionError, thrown when:
- an attempt is made to start a transaction on a connection which is already in transaction
- a connection is released without committing or rolling back an active transaction
- QueryError, thrown when:
- executing of a query fails
- ParseError, thrown when:
- parsing of query results fails
- StoreError, thrown when:
- An attempt to change the state of in-memory models is made illegally (e.g. inserting the same model twice)
- SyncError, thrown when:
- a change to an immutable model is detected
- a connection is released without committing or rolling back pending model changes
- ModelError, throw when:
- an inconsistent model is detected (e.g. model without
[handler]
property)
- an inconsistent model is detected (e.g. model without
If an error is thrown during query execution, query result parsing, or model syncing, the DAO session will be immediately closed. If DAO is in transaction, then the transaction will be rolled back. Basically, any error generated within dao.execute()
and dao.sync()
method will render the session useless and no further communication with the database through this DOA object will be possible. The connection itself will be released to the pool so that it can be used by other clients.
Copyright (c) 2016 Hercules Inc.
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.