SQL specification
The Tableland SQL language specification offers a unique set of SQL that is a subset of the SQLite language.
Author(s): @carsonfarmer, @brunocalza, @jsign
Synopsis
Tableland understands a small subset of the standard SQL language. It does omit many features while at the same time adding a few features of its own. This document attempts to describe precisely what parts of the SQL language Tableland does and does not support. A list of supported data types is also provided. The SQL language supported by Tableland is a subset of the SQLite SQL language specification (and as such, we borrow heavily from their documentation with attribution), with additional constraints specific to Tableland operations.
This general SQL specification is broken down into two core sub-documents (which are linked below). This specification is a living document, and as such, may be updated over time. Proposals for the addition of SQL language features and data types may be submitted by the Tableland community over time. These proposals will be evaluated for technical feasibility, utility to the community, and longer-term sustainability.
Table of Contents
Statement Types
The core Tableland SQL parser accepts an SQL statement list which is a
semicolon-separated list of statements. Each SQL statement in the
statement list is an instance of one of the following specific statement
types. All other standard SQL statement types are unavailable (at the
moment). Each statement type is associated with a well-known SQL command
(see following sections). In general, the entire Tableland SQL API can
be summarized in eight command/statement types: CREATE TABLE
, ALTER TABLE
,
INSERT
, UPDATE
, DELETE
, SELECT
, GRANT
, REVOKE
.
⚠️ The statement and data types provided here are part of the official minimal Tableland SQL specification. Additional functionality may be available in practice. However, it is not recommended that developers rely on SQL features outside of this minimal specification in the long-term.
CREATE TABLE
The CREATE TABLE
command is used to create a new table on Tableland. A
CREATE TABLE
command specifies the following attributes of the new
table:
- The name of the new table (
table_name
). - The name of each column in the table (
column_name
). - The declared type of each column in the table
(
data_type
). - A default value or expression for each column in the table.
- Optionally, a
PRIMARY KEY
for the table. Both single column and composite (multiple column) primary keys are supported. - A set of SQL constraints for the
table. Tableland supports
UNIQUE
,NOT NULL
,CHECK
andPRIMARY KEY
constraints (see previous bullet). - Optionally, a generated column constraint.
Structure
CREATE TABLE *table_name* ( [
{ *column_name* *data_type* [ *column_constraint* [, ... ] ]
| table_constraint }
[, ...]
] );
where column_constraint
has structure
[ CONSTRAINT constraint_name ]
{ NOT NULL |
CHECK ( expression ) |
DEFAULT default_expr |
UNIQUE index_parameters |
PRIMARY KEY index_parameters |
}
and table_constraint
has structure
[ CONSTRAINT constraint_name ]
{ CHECK ( expression ) |
UNIQUE ( column_name [, ... ] ) |
PRIMARY KEY ( column_name [, ... ] )
Details
Table Identifiers/Names
Every CREATE TABLE
statement must specify a fully-qualified table
name (name) as the name of the new table. The fully-qualified table
name has the following structure:
Where is optional. When a prefix is included, it must
start with a letter and be followed by any combination of (zero or more)
letters, numbers, and/or underscores. A prefix string may be up to 32
bytes in length. In practice, long names with spaces must be slug-ified
with underscores. For example, "my amazing table"
would become
"my_amazing_table"
. The last two components of the table name, must
be the chain id and the table id, which are numeric values separated by
an underscore. For example, a valid table name without a prefix looks
like _42_0
(or 42_1
), whereas a valid table name with a prefix
might look like dogs_42_0
.
⚠️ It is not up to the caller to determine what table id to use in a
CREATE TABLE
statement. The table id is a monotonically-increasing numeric value which is provided by the smart contract that is processing the create statements. See the On-Chain API Specification for details on the smart contract calls used to generateCREATE TABLE
statements in practice.
Table names are globally unique. The combination of chain id and monotonically increasing table id ensures this is the case in practice. As such, the addition of a user-defined prefix string is an aesthetic feature that most developers will find useful (but is not required). The maximum (slug-ified) prefix length is 32 bytes.
ℹ️ Tableland also supports quoted identifiers (for table names, column names, etc). This allows callers to use SQL Keywords (see next section) as part of identifiers, etc. There are some limitations to this, and it does not circumvent any other naming constraints.
Reserved Keywords
The SQL standard specifies a large number of keywords which may not be used as the names of tables, indices, columns, databases, or any other named object. The list of keywords is often so long that few people can remember them all. For most SQL code, your safest bet is to never use any English language word as the name of a user-defined object.
If you want to use a keyword as a name, you need to quote it. There are four ways of quoting keywords in SQLite:
'keyword'
— A keyword in single quotes is a string literal."keyword"
— A keyword in double-quotes is an identifier.[keyword]
— A keyword enclosed in square brackets is an identifier. This is not standard SQL, it is included in Tableland for compatibility.keyword
— A keyword enclosed in grave accents (ASCII code 96) is an identifier. This is not standard SQL, it is included in Tableland for compatibility.
The list below shows all possible reserved keywords used by Tableland (or SQLite). Any identifier that is not on the following element list is not considered a keyword to the SQL parser in Tableland:
TRUE
, FALSE
, AND
, OR
, NOT
, NULL
, NONE
, INTEGER
,
NUMERIC
, REAL
, TEXT
, CAST
, AS
, IS
, ISNULL
, NOTNULL
,
COLLATE
, LIKE
, IN
, REGEXP
, GLOB
, MATCH
, ESCAPE
, BETWEEN
,
CASE
, WHEN
, THEN
, ELSE
, END
, SELECT
, FROM
, WHERE
,
GROUP
, BY
, HAVING
, LIMIT
, OFFSET
, ORDER
, ASC
, DESC
,
NULLS
, FIRST
, LAST
, DISTINCT
, ALL
, JOIN
, ON
, USING
,
EXISTS
, FILTER
, BLOB
, INT
, ANY
, CREATE
, TABLE
, PRIMARY
,
KEY
, UNIQUE
, CHECK
, DEFAULT
, GENERATED
, ALWAYS
, STORED
,
VIRTUAL
, CONSTRAINT
, INSERT
, VALUES
, INTO
, DELETE
, UPDATE
,
SET
, GRANT
, TO
, REVOKE
, CONFLICT
, DO
, NOTHING
ℹ️ You can also find the most up to date list of keywords used by Tableland in the reference parser implementation. See Implementation
⚠️ Table names that begin with
sqlite
,system
orregistry
are also reserved for internal use. It is an error to attempt to create a table with a name that starts with these reserved names.
Column Definitions and Constraints
Every CREATE TABLE
statement includes one or more column definitions,
optionally followed by a list of table constraints. Each column
definition consists of the name of the column, followed by the declared
type of the column (see Data Types), then one or more
optional column constraints. Included in the definition of column
constraints for the purposes of the previous statement is the DEFAULT
clause, even though this is not really a constraint
in the sense that it does not restrict the data that the table may
contain. The other constraints, NOT NULL
, CHECK
, UNIQUE
, and
PRIMARY KEY
constraints, impose restrictions on the table data.
⚠️ The number of columns in a table is limited by the
MaxColumns
validator configuration parameters (defaults to 24). A single character fields in a table cannot store more thanMaxTextLength
bytes of data (defaults to 1024). The number of rows in a table is limited by theMaxRowCount
validator configuration parameter (defaults to 100,000). This values are all configurable at the network-level, and may change in the future.
⚠️ In practice, a
CREATE TABLE
statement must be sent as a single top-level statement (i.e., it must be provided in a statement list of length one).
🚧 Feature At Risk:
FOREIGN KEY
constraints of the formFOREIGN KEY(column_name) REFERENCES table_id(column_name)
are currently not supported across Tableland tables. Instead, dynamicJOIN
s can be used to reference columns in remote tables. However, inclusion ofFOREIGN KEY
constraints are being considered for inclusion in the Tableland SQL specification with some specific limitations. In particular, key constraint actions would be restricted toSET NULL
orSET DEFAULT
(see the section called SQLite foreign key constraint actions at the link below). See SQLite Foreign Key
Column Defaults
The DEFAULT
clause specifies a default value to use for the column if
no value is explicitly provided by the user when doing an INSERT
. If
there is no explicit DEFAULT
clause attached to a column definition,
then the default value of the column is NULL
. An explicit DEFAULT
clause may specify that the default value is NULL
, a string constant,
a blob constant, a signed-number, or any constant expression enclosed in
parentheses. For the purposes of the DEFAULT
clause, an expression is
considered constant if it contains no sub-queries, column, or table
references, or string literals enclosed in double-quotes instead of
single-quotes.
Each time a row is inserted into the table by an INSERT
statement that
does not provide explicit values for all table columns the values stored
in the new row are determined by their default values, as follows:
- If the default value of the column is a constant
NULL
, text, blob or signed-number value, then that value is used directly in the new row. - If the default value of a column is an expression in parentheses, then the expression is evaluated once for each row inserted and the results used in the new row.
Generated Columns
A column that includes a GENERATED ALWAYS AS
clause is a generated
column:
CREATE TABLE table_id (
...,
column_name data_type { GENERATED ALWAYS } AS (*expression*) { STORED | VIRTUAL }
);
Generated columns (also sometimes called "computed columns") are columns of a table whose values are a function of other columns in the same row. Generated columns can be read, but their values can not be directly written. The only way to change the value of a generated column is to modify the values of the other columns used to calculate the generated column.
The GENERATED ALWAYS
keywords at the beginning of the constraint and
the VIRTUAL
or STORED
keyword at the end are all optional. Only the
AS
keyword and the parenthesized expression are required. If the
trailing VIRTUAL
or STORED
keyword is omitted, then VIRTUAL
is the
default.
The value of a VIRTUAL
column is computed when read, whereas the value
of a STORED
column is computed when the row is written. STORED
columns take up space in the database file, whereas VIRTUAL
columns
use more CPU cycles when being read.
Features and Limitations
- Generated columns must also have a defined data type (just like all columns in Tableland). Tableland will attempt to transform the result of the generating expression into that data type using the same affinity rules as for ordinary columns.
- Generated columns may have
NOT NULL
,CHECK
, andUNIQUE
constraints, just like ordinary columns. - The expression of a generated column can refer to any of the other declared columns in the table, including other generated columns, as long as the expression does not directly or indirectly refer back to itself.
- Generated columns may not have a
DEFAULT
clause. The value of a generated column is always the value specified by the expression that follows theAS
keyword. - Generated columns may not be used as part of the
PRIMARY KEY
. - The expression of a generated column may only reference constant literals and columns within the same row, and may only use scalar deterministic functions. The expression may not use sub-queries, aggregate functions, etc.
- The expression of a generated column may refer to other generated columns in the same row, but no generated column can depend upon itself, either directly or indirectly.
- Every table must have at least one non-generated column.
- The data type of the generated column is determined only by the
declared data type from the column definition. The datatype of the
GENERATED ALWAYS AS
expression has no affect on the data type of the column data itself.
Primary Key
Each table in Tableland may have at most one PRIMARY KEY
. If the
keywords PRIMARY KEY
are added to a column definition, then the
primary key for the table consists of that single column. Or, if a
PRIMARY KEY
clause is specified as a separate table constraint, then
the primary key of the table consists of the list of columns specified
as part of the PRIMARY KEY
clause. The PRIMARY KEY
clause must
contain only column names. An error is raised if more than one
PRIMARY KEY
clause appears in a CREATE TABLE
statement. The
PRIMARY KEY
is optional.
If a table has a single column primary key and the declared type of that
column is INTEGER
, then the column is known as an
INTEGER PRIMARY KEY
. See below for a description of the special
properties and behaviors associated with an
INTEGER PRIMARY KEY
.
Each row in a table with a primary key must have a unique combination of
values in its primary key columns. If an INSERT
or UPDATE
statement
attempts to modify the table content so that two or more rows have
identical primary key values, that is a constraint violation. Related,
the SQL standard is that a PRIMARY KEY
should always be NOT NULL
, so
Tableland enforces this constraint.
Integer Primary Key
All rows within Tableland tables have a 64-bit signed integer key that
uniquely identifies the row within its table. This integer is usually
called the ROWID
. The ROWID
value can be accessed using one of the
special case-independent names "rowid"
, "oid"
, or "_rowid_"
in
place of a column name. As such, these values are not allowed as
identifiers for columns in a CREATE TABLE
statement.
The data for Tableland tables are stored in sorted order, by ROWID
.
This means that retrieving or sorting records by ROWID
is fast.
Searching for a record with a specific ROWID
, or for all records with
ROWID
s within a specified range is around twice as fast as a similar
search made by specifying any other PRIMARY KEY
. This ROWID
sorting
is also required for sub-queries and other SQL features on Tableland, to
ensure deterministic ordering of results.
With one exception noted below, if a table has a primary key that
consists of a single column and the declared type of that column is
INTEGER
, then the column becomes an alias for the ROWID
. Such a
column is usually referred to as an "integer primary key". A
PRIMARY KEY
column only becomes an integer primary key if the declared
type name is exactly INTEGER
. Other integer type names like INT
causes the primary key column to behave as an ordinary table column with
integer affinity and a unique index, not as an alias for the ROWID
.
In the above case of an integer primary key, there is an additional
implied AUTOINCREMENT
constraint, which forces the integer primary key
to behave as if it were specified with
INTEGER PRIMARY KEY AUTOINCREMENT
. See Autoincrement
for further details.
The exception mentioned above is that if the declaration of a column
with declared type INTEGER
includes an PRIMARY KEY DESC
clause, it
does not become an alias for the ROWID
and is not classified as an
integer primary key. In practice, this means that the following three
table declarations all cause the column "x" to be an alias for the
ROWID
(an integer primary key):
CREATE TABLE t(x INTEGER PRIMARY KEY ASC, y, z);
CREATE TABLE t(x INTEGER, y, z, PRIMARY KEY(x ASC));
CREATE TABLE t(x INTEGER, y, z, PRIMARY KEY(x DESC));
But the following declaration does not result in "x" being an alias for
the ROWID
:
CREATE TABLE t(x INTEGER PRIMARY KEY DESC, y, z);
Given this, and the implied autoincrement behavior, the following
transformation rules are enforced by the Tableland Parser to maintain
the correct ROWID
alias behavior:
Statement | Transformation |
---|---|
CREATE TABLE (a INTEGER PRIMARY KEY); | Inject AUTOINCREMENT |
CREATE TABLE (a INTEGER PRIMARY KEY DESC); | Unchanged and not an alias |
CREATE TABLE (a INTEGER, PRIMARY KEY(x ASC); | Transformed to first row version with injected AUTOINCREMENT |
CREATE TABLE (a INTEGER, PRIMARY KEY(x DESC); | Transformed to second row version and no longer an alias |
These transformations to the more "canonical" direct constraint on the
primary key are required to enforce the implied AUTOINCREMENT
behavior
on the special integer primary keys.
Rowid values may not be modified using an UPDATE
statement by
attempting to assign to one of the built-in aliases ("rowid"
, "oid"
or "_rowid_"
). However, it is possible to UPDATE
an integer
primary key value (which is an alias to ROWID
) by specifying a value
directly. Similarly, an INSERT
statement may be used to directly
provide a value to use as the ROWID
for any row inserted. For example,
the following statements are allowed, and will update/set the ROWID
value directly:
INSERT INTO a VALUES (2, 'Hello');
UPDATE a SET a = 10 WHERE b = 'Hello';
⚠️ If an
UPDATE
orINSERT
sets a givenROWID
to the largest possible value, then newINSERT
s are not allowed and any attempt to insert a new row will fail with an error. As such, use caution when directly assigning values to aROWID
alias in the form of an integer primary key.
Autoincrement
In Tableland, a column with type INTEGER PRIMARY KEY
is an alias for
the ROWID
which is always a 64-bit signed integer. It is implied that
this column will behave as INTEGER PRIMARY KEY AUTOINCREMENT
. This is
a special feature of the Tableland SQL Specification, and helps to
ensure deterministic ordering of values within a table.
ℹ️ While the
AUTOINCREMENT
keyword is implied withINTEGER PRIMARY KEY
columns, the keyword itself is not allowed in this specification. Any attempt to use theAUTOINCREMENT
keyword on any column results in an error.
On an INSERT
, the ROWID
or INTEGER PRIMARY KEY
column will be
filled automatically with a monotonically increasing integer value,
usually one more than the largest ROWID
currently in use.
In practice, this prevents the reuse of ROWID
s over the lifetime of
the table. In other words, the purpose of the implied AUTOINCREMENT
is
to prevent the reuse of ROWID
s from previously deleted rows.
The ROWID
chosen for the new row is at least one larger than the
largest ROWID
that has ever before existed in that same table. If the
table has never before contained any data, then a ROWID
of 1 is used.
If the largest possible ROWID
has previously been inserted, then new
INSERT
s are not allowed and any attempt to insert a new row will
fail with an error. Only ROWID
values from previous transactions that
were committed are considered. ROWID
values that were rolled back are
ignored and can be reused.
Rows with automatically selected ROWID
s are guaranteed to have
ROWID
s that have never been used before by the same table. And the
automatically generated ROWID
s are guaranteed to be monotonically
increasing. These are important properties for blockchain applications.
Note that "monotonically increasing" does not imply that the ROWID
always increases by exactly one. One is the usual increment. However, if
an insert fails due to (for example) a uniqueness constraint, the
ROWID
of the failed insertion attempt might not be reused on
subsequent inserts, resulting in gaps in the ROWID
sequence. Tableland
guarantees that automatically chosen ROWID
s will be increasing but not
that they will be sequential.
ALTER TABLE
The ALTER TABLE
command allows the following alterations of an
existing table: renaming a column, adding a column, and dropping a
column.
Structure
ALTER TABLE table_name *action*
where action is one of:
-- For renaming a column
RENAME [ COLUMN ] *column_name* TO *new_column_name*
-- For adding a column
ADD [ COLUMN ] *column_name* *data_type* [ *column_constraint* [, ... ] ]
-- For dropping a dolumn
DROP [ COLUMN ] *column_name*
Details
The ADD COLUMN
syntax is used to add a new column to an existing
table. The new column is always appended to the end of the list of
existing columns. The new column may take any of the forms permissible
in a CREATE TABLE
statement, with the following
restrictions:
- The column may not have a PRIMARY KEY or UNIQUE constraint.
- If a NOT NULL constraint is specified, then the column must have a default value other than NULL.
- The column may not be GENERATED ALWAYS ... STORED, though VIRTUAL columns are allowed.
The DROP COLUMN
syntax is used to remove an existing column from a
table. The DROP COLUMN
command removes the named column from the
table, and rewrites its content to purge the data associated with that
column. The DROP COLUMN
command only works if the column is not
referenced by any other parts of the schema and is not a PRIMARY KEY
and does not have a UNIQUE
constraint. Possible reasons why the
DROP COLUMN
command can fail include:
- The column is a
PRIMARY KEY
or part of one. - The column has a
UNIQUE
constraint. - The column is named in a table or column
CHECK
constraint not associated with the column being dropped. - The column is used in the expression of a generated column.
DELETE
The DELETE
command removes records from the table identified by
the table id.
Structure
DELETE FROM table_name [ WHERE condition ]
Details
If the WHERE
clause is not present, all records in
the table are deleted. If a WHERE
clause is supplied, then only those
rows for which the WHERE
clause boolean expression is true are
deleted. Rows for which the expression is false or NULL
are retained.
INSERT
The INSERT
command creates new rows in a table identified by the table
name.
Structure
INSERT INTO table_name [ ( *column_name* [, ...] ) ] VALUES (
{ expression } [, ...]
);
or
INSERT INTO table_name DEFAULT VALUES;
or, the following limited sub-query syntax
INSERT INTO table_name [ ( *column_name* [, ...] ) ] SELECT [ * | expression [, ...] ]
[ FROM from_clause [, ...] ]
[ WHERE where_clause ];
Details
An INSERT
statement creates one or more new rows in an existing table.
If the column_name
list after table_name
is omitted then the number
of values inserted into each row must be the same as the number of
columns in the table. In this case the result of evaluating the
left-most expression from each term of the VALUES
list is inserted
into the left-most column of each new row, and so forth for each
subsequent expression. If a column_name
list is specified, then the
number of values in each term of the VALUE
list must match the number
of specified columns. Each of the named columns of the new row is
populated with the results of evaluating the corresponding VALUES
expression. Table columns that do not appear in the column list are
populated with the default column value (specified as part of
the CREATE TABLE
statement), or with NULL
if no default value is
specified.
The alternative INSERT ... DEFAULT VALUES
statement inserts a single
new row into the named table. Each column of the new row is populated
with its default value, or with a NULL
if no default value is
specified as part of the column definition in the CREATE TABLE
statement.
The last form of the INSERT
statement contains a SELECT
statement
instead of a VALUES
clause. A new entry is inserted into the table for
each row of data returned by executing the SELECT
statement. If a
column name list is specified, the number of columns in the result of
the SELECT
must be the same as the number of items in the column name
list. Otherwise, if no column name list is specified, the number of
columns in the result of the SELECT
must be the same as the number of
columns in the table. Only simple (flattened) SELECT
statement may be
used in an INSERT
statement of this form. This means the SELECT
statements cannot include UNIONs, JOINs, or further sub-queries.
Additionally, only direct references to tables on the same chain are
supported.
⚠️ Although the
GROUP BY
clause is supported,HAVING
is not allowed in anySELECT
statements within anINSERT
. Additionally, under the hood, the Tableland Specification forces an implicitORDER BY rowid
clause on theSELECT
statement.
UPSERT
UPSERT
is a special syntax addition to INSERT
that causes the
INSERT
to behave as an UPDATE
or a no-op if the INSERT
would
violate a uniqueness constraint. UPSERT
is not standard SQL. UPSERT
in Tableland follows the syntax used in
SQLite.
Structure
INSERT INTO table_name [ ( *column_name* [, ...] ) ] VALUES (
{ expression } [, ...]
) [upsert_clause];
where upsert_clause
has structure
ON CONFLICT [ conflict_target ] conflict_action
where conflict_target
has structure
[ ( *column_name* [, ...] ) ] [ WHERE condition ]
and conflict_action
can be one of
DO NOTHING
or
DO UPDATE SET { column_name = { expression | DEFAULT } } [, ...]
[ WHERE condition ];
Details
An UPSERT
is an ordinary INSERT
statement that is followed by the
special ON CONFLICT
clause shown above.
The syntax that occurs in between the "ON CONFLICT
" and "DO
"
keywords is called the "conflict target". The conflict target specifies
a specific uniqueness constraint that will trigger the upsert. The
conflict target is required for DO UPDATE
upserts, but is optional for
DO NOTHING
. When the conflict target is omitted, the upsert behavior
is triggered by a violation of any uniqueness constraint on the table of
the INSERT
.
If the insert operation would cause the uniqueness constraint identified
by the conflict_target
clause to fail, then the insert is omitted and
either the DO NOTHING
or DO UPDATE
operation is performed instead.
In the case of a multi-row insert, this decision is made separately for
each row of the insert.
The special UPSERT
processing happens only for uniqueness constraint
on the table that is receiving the INSERT
. A "uniqueness constraint"
is an explicit UNIQUE
or PRIMARY KEY
constraint within the
CREATE TABLE
statement, or a unique index. UPSERT
does not intervene
for failed NOT NULL
constraints.
Column names in the expressions of a DO UPDATE
refer to the original
unchanged value of the column, before the attempted INSERT
.
Note that the DO UPDATE
clause acts only on the single row that
experienced the constraint error during INSERT
. It is not necessary to
include a WHERE
clause that restricts the action to that one row. The
only use for the WHERE
clause at the end of the DO UPDATE
is to
optionally change the DO UPDATE
into a no-op depending on the original
and/or new values.
UPDATE
An UPDATE
statement is used to modify a subset of the values stored in
zero or more rows of the database table identified by the table name.
Structure
UPDATE table_name
SET { column_name = { expression | DEFAULT } } [, ...]
[ WHERE condition ];
Details
If the UPDATE
statement does not have a WHERE
clause, all rows in the table are modified by the
UPDATE
. Otherwise, the UPDATE
affects only those rows for which the
WHERE
clause boolean expression is true. It is not an error if the
WHERE
clause does not evaluate to true for any row in the table; this
just means that the UPDATE
statement affects zero rows.
The modifications made to each row affected by an UPDATE
statement are
determined by the list of assignments following the SET
keyword. Each
assignment specifies a column-name
to the left of the equals sign and
a scalar expression to the right. For each affected row, the named
columns are set to the values found by evaluating the corresponding
scalar expressions. If a single column-name
appears more than once in
the list of assignment expressions, all but the rightmost occurrence is
ignored. Columns that do not appear in the list of assignments are left
unmodified. The scalar expressions may refer to columns of the row being
updated. In this case all scalar expressions are evaluated before any
assignments are made.
ℹ️ An assignment in the
SET
clause can be a parenthesized list of column names on the left and aROW
value of the same size on the right. For example, consider the following two “styles” ofUPDATE
statements:UPDATE table_id SET (a,b)=(b,a);
orUPDATE table_id SET a=b, b=a;
.
GRANT/REVOKE
The GRANT
and REVOKE
commands are used to define low-level access
privileges for a table identified by table name and id.
Structure
GRANT { INSERT | UPDATE | DELETE } [, ...]
ON { [ TABLE ] table_name [, ...] }
TO role [, ...]
REVOKE { INSERT | UPDATE | DELETE } [, ...]
ON { [ TABLE ] table_name [, ...] }
FROM role [, ...]
Details
The GRANT
command gives specific privileges on a table to one or more
role
. These privileges are added to those already granted, if any. By
default, the creator of a table (as specified by a public ETH address)
has all (valid) privileges on creation. The owner could, however, choose
to revoke some of their own privileges for safety reasons.
Related, if a table is created with an access controller contract
specified, or if an address with sufficient privileges updates a table’s
access control rules to use a controller contract, then all
command-based access control rules are ignored in favor of the
controller contract access control. In other words, if a controller
contract is set, GRANT
/REVOKE
is disabled. See On-Chain API
Specification for further details on specifying and controlling access
via a controller smart contract.
⚠️ Currently, the only allowable privileges for granting are
INSERT
,UPDATE
andDELETE
. Note thatSELECT
privileges are not required at this time, asSELECT
statements are not access controlled (all reads are allowed). See theSELECT
section for further details.
Roles (role
) in Tableland are defined by an Ethereum public-key based
address. Any (hex string encoded) ETH address is a valid Tableland role,
and as such, privileges can be granted to any valid ETH address. In
practice, ETH address strings must be specified as string literals using
single quotes (e.g., '0x181Ec6E8f49A1eEbcf8969e88189EA2EFC9108dD'
).
ℹ️ Only a table owner has permission to
GRANT
orREVOKE
access privileges to other roles/accounts.
Conversely to the GRANT
command, the REVOKE
command removes
specific, previously granted access privileges on a table from one or
more roles. All role definitions and allowable privileges associated
with granting privileges also apply to revoking them.
SELECT
The SELECT
statement is used to query the database. The result of a
SELECT
is zero or more rows of data where each row has a fixed number
of columns. A SELECT
statement does not make any changes to the
database.
Structure
The SELECT
statement is the work-house of the SQL query model, and as
such, the available syntax is extremely complex. In practice, most
SELECT
statements are simple SELECT
statements of the form:
SELECT [ ALL | DISTINCT ]
[ * | expression [, ...] ]
[ FROM from_clause [, ...] ]
[ WHERE where_clause ]
[ GROUP BY [ expression [, ...] ]
[ ORDER BY expression [ ASC | DESC ]
[ LIMIT { count | ALL } ]
[ OFFSET { number } ]
See the standalone sections on WHERE
,
FROM
, and JOIN
clauses for further
details on query structure.
Details
Generating the results of a simple SELECT
statement is presented as a
four step process in the description below:
FROM
clause processing: The input data for the simpleSELECT
is determined. The input data is either implicitly a single row with 0 columns (if there is noFROM
clause) or is determined by theFROM
clause.WHERE
clause processing: The input data is filtered using theWHERE
clause expression.- Result set processing (
GROUP BY
and result expression processing): The set of result rows is computed by aggregating the data according to anyGROUP BY
clause and calculating the result set expressions for the rows of the filtered input dataset. DISTINCT/ALL
keyword processing: If the query is aSELECT DISTINCT
query (see further details below), duplicate rows are removed from the set of result rows.
There are two types of simple SELECT
statement — aggregate and
non-aggregate queries. A simple SELECT
statement is an aggregate query
if it contains either a GROUP BY
clause or one or more aggregate
functions in the result set. Otherwise, if a simple SELECT
contains
no aggregate functions or a GROUP BY
clause, it is a non-aggregate
query.
Once the input data from the FROM
clause has been filtered by the
WHERE
clause expression (if any), the set of result rows for the
simple SELECT
are calculated. Exactly how this is
done depends on whether
the simple SELECT
is an aggregate or non-aggregate query, and whether
or not a GROUP BY
clause was specified.
One of the ALL
or DISTINCT
keywords may follow the SELECT
keyword
in a simple SELECT
statement. If the simple SELECT
is a
SELECT ALL
, then the entire set of result rows are returned by the
SELECT
. If neither ALL
or DISTINCT
are present, then the behavior
is as if ALL
were specified. If the simple SELECT
is a
SELECT DISTINCT
, then duplicate rows are removed from the set of
result rows before it is returned. For the purposes of detecting
duplicate rows, two NULL
values are considered to be equal.
WHERE
clause
Structure
WHERE condition
Details
The SQL WHERE
clause is an optional clause of the SELECT
, DELETE
,
and/or UPDATE
statements. It appears after the primary clauses of the
corresponding statement. For example in a SELECT
statement, the
WHERE
clause can be added after the FROM
clause to filter rows
returned by the query. Only rows for which the WHERE
clause expression
evaluates to true are included from the dataset before continuing. Rows
are excluded from the result if the WHERE
clause evaluates to either
false or NULL
.
When evaluating a SELECT
statement with a WHERE
clause, Tableland
uses the following steps:
- Determine the table(s) in the
FROM
clause, - Evaluate the conditions in the
WHERE
clause to determine the rows that meet the given condition, - Generate the final result set based on the rows in the previous
step, with columns subset to match the
SELECT
statement.
ℹ️ The search condition in the
WHERE
clause is made up of any number of comparisons (=, \<, >, LIKE, IN, etc), combined using a range of logical operators (e.g., OR, AND, ALL, ANY, etc).
FROM
clause
Structure
FROM { table_name [ * ] [ [ AS ] alias ] | ( sub_select ) [ AS ] alias }
Details
The input data used by a simple SELECT
query is a set of N rows
each M columns wide. If the FROM
clause is omitted from a simple
SELECT
statement, then the input data is implicitly a single row zero
columns wide (i.e. N=1 and M=0).
If a FROM
clause is specified, the data on which a simple SELECT
query operates comes from the one or more tables or sub-queries
(SELECT
statements in parentheses) specified following the FROM
keyword. A sub-query specified in the table or sub-query
clause following the FROM
clause in a simple SELECT
statement is
handled as if it was a table containing the data returned by executing
the sub-query statement.
If there is only a single table or sub-query in the FROM
clause (a
common case), then the input data used by the SELECT
statement is the
contents of the named table. If there is more than one table or
sub-query in FROM
clause, then the contents of all tables and/or
sub-queries are joined into a single dataset for the simple SELECT
statement to operate on. Exactly how the data is combined depends on the
specific JOIN
clause (i.e., the combination of join
operator and join constraint) used to connect the tables or sub-queries
together.
JOIN
clause
Structure
[ NATURAL ] join_type table_or_subquery [ ON on_expression | USING ( column_name [, ...] ) ]
where join_type
is one of
[ INNER ] JOIN
LEFT [ OUTER ] JOIN
RIGHT [ OUTER ] JOIN
FULL [ OUTER ] JOIN
ℹ️ There is no difference between the "
INNER JOIN
", "JOIN"
and ",
" join operators. They are completely interchangeable in Tableland.
or,
CROSS JOIN table_or_subquery [ ON on_expression | USING ( column_name [, ...] ) ]
or, a series of comma-separated tables or sub-queries followed by an optional join constraint as in above.
The table_or_subquery
is a table or sub-query of the form:
{ table_name [ [ AS ] alias ] | ( sub_select ) [ AS ] alias }
Details
All joins in Tableland are based on the cartesian product of the left- and right-want databsets. The columns of the cartesian product dataset are, in order, all the columns of the left-hand dataset followed by all the columns of the right-hand dataset. This is a row in the cartesian product dataset formed by combining each unique combination of a raw from the left-hand and right-hand datasets. In other words, if the left-hand dataset consists of rows and columns, and the right-hand dataset of rows of columns, then the cartesian product is a dataset of rows, each containing columns.
If the join operator is "CROSS JOIN
", "INNER JOIN
", "JOIN
" or a
comma (",
") and there is no ON
or USING
clause, then the result of
the join is simply the cartesian product of the left and right-hand
datasets. If join operator does have ON
or USING
clauses, those are
handled according to the following bullet points:
- If there is an
ON
clause then theON
expression is evaluated for each row of the cartesian product as a boolean expression. Only rows for which the expression evaluates to true are included from the dataset. - If there is a
USING
clause then each of the column names specified must exist in the datasets to both the left and right of the join operator. For each pair of named columns, the expression is evaluated for each row of the cartesian product as a boolean expression. Only rows for which all such expressions evaluates to true are included from the result set. When comparing values as a result of aUSING
clause, the normal rules for handling affinities, collation sequences andNULL
values in comparisons apply. The column from the dataset on the left-hand side of the join operator is considered to be on the left-hand side of the comparison operator (=
) for the purposes of collation sequence and affinity precedence. - For each pair of columns identified by a
USING
clause, the column from the right-hand dataset is omitted from the joined dataset. This is the only difference between aUSING
clause and its equivalentON
constraint. - If the
NATURAL
keyword is in the join operator then an implicitUSING
clause is added to the join constraints. The implicitUSING
clause contains each of the column names that appear in both the left and right-hand input datasets. If the left and right-hand input datasets feature no common column names, then theNATURAL
keyword has no effect on the results of the join. AUSING
orON
clause may not be added to a join that specifies theNATURAL
keyword. - If the join operator is a "
LEFT JOIN
" or "LEFT OUTER JOIN
", then after theON
orUSING
filtering clauses have been applied, an extra row is added to the output for each row in the original left-hand input dataset that does not match any row in the right-hand dataset. The added rows containNULL
values in the columns that would normally contain values copied from the right-hand input dataset - If the join operator is a "
RIGHT JOIN
" or "RIGHT OUTER JOIN
", then after theON
orUSING
filtering clauses have been applied, an extra row is added to the output for each row in the original right-hand input dataset that does not match any row in the left-hand dataset. The added rows containNULL
values in the columns that would normally contain values copied from the left-hand input dataset. - A "
FULL JOIN
" or "FULL OUTER JOIN
" is a combination of a "LEFT JOIN
" and a "RIGHT JOIN
". Extra rows of output are added for each row in left dataset that matches no rows in the right, and for each row in the right dataset that matches no rows in the left. Unmatched columns are filled in withNULL
.
When more than two tables are joined together as part of a FROM
clause, the join operations are processed in order from left to right.
In other words, the FROM
clause is computed as
.
⚠️ The "
CROSS JOIN
" join operator produces the same result as the "INNER JOIN"
, "JOIN
" and ",
" operators, but is handled differently by the query optimizer in that it prevents the query optimizer from reordering the tables in the join. An application programmer can use theCROSS JOIN
operator to directly influence the algorithm that is chosen to implement theSELECT
statement. Avoid usingCROSS JOIN
except in specific situations where manual control of the query optimizer is desired. Avoid usingCROSS JOIN
early in the development of an application as doing so is a premature optimization. The special handling ofCROSS JOIN
is an implementation detail. It is not a part of standard SQL, and should not be relied upon.
Compound Select Statements
Two or more simple SELECT
statements may be connected together to form
a compound SELECT
using the UNION
, UNION ALL
, INTERSECT
or
EXCEPT
operator.
In a compound SELECT
, all the constituent SELECT
s must return the
same number of result columns. As the components of a compound SELECT
must be simple SELECT
statements, they may not contain ORDER BY
or
LIMIT
clauses. ORDER BY
and LIMIT
clauses may only occur at the
end of the entire compound SELECT
, and then only if the final element
of the compound is not a VALUES
clause.
A compound SELECT
created using UNION ALL
operator returns all the
rows from the SELECT
to the left of the UNION ALL
operator, and all
the rows from the SELECT
to the right of it. The UNION
operator
works the same way as UNION ALL
, except that duplicate rows are
removed from the final result set. The INTERSECT
operator returns the
intersection of the results of the left and right SELECT
s. The
EXCEPT
operator returns the subset of rows returned by the left
SELECT
that are not also returned by the right-hand SELECT
.
Duplicate rows are removed from the results of INTERSECT
and EXCEPT
operators before the result set is returned.
For the purposes of determining duplicate rows for the results of
compound SELECT
operators, NULL
values are considered equal to other
NULL
values and distinct from all non-NULL
values. The collation
sequence used to compare two text values is determined as if the columns
of the left and right-hand SELECT
statements were the left and
right-hand operands of the equals (=
) operator, except that greater
precedence is not assigned to a collation sequence specified with the
postfix COLLATE
operator. No affinity transformations are applied to
any values when comparing rows as part of a compound SELECT
.
When three or more simple SELECT
s are connected into a compound
SELECT
, they group from left to right. In other words, if , and
are all simple SELECT
statements, is processed as
.
Custom functions
The Tableland SQL Specification includes several web3 native functions that simplify working with blockchain transactions. The list of custom functions may grow over time.
TXN_HASH()
The Validator will replace this text with the hash of the transaction that delivered the SQL event (only available in write queries).
INSERT INTO {table_name} VALUES (TXN_HASH());
BLOCK_NUM()
The Validator will replace this text with the number of the block that delivered the SQL event (only available in write queries).
INSERT INTO {table_name} VALUES (BLOCK_NUM());
If BLOCK_NUM
is called with an integer argument (i.e.,
BLOCK_NUM(<chain_id>)
), the Validator will replace this text with the
number of the last seen block for the given chain (only available to
read queries).
Data Types
Tableland supports a small set of accepted column types in user-defined tables. The currently supported types are listed below and can be used to represent most, if not all, common SQL types:
Type | Description |
---|---|
INT | Signed integer values, stored in 0, 1, 2, 3, 4, 6, or 8 bytes depending on the magnitude of the value. |
INTEGER | Same as INT , except it may also be used to represent an auto-incrementing PRIMARY KEY field. |
TEXT | Text string, stored using the database encoding (UTF-8). |
BLOB | A blob of data, stored exactly as it was input. Useful for byte slices etc. |
Details
When creating tables, every column definition must specify a data type for that column, and the data type must be one of the above types. No other data type names are allowed, though new types might be added in future versions of the Tableland SQL specification.
Content inserted into a column must be either a NULL
(assuming there
is no NOT NULL
constraint on the column) or the type specified.
Tableland will attempt to coerce input data into the appropriate type
using the usual affinity rules, as most SQL engines all do. However, if
the value cannot be losslessly converted in the specified datatype, then
an error will be raised.
Common Types
For users looking for more nuanced data types in tables, the following set of recommendations will help guide table schema design. Additionally, new types might be added in future versions of the Tableland SQL Specification, and users are able to make requests/suggestions via Tableland TIPs.
Character
Tableland represents all character/text types using the single
variable-length TEXT
type. Although the type TEXT
is not in any SQL
standard, several other SQL database management systems have it as well.
You can store any text/character-based data as TEXT
. Additionally,
more complex data types such as dates, timestamps, JSON strings, and
more can be represented using TEXT
(or in some cases BLOB
).
Integers
Numeric types often consist of integer and floating-point (float/real)
numbers. On Tableland, two-, four-, and eight-byte integers are all
represented by the INTEGER
type, and their storage size depends on the
magnitude of the value itself.
Floats
Tableland does not have a separate data type to represent float/real types. This is because in practice floating point values are approximate, which may lead to non-deterministic behavior across compute platforms. If you need an exact answer, you should not use floating-point values, in Tableland or in any other software. This is not a Tableland limitation per se, but a mathematical limitation inherent in the design of floating-point numbers.
See the SQLite documentation about issues with floating-point numbers, or learn more about why floating-point math is hard.
⚠️ In addition to not supporting floating point values (REAL
) as a
storage data type in create statements, the Tableland specification also
does not allow REAL
value literals in read or write queries.
Boolean
Tableland does not have a separate data type to represent boolean
values. Instead, Tableland users should represent true and false values
using the integers 1
(true) and 0
(false).
Date/Time
Tableland does not have a storage class set aside for storing dates
and/or times. Instead, users of Tableland can store dates and times as
TEXT
or INTEGER
values:
TEXT
as ISO-8601 strings.INTEGER
as Unix Time (number of seconds since (or before) 1970-01-01 00:00:00 UTC).
Tableland does not support any of the date nor time functions provided by the SQLite database engine. Namely, these functions can lead to non-deterministic behavior, so they are not available.
JSON
JSON data types are for storing JSON (JavaScript Object Notation) data,
as specified in RFC 7159. In
practice, Tableland stores JSON as ordinary TEXT
. Users are able to
manipulate JSON data using a number of functions that make working with
JSON data much easier. For example, the json(X)
function verifies that
its argument X
is a valid JSON string and returns a minified version
of that JSON string (with all unnecessary whitespace removed). If X
is
not a well-formed JSON string, then this routine throws an error. The
JSON manipulation functions supported by Tableland is derived from
SQLite, which in turn is generally compatible (in terms of syntax) with
PostgresQL. Tableland currently supports the 15 scalar functions and
operators and two aggregate SQL functions for JSON
data provided by the SQLite database
engine.
🚧 Feature At Risk: Note that these JSON scalar functions, operators, and aggregate functions have not yet been formalized into the Tableland SQL language specification. You are welcome to use them for now, but they should be considered unstable features.
Solidity
To prevent overflows while working with Solidity numbers, it is
recommended to use a text
type in certain scenarios. Anything larger
than a uint64
/ int32
could lead to an overflow in the Tableland
database. Note that in many use cases, it is unlikely overflows will
happen due to the extremely large size of these numbers.
Alternatively, consider casting the overflow-able numbers to or simply
use a int64
in smart contracts if it makes sense for the use case. See
the following tables for how each Solidity number should be defined in
Tableland schemas:
Solidity Type | SQL Type |
---|---|
uint256 | text |
uint128 | text |
uint64 | text |
uint32 | integer |
uint16 | integer |
uint8 | integer |
int256 | text |
int128 | text |
int64 | integer |
int32 | integer |
int16 | integer |
int8 | integer |
Other best practices have also been defined below:
Solidity Type | SQL Type |
---|---|
string | text |
address | text |
bytes | blob |
bool | integer |
⚠️ Tableland doesn’t support boolean values, but TRUE/FALSE keywords are supported since they're aliased to a
1
or0
. Thus, a Soliditybool
should be defined as aninteger
in Tableland.
Encoding
As mentioned in the section on Statement Types, the core Tableland SQL parser accepts a semicolon-separated list of statements, which are then parsed and evaluated according to this Tableland SQL Specification. Internally, the statements are represented using an abstract syntax tree (AST). The internal representation of the nodes of this AST is outside the bounds of the Tableland SQL Specification, however, further details can be found in the Go Tableland SQL Parser reference implementation.
With the above caveat in mind, the Tableland SQL Specification does define a canonical string encoding of a set of (compliant) SQL statements that have passed through the Tableland SQL Parser (and have been represented via the Parser's AST). That is, this Specification outlines — in general terms — the string encoding produced by parsing a set of Tableland SQL Specification compliant statements and re-encoding them into a canonical (string) format.
There are some nuances and corner cases that affect the final encoded
string. For example, the statement UPDATE t SET (A, b) = (1, 2);
is
ultimately encoded as update t set A = 1, b = 2
. This is because the
two statements are equivalent, and their representation within the AST
is identical. As such, when producing the canonical string for such a
statement, the parser outputs the most conventional form.
In general, any (set of) statement(s) processed by the parser should be encoded such that,
- All SQL language components are specified using lower case ASCII characters,
- The execution of the (set of) statement(s) after encoding is equivalent to the original set of statements, and
- The encoding of a (set of) statements(s) is as close as possible to the original (set of) statement(s).
Any further guarantees are left outside the scope of this specification.