IBM Cloud Docs
DataPrime reference

DataPrime reference

This guide provides a full glossary of all available DataPrime operators and expressions.

Operators

DataPrime provides the following operators.

block

The negation of filter. Filters-out all events where the condition is true. The same effect can be achieved by using filter with !(condition).

block $d.status_code >= 200 && $d.status_code <= 299         # Leave all events which don't have a status code of 2xx

The data is exposed using the following fields:

  • $m - Event metadata

    • timestamp
    • severity – Possible values are Verbose, Debug, Info, Warning, Error, Critical
    • priorityclass – Possible values are high, medium, low
    • logid
  • $l - Event labels

    • applicationname
    • subsystemname
    • category
    • classname
    • computername
    • methodname
    • threadid
    • ipaddress
  • $d -The user’s data

bottom

No grouping variation: Limits the rows returned to a specified number and order the result by a set of expressions.

order_direction := "descending"/"ascending" according to top/bottom
bottom <limit> <result_expression1> [as <alias>] [, <result_expression2> [as <alias2>], ...] by <orderby_expression> [as alias>]

For example, the following query:

bottom 5 $m.severity as $d.log_severity by $d.duration

Will result in logs of the following form:

[
   { "log_severity": "Debug", "duration":  1000 }
   { "log_severity": "Warning", "duration": 2000 },
   ...
]

Grouping variation: Limits the rows returned to a specified number and groups them by a set of aggregation expressions and orders them by a set of expressions.

order_direction := "descending"/"ascending" according to top/bottom

bottom <limit> <(groupby_expression1|aggregate_function1)> [as <alias>] [, <(groupby_expression2|aggregate_function2)> [as <alias2>], ...] by <(groupby_expression1|aggregate_function1)> [as <alias>]

For example, the following query:

bottom 10 $m.severity, count() as $d.number_of_severities by avg($d.duration) as $d.avg_duration

Will result in logs of the following form:

[
   { "severity": "Warning", "number_of_severities": 50, avg_duration: 1000 },
   { "severity": "Debug", "number_of_severities":  10, avg_duration: 2000 }
   ...
]

Supported aggregation functions are listed in “Aggregation Functions” section.

choose

Leave only the keypaths provided, discarding all other keys. Fully supports nested keypaths in the output.

(choose|select) <keypath1> [as <new_keypath>],<keypath2> [as <new_keypath>],...

Examples:

choose $d.mysuperkey.myfield
choose $d.my_superkey.mykey as $d.important_value, 10 as $d.the_value_ten

convert

Convert the data types of keys.

The datatypes keyword is optional and can be used for readability.

(conv|convert) [datatypes] <keypath1>:<datatype1>,<keypath2>:<datatype2>,...

Examples:

convert $d.level:number
conv datatypes $d.long:number,$d.lat:number
convert $d.data.color:number,$d.item:string

count

Returns a single row containing the number of rows produced by the preceding operators.

count [into <keypath>]

An alias can be provided to override the keypath where the result will be written.

For example, the following part of a query:

count into $d.num_rows

Will result in a single row of the following form:

{ "num_rows": 7532 }

countby

Returns a row counting all the rows grouped by the expression.

countby <expression> [as <alias>] [into <keypath>]

An alias can be provided to override the keypath the result will be written into.

For example, the following part of a query

countby $d.verb into $d.verb_count

Will result in a row for each group.

It is functionally identical to

groupby $data.verb calculate count() as $d.verb_count

create

Create a new key and set its value to the result of the expression. Key creation is granular, meaning that parent keys in the path are not overwritten.

  (a|add|c|create) <keypath> from <expression> [on keypath exists (fail|skip|overwrite)] [on keypath missing (fail|create|skip)] [on datatype change (skip|fail|overwrite)

The creation can be controlled by adding the following clauses:

  • Adding keypath exists allows to choose what to do when the keypath already exists.

    • overwrite – Overwrites the old value. This is the default value

    • fail – Fails the query

    • skip – Skips the creation of the key

  • Adding keypath missing chooses what to do when the new keypath does not exist.

    • create – Creates the key. This is the default value

    • fail – Fails the query

    • skip – Skips the creation of the new key

  • Adding on datatype changed chooses what to do if the key already exists and the new data changes the datatype of the value.

    • overwrite – Overwrites the value. This is the default value.

    • fail – Fails the query

    • skip – Leaves the key with the original value (and type)

Examples:

create $d.radius from 100+23
c $d.log_data.truncated_message from $d.message.substring(1,50)
c $data.trimmed_name from $data.username.trim()

create $d.temperature from 100*23 on datatype changed skip

distinct

Returns one row for each distinct combination of the provided expressions.

distinct <expression> [as <alias>] [, <expression_2> [as <alias_2>], ...]

This operator is functionally identical to groupby without any aggregate functions.

enrich

Enrich your logs using additional context from a lookup table.

Upload your lookup table using Data flow Data Flow icon > Data Enrichment > Custom Enrichment.

enrich <value_to_lookup> into <enriched_key> using <lookup_table>
  • value_to_lookup – A string expression that will be looked up in the lookup table.

  • enriched_key – Destination key to store the enrichment result.

  • lookup_table – The name of the Custom Enrichment table to be used.

The table’s columns will be added as sub-keys to the destination key. If value_to_lookup is not found, the destination key will be null. You can then filter the results using the DataPrime capabilities, such as filtering logs by specific value in the enriched field.

Example:

The original log:

{
    "userid": "111",
    ...
}

The Custom Enrichment lookup table called my_users:

Sample lookup table
ID Name Department
111 John Finance
222 Emily IT

Running the following query:

enrich $d.userid into $d.user_enriched using my_users

Will result in the following enriched log:

{
    "userid": "111",
    "user_enriched": {
        "ID": "111",
        "Name": "John",
        "Department": "Finance"
    },
    ...
}

Consider the following when using enrich:

  • Run the DataPrime query source lookup_table to view the enrichment table.

  • If the original log already contains the enriched key:

    • If value_to_lookup exists in the lookup_table, the sub-keys will be updated with the new value. If the value_to_lookup does not exist, their current value will remain.

    • Any other sub-keys which are not columns in the lookup_table will remain with their existing values.

  • All values in the lookup_table are considered to be strings. This means that:

    • The value_to_lookup must be in a string format.

    • All values are enriched in a string format. You can then convert them to your preferred format (for example, JSON, timestamp) using the appropriate functions.

extract

Extract data from some string value into a new object. Multiple extraction methods are supported.

(e|extract) <expression> into <keypath> using <extraction-type>(<extraction-params>) [datatypes keypath:datatype,keypath:datatype,...]

The following are the supported extraction methods and their parameters:

  • regexp – Create a new object based on regexp capture-groups

  • e – A regular expression with names capture-groups.

Example:

extract $d.my_text into $d.my_data using regexp(e=/user (?<user>.*) has logged in/)
  • kv – Extract a new object from a string that contains key=value key=value... pairs

  • pair_delimiter – The delimiter to expect between pairs. Default is (a space)

  • key_delimiter – The delimiter to expect separating between a key and a value. Default is =.

Examples:

extract $d.text into $d.my_kvs using kv()
e $d.text into $d.my_kvs using kv(pair_delimiter=' ',key_delimiter='=')
  • jsonobject – Extract a new object from a string contains an encoded json object, potentially attempting to unescape the string before decoding it into a json

  • max_unescape_count – Max number of escaping levels to unescape before parsing the json. Default is 1. When set to 1 or more, the engine will detect whether the value contains an escaped JSON string and unescape it until its parsable or max unescape count is exceeded.

Example:

e $d.json_message_as_str into $d.json_message using jsonobject(max_unescape_count=1)

It is possible to provide datatype information as part of the extraction, by using the datatypes clause. For example, adding datatypes my_field:number to an extraction would cause the extract my_field keypath to be a number instead of a string. For example:

extract $d.my_msg into $d.data using kv() datatypes my_field:number

Extracted data always goes into a new keypath as an object, allowing further processing of the new keys inside that new object. For example:

# Assuming a dataset which look like that:
{ "msg": "query_type=fetch query_id=100 query_results_duration_ms=232" }
{ "msg": "query_type=fetch query_id=200 query_results_duration_ms=1001" }

# And the following DataPrime query:
source logs
  | extract $d.msg into $d.query_data using kv() datatypes
query_results_duration_ms:number
  | filter $d.query_data.query_results_duration_ms > 500

# The results will contain only the second message, in which the duration is greater than 500 ms

filter

Filter events, leaving only events for which the condition evaluates to true.

(f|filter|where) <condition-expression>

Examples:

f $d.radius > 10
filter $m.severity.toUpperCase() == 'INFO'
filter $l.applicationname == 'recommender'
filter $l.applicationname == 'myapp' && $d.msg.contains('failure')

Comparison with null works only for scalar values and will always return null on JSON subtrees.

groupby

Groups the results of the preceding operators by the specified grouping expressions and calculates aggregate functions for every group created.

groupby <grouping_expression> [as <alias>] [, <grouping_expression_2> [as <alias_2>], ...] [calculate]
  <aggregate_function> [as <result_keypath>]
  [, <aggregate_function_2> [as <result_keypath_2], ...]

For example, the following query:

groupby $m.severity calculate sum($d.duration)

Will result in logs of the following form:

{ "severity": "Warning", "_sum": 17045 }

The keypaths for the grouping expressions will always be under $d. Using the as keyword, we can rename the keypath for the grouping expressions and aggregation functions. For example:

groupby $l.applicationname as $d.app calculate sum($d.duration) as $d.sum_duration

Will result in logs of the following form:

{ "app": "web-api", "sum_duration": 17045 }

When querying with the groupby operator, you can apply an aggregation function (such asavg, max, sum) to the bucket of results. This feature gives you the power to manipulate an aggregation expression inside the expression itself, allowing you to calculate and manipulate your data simultaneously.

join

Join merges the results of the current (left) query with a second (right) query based on a specified condition. It offers multiple forms to control how the data is combined and supports nesting, allowing the right query to include its own join command.

Join supports three variations:

join left|join
For each event in the left query, the command selects a matching event from the right query based on the specified condition. If no match is found, rows are included for all events in the left query. Unmatched rows from the right query are set to null.
join full
Returns a row for every event, including those that may not have a match in either query (left or right), filling missing values with null.
join inner
Only returns rows where there are non-null results from both queries.

Syntax:

 <left_side_query> | join (<right_side_query>) on <condition> into <right_side_target>
 <left_side_query> | join (<right_side_query>) using <join_keypath_1> [, <join_keypath_2>, ...] into <right_side_target>

Where:

  • <right_side_query> - The <right_side_query> denotes the new query to be joined to.

  • <left_side_query> - The <left_side_query> denotes the initial query, for example, in the query source logs | filter x != null | join ..., the left hand side query is source logs | filter x != null.

  • <condition> - The condition if results from both queries should be joined.

  • <join_keypath_n> - <join_keypath_n> as a join key means to join results where a given keypath is equal in the results of both the left side query and the right side query.

  • <right_side_target> - The keypath where the joined data will be added to the current query.

join example

You have this custom enrichment table named users providing information on IDs related to names:

{ "id": "111", "name": "John" }
{ "id": "222", "name": "Emily" }
{ "id": "333", "name": "Alice" }

And this data providing login events and user IDs, but not the user name associated with the user IDs.

{ "userid": "111", "timestamp": "2022-01-01T12:00:00Z" }
{ "userid": "111", "timestamp": "2022-01-01T12:30:00Z" }
{ "userid": "222", "timestamp": "2022-01-01T13:00:00Z" }
{ "userid": "222", "timestamp": "2022-01-01T13:00:00Z" }
{ "userid": "222", "timestamp": "2022-01-01T13:00:00Z" }

Using join you can use a query to return data including the desired data.

source users | join (source logs | countby userid) on id == userid into logins

This query is processed as follows:

  • The source is the custom enrichment table (users).

  • In the join, a count by the userid field is generated. This gives us our count statistics.

  • The id field in the custom enrichment table is compared with the userid field in the logs.

  • The result is pushed into the logins key. If the logins key already exists on the left hand query, it will be overwritten.

For example:

{ "id": "111", "name": "John", "logins": { "userid": "111", "_count": 2 } }
{ "id": "222", "name": "Emily", "logins": { "userid": "222", "_count": 3 } }
{ "id": "333", "name": "Alice", "logins": null }

The result of the right side query now is inside the logins field. Note that there were no logins for user ID 333 (Alice), so the logins field is null because there was no result matched by the join condition.

join example with the using keyword

Consider if our login dataset is:

{ "id": "111", "timestamp": "2022-01-01T12:00:00Z" }
{ "id": "111", "timestamp": "2022-01-01T12:30:00Z" }
{ "id": "222", "timestamp": "2022-01-01T13:00:00Z" }
{ "id": "222", "timestamp": "2022-01-01T13:00:00Z" }
{ "id": "222", "timestamp": "2022-01-01T13:00:00Z" }

In this case data on both sides of the join include the field id. In this case the using keyword can be used to take advantage of the common data:

source users | join (source logins | countby id) using id into logins

The result will be similar, but instead of userid, the field id is returned.

{ "id": "111", "name": "John", "logins": { "id": "111", "_count": 2 } }
{ "id": "222", "name": "Emily", "logins": { "id": "222", "_count": 3 } }
{ "id": "333", "name": "Alice", "logins": null }

If you have two fields that are named differently but would simplify your join query, you can use move to move one of the fields so keypaths match on both sides.

join example using left=> and right=> keywords

You can use the left=> and right=> prefixes to refer to the events of the left and right queries. However, it is not required if a keypath exists in only one of the queries.

Using the data from the previous example, consider the query:

source users | join (source logins | countby id) on left=>id == right=>id into logins

This is required because both data sets contain a field with the same name (id). For DataPrime to uniquely identify a field, it must know which side of the query we are referring. This query will result in the same output as the previous that used the using keyword.

When using the == (equality) operator in your condition, it must compare a keypath from the left query with a keypath from the right query. However, given that the keypaths must be unique or prefixed with left=> or right=>, the ordering of the operands is not important.

join full example

You have this custom enrichment table named users providing information on IDs related to names:

{ "id": "111", "name": "John" }
{ "id": "222", "name": "Emily" }
{ "id": "333", "name": "Alice" }

And consider this dataset:

{ "id": "001", "timestamp": "2022-01-01T12:00:00Z" }
{ "id": "111", "timestamp": "2022-01-01T12:00:00Z" }
{ "id": "111", "timestamp": "2022-01-01T12:30:00Z" }
{ "id": "222", "timestamp": "2022-01-01T13:00:00Z" }
{ "id": "222", "timestamp": "2022-01-01T13:00:00Z" }
{ "id": "222", "timestamp": "2022-01-01T13:00:00Z" }

The second set of documents (right query) includes a log entry with "id": "001" that does not exist in the first set of documents (left query). If you use a standard join, this entry from the right query will be ignored and will not appear in the result. To ensure that every id field is included in the output, whether or not it appears in the left or right query, you can use join full:

source users | join full (source logins | countby id) using id into logins

This query results in:

{ "id": "001", "name": "null", "logins": { "id": "001", "_count": 1 } }
{ "id": "111", "name": "John", "logins": { "id": "111", "_count": 2 } }
{ "id": "222", "name": "Emily", "logins": { "id": "222", "_count": 3 } }
{ "id": "333", "name": "Alice", "logins": null }

By using join full, all id fields from both datasets are preserved, and any missing values are set to null.

join full is particularly useful when the results of both queries include time buckets. For example, if the left query's results are missing a time bucket for a specific hour (for example, XX:XX:XX), with join full this data point is included. This is especially useful for comparing two time series on a graph.

join inner example

If you want to remove any rows if column results for the left or right queries that produce a null value, use join inner.

Using the previous data, this query removes rows with unmatched data from either side:

source users | join inner (source logins | countby id) using id into logins
  • The left query source users retrieves the users dataset containing the fields id and name.

  • The right query (source logins | countby id) retrieves the logins dataset, grouping by id and counting occurrences for each id.

  • join inner matches rows where id exists in both datasets and merges the data into a single record.

  • Rows with no match in either dataset are excluded from the final results.

In this case, for the two document sets above, results will be as follows:

{ "id": "111", "name": "John", "logins": { "id": "111", "_count": 2 } }
{ "id": "222", "name": "Emily", "logins": { "id": "222", "_count": 3 } }

Limitations and considerations

There are limitations and considerations when including join in a query:

  • The join condition only supports keypath equality (==). If multiple equality conditions are needed, they can be combined with && (logical and).

  • One side of the join (either current query or join query) must be small (< 200MB). You can use filter and remove to reduce the size of the query.

  • Left outer joins require all columns in the condition to be non-null. Any null columns will not be joined. To include right hand query null columns, use join full. To exclude all null columns produced by the left and right joins, use join inner.

limit

Limits the output to the first event-count events.

limit <event-count>

Example:

limit 100

move

Move a key (including its child keys, if any) to a new location.

(m|move) <source-keypath> to <target-keypath>

Examples:

move $d.my_data.hostname to $d.my_new_data.host
m $d.kubernetes.labels to $d.my_labels

orderby / sortby / order by / sort by

Sort the data by ascending/descending order of the expression value. Ordering by multiple expressions is supported.

(orderby|sortby|order by|sort by) <expression> [(asc|desc)] , ...

Examples:

orderby $d.myfield.myfield
orderby $d.myfield.myfield:number desc
sortby $d.myfield desc

Sorting numeric values can be done by casting expression to the type:, for example, <expression>: number. In some cases, this will be inferred automatically by the engine.

redact

Replace all substrings matching a regexp pattern from some keypath value, effectively hiding the original content.

The matching keyword is optional and can be used to increase readability.

redact <keypath> [matching] /<regular-expression>/ to '<redacted_str>'
redact <keypath> [matching] <string> to '<redacted_str>'

Examples:

redact $d.mykey /[0-9]+/ to 'SOME_INTEGER'
redact $d.mysuperkey.user_id 'root' to 'UNKNOWN_USER'
redact $d.mysuperkey.user_id matchingn 'root' to 'UNKNOWN_USER'

remove

Remove a keypath from the object.

r|remove <keypath1> [ "," <keypath2> ]...

Examples:

r $d.mydata.unneeded_key
remove $d.mysuperkey.service_name, $d.mysuperkey.unneeded_key

replace

Replace the value of some key with a new value.

If the replacement value changes the datatype of the keypath, the following options are available:

  • skip – The replacement will be ignored

  • fail – The query will fail

  • overwrite – The new value will overwrite the previous one, changing the datatype of the keypath

replace <keypath> with <expression> [on datatype changed skip/fail/overwrite]

Examples:

replace $d.message with null
replace $d.some_superkey.log_length_plus_10 with $d.original_log.length()+10 on datatype changed overwrite

roundtime

Rounds the time of the event into some time interval, possibly creating a new key for the result.

  • If source-timestamp is not provided, then $m.timestamp is used as the source timestamp.

  • If source-timestamp is provided, it should be of type (or cast to) timestamp.

By default, the rounded result is written back to the source keypath source-timestamp. If into target-keypath is provided, then source-timestamp is not modified, and the result is written to a new target-keypath.

Supported time intervals are:

  • Xns – X nanoseconds (Be cautious of the source-timestamp‘s resolution)
  • Xms – X milliseconds
  • Xs – X seconds
  • Xm – X minutes
  • Xh – X hours
  • Xd – X days

And any combination from greater to lesser time units, for example, 1h30m15s.

roundtime [source-timestamp] to <time-interval> [into <target-keypath>]

Examples:

roundtime to 1h into $d.tm
roundtime $d.timestamp to 1h
roundtime $d.my_timestamp: timestamp to 60m
roundtime to 60s into $d.rounded_ts_to_the_minute

source

Set the data source that your DataPrime query is based on.

(source|from) <data_store>

Where data_store can be either:

  • logs

  • The name of the custom enrichment. In this case, the command will display the custom enrichment table.

Examples:

source logs

stitch

The stitch command performs a horizontal union of two datasets, combining them side-by-side. It aligns rows from one dataset with rows from another and concatenates their columns, creating a single, unified dataset.

When using the stitch command:

  • Datasets must be ordered, since rows are combined in sequence (that is, row 1 of dataset A is stitched with row 1 of dataset B).

  • If one dataset has more rows than the other, unmatched rows will have null values in the stitched columns.

  • The resulting dataset will contain all columns from both datasets.

... | stitch (<subquery>) into <target-keypath>

Example:

You have these custom enrichment tables:

sales dataset:

{ "product": "Widget", "sales": 100 }
{ "product": "Gadget", "sales": 200 }
{ "product": "Dashboard", "sales": 150 }

revenue dataset:

{ "product": "Widget", "revenue": 5000 }
{ "product": "Gadget", "revenue": 8000 }
{ "product": "Dashboard", "revenue": 6000 }

In this query you will combine these datasets side-by-side, ensuring that each row from one dataset aligns with the corresponding row from the other:

source sales | orderby product
| stitch (source revenue | orderby product) into combined_data
  • source sales fetches all rows from the sales dataset, which contains products and their corresponding sales figures.

  • orderby product sorts the sales dataset by the product field to creae a consistent order for row alignment.

  • stitch (source revenue | orderby product) fetches rows from the revenue dataset and sorts them by the product field. The sales and revenue datasets are combined horizontally, aligning rows based on their order after sorting.

  • into combined_data stores the combined dataset into a variable called combined_data.

The result from the query is:

{ "product": "Widget", "sales": 100, "combined_data": { "product": "Widget", "revenue": 5000 } }
{ "product": "Gadget", "sales": 200, "combined_data": { "product": "Gadget", "revenue": 8000 } }
{ "product": "Dashboard", "sales": 150, "combined_data": { "product": "Dashboard", "revenue": 6000 } }

If the datasets have unequal rows, the stitch command fills the missing values with null.

For example, consider the following datasets:

sales dataset (3 rows):

{ "product": "Widget", "sales": 100 }
{ "product": "Gadget", "sales": 200 }
{ "product": "Dashboard", "sales": 150 }

revenue dataset (2 rows):

{ "product": "Widget", "revenue": 5000 }
{ "product": "Gadget", "revenue": 8000 }

Running this query:

source sales | orderby product
| stitch (source revenue | orderby product) into combined_data

Results in:

{ "product": "Widget", "sales": 100, "combined_data": { "product": "Widget", "revenue": 5000 } }
{ "product": "Gadget", "sales": 200, "combined_data": { "product": "Gadget", "revenue": 8000 } }
{ "product": "Dashboard", "sales": 150, "combined_data": { "product": "Dashboard", "revenue": null } }

stitch usage notes

  • Rows must correlate logically for stitching to produce meaningful results. Make sure that rows from both datasets represent the same entities and are in the same order. For example, if the product field in the sales dataset does not match the product field in the revenue dataset for corresponding rows, stitching will not work as expected.

  • If datasets differ in row count, the result will include null values for missing data in the shorter dataset.

top

No grouping variation: Limits the rows returned to a specified number and order the result by a set of expressions.

order_direction := "descending"/"ascending" according to top/bottom

top <limit> <result_expression1> [as <alias>] [, <result_expression2> [as <alias2>], ...] by <orderby_expression> [as alias>]

For example, the following query:

top 5 $m.severity as $d.log_severity by $d.duration

Will result in logs of the following form:

[
   { "log_severity": "Warning", "duration": 2000 },
   { "log_severity": "Debug", "duration":  1000 }
   ...
]

Grouping variation: Limits the rows returned to a specified number and groups them by a set of aggregation expressions and orders them by a set of expressions.

order_direction := "descending"/"ascending" according to top/bottom

top <limit> <(groupby_expression1|aggregate_function1)> [as <alias>] [, <(groupby_expression2|aggregate_function2)> [as <alias2>], ...] by <(groupby_expression1|aggregate_function1)> [as <alias>]

For example, the following query:

top 10 $m.severity, count() as $d.number_of_severities by avg($d.duration) as $d.avg_duration

Will result in logs of the following form:

[
   { "severity": "Debug", "number_of_severities":  10, avg_duration: 2000 }
   { "severity": "Warning", "number_of_severities": 50, avg_duration: 1000 },
   ...
]

You can apply an aggregation function.

Expressions

DataPrime supports a limited set of JavaScript constructs that can be used in expressions.

The data is exposed using the following fields:

  • $m – Event metadata

    • timestamp

    • severity – Possible values are Verbose, Debug, Info, Warning, Error, and Critical.

    • priorityclass – Possible values are high, medium, low.

    • logid

  • $l – Event labels

    • applicationname

    • subsystemname

    • category

    • classname

    • computername

    • methodname

    • threadid

    • ipaddress

  • $d – The user’s data

Field Access

Accessing nested data is done by using a keypath, similar to any programming language or JSON tool. Keys with special characters can be accessed using a map-like syntax, with the key string as the map index, for example, $d.my_superkey['my_field_with_a_special/character'].

Examples:

$m.timestamp
$d.my_superkey.myfield
$d.my_superkey['my_field_with_a_special/character']
$l.applicationname

Language Constructs

All standard language constructs are supported:

  • Constants

  • Nested field access

  • Basic math operations between numbers, including modulo (%)

  • Boolean operations &&, ||, !

  • Comparisons

  • String concatenations through concat.

  • casting – A simple notation for casting data types: for example, $d.temperature:number. Type inference is automatically applied when possible.

Scalar Functions

Various functions can be used to transform values. All functions can be called as methods as well, for example, $d.msg.contains('x') is equivalent to contains($d.msg,'x').

String Functions

chr

Returns the Unicode code point number as a single character string.

chr(number: number): string

codepoint

Returns the Unicode code point of the only character of string.

codepoint(string: string): number

concat

Concatenates multiple strings into one.

concat(value: string, ...values: string): string

contains

Returns true if substring is contained in string

contains(string: string, substring: string): bool

endsWith

Returns true if string ends with suffix

endsWith(string: string, suffix: string): bool

indexOf

Returns the position of substring in string, or -1 if not found.

indexOf(string: string, substring: string): number

length

Returns the length of value

length(value: string): number

ltrim

Removes whitespace to the left of the string value

ltrim(value: string): string

matches

Evaluates the regular expression pattern and determines if it is contained within string.

matches(string: string, regexp: regexp): bool

pad

Left pads string to charCount. If size < fillWith.length() of string, result is truncated. See padLeft for more details. Alias for padLeft.

pad(value: string, charCount: number, fillWith: string): string

padLeft

padLeft(value: string, charCount: number, fillWith: string): string

Left pads string to charCount. If size < fillWith.length() of string, result is truncated.

padRight

Right pads string to charCount. If size < fillWith.length() of string, result is truncated.

padRight(value: string, charCount: number, fillWith: string): string

regexpSplitParts

Splits string on regexp-delimiter, returns the field at index. Indexes start with 1.

regexpSplitParts(string: string, delimiter: regexp, index: number): string

rtrim

Removes whitespace to the right of the string value.

rtrim(value: string): string

splitParts

Splits string on delimiter, returns the field at index. Indexes start with 1.

splitParts(string: string, delimiter: string, index: number): string

startsWith

Returns true if string starts with prefix.

startsWith(string: string, prefix: string): bool

substr

Returns the substring in value, from position from and up to length length.

substr(value: string, from: number, length: number?): string

toLowerCase

Converts value to lowercase.

toLowerCase(value: string): string

toUpperCase

Converts value to uppercase.

toUpperCase(value: string): string

trim

Removes whitespace from the edges of a string value.

trim(value: string): string

IP Functions

ipInSubnet

Returns true if IP is in the subnet of ipPrefix.

ipInSubnet(ip: string, ipPrefix: string): bool

ipPrefix

Returns the IP prefix of a given IP address with subnetSize bits (for example, 192.128.0.0/9).

ipPrefix(ip: string, subnetSize: number): string

String interpolation

  • this is an interpolated {$d.some_keypath} string{$d.some_keypath} will be replaced with the evaluated expression that is enclosed by the brackets.

  • this is how you escape \{ and \} and \` – The backward slash \ is used to escape characters such as { and } that are used for keypaths.

UUID Functions

isUuid

Returns true if UUID is valid.

isUuid(uuid: string): bool

randomUuid

Returns a random UUIDv4.

randomUuid(): string

General Functions

firstNonNull

Returns the first non-null value from the parameters. Works only on scalars.

firstNonNull(value: any, ...values: any): any

if

Return either the then or else according to the result of condition.

if(condition: bool, then: any, else: any?): any

in

Tests if the comparand is equal to any of the values in a set v1 ... vN.

in(comparand: any, value: any, ...values: any): bool

recordLocation

Returns the location of the record (for example, s3 URL).

recordLocation(): string

Number Functions

abs

Returns the absolute value of number.

abs(number: number): number

ceil

Rounds the value up to the nearest integer.

ceil(number: number): number

e

Returns the constant Euler’s number.

e(): number

floor

Rounds the value down to the nearest integer.

floor(number: number): number

fromBase

Returns the value of string interpreted as a base-radix number.

fromBase(string: string, radix: number): number

ln

Returns the natural log of number.

ln(number: number): number

log

Returns the log of number in base base.

log(base: number, number: number): number

log2

Returns the log of number in base 2. Equivalent to log(2, number).

log2(number: number): number

max

Returns the maximum number of all the numbers passed to the function.

max(value: number, ...values: number): number

min

Returns the least number of all the numbers passed to the function.

min(value: number, ...values: number): number

mod

Returns the modulus (remainder) of number divided by divisor.

mod(number: number, divisor: number): number

pi

Returns the constant Pi.

pi(): number

power

Returns number^exponent.

power(number: number, exponent: number): number

random

Returns a pseudo-random value in the range 0.0 <= x < 1.0.

random(): number

randomInt

Returns a pseudo-random integer number between 0 and n (exclusive).

randomInt(upperBound: number): number

round

Round number to digits decimal places.

round(number: number, digits: number?): number

sqrt

Returns square root of a number.

sqrt(number: number): number

toBase

Returns the base-radix representation of number.

toBase(number: number, radix: number): string

URL Functions

urlDecode

Unescapes the URL encoded in string. Contrast with urlEncode.

urlDecode(string: string): string

urlEncode

Escapes string by encoding it so that it can be safely included in URL. Contrast with urlDecode.

urlEncode(string: string): string

Date / Time Functions

Functions for processing timestamps, intervals and other time-related constructs.

Time Units

Many date/time functions accept a time unit argument to modify their behavior. DataPrime supports time units from nanoseconds to days. They are represented as literal strings of the time unit name in either long or short notation:

  • long notation: day, hour, minute, second, milli, micro, nano

  • short notation: d, h, m, s, ms, us, ns

Time Zones

DataPrime timestamps are always stored in the UTC time zone, but some date/time functions accept a time zone argument to modify their behavior. Time zone arguments are strings that specify a time zone offset, shorthand or identifier:

  • time zone offset in hours (for example, '+01' or '-02')
  • time zone offset in hours and minutes (for example, '+0130' or '-0230')
  • time zone offset in hours and minutes with separator (for example '+01:30' or '-02:30')
  • time zone shorthand (for example, 'UTC', 'GMT', 'EST', and so on)
  • time zone identifier (for example, 'Asia/Yerevan', 'Europe/Zurich', 'America/Winnipeg', and so on)

addInterval

Adds two intervals together. Also works with negative intervals. Equivalent to left + right.

addInterval(left: interval, right: interval): interval

addTime

Adds an interval to a timestamp. Also works with negative intervals. Equivalent to t + i.

addTime(t: timestamp, i: interval): timestamp

diffTime

Calculates the duration between two timestamps. Positive if to > from, negative if to < from. Equivalent to to - from.

diffTime(to: timestamp, from: timestamp): interval

extractTime

Extracts either a date or time unit from a timestamp. Returns a floating point number for time units less than a minute, otherwise an integer. Date units such as month or week start from 1 (not from 0).

extractTime(timestamp: timestamp, unit: dateunit | timeunit, tz: string?): number

Function parameters:

  • timestamp (required) – the timestamp to extract from.
  • unit (required) – the date or time unit to extract. Must be a string literal and one of:
    • any time unit in either long or short notation
    • a date unit in long notation: year, month, week, day_of_year, day_of_week
    • a date unit in short notation: Y, M, W, doy, dow
  • tz (optional) – a time zone to convert the timestamp before extracting.

Example 1: extract the hour in Tokyo

limit 1 | choose $m.timestamp.extractTime('h', 'Asia/Tokyo') as h # Result 1: 11pm { "h": 23 }

Example 2: extract the number of seconds

limit 1 | choose $m.timestamp.extractTime('second') as s # Result 2: 38.35 seconds { "s": 38.3510265 }

Example 3: extract the timestamp’s month

limit 1 | choose $m.timestamp.extractTime('month') as m # Result 3: August { "m": 8 }

Example 4: extract the day of the week

limit 1 | choose $m.timestamp.extractTime('dow') as d # Result 4: Tuesday { "d": 2 }

formatInterval

Formats interval to a string with an optional time unit scale.

formatInterval(interval: interval, scale: timeunit?): string

Function parameters:

  • interval (required) – the interval to format.
  • scale (optional) – the maximum time unit of the interval to show. Defaults to nano.

Example:

limit 3 | choose formatInterval(now() - $m.timestamp, 's') as i # Results: { "i": "122s261ms466us27ns" } { "i": "122s359ms197us227ns" } { "i": "122s359ms197us227ns" }

formatTimestamp

Formats a timestamp to a string with an optional format specification and destination time zone.

formatTimestamp(timestamp: timestamp, format: string?, tz: string?): string

Function parameters:

  • timestamp (required) – the timestamp to format.
  • format (optional) – a date/time format specification for parsing timestamps. Defaults to 'iso8601'. The format can be any string with embedded date/time formatters, or one of several shorthands. Here are a few samples:
    • '%Y-%m-%d' – print the date only, for example '2023-04-05'
    • '%H:%M:%S' – print the time only, for example '16:07:33'
    • '%F %H:%M:%S' – print both date and time, for example '2023-04-05 16:07:33'
    • 'iso8601' – print a timestamp in ISO 8601 format, for example '2023-04-05T16:07:33.123Z'
    • 'timestamp_milli' – print a timestamp in milliseconds (13 digits), for example '1680710853123'

tz (optional) – the destination time zone to convert the timestamp before formatting.

Example 1: print a timestamp with default format and +5h offset

limit 1 | choose $m.timestamp.formatTimestamp(tz='+05') as ts # Result 1: { "ts": "2023-08-29T19:08:37.405937400+0500" }

Example 2: print only the year and month

limit 1 | choose $m.timestamp.formatTimestamp('%Y-%m') as ym # Result 2: { "ym": "2023-08" }

Example 3: print only the hours and minutes

limit 1 | choose $m.timestamp.formatTimestamp('%H:%M') as hm # Result 3: { "hm": "14:11" }

Example 4: print a timestamp in milliseconds (13 digits)

limit 1 | choose $m.timestamp.formatTimestamp('timestamp_milli') as ms # Result 4: { "ms": "1693318678696" }

fromUnixTime

Converts a number of a specific time units since the UNIX epoch to a timestamp (in UTC). The UNIX epoch starts on January 1, 1970 – earlier timestamps are represented by negative numbers.

fromUnixTime(unixTime: number, timeUnit: timeunit?): timestamp

Function parameters:

  • unixTime (required) – the amount of time units to convert. Can be either positive or negative and will be rounded down to an integer.
  • timeUnit (optional) – the time units to convert. Defaults to 'milli'.

Example:

limit 1 | choose fromUnixTime(1658958157515, 'ms') as ts # Result: { "ts": 1658958157515000000 }

multiplyInterval

Multiplies an interval by a numeric factor. Works both with integer and fractional numbers. Equivalent to i * factor.

multiplyInterval(i: interval, factor: number): interval

now

Returns the current time at query execution time. Stable across all rows and within the entire query, even when used multiple times. Nanosecond resolution if the runtime supports it, otherwise millisecond resolution.

now(): timestamp

Example:

limit 3 | choose now() as now, now() - $m.timestamp as since # Results: { "now": 1693312549105874700, "since": "14m954ms329us764ns" } { "now": 1693312549105874700, "since": "14m954ms329us764ns" } { "now": 1693312549105874700, "since": "14m960ms519us564ns" }

parseInterval

Parses an interval from a string with format NdNhNmNsNmsNusNns where N is the amount of each time unit. Returns null when the input does not match the expected format:

  • It consists of time unit components – a non-negative integer followed by the short time unit name. Supported time units are: 'd', 'h', 'm', 's', 'ms', 'us', 'ns'.
  • There must be at least one time unit component.
  • The same time unit cannot appear more than once.
  • Components must be decreasing in time unit order – from days to nanoseconds.
  • It can start with - to represent negative intervals.
parseInterval(string: string): interval

Example 1: parse a zero interval

limit 1 | choose '0s'.parseInterval() as i # Result 1: { "i": "0ns" }

Example 2: parse a positive interval

limit 1 | choose '1d48h0m'.parseInterval() as i # Result 2: { "i": "3d" }

Example 3: parse a negative interval

limit 1 | choose '-5m45s'.parseInterval() as i # Result 3: { "i": "-5m45s" }

parseTimestamp

Parses a timestamp from string with an optional format specification and time zone override. Returns null when the input does not match the expected format.

parseTimestamp(string: string, format: string?, tz: string?): timestamp

Function parameters:

  • string (required) – the input from which the timestamp will be extracted.
  • format (optional) – a date/time format specification for parsing timestamps. Defaults to 'auto'. The format can be any string with embedded date/time extractors, one of several shorthands, or a cascade of formats to be attempted in sequence. Here are a few samples:
    • '%Y-%m-%d' – parse date only, for example '2023-04-05'
    • '%F %H:%M:%S' – parse date and time, for example '2023-04-05 16:07:33'
    • 'iso8601' – parse a timestamp in ISO 8601 format, for example '2023-04-05T16:07:33.123Z'
    • 'timestamp_milli' – parse a timestamp in milliseconds (13 digits), for example '1680710853123'
    • '%m/%d/%Y|timestamp_second' – parse either a date or a timestamp in seconds, in that order
  • tz (optional) – a time zone override to convert the timestamp while parsing. This parameter will override any time zone present in the input. A time zone can be extracted from the string by using an appropriate format and omitting this parameter.

Example 1: parse a date with the default format

limit 1 | choose '2023-04-05'.parseTimestamp() as ts # Result 1: { "ts": 1680652800000000000 }

Example 2: parse a date in US format

limit 1 | choose '04/05/23'.parseTimestamp('%D') as ts # Result 2: { "ts": 1680652800000000000 }

Example 3: parse date and time with units

limit 1 | choose '2023-04-05 16h07m'.parseTimestamp('%F %Hh%Mm') as ts # Result 3: { "ts": 1680710820000000000 }

Example 4: parse a timestamp in seconds (10 digits)

limit 1 | choose '1680710853'.parseTimestamp('timestamp_second') as ts # Result 4: { "ts": 1680710853000000000 }

roundInterval

Rounds an interval to a time unit scale. Lesser time units will be zeroed out.

roundInterval(interval: interval, scale: timeunit): interval

Function parameters:

  • interval (required) – the interval to round.
  • scale (required) – the maximium time unit of the interval to keep.

Example:

limit 1 | choose 2h5m45s.roundInterval('m') as i # Result: { "i": "2h5m" }

roundTime

Rounds a timestamp to the given interval. Useful for bucketing, for example, rounding to 1h for hourly buckets. Equivalent to date / interval.

roundTime(date: timestamp, interval: interval): timestamp

Example:

groupby $m.timestamp.roundTime(1h) as bucket count() as n # Results: { "bucket": "29/08/2023 15:00:00.000 pm", "n": 40653715 } { "bucket": "29/08/2023 14:00:00.000 pm", "n": 1779386 }

subtractInterval

Subtracts one interval from another. Equivalent to addInterval(left, -right) and left - right.

subtractInterval(left: interval, right: interval): interval

subtractTime

Subtracts an interval from a timestamp. Equivalent to addTime(t, -i) and t - i.

subtractTime(t: timestamp, i: interval): timestamp

toInterval

Converts a number of specific time units to an interval. Works with both integer and floating point and positive and negative numbers.

toInterval(number: number, timeUnit: timeunit?): interval

Function parameters:

  • number (required) – the amount of time units to convert.
  • timeUnit (optional) – Time units to convert. Defaults to nano.

Example 1: convert a floating point number

limit 1 | choose 2.5.toInterval('h') as i # Result 1: { "i": "2h30m" } # Example 2: convert an integer number limit 1 | choose -9000.toInterval() as i # Result 2: { "i": "-9us" }

toUnixTime

Converts timestamp to a number of specific time units since the UNIX epoch (in UTC). The UNIX epoch starts on January 1, 1970 – earlier timestamps are represented by negative numbers.

toUnixTime(timestamp: timestamp, timeUnit: timeunit?): number

Function parameters:

  • timestamp (required) – the timestamp to convert.
  • timeUnit (optional) – the time units to convert to. Defaults to 'milli'.

Example:

limit 1 | choose $m.timestamp.toUnixTime('hour') as hr # Result: { "hr": 470363 }

Encoding / Decoding Functions

decodeBase64

Decode a base-64 encoded string.

decodeBase64(value: string): string

encodeBase64

Encode a string into base-64.

encodeBase64(value: string): string

Case Expressions

Case expressions are special constructs in the language that allow choosing between multiple options in a readable way. They can be wherever an expression is expected.

case

Choose between multiple values based on several generic conditions. Uses a default-value if no condition is met.

case {
  condition1 -> value1,
  condition2 -> value2,
  ...
  conditionN -> valueN,
  _          -> <default-value>
}

Example:

case {
  $d.status_code == 200 -> 'success',
  $d.status_code == 201 -> 'created',
  $d.status_code == 404 -> 'not-found',
  _ -> 'other'
}

# Here's the same example inside the context of a query. A new field is created with the `case` result,
# and then a filter will be applied, leaving only non-successful responses.

source logs | ... | create $d.http_response_outcome from case {
  $d.status_code == 200 -> 'success',
  $d.status_code == 201 -> 'created',
  $d.status_code == 404 -> 'not-found',
  _                     -> 'other'
} | filter $d.http_response_outcome != 'success'

case_contains

Shorthand for case which allows checking if a string s contains one of several substrings without repeating the expression leading to s. The chosen value is the first which matches s.contains(substring).

case_contains {
  s: string,
  substring1 -> result1,
  substring2 -> result2,
  ...
  substring3 -> resultN
}

Example:

case_contains {
  $l.subsystemname,
  '-prod-' -> 'production',
  '-dev-'  -> 'development',
  '-stg-'  -> 'staging',
  _        -> 'test'
}

case_equals

Shorthand for case which allows comparing some expression e to several results without repeating the expression. The chosen value is the first which matches s == value.

case_equals {
  e: any,
  value1 -> result1,
  value2 -> result2,
  ...
  valueN -> resultN
}

Example:

case_equals {
  $m.severity,
  'info'   -> true,
  'warning -> true,
  _        -> false
}

case_greaterthan

Shorthand for case which allows comparing n to multiple values without repeating the expression leading to n. The chosen value is the first which matches expression > value.

case_greaterthan {
  n: number,
  value1: number -> result1,
  value2: number -> result2,
  ...
  valueN: number -> resultN,
  _              -> <default-value>
}

Example:

case_greaterthan {
  $d.status_code,
  500 -> 'server-error',
  400 -> 'client-error',
  300 -> 'redirection',
  200 -> 'success',
  100 -> 'information',
  _   -> 'other'
}

case_lessthan

Shorthand for case which allows comparing a number n to multiple values without repeating the expression leading to n. The chosen value is the first which matches expression < value.

case_lessthan {
  n: number,
  value1: number -> result1,
  value2: number -> result2,
  ...
  valueN: number -> resultN,
  _              -> <default-value>
}

Example:

case_lessthan {
  $d.temperature_celsius,
  10 -> 'freezing',
  20 -> 'cold',
  30 -> 'fun',
  45 -> 'hot',
  _  -> 'burning'
}

Aggregation Functions

any_value

Returns any non-null expression value in the group. If expression is not defined, it defaults to the $data object.

any_value(expression: any?)

Returns null if all expression values in the group are null.

Example:

groupby $m.severity calculate any_value($d.url)

avg

Calculates the average value of a numerical expression in the group.

avg(expression: number)

Example:

groupby $m.severity calculate avg($d.duration) as average_duration

count

Counts non-null expression values. If expression is not defined, all rows will be counted.

count(expression: any?) [into <keypath>]

An alias can be provided to override the keypath the result will be written to.

For example, the following part of a query

count() into $d.num_rows

will result in a single row of the following form:

{ "num_rows": 7532 }

count_if

Counts non-null expression values on rows which satisfy the condition. If expression is not defined, all rows that satisfy condition will be counted.

count_if(condition: bool, expression: any?)

Example:

groupby $m.severity calculate count_if($d.duration > 500) as $d.high_duration_logs
groupby $m.severity calculate count_if($d.duration > 500, $d.company_id) as $d.high_duration_logs

distinct_count

Counts non-null distinct expression values.

distinct_count(expression: any)

Example:

groupby $l.applicationname calculate distinct_count($d.username) as active_users

distinct_count_if

Counts non-null distinct expression values on rows which satisfy condition.

distinct_count_if(condition: bool, expression: any)

Example:

groupby $l.applicationname calculate distinct_count_if($m.severity == 'Error', $d.username) as users_with_errors

max

Calculates the maximum value of a numerical expression in the group.

max(expression: number | timestamp)

Example:

groupby $m.severity calculate max($d.duration)

min

Calculates the minimum value of a numerical expression in the group.

min(expression: number | timestamp)

Example:

groupby $m.severity calculate min($d.duration)

percentile

Calculates the approximate n-th percentile value of a numerical expression in the group.

percentile(percentile: number, expression: number, error_threshold: number?)

Since the percentile calculation is approximate, the accuracy may be controlled with the error_threshold parameter which ranges from 0 to 1 (defaults to 0.01). A lower value will result in better accuracy at the cost of longer query times.

Example:

groupby $m.severity calculate percentile(0.99, $d.duration) as p99_latency

sample_stddev

Computes the sample standard deviation of a numerical expression in the group.

sample_stddev(expression: number)

Example:

groupby $m.severity calculate sample_stddev($d.duration)

sample_variance

Computes the variance of a numerical expression in the group.

sample_variance(expression: number)

Example:

groupby $m.severity calculate sample_variance($d.duration)

stddev

Computes the standard deviation of a numerical expression in the group.

stddev(expression: number)

Example:

groupby $m.severity calculate stddev($d.duration)

sum

Calculates the sum of a numerical expression in the group.

sum(expression: number)

Example:

groupby $m.severity calculate sum($d.duration) as total_duration

variance

Computes the variance of a numerical expression in the group.

variance(expression: number)

Example:

groupby $m.severity calculate variance($d.duration)

DataPrime Expressions in Aggregations

When querying with the groupby operator, you can apply an aggregation function (such asavg, max, sum) to the bucket of results. This feature gives you the ability to manipulate an aggregation expression inside the expression itself, allowing you to calculate and manipulate your data simultaneously.

Example 1

This examples takes logs which have some connect_duration and batch_duration fields, and calculates the ratio between the averages of those durations, per region.

# Query
source logs
  | groupby region aggregate avg(connect_duration) / avg(batch_duration)

Example 2

This query calculates the percentage of logs which don’t have a kubernetes_pod_name out of the total number of logs. The calculation is done per subsystem.

# Query
source logs
| groupby $l.subsystemname aggregate
  sum(if(kubernetes.pod_name != null,1,0)) / count() as pct_without_pod_name

Example 3

This query calculates the ratio between the maximum and minimum salary per department, and provides a Based on N Employees string as an additional column per row.

# Query
source logs
| groupby department_id aggregate
    max(salary) / min(salary) as salary_ratio
    `Based on {count()} Employees`

Example 4

This query calculates the ratio between error logs and info logs.

source logs
| groupby $m.timestamp / 1h as hour aggregate
    count_if($m.severity == '5') / count_if($m.severity == '3') as error_to_info_ratio