Archived documentation version rendered and hosted by DevNetExpertTraining.com
Documentation

Annotated CSV

InfluxDB and Flux return query results in annotated CSV format. You can also read annotated CSV directly from Flux with the csv.from() function, write data to InfluxDB using annotated CSV and the influx write command, or upload a CSV file in the UI.

CSV tables must be encoded in UTF-8 and Unicode Normal Form C as defined in UAX15. InfluxDB removes carriage returns before newline characters.

Examples

In this topic, you’ll find examples of valid CSV syntax for responses to the following query:

from(bucket:"mydb/autogen")
    |> range(start:2018-05-08T20:50:00Z, stop:2018-05-08T20:51:00Z)
    |> group(columns:["_start","_stop", "region", "host"])
    |> yield(name:"my-result")

CSV response format

Flux supports encodings listed below.

Tables

A table may have the following rows and columns.

Rows

  • Annotation rows: describe column properties.

  • Header row: defines column labels (one header row per table).

  • Record row: describes data in the table (one record per row).

Example

Encoding of a table with and without a header row.

result,table,_start,_stop,_time,region,host,_value
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,east,A,15.43
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,east,B,59.25
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,east,C,52.62
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,east,A,15.43
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,east,B,59.25
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,east,C,52.62

Columns

In addition to the data columns, a table may include the following columns:

  • Annotation column: Only used in annotation rows. Always the first column. Displays the name of an annotation. Value can be empty or a supported annotation. You’ll notice a space for this column for the entire length of the table, so rows appear to start with ,.

  • Result column: Contains the name of the result specified by the query.

  • Table column: Contains a unique ID for each table in a result.

Multiple tables and results

If a file or data stream contains multiple tables or results, the following requirements must be met:

  • A table column indicates which table a row belongs to.
  • All rows in a table are contiguous.
  • An empty row delimits a new table boundary in the following cases:
    • Between tables in the same result that do not share a common table schema.
    • Between concatenated CSV files.
  • Each new table boundary starts with new annotation and header rows.
Example

Encoding of two tables in the same result with the same schema (header row) and different schema.

result,table,_start,_stop,_time,region,host,_value
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,east,A,15.43
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,east,B,59.25
my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,east,C,52.62
my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,west,A,62.73
my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,west,B,12.83
my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,west,C,51.62

,result,table,_start,_stop,_time,region,host,_value
,my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,east,A,15.43
,my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,east,B,59.25
,my-result,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,east,C,52.62

,result,table,_start,_stop,_time,location,device,min,max
,my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,USA,5825,62.73,68.42
,my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,USA,2175,12.83,56.12
,my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,USA,6913,51.62,54.25

Dialect options

Flux supports the following dialect options for text/csv format.

Option Description Default
header If true, the header row is included. true
delimiter Character used to delimit columns. ,
quoteChar Character used to quote values containing the delimiter. "
annotations List of annotations to encode (datatype, group, or default). empty
commentPrefix String prefix to identify a comment. Always added to annotations. #

Annotations

Annotation rows describe column properties, and start with # (or commentPrefix value). The first column in an annotation row always contains the annotation name. Subsequent columns contain annotation values as shown in the table below.

Annotation name Values Description
datatype a data type or line protocol element Describes the type of data or which line protocol element the column represents.
group boolean flag true or false Indicates the column is part of the group key.
default a value of the column’s data type Value to use for rows with an empty value.

To encode a table with its group key, the datatype, group, and default annotations must be included. If a table has no rows, the default annotation provides the group key values.

Data types

Datatype Flux type Description
boolean bool “true” or “false”
unsignedLong uint unsigned 64-bit integer
long int signed 64-bit integer
double float IEEE-754 64-bit floating-point number
string string UTF-8 encoded string
base64Binary bytes base64 encoded sequence of bytes as defined in RFC 4648
dateTime time instant in time, may be followed with a colon : and a description of the format (number, RFC3339, RFC3339Nano)
duration duration length of time represented as an unsigned 64-bit integer number of nanoseconds

Line protocol elements

The datatype annotation accepts data types and line protocol elements. Line protocol elements identify how columns are converted into line protocol when using the influx write command to write annotated CSV to InfluxDB.

Line protocol element Description
measurement column value is the measurement
field (default) column header is the field key, column value is the field value
tag column header is the tag key, column value is the tag value
time column value is the timestamp (alias for dateTime)
ignore orignored column is ignored and not included in line protocol

Mixing data types and line protocol elements

Columns with data types (other than dateTime) in the #datatype annotation are treated as fields when converted to line protocol. Columns without a specified data type default to field when converted to line protocol and column values are left unmodified in line protocol. See an example below and line protocol data types and format.

Time columns

A column with time or dateTime #datatype annotations are used as the timestamp when converted to line protocol. If there are multiple time or dateTime columns, the last column (on the right) is used as the timestamp in line protocol. Other time columns are ignored and the influx write command outputs a warning.

Time column values should be Unix timestamps (in an accepted timestamp precision), RFC3339, or RFC3339Nano.

Example line protocol elements in datatype annotation
#datatype measurement,tag,tag,field,field,ignored,time
m,cpu,host,time_steal,usage_user,nothing,time
cpu,cpu1,host1,0,2.7,a,1482669077000000000
cpu,cpu1,host2,0,2.2,b,1482669087000000000

Resulting line protocol:

cpu,cpu=cpu1,host=host1 time_steal=0,usage_user=2.7 1482669077000000000
cpu,cpu=cpu1,host=host2 time_steal=0,usage_user=2.2 1482669087000000000
Example of mixing data types line protocol elements
#datatype measurement,tag,string,double,boolean,long,unsignedLong,duration,dateTime
#default test,annotatedDatatypes,,,,,,
m,name,s,d,b,l,ul,dur,time
,,str1,1.0,true,1,1,1ms,1
,,str2,2.0,false,2,2,2us,2020-01-11T10:10:10Z

Resulting line protocol:

test,name=annotatedDatatypes s="str1",d=1,b=true,l=1i,ul=1u,dur=1000000i 1
test,name=annotatedDatatypes s="str2",d=2,b=false,l=2i,ul=2u,dur=2000i 1578737410000000000

Annotated CSV in Flux

Flux requires all annotation and header rows in annotated CSV. The example below illustrates how to use the csv.from() function to read annotated CSV in Flux:

import "csv"

csvData = "#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,dateTime:RFC3339,string,string,double,string
#group,false,false,true,true,false,true,false,false,true
#default,,,,,,,,,
,result,table,_start,_stop,_time,region,host,_value,_measurement
,,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,east,A,15.43,cpu
,,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,east,B,59.25,cpu
,,0,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,east,C,52.62,cpu
,,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,west,A,62.73,cpu
,,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,west,B,12.83,cpu
,,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,west,C,51.62,cpu
"

csv.from(csv: csvData)

Flux only supports data types in the #datatype annotation. It does does not support line protocol elements.

Errors

If an error occurs during execution, a table returns with:

  • An error column that contains an error message.
  • A reference column with a unique reference code to identify more information about the error.
  • A second row with error properties.

If an error occurs:

  • Before results materialize, the HTTP status code indicates an error. Error details are encoded in the csv table.
  • After partial results are sent to the client, the error is encoded as the next table and remaining results are discarded. In this case, the HTTP status code remains 200 OK.
Example

Encoding for an error with the datatype annotation:

#datatype,string,long
,error,reference
,Failed to parse query,897

Encoding for an error that occurs after a valid table has been encoded:

#datatype,string,long,dateTime:RFC3339,dateTime:RFC3339,dateTime:RFC3339,string,string,double
,result,table,_start,_stop,_time,region,host,_value
,my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:00Z,west,A,62.73
,my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:20Z,west,B,12.83
,my-result,1,2018-05-08T20:50:00Z,2018-05-08T20:51:00Z,2018-05-08T20:50:40Z,west,C,51.62
#datatype,string,long
,error,reference,query terminated: reached maximum allowed memory limits,576

Select your region

Upgrade to InfluxDB Cloud or InfluxDB 2.0!

InfluxDB Cloud and InfluxDB OSS 2.0 ready for production.