schema
stringclasses 471
values | key
stringlengths 0
203
| description
stringlengths 0
4.37k
| object
stringlengths 2
322k
|
---|---|---|---|
comet.json
|
domain
|
Output domain in output Area (Will be the Database name in Hive or Dataset in BigQuery)
|
{"type": "string"}
|
comet.json
|
table
|
Dataset Name in output Area (Will be the Table name in Hive & BigQuery)
|
{"type": "string"}
|
comet.json
|
partition
|
List of columns used for partitioning the output.
|
{"type": "array", "items": {"type": "string"}}
|
comet.json
|
presql
|
List of SQL requests to executed before the main SQL request is run
|
{"type": "array", "items": {"type": "string"}}
|
comet.json
|
postsql
|
List of SQL requests to executed after the main SQL request is run
|
{"type": "array", "items": {"type": "string"}}
|
comet.json
|
expectations
|
Expectations to check after Load / Transform has succeeded
|
{}
|
comet.json
|
acl
|
Map of rolename -> List[Users].
|
{"type": "array", "items": {}}
|
comet.json
|
comment
|
Output table description
|
{"type": "string"}
|
comet.json
|
freshness
|
Configure freshness checks on the output table
|
{}
|
comet.json
|
attributesDesc
|
Attributes comments and access policies
|
{"type": "array", "items": {}}
|
comet.json
|
python
|
Python script URI to execute instead of the SQL request
|
{"type": "string"}
|
comet.json
|
tags
|
Set of string to attach to the output table
|
{"type": "array", "items": {"type": "string"}}
|
comet.json
|
schedule
|
Cron expression to use for this task
|
{"type": "string"}
|
comet.json
|
dagRef
|
Cron expression to use for this domain/table
|
{"type": "string"}
|
comet.json
|
path
|
Name of the lock
|
{"type": "string"}
|
comet.json
|
timeout
|
Name of the lock
|
{"type": "integer"}
|
comet.json
|
pollTime
|
Name of the lock
|
{"type": "string"}
|
comet.json
|
refreshTime
|
Name of the lock
|
{"type": "string"}
|
comet.json
|
path
|
Main SQL request to execute (do not forget to prefix table names with the database name to avoid conflicts)
|
{"type": "string"}
|
comet.json
|
sink
|
Output Database (refer to a project id in BigQuery). Default to SL_DATABASE env var if set.
|
{}
|
comet.json
|
maxErrors
|
Output domain in output Area (Will be the Database name in Hive or Dataset in BigQuery)
|
{"type": "string"}
|
comet.json
|
database
|
Dataset Name in output Area (Will be the Table name in Hive & BigQuery)
|
{"type": "string"}
|
comet.json
|
active
|
Output table description
|
{"type": "boolean"}
|
comet.json
|
name
|
Column name
|
{"type": "string"}
|
comet.json
|
comment
|
Column description
|
{"type": "string"}
|
comet.json
|
accessPolicy
|
Access policy to apply to this column
|
{"type": "string"}
|
comet.json
|
Domain
|
A schema in JDBC database or a folder in HDFS or a dataset in BigQuery.
|
{"type": "object", "properties": {"name": {"type": "string"}, "rename": {"type": "string"}, "metadata": {}, "tables": {"type": "array", "items": {}}, "comment": {"type": "string"}, "database": {"type": "string"}, "tags": {"type": "array", "items": {"type": "string"}}}}
|
comet.json
|
name
|
Domain name. Make sure you use a name that may be used as a folder name on the target storage.
- When using HDFS or Cloud Storage, files once ingested are stored in a sub-directory named after the domain name.
- When used with BigQuery, files are ingested and sorted in tables under a dataset named after the domain name.
|
{"type": "string"}
|
comet.json
|
rename
|
If present, the attribute is renamed with this name
|
{"type": "string"}
|
comet.json
|
tables
|
List of schemas for each dataset in this domain.
A domain usually contains multiple schemas. Each schema defining how the contents of the input file should be parsed.
See Schema for more details.
|
{"type": "array", "items": {}}
|
comet.json
|
comment
|
Domain Description (free text)
|
{"type": "string"}
|
comet.json
|
database
|
Output Database (refer to a project id in BigQuery). Default to SL_DATABASE env var if set.
|
{"type": "string"}
|
comet.json
|
tags
|
Set of string to attach to this domain
|
{"type": "array", "items": {"type": "string"}}
|
comet.json
|
name
|
Optional name. If not specified, the name of the file without the extension is used.
|
{"type": "string"}
|
comet.json
|
default
|
Default task properties to apply to all tasks defined in tasks section and in included files
|
{}
|
comet.json
|
items
|
List of transform tasks to execute
|
{}
|
comet.json
|
name
|
table name. Set to '*' to extract all tables
|
{"type": "string"}
|
comet.json
|
columns
|
List of columns to extract. All columns by default.
|
{"type": "array", "items": {"type": "string"}}
|
comet.json
|
catalog
|
Optional catalog name in the source database
|
{"type": "string"}
|
comet.json
|
schema
|
Database schema where source tables are located
|
{"type": "string"}
|
comet.json
|
template
|
Metadata to use for the generated YAML file.
|
{"type": "string"}
|
comet.json
|
tables
|
List of tables to extract
|
{"type": "array", "items": {}}
|
comet.json
|
tableTypes
|
One or many of the predefined table types
|
{"type": "array", "items": {"type": "string"}}
|
comet.json
|
jdbcSchemas
|
List database connections to use to extract the data
|
{"type": "array", "items": {}}
|
comet.json
|
connectionRef
|
Connection name as defined in the connections section of the application.conf file
|
{"type": "string"}
|
comet.json
|
connection
|
JDBC connection options: url, user, password ...
|
{}
|
comet.json
|
RefInput
|
Input for ref object
|
{"type": "object", "properties": {"database": {"type": "string"}, "domain": {"type": "string"}, "table": {"type": "string"}}, "required": ["table"]}
|
comet.json
|
database
|
Database pattern to match, none if any database
|
{"type": "string"}
|
comet.json
|
domain
|
Domain pattern to match, none if any domain match
|
{"type": "string"}
|
comet.json
|
table
|
Table pattern to match
|
{"type": "string"}
|
comet.json
|
RefOutput
|
Output for ref object
|
{"type": "object", "properties": {"database": {"type": "string"}, "domain": {"type": "string"}, "table": {"type": "string"}}, "required": ["table", "domain", "database"]}
|
comet.json
|
database
|
{"type": "string"}
|
|
comet.json
|
domain
|
{"type": "string"}
|
|
comet.json
|
table
|
{"type": "string"}
|
|
comet.json
|
Ref
|
Describe how to resolve a reference in a transform task
|
{"type": "object", "properties": {"input": {}, "output": {}}, "required": ["input", "output"]}
|
comet.json
|
input
|
The input table to resolve
|
{}
|
comet.json
|
output
|
The output table resolved with the domain and database
|
{}
|
comet.json
|
env
|
Default environment to use. May be also set using the SL_ENV environment variable
|
{"type": "string"}
|
comet.json
|
datasets
|
When using filesystem storage, default path to store the datasets
|
{"type": "string"}
|
comet.json
|
dags
|
{"type": "string"}
|
|
comet.json
|
metadata
|
default metadata folder name. May be also set using the SL_METADATA environment variable
|
{"type": "string"}
|
comet.json
|
validateOnLoad
|
Validate the YAML file when loading it. If set to true fails on any error
|
{"type": "boolean"}
|
comet.json
|
archive
|
Should ingested files be archived after ingestion ?
|
{"type": "boolean"}
|
comet.json
|
sinkReplayToFile
|
Should invalid records be stored in a replay file ?
|
{"type": "boolean"}
|
comet.json
|
defaultFormat
|
Default write format in Spark. parquet is the default
|
{"type": "string"}
|
comet.json
|
defaultRejectedWriteFormat
|
Default write format in Spark for rejected records. parquet is the default
|
{"type": "string"}
|
comet.json
|
defaultAuditWriteFormat
|
Default write format in Spark for audit records. parquet is the default
|
{"type": "string"}
|
comet.json
|
csvOutput
|
output files in CSV format ? Default is false
|
{"type": "boolean"}
|
comet.json
|
csvOutputExt
|
CSV file extension when csvOutput is true. Default is .csv
|
{"type": "string"}
|
comet.json
|
privacyOnly
|
Only generate privacy tasks. Reserved for internal use
|
{"type": "boolean"}
|
comet.json
|
emptyIsNull
|
Should empty strings be considered as null values ?
|
{"type": "boolean"}
|
comet.json
|
validator
|
Default validator to use when none is specified in the schema. Valid values are 'spark' or 'native'. Default is 'spark'
|
{"type": "string"}
|
comet.json
|
rowValidatorClass
|
{"type": "string"}
|
|
comet.json
|
treeValidatorClass
|
{"type": "string"}
|
|
comet.json
|
loadStrategyClass
|
In what order should the files for a same table be loaded ? By time (default) or by or name ?
|
{"type": "string", "enum": ["ai.starlake.job.load.IngestionNameStrategy", "ai.starlake.job.load.IngestionTimeStrategy"]}
|
comet.json
|
analyze
|
Should we analyze the result and generate HIVE statistics ? (useful for Spark / Databricks)
|
{"type": "boolean"}
|
comet.json
|
hive
|
Should we create the table in Hive ? (useful for Spark / Databricks)
|
{"type": "boolean"}
|
comet.json
|
grouped
|
Should we load of the files to be stored in the same table in a single task or one by one ?
|
{"type": "boolean"}
|
comet.json
|
groupedMax
|
Maximum number of files to be stored in the same table in a single task
|
{"type": "integer"}
|
comet.json
|
mergeForceDistinct
|
Should we force a distinct on the merge ?
|
{"type": "boolean"}
|
comet.json
|
mergeOptimizePartitionWrite
|
Should we optimize the partition write on the merge ?
|
{"type": "boolean"}
|
comet.json
|
area
|
pending, accepted ... areas configuration
|
{}
|
comet.json
|
airflow
|
Airflow endpoint configuration when using Airflow Launcher
|
{}
|
comet.json
|
hadoop
|
Hadoop configuration if applicable
|
{}
|
comet.json
|
connections
|
Connection configurations
|
{}
|
comet.json
|
jdbcEngines
|
JDBC engine configurations
|
{}
|
comet.json
|
privacy
|
Privacy algorithms
|
{}
|
comet.json
|
root
|
Root folder for the application. May be also set using the SL_ROOT environment variable
|
{"type": "string"}
|
comet.json
|
internal
|
Internal configuration
|
{}
|
comet.json
|
accessPolicies
|
Access policies configuration
|
{}
|
comet.json
|
scheduling
|
Spark Job scheduling configuration
|
{}
|
comet.json
|
udfs
|
Coma separated list of UDF to register in Spark jobs. May be also set using the SL_UDFS environment variable
|
{"type": "string"}
|
comet.json
|
expectations
|
Expectations configuration
|
{}
|
comet.json
|
sqlParameterPattern
|
Pattern to use to replace parameters in SQL queries in addition to the jinja syntax {{param}}. Default is ${param}
|
{"type": "string"}
|
comet.json
|
rejectAllOnError
|
Should we reject all records when an error occurs ? Default is false
|
{"type": "string"}
|
comet.json
|
rejectMaxRecords
|
Maximum number of records to reject when an error occurs. Default is 100
|
{"type": "integer"}
|
comet.json
|
maxParCopy
|
{"type": "integer"}
|
|
comet.json
|
dsvOptions
|
DSV ingestion extra options
|
{}
|
comet.json
|
forceViewPattern
|
reserved
|
{"type": "string"}
|
comet.json
|
forceDomainPattern
|
reserved
|
{"type": "string"}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.