commit_message
stringlengths
3
2.32k
diff
stringlengths
186
49.5k
concern_count
int64
1
5
shas
stringlengths
44
220
types
stringlengths
6
45
support custom style by class for mini-map affects: @logicflow/extension,add gitignore.nix to dep update matrix
["diff --git a/packages/extension/src/components/mini-map/index.ts b/packages/extension/src/components/mini-map/index.ts\nindex 35cd047..ad5194d 100644\n--- a/packages/extension/src/components/mini-map/index.ts\n+++ b/packages/extension/src/components/mini-map/index.ts\n@@ -2,7 +2,7 @@ import { Extension } from '@logicflow/core';\n \n interface MiniMapPlugin extends Extension {\n init: (option) => void;\n- show: (leftPosition, topPosition) => void;\n+ show: (leftPosition?: number, topPosition?: number) => void;\n hide: () => void;\n [x: string]: any;\n }\n@@ -96,12 +96,13 @@ const MiniMap: MiniMapPlugin = {\n const miniMapContainer = document.createElement('div');\n const miniMapWrap = MiniMap.__miniMapWrap;\n miniMapContainer.appendChild(miniMapWrap);\n- miniMapContainer.style.left = `${left}px`;\n- miniMapContainer.style.top = `${top}px`;\n+ if (typeof left !== 'undefined' && typeof top !== 'undefined') {\n+ miniMapContainer.style.left = `${left}px`;\n+ miniMapContainer.style.top = `${top}px`;\n+ }\n miniMapContainer.style.position = 'absolute';\n miniMapContainer.className = 'lf-mini-map';\n MiniMap.__container.appendChild(miniMapContainer);\n- \n MiniMap.__miniMapWrap.appendChild(MiniMap.__viewport);\n \n const header = document.createElement('div');\n", "diff --git a/.github/workflows/update-deps.yml b/.github/workflows/update-deps.yml\nindex e0ccd62..1236f58 100644\n--- a/.github/workflows/update-deps.yml\n+++ b/.github/workflows/update-deps.yml\n@@ -13,6 +13,7 @@ jobs:\n - nixpkgs\n - poetry2nix\n - pre-commit-hooks\n+ - gitignore.nix\n steps:\n - name: Checkout\n uses: actions/checkout@v2\n"]
2
["4c97625774f65ed3d59caefc5c691fabf0adc499", "c444fdb9e85ce44c5c0c99addc777dd7b6085153"]
["feat", "ci"]
export order,apply element migrated events This is a very straightforward event applier. All it needs to do is update the persisted data for the element instance using the data in the event.
["diff --git a/src/Object/_api.ts b/src/Object/_api.ts\nindex a1e5c8c..70e9fea 100644\n--- a/src/Object/_api.ts\n+++ b/src/Object/_api.ts\n@@ -38,9 +38,9 @@ export {Optional} from './Optional'\n export {OptionalKeys} from './OptionalKeys'\n export {Overwrite} from './Overwrite'\n export {Partial} from './Partial'\n+export {Path} from './Path'\n export {Paths} from './Paths'\n export {PathValid} from './PathValid'\n-export {Path} from './Path'\n export {Pick} from './Pick'\n export {Readonly} from './Readonly'\n export {ReadonlyKeys} from './ReadonlyKeys'\n", "diff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\nindex da05e13..9231df3 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n@@ -154,6 +154,9 @@ public final class EventAppliers implements EventApplier {\n register(\n ProcessInstanceIntent.SEQUENCE_FLOW_TAKEN,\n new ProcessInstanceSequenceFlowTakenApplier(elementInstanceState, processState));\n+ register(\n+ ProcessInstanceIntent.ELEMENT_MIGRATED,\n+ new ProcessInstanceElementMigratedApplier(elementInstanceState));\n }\n \n private void registerProcessInstanceCreationAppliers(final MutableProcessingState state) {\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\nindex e5a0f3a..d38358f 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\n@@ -24,5 +24,16 @@ final class ProcessInstanceElementMigratedApplier\n }\n \n @Override\n- public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {}\n+ public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {\n+ elementInstanceState.updateInstance(\n+ elementInstanceKey,\n+ elementInstance ->\n+ elementInstance\n+ .getValue()\n+ .setProcessDefinitionKey(value.getProcessDefinitionKey())\n+ .setBpmnProcessId(value.getBpmnProcessId())\n+ .setVersion(value.getVersion())\n+ .setElementId(value.getElementId())\n+ .setFlowScopeKey(value.getFlowScopeKey()));\n+ }\n }\n"]
2
["879edb6ed90f88b9ae6a3c2e8878ae1be48e0c88", "39d5d1cfe8d2210305df2c8fab4a4ae430732cf7"]
["refactor", "feat"]
remove unnecessary lines from verify-wal test,exclude github.io from link checking to avoid rate limiting
["diff --git a/storage/wal/verifier_test.go b/storage/wal/verifier_test.go\nindex 61e1536..a44755f 100644\n--- a/storage/wal/verifier_test.go\n+++ b/storage/wal/verifier_test.go\n@@ -138,22 +138,13 @@ func writeCorruptEntries(file *os.File, t *testing.T, n int) {\n \t\t}\n \t}\n \n-\n \t// Write some random bytes to the file to simulate corruption.\n \tif _, err := file.Write(corruption); err != nil {\n \t\tfatal(t, \"corrupt WAL segment\", err)\n \t}\n-\tcorrupt := []byte{1, 255, 0, 3, 45, 26, 110}\n-\n-\twrote, err := file.Write(corrupt)\n-\tif err != nil {\n-\t\tt.Fatal(err)\n-\t} else if wrote != len(corrupt) {\n-\t\tt.Fatal(\"Error writing corrupt data to file\")\n-\t}\n \n \tif err := file.Close(); err != nil {\n-\t\tt.Fatalf(\"Error: filed to close file: %v\\n\", err)\n+\t\tt.Fatalf(\"Error: failed to close file: %v\\n\", err)\n \t}\n }\n \n", "diff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\nindex 90c5a27..db6457b 100644\n--- a/.github/workflows/ibis-docs-lint.yml\n+++ b/.github/workflows/ibis-docs-lint.yml\n@@ -101,6 +101,7 @@ jobs:\n --exclude-mail \\\n --exclude fonts.gstatic.com \\\n --exclude github.com \\\n+ --exclude github.io \\\n --no-progress \\\n --github-token ${{ steps.generate_token.outputs.token }}\n \n"]
2
["fba4326c72fc22d81aba6976a9fef1e4b6154fd9", "ce0539a32b927a3559feebf8f5307e3863e992a1"]
["refactor", "ci"]
simplify loadFiles code,deploy dmn using java client This test is an acceptance test that verifies that the java client can deploy a dmn decision model using the newDeployCommand client method. It verifies that the model was resource was parsed and deployed, resulting in a response that contains metadata of the deployed decision requirements graph and the decisions it contains.
["diff --git a/frontend/app/player/web/network/loadFiles.ts b/frontend/app/player/web/network/loadFiles.ts\nindex ec174fc..d164333 100644\n--- a/frontend/app/player/web/network/loadFiles.ts\n+++ b/frontend/app/player/web/network/loadFiles.ts\n@@ -1,43 +1,33 @@\n import APIClient from 'App/api_client';\n \n-const NO_NTH_FILE = \"nnf\"\n-const NO_UNPROCESSED_FILES = \"nuf\"\n+const NO_FILE_OK = \"No-file-but-this-is-ok\"\n+const NO_BACKUP_FILE = \"No-efs-file\"\n \n export const loadFiles = (\n urls: string[],\n onData: (data: Uint8Array) => void,\n ): Promise<void> => {\n- const firstFileURL = urls[0]\n- urls = urls.slice(1)\n- if (!firstFileURL) {\n+ if (!urls.length) {\n return Promise.reject(\"No urls provided\")\n }\n- return window.fetch(firstFileURL)\n- .then(r => {\n- return processAPIStreamResponse(r, true)\n- })\n- .then(onData)\n- .then(() =>\n- urls.reduce((p, url) =>\n- p.then(() =>\n- window.fetch(url)\n- .then(r => {\n- return processAPIStreamResponse(r, false)\n- })\n- .then(onData)\n- ),\n- Promise.resolve(),\n- )\n+ return urls.reduce((p, url, index) =>\n+ p.then(() =>\n+ window.fetch(url)\n+ .then(r => {\n+ return processAPIStreamResponse(r, index===0)\n+ })\n+ .then(onData)\n+ ),\n+ Promise.resolve(),\n )\n .catch(e => {\n- if (e === NO_NTH_FILE) {\n+ if (e === NO_FILE_OK) {\n return\n }\n throw e\n })\n }\n \n-\n export async function requestEFSDom(sessionId: string) {\n return await requestEFSMobFile(sessionId + \"/dom.mob\")\n }\n@@ -50,21 +40,18 @@ async function requestEFSMobFile(filename: string) {\n const api = new APIClient()\n const res = await api.fetch('/unprocessed/' + filename)\n if (res.status >= 400) {\n- throw NO_UNPROCESSED_FILES\n+ throw NO_BACKUP_FILE\n }\n return await processAPIStreamResponse(res, false)\n }\n \n-const processAPIStreamResponse = (response: Response, isFirstFile: boolean) => {\n+const processAPIStreamResponse = (response: Response, canBeMissed: boolean) => {\n return new Promise<ArrayBuffer>((res, rej) => {\n- if (response.status === 404 && !isFirstFile) {\n- return rej(NO_NTH_FILE)\n+ if (response.status === 404 && canBeMissed) {\n+ return rej(NO_FILE_OK)\n }\n if (response.status >= 400) {\n- return rej(\n- isFirstFile ? `no start file. status code ${ response.status }`\n- : `Bad endfile status code ${response.status}`\n- )\n+ return rej(`Bad file status code ${response.status}. Url: ${response.url}`)\n }\n res(response.arrayBuffer())\n }).then(buffer => new Uint8Array(buffer))\n", "diff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java\nindex f36465b..6b6ab48 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java\n@@ -67,6 +67,49 @@ public final class CreateDeploymentTest {\n }\n \n @Test\n+ public void shouldDeployDecisionModel() {\n+ // given\n+ final String resourceName = \"dmn/drg-force-user.dmn\";\n+\n+ // when\n+ final DeploymentEvent result =\n+ CLIENT_RULE\n+ .getClient()\n+ .newDeployCommand()\n+ .addResourceFromClasspath(resourceName)\n+ .send()\n+ .join();\n+\n+ // then\n+ assertThat(result.getKey()).isPositive();\n+ assertThat(result.getDecisionRequirements()).hasSize(1);\n+ assertThat(result.getDecisions()).hasSize(2);\n+\n+ final var decisionRequirements = result.getDecisionRequirements().get(0);\n+ assertThat(decisionRequirements.getDmnDecisionRequirementsId()).isEqualTo(\"force_users\");\n+ assertThat(decisionRequirements.getDmnDecisionRequirementsName()).isEqualTo(\"Force Users\");\n+ assertThat(decisionRequirements.getVersion()).isEqualTo(1);\n+ assertThat(decisionRequirements.getDecisionRequirementsKey()).isPositive();\n+ assertThat(decisionRequirements.getResourceName()).isEqualTo(resourceName);\n+\n+ final var decision1 = result.getDecisions().get(0);\n+ assertThat(decision1.getDmnDecisionId()).isEqualTo(\"jedi_or_sith\");\n+ assertThat(decision1.getDmnDecisionName()).isEqualTo(\"Jedi or Sith\");\n+ assertThat(decision1.getVersion()).isEqualTo(1);\n+ assertThat(decision1.getDecisionKey()).isPositive();\n+ assertThat(decision1.getDmnDecisionRequirementsId()).isEqualTo(\"force_users\");\n+ assertThat(decision1.getDecisionRequirementsKey()).isPositive();\n+\n+ final var decision2 = result.getDecisions().get(1);\n+ assertThat(decision2.getDmnDecisionId()).isEqualTo(\"force_user\");\n+ assertThat(decision2.getDmnDecisionName()).isEqualTo(\"Which force user?\");\n+ assertThat(decision2.getVersion()).isEqualTo(1);\n+ assertThat(decision2.getDecisionKey()).isPositive();\n+ assertThat(decision2.getDmnDecisionRequirementsId()).isEqualTo(\"force_users\");\n+ assertThat(decision2.getDecisionRequirementsKey()).isPositive();\n+ }\n+\n+ @Test\n public void shouldRejectDeployIfProcessIsInvalid() {\n // given\n final BpmnModelInstance process =\ndiff --git a/qa/integration-tests/src/test/resources/dmn/drg-force-user.dmn b/qa/integration-tests/src/test/resources/dmn/drg-force-user.dmn\nnew file mode 100644\nindex 0000000..8d55c55\n--- /dev/null\n+++ b/qa/integration-tests/src/test/resources/dmn/drg-force-user.dmn\n@@ -0,0 +1,144 @@\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n+<definitions xmlns=\"https://www.omg.org/spec/DMN/20191111/MODEL/\" xmlns:dmndi=\"https://www.omg.org/spec/DMN/20191111/DMNDI/\" xmlns:dc=\"http://www.omg.org/spec/DMN/20180521/DC/\" xmlns:biodi=\"http://bpmn.io/schema/dmn/biodi/2.0\" xmlns:di=\"http://www.omg.org/spec/DMN/20180521/DI/\" id=\"force_users\" name=\"Force Users\" namespace=\"http://camunda.org/schema/1.0/dmn\" exporter=\"Camunda Modeler\" exporterVersion=\"4.12.0\">\n+ <decision id=\"jedi_or_sith\" name=\"Jedi or Sith\">\n+ <decisionTable id=\"DecisionTable_14n3bxx\">\n+ <input id=\"Input_1\" label=\"Lightsaber color\" biodi:width=\"192\">\n+ <inputExpression id=\"InputExpression_1\" typeRef=\"string\">\n+ <text>lightsaberColor</text>\n+ </inputExpression>\n+ </input>\n+ <output id=\"Output_1\" label=\"Jedi or Sith\" name=\"jedi_or_sith\" typeRef=\"string\" biodi:width=\"192\">\n+ <outputValues id=\"UnaryTests_0hj346a\">\n+ <text>\"Jedi\",\"Sith\"</text>\n+ </outputValues>\n+ </output>\n+ <rule id=\"DecisionRule_0zumznl\">\n+ <inputEntry id=\"UnaryTests_0leuxqi\">\n+ <text>\"blue\"</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0c9vpz8\">\n+ <text>\"Jedi\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_1utwb1e\">\n+ <inputEntry id=\"UnaryTests_1v3sd4m\">\n+ <text>\"green\"</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0tgh8k1\">\n+ <text>\"Jedi\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_1bwgcym\">\n+ <inputEntry id=\"UnaryTests_0n1ewm3\">\n+ <text>\"red\"</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_19xnlkw\">\n+ <text>\"Sith\"</text>\n+ </outputEntry>\n+ </rule>\n+ </decisionTable>\n+ </decision>\n+ <decision id=\"force_user\" name=\"Which force user?\">\n+ <informationRequirement id=\"InformationRequirement_1o8esai\">\n+ <requiredDecision href=\"#jedi_or_sith\" />\n+ </informationRequirement>\n+ <decisionTable id=\"DecisionTable_07g94t1\" hitPolicy=\"FIRST\">\n+ <input id=\"InputClause_0qnqj25\" label=\"Jedi or Sith\">\n+ <inputExpression id=\"LiteralExpression_00lcyt5\" typeRef=\"string\">\n+ <text>jedi_or_sith</text>\n+ </inputExpression>\n+ <inputValues id=\"UnaryTests_1xjidd8\">\n+ <text>\"Jedi\",\"Sith\"</text>\n+ </inputValues>\n+ </input>\n+ <input id=\"InputClause_0k64hys\" label=\"Body height\">\n+ <inputExpression id=\"LiteralExpression_0ib6fnk\" typeRef=\"number\">\n+ <text>height</text>\n+ </inputExpression>\n+ </input>\n+ <output id=\"OutputClause_0hhe1yo\" label=\"Force user\" name=\"force_user\" typeRef=\"string\" />\n+ <rule id=\"DecisionRule_13zidc5\">\n+ <inputEntry id=\"UnaryTests_056skcq\">\n+ <text>\"Jedi\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_0l4xksq\">\n+ <text>&gt; 190</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0hclhw3\">\n+ <text>\"Mace Windu\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_0uin2hk\">\n+ <description></description>\n+ <inputEntry id=\"UnaryTests_16maepk\">\n+ <text>\"Jedi\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_0rv0nwf\">\n+ <text>&gt; 180</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0t82c11\">\n+ <text>\"Obi-Wan Kenobi\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_0mpio0p\">\n+ <inputEntry id=\"UnaryTests_09eicyc\">\n+ <text>\"Jedi\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_1bekl8k\">\n+ <text>&lt; 70</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0brx3vt\">\n+ <text>\"Yoda\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_06paffx\">\n+ <inputEntry id=\"UnaryTests_1baiid4\">\n+ <text>\"Sith\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_0fcdq0i\">\n+ <text>&gt; 200</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_02oibi4\">\n+ <text>\"Darth Vader\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_1ua4pcl\">\n+ <inputEntry id=\"UnaryTests_1s1h3nm\">\n+ <text>\"Sith\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_1pnvw8p\">\n+ <text>&gt; 170</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_1w1n2rc\">\n+ <text>\"Darth Sidius\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_00ew25e\">\n+ <inputEntry id=\"UnaryTests_07uxyug\">\n+ <text></text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_1he6fym\">\n+ <text></text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_07i3sc8\">\n+ <text>\"unknown\"</text>\n+ </outputEntry>\n+ </rule>\n+ </decisionTable>\n+ </decision>\n+ <dmndi:DMNDI>\n+ <dmndi:DMNDiagram>\n+ <dmndi:DMNShape dmnElementRef=\"jedi_or_sith\">\n+ <dc:Bounds height=\"80\" width=\"180\" x=\"160\" y=\"280\" />\n+ </dmndi:DMNShape>\n+ <dmndi:DMNShape id=\"DMNShape_1sb3tre\" dmnElementRef=\"force_user\">\n+ <dc:Bounds height=\"80\" width=\"180\" x=\"280\" y=\"80\" />\n+ </dmndi:DMNShape>\n+ <dmndi:DMNEdge id=\"DMNEdge_0gt1p1u\" dmnElementRef=\"InformationRequirement_1o8esai\">\n+ <di:waypoint x=\"250\" y=\"280\" />\n+ <di:waypoint x=\"370\" y=\"180\" />\n+ <di:waypoint x=\"370\" y=\"160\" />\n+ </dmndi:DMNEdge>\n+ </dmndi:DMNDiagram>\n+ </dmndi:DMNDI>\n+</definitions>\n"]
2
["983fef55ef08ca2ca25349bb2d5bdff10ecf89f4", "73eac947689e3fc6b53bf626a6b4604056166d6e"]
["refactor", "test"]
set cursor position in setHorizontalRule correctly, fix #2429,add gitignore.nix to dep update matrix
["diff --git a/packages/extension-horizontal-rule/src/horizontal-rule.ts b/packages/extension-horizontal-rule/src/horizontal-rule.ts\nindex 6f583e1..c905b63 100644\n--- a/packages/extension-horizontal-rule/src/horizontal-rule.ts\n+++ b/packages/extension-horizontal-rule/src/horizontal-rule.ts\n@@ -49,15 +49,14 @@ export const HorizontalRule = Node.create<HorizontalRuleOptions>({\n // set cursor after horizontal rule\n .command(({ tr, dispatch }) => {\n if (dispatch) {\n- const { parent, pos } = tr.selection.$from\n- const posAfter = pos + 1\n- const nodeAfter = tr.doc.nodeAt(posAfter)\n+ const { $to } = tr.selection\n+ const posAfter = $to.end()\n \n- if (nodeAfter) {\n- tr.setSelection(TextSelection.create(tr.doc, posAfter))\n+ if ($to.nodeAfter) {\n+ tr.setSelection(TextSelection.create(tr.doc, $to.pos))\n } else {\n // add node after horizontal rule if it\u2019s the end of the document\n- const node = parent.type.contentMatch.defaultType?.create()\n+ const node = $to.parent.type.contentMatch.defaultType?.create()\n \n if (node) {\n tr.insert(posAfter, node)\n", "diff --git a/.github/workflows/update-deps.yml b/.github/workflows/update-deps.yml\nindex e0ccd62..1236f58 100644\n--- a/.github/workflows/update-deps.yml\n+++ b/.github/workflows/update-deps.yml\n@@ -13,6 +13,7 @@ jobs:\n - nixpkgs\n - poetry2nix\n - pre-commit-hooks\n+ - gitignore.nix\n steps:\n - name: Checkout\n uses: actions/checkout@v2\n"]
2
["34d80114704679118e9bb6058e0d6c7aa03fd4b5", "c444fdb9e85ce44c5c0c99addc777dd7b6085153"]
["fix", "ci"]
typo,svg helper
["diff --git a/README.md b/README.md\nindex b823d75..53f265d 100644\n--- a/README.md\n+++ b/README.md\n@@ -74,7 +74,7 @@ With oclif you can create 2 different CLI types, single and multi.\n \n Single CLIs are like `ls` or `cat`. They can accept arguments and flags. Single CLIs can optionally be just be a single file.\n \n-Multi CLIs are like `git` or `heroku`. They have subcommands that are themselves single CLIs commands. In the `package.json` there is a field `oclif.commands` that points to a directory. This directory contains all the subcommands for the CLI. For example, if you wanted a CLI called `mycli` with the commands `mycli create` and `mycli destroy`, you would have a project like the following:\n+Multi CLIs are like `git` or `heroku`. They have subcommands that are themselves single CLI commands. In the `package.json` there is a field `oclif.commands` that points to a directory. This directory contains all the subcommands for the CLI. For example, if you wanted a CLI called `mycli` with the commands `mycli create` and `mycli destroy`, you would have a project like the following:\n \n ```\n package.json\n", "diff --git a/src/__test__/m.spec.ts b/src/__test__/m.spec.ts\nindex 0f9cb63..ff5e5f0 100644\n--- a/src/__test__/m.spec.ts\n+++ b/src/__test__/m.spec.ts\n@@ -1,4 +1,4 @@\n-import { className, m, ns, style } from '../m';\n+import { className, m, ns, svg, style } from '../m';\n import { VNode, VProps } from '../structs';\n \n const h = (tag: string, props?: VProps, ...children: VNode[]) =>\n@@ -173,6 +173,28 @@ describe('.m', () => {\n });\n });\n \n+ it('should attach ns to props using svg helper', () => {\n+ const vnode = {\n+ tag: 'svg',\n+ children: [\n+ 'foo',\n+ {\n+ tag: 'div',\n+ },\n+ ],\n+ };\n+ expect(svg(vnode)).toEqual({\n+ tag: 'svg',\n+ props: { ns: 'http://www.w3.org/2000/svg' },\n+ children: [\n+ 'foo',\n+ {\n+ tag: 'div',\n+ },\n+ ],\n+ });\n+ });\n+\n it('should move key to distinct property', () => {\n expect(h('div', { key: 'foo' }, 'foo', h('div'))).toEqual({\n tag: 'div',\n"]
2
["06c12fb603e3a38eca0340a92719ee59d34a9f47", "4aa3e4c438742ef0fe694ffaf6a181874366d777"]
["docs", "test"]
Improved Config Loading #423,update version (nightly.0)
["diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex 76dd749..2087803 100644\n--- a/CHANGELOG.md\n+++ b/CHANGELOG.md\n@@ -7,6 +7,7 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/)\n ## [Unreleased]\n * Fix the tab '(Sync)' suffix in named tabs (https://github.com/zellij-org/zellij/pull/410)\n * Improve performance when multiple panes are open (https://github.com/zellij-org/zellij/pull/318)\n+* Improve error reporting and tests of configuration (https://github.com/zellij-org/zellij/pull/423)\n \n ## [0.6.0] - 2021-04-29\n * Doesn't quit anymore on single `q` press while in tab mode (https://github.com/zellij-org/zellij/pull/342)\n", "diff --git a/Cargo.lock b/Cargo.lock\nindex f949506..6a10219 100644\n--- a/Cargo.lock\n+++ b/Cargo.lock\n@@ -94,7 +94,7 @@ dependencies = [\n \n [[package]]\n name = \"els\"\n-version = \"0.1.22\"\n+version = \"0.1.23-nightly.0\"\n dependencies = [\n \"erg_common\",\n \"erg_compiler\",\n@@ -105,7 +105,7 @@ dependencies = [\n \n [[package]]\n name = \"erg\"\n-version = \"0.6.10\"\n+version = \"0.6.11-nightly.0\"\n dependencies = [\n \"els\",\n \"erg_common\",\n@@ -115,7 +115,7 @@ dependencies = [\n \n [[package]]\n name = \"erg_common\"\n-version = \"0.6.10\"\n+version = \"0.6.11-nightly.0\"\n dependencies = [\n \"backtrace-on-stack-overflow\",\n \"crossterm\",\n@@ -126,7 +126,7 @@ dependencies = [\n \n [[package]]\n name = \"erg_compiler\"\n-version = \"0.6.10\"\n+version = \"0.6.11-nightly.0\"\n dependencies = [\n \"erg_common\",\n \"erg_parser\",\n@@ -134,7 +134,7 @@ dependencies = [\n \n [[package]]\n name = \"erg_parser\"\n-version = \"0.6.10\"\n+version = \"0.6.11-nightly.0\"\n dependencies = [\n \"erg_common\",\n \"unicode-xid\",\ndiff --git a/Cargo.toml b/Cargo.toml\nindex 04fdad7..ecc45e5 100644\n--- a/Cargo.toml\n+++ b/Cargo.toml\n@@ -20,7 +20,7 @@ members = [\n ]\n \n [workspace.package]\n-version = \"0.6.10\"\n+version = \"0.6.11-nightly.0\"\n authors = [\"erg-lang team <[email protected]>\"]\n license = \"MIT OR Apache-2.0\"\n edition = \"2021\"\n@@ -64,10 +64,10 @@ full-repl = [\"erg_common/full-repl\"]\n full = [\"els\", \"full-repl\", \"unicode\", \"pretty\"]\n \n [workspace.dependencies]\n-erg_common = { version = \"0.6.10\", path = \"./crates/erg_common\" }\n-erg_parser = { version = \"0.6.10\", path = \"./crates/erg_parser\" }\n-erg_compiler = { version = \"0.6.10\", path = \"./crates/erg_compiler\" }\n-els = { version = \"0.1.22\", path = \"./crates/els\" }\n+erg_common = { version = \"0.6.11-nightly.0\", path = \"./crates/erg_common\" }\n+erg_parser = { version = \"0.6.11-nightly.0\", path = \"./crates/erg_parser\" }\n+erg_compiler = { version = \"0.6.11-nightly.0\", path = \"./crates/erg_compiler\" }\n+els = { version = \"0.1.23-nightly.0\", path = \"./crates/els\" }\n \n [dependencies]\n erg_common = { workspace = true }\ndiff --git a/crates/els/Cargo.toml b/crates/els/Cargo.toml\nindex bc031e6..7c9455f 100644\n--- a/crates/els/Cargo.toml\n+++ b/crates/els/Cargo.toml\n@@ -2,7 +2,7 @@\n name = \"els\"\n description = \"An Erg compiler frontend for IDEs, implements LSP.\"\n documentation = \"http://docs.rs/els\"\n-version = \"0.1.22\"\n+version = \"0.1.23-nightly.0\"\n authors.workspace = true\n license.workspace = true\n edition.workspace = true\n"]
2
["099861ff5b0f83773ca0af4c70e6e39be3b0336c", "607ecc92b5f8c084304e406eec725b7dcfa0a562"]
["docs", "build"]
Fix Cube Store build on Windows,path correction Signed-off-by: Pranav C <[email protected]>
["diff --git a/.github/workflows/publish.yml b/.github/workflows/publish.yml\nindex a986501..db33097 100644\n--- a/.github/workflows/publish.yml\n+++ b/.github/workflows/publish.yml\n@@ -409,7 +409,7 @@ jobs:\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\n - name: Instal LLVM for Windows\n if: ${{ startsWith(matrix.os, 'windows') }}\n- run: choco install -y llvm --version 9.0.1\n+ run: choco install -y --force llvm --version 9.0.1\n - name: Set Env Variables for Windows\n uses: allenevans/[email protected]\n if: ${{ startsWith(matrix.os, 'windows') }}\ndiff --git a/.github/workflows/rust.yml b/.github/workflows/rust.yml\nindex d45cdf9..8d87ed6 100644\n--- a/.github/workflows/rust.yml\n+++ b/.github/workflows/rust.yml\n@@ -158,7 +158,7 @@ jobs:\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\n - name: Instal LLVM for Windows\n if: ${{ startsWith(matrix.os, 'windows') }}\n- run: choco install -y llvm --version 9.0.1\n+ run: choco install -y --force llvm --version 9.0.1\n - name: Set Env Variables for Windows\n uses: allenevans/[email protected]\n if: ${{ startsWith(matrix.os, 'windows') }}\n", "diff --git a/packages/nocodb-nest/src/modules/test/TestResetService/resetMetaSakilaSqliteProject.ts b/packages/nocodb-nest/src/modules/test/TestResetService/resetMetaSakilaSqliteProject.ts\nindex 3afce9b..8425b00 100644\n--- a/packages/nocodb-nest/src/modules/test/TestResetService/resetMetaSakilaSqliteProject.ts\n+++ b/packages/nocodb-nest/src/modules/test/TestResetService/resetMetaSakilaSqliteProject.ts\n@@ -1,11 +1,9 @@\n import { promises as fs } from 'fs';\n import axios from 'axios';\n+import path from 'path'\n \n const sqliteFilePath = (parallelId: string) => {\n- const rootDir = __dirname.replace(\n- '/src/modules/test/TestResetService',\n- '',\n- );\n+ const rootDir = process.cwd()\n \n return `${rootDir}/test_sakila_${parallelId}.db`;\n };\n@@ -78,10 +76,10 @@ const deleteSqliteFileIfExists = async (parallelId: string) => {\n };\n \n const seedSakilaSqliteFile = async (parallelId: string) => {\n- const testsDir = __dirname.replace(\n- '/src/modules/test/TestResetService',\n- '/tests',\n- );\n+ const testsDir = path.join(\n+ process.cwd(),\n+ 'tests'\n+ );;\n \n await fs.copyFile(\n `${testsDir}/sqlite-sakila-db/sakila.db`,\ndiff --git a/packages/nocodb-nest/src/modules/test/TestResetService/resetMysqlSakilaProject.ts b/packages/nocodb-nest/src/modules/test/TestResetService/resetMysqlSakilaProject.ts\nindex 6bcd3f1..e4ed112 100644\n--- a/packages/nocodb-nest/src/modules/test/TestResetService/resetMysqlSakilaProject.ts\n+++ b/packages/nocodb-nest/src/modules/test/TestResetService/resetMysqlSakilaProject.ts\n@@ -1,4 +1,5 @@\n import { promises as fs } from 'fs';\n+import path from 'path';\n import axios from 'axios';\n import { knex } from 'knex';\n import Audit from '../../../models/Audit';\n@@ -85,10 +86,7 @@ const resetSakilaMysql = async (\n parallelId: string,\n isEmptyProject: boolean,\n ) => {\n- const testsDir = __dirname.replace(\n- '/src/modules/test/TestResetService',\n- '/tests',\n- );\n+ const testsDir = path.join(process.cwd(), '/tests');\n \n try {\n await knex.raw(`DROP DATABASE test_sakila_${parallelId}`);\ndiff --git a/packages/nocodb-nest/src/modules/test/TestResetService/resetPgSakilaProject.ts b/packages/nocodb-nest/src/modules/test/TestResetService/resetPgSakilaProject.ts\nindex 1a042c3..73923ef 100644\n--- a/packages/nocodb-nest/src/modules/test/TestResetService/resetPgSakilaProject.ts\n+++ b/packages/nocodb-nest/src/modules/test/TestResetService/resetPgSakilaProject.ts\n@@ -1,6 +1,7 @@\n import { promises as fs } from 'fs';\n import axios from 'axios';\n import { knex } from 'knex';\n+import path from 'path'\n import Audit from '../../../models/Audit';\n import type Project from '../../../models/Project';\n \n@@ -78,8 +79,8 @@ const isSakilaPgToBeReset = async (parallelId: string, project?: Project) => {\n };\n \n const resetSakilaPg = async (parallelId: string, isEmptyProject: boolean) => {\n- const testsDir = __dirname.replace(\n- '/src/modules/test/TestResetService',\n+ const testsDir = path.join(\n+ process.cwd(),\n '/tests',\n );\n \n"]
2
["e34bb04baac7574e67bc566d13dea72092e0cfa3", "974e033a3ca7484290a04201ee33856a25da0942"]
["ci", "fix"]
missing transformation for T,make jq use compact json for rebase branch query
["diff --git a/src/Tuple/Merge.ts b/src/Tuple/Merge.ts\nindex dfa7ce5..5ba44b7 100644\n--- a/src/Tuple/Merge.ts\n+++ b/src/Tuple/Merge.ts\n@@ -30,7 +30,7 @@ type _MergeFlat<O extends object, O1P extends object> = {\n }\n \n type MergeDeep<T extends any[], T1 extends any[]> =\n- TupleOf<Compute<_MergeDeep<T, Omit<ObjectOf<T1>, keyof T>, ObjectOf<T1>>>>\n+ TupleOf<Compute<_MergeDeep<ObjectOf<T>, Omit<ObjectOf<T1>, keyof T>, ObjectOf<T1>>>>\n // same principle as above, but with a little tweak\n // we keep the original `O1` to know if we can merge\n // => if `O` and `O1` have `object` fields of same name\n", "diff --git a/.github/workflows/ibis-rebase-nightly.yml b/.github/workflows/ibis-rebase-nightly.yml\nindex 0e284b0..4a3ec7a 100644\n--- a/.github/workflows/ibis-rebase-nightly.yml\n+++ b/.github/workflows/ibis-rebase-nightly.yml\n@@ -22,7 +22,7 @@ jobs:\n | cut -d ' ' -f2 \\\n | grep -P '\\d+\\.x\\.x' \\\n | xargs printf '\"%s\"' \\\n- | jq -s '{branch: .}')\n+ | jq -rcMs '{branch: .}')\n \n echo \"::set-output name=matrix::$branches\"\n \n"]
2
["c4d9e5023fa0f88ba283b37da27677ceda1cbfbb", "4638dcdf7011e8e42d11fde04f068f22ee20fa1d"]
["fix", "ci"]
add descriptions to buttons on hover,change notice from 'danger' > 'info' Signed-off-by: Raju Udava <[email protected]>
["diff --git a/benchmarks/main.mjs b/benchmarks/main.mjs\nindex 0c2dc6b..e2f79d4 100644\n--- a/benchmarks/main.mjs\n+++ b/benchmarks/main.mjs\n@@ -65,8 +65,9 @@ const vnode = () =>\n },\n style: style({ margin: '5px' }),\n disabled,\n+ title: suite.name.split(' | ')[1],\n },\n- [suite.name],\n+ [suite.name.split(' | ')[0]],\n ),\n ),\n m(\ndiff --git a/benchmarks/suites/appendManyRowsToLargeTable.mjs b/benchmarks/suites/appendManyRowsToLargeTable.mjs\nindex e6a034e..7e34ca3 100644\n--- a/benchmarks/suites/appendManyRowsToLargeTable.mjs\n+++ b/benchmarks/suites/appendManyRowsToLargeTable.mjs\n@@ -31,7 +31,9 @@ data2.forEach(({ id, label }) => {\n el2.appendChild(tr);\n });\n \n-const suite = new benchmark.Suite('append many rows to large table');\n+const suite = new benchmark.Suite(\n+ 'append many rows to large table | appending 1,000 to a table of 10,000 rows.',\n+);\n \n const hoistedVNode = m(\n 'table',\ndiff --git a/benchmarks/suites/clearRows.mjs b/benchmarks/suites/clearRows.mjs\nindex ad47036..2a7711b 100644\n--- a/benchmarks/suites/clearRows.mjs\n+++ b/benchmarks/suites/clearRows.mjs\n@@ -27,7 +27,7 @@ data2.forEach(({ id, label }) => {\n el2.appendChild(row);\n });\n \n-const suite = new benchmark.Suite('clear rows');\n+const suite = new benchmark.Suite('clear rows | clearing a table with 1,000 rows');\n \n const hoistedVNode = m('table', undefined, [], VFlags.NO_CHILDREN);\n \ndiff --git a/benchmarks/suites/createManyRows.mjs b/benchmarks/suites/createManyRows.mjs\nindex 578f511..96c7b02 100644\n--- a/benchmarks/suites/createManyRows.mjs\n+++ b/benchmarks/suites/createManyRows.mjs\n@@ -7,7 +7,7 @@ import benchmark from '../benchmark';\n import { m, patch } from '../../src/index';\n import { buildData } from '../data';\n \n-const suite = new benchmark.Suite('create many rows');\n+const suite = new benchmark.Suite('create many rows | creating 10,000 rows');\n \n const hoistedVNode = m(\n 'div',\ndiff --git a/benchmarks/suites/createRows.mjs b/benchmarks/suites/createRows.mjs\nindex bfcc876..4d9ff57 100644\n--- a/benchmarks/suites/createRows.mjs\n+++ b/benchmarks/suites/createRows.mjs\n@@ -7,7 +7,7 @@ import benchmark from '../benchmark';\n import { m, patch } from '../../src/index';\n import { buildData } from '../data';\n \n-const suite = new benchmark.Suite('create rows');\n+const suite = new benchmark.Suite('create rows | creating 1,000 rows');\n \n const hoistedVNode = m(\n 'table',\ndiff --git a/benchmarks/suites/partialUpdate.mjs b/benchmarks/suites/partialUpdate.mjs\nindex 55948a9..c5f1de3 100644\n--- a/benchmarks/suites/partialUpdate.mjs\n+++ b/benchmarks/suites/partialUpdate.mjs\n@@ -34,7 +34,7 @@ data2.forEach(({ id, label }) => {\n el2.appendChild(tr);\n });\n \n-const suite = new benchmark.Suite('partial update');\n+const suite = new benchmark.Suite('partial update | updating every 10th row for 1,000 rows');\n \n const hoistedVNode = m(\n 'table',\ndiff --git a/benchmarks/suites/removeRow.mjs b/benchmarks/suites/removeRow.mjs\nindex aeb1e9a..31c7599 100644\n--- a/benchmarks/suites/removeRow.mjs\n+++ b/benchmarks/suites/removeRow.mjs\n@@ -30,7 +30,7 @@ data2.forEach(({ id, label }) => {\n el2.appendChild(tr);\n });\n \n-const suite = new benchmark.Suite('remove row');\n+const suite = new benchmark.Suite('remove row | removing one row');\n \n const hoistedVNode = m(\n 'table',\ndiff --git a/benchmarks/suites/replaceAllRows.mjs b/benchmarks/suites/replaceAllRows.mjs\nindex 9555ae4..7001667 100644\n--- a/benchmarks/suites/replaceAllRows.mjs\n+++ b/benchmarks/suites/replaceAllRows.mjs\n@@ -41,7 +41,7 @@ data2.forEach(({ id, label }) => {\n \n shuffleArray(data2);\n \n-const suite = new benchmark.Suite('replace all rows');\n+const suite = new benchmark.Suite('replace all rows | updating all 1,000 rows');\n \n const hoistedVNode = m(\n 'table',\ndiff --git a/benchmarks/suites/selectRow.mjs b/benchmarks/suites/selectRow.mjs\nindex 76be216..de69359 100644\n--- a/benchmarks/suites/selectRow.mjs\n+++ b/benchmarks/suites/selectRow.mjs\n@@ -30,7 +30,7 @@ data2.forEach(({ id, label }) => {\n el2.appendChild(tr);\n });\n \n-const suite = new benchmark.Suite('select row');\n+const suite = new benchmark.Suite('select row | highlighting a selected row');\n \n const hoistedVNode = m(\n 'table',\ndiff --git a/benchmarks/suites/swapRows.mjs b/benchmarks/suites/swapRows.mjs\nindex 2a91e74..ce52036 100644\n--- a/benchmarks/suites/swapRows.mjs\n+++ b/benchmarks/suites/swapRows.mjs\n@@ -36,7 +36,7 @@ data2.forEach(({ id, label }) => {\n el2.appendChild(tr);\n });\n \n-const suite = new benchmark.Suite('swap rows');\n+const suite = new benchmark.Suite('swap rows | swap 2 rows for table with 1,000 rows');\n \n const hoistedVNode = m(\n 'table',\n", "diff --git a/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md b/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md\nindex 17a1d85..b8c3f52 100644\n--- a/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md\n+++ b/packages/noco-docs/docs/030.workspaces/040.actions-on-workspace.md\n@@ -20,7 +20,7 @@ To update the workspace name:\n ## Delete workspace\n If you determine that a workspace is no longer necessary, you have the option to permanently remove it from your settings. Deleting a workspace will delete all the bases and data associated with it.\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/040.bases/070.actions-on-base.md b/packages/noco-docs/docs/040.bases/070.actions-on-base.md\nindex b8e5723..7207971 100644\n--- a/packages/noco-docs/docs/040.bases/070.actions-on-base.md\n+++ b/packages/noco-docs/docs/040.bases/070.actions-on-base.md\n@@ -69,7 +69,7 @@ To duplicate a base, you can follow these straightforward steps:\n \n If you determine that a base is no longer necessary, you have the option to permanently remove it from your workspace. Deleting a base will delete all the tables and data associated with it.\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/050.tables/060.actions-on-table.md b/packages/noco-docs/docs/050.tables/060.actions-on-table.md\nindex 3cf03d3..8ae9ade 100644\n--- a/packages/noco-docs/docs/050.tables/060.actions-on-table.md\n+++ b/packages/noco-docs/docs/050.tables/060.actions-on-table.md\n@@ -46,7 +46,7 @@ A new table will be generated, mirroring the original table's schema and content\n \n ## Delete table\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/070.fields/060.actions-on-field.md b/packages/noco-docs/docs/070.fields/060.actions-on-field.md\nindex 600c6fd..fe2cfa8 100644\n--- a/packages/noco-docs/docs/070.fields/060.actions-on-field.md\n+++ b/packages/noco-docs/docs/070.fields/060.actions-on-field.md\n@@ -83,7 +83,7 @@ New field will be created to the right of the original field.\n New field will be created to the left of the original field.\n \n ### Delete field\n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \ndiff --git a/packages/noco-docs/docs/080.records/070.actions-on-record.md b/packages/noco-docs/docs/080.records/070.actions-on-record.md\nindex a9245ff..6d4774a 100644\n--- a/packages/noco-docs/docs/080.records/070.actions-on-record.md\n+++ b/packages/noco-docs/docs/080.records/070.actions-on-record.md\n@@ -54,8 +54,8 @@ On the bulk update modal,\n 5. Click on the `Bulk Update all` button\n 6. A confirmation dialog will be displayed. Click on `Confirm` to update the records.\n \n-:::danger\n-This operation cannot be undone.\n+:::info\n+**This action cannot be undone.**\n :::\n \n ![Bulk Update](/img/v2/records/bulk-update-1.png)\ndiff --git a/packages/noco-docs/docs/090.views/090.actions-on-view.md b/packages/noco-docs/docs/090.views/090.actions-on-view.md\nindex c6c6ab2..7d23959 100644\n--- a/packages/noco-docs/docs/090.views/090.actions-on-view.md\n+++ b/packages/noco-docs/docs/090.views/090.actions-on-view.md\n@@ -41,7 +41,7 @@ The view context menu provides a set of tools to interact with the view. The vie\n \n ## Delete view\n \n-:::danger\n+:::info\n **This action cannot be undone.**\n :::\n \n"]
2
["d8d0ba8ea17ed43a04f90213851d2f27056d8cf0", "2ba752d45350a676babe553dd68f019af81b512b"]
["feat", "docs"]
fixing deploying to kubernetes Signed-off-by: Rajesh Rajendran <[email protected]>,create mock img server
["diff --git a/.github/workflows/frontend.yaml b/.github/workflows/frontend.yaml\nindex 7e42967..77e4abf 100644\n--- a/.github/workflows/frontend.yaml\n+++ b/.github/workflows/frontend.yaml\n@@ -22,26 +22,22 @@ jobs:\n ${{ runner.OS }}-build-\n ${{ runner.OS }}-\n \n+ - uses: azure/k8s-set-context@v1\n+ with:\n+ method: kubeconfig\n+ kubeconfig: ${{ secrets.OSS_KUBECONFIG }} # Use content of kubeconfig in secret.\n+ id: setcontext\n - name: Install\n run: npm install\n \n- - name: Build\n- run: npm run build:staging\n- env:\n- ENVIRONMENT: staging\n-\n- - name: Deploy\n- env:\n- AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }}\n- AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }}\n- AWS_REGION: eu-central-1\n- AWS_S3_BUCKET_NAME: ${{ secrets.AWS_S3_BUCKET_NAME }}\n+ - name: Build and deploy\n run: |\n- aws configure set default.s3.signature_version s3v4\n- aws --endpoint-url https://${{secrets.DOMAIN_NAME}}/frontend/ s3 cp \\\n- --recursive \\\n- --region \"$AWS_REGION\" \\\n- public s3://$AWS_S3_BUCKET_NAME\n+ cd frontend\n+ bash build.sh\n+ cp -arl public frontend\n+ minio_pod=$(kubectl get po -n db -l app.kubernetes.io/name=minio -n db --output custom-columns=name:.metadata.name | tail -n+2)\n+ kubectl -n db cp frontend $minio_pod:/data/\n+ rm -rf frontend\n \n # - name: Debug Job\n # if: ${{ failure() }}\n", "diff --git a/scripts/gulp/tasks/test.ts b/scripts/gulp/tasks/test.ts\nindex 8014b12..d10c1aa 100644\n--- a/scripts/gulp/tasks/test.ts\n+++ b/scripts/gulp/tasks/test.ts\n@@ -26,12 +26,18 @@ task('test.imageserver', () => {\n function handleRequest(req, res) {\n const urlParse = url.parse(req.url, true);\n \n+ res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Access-Control-Allow-Methods', 'GET');\n+ res.setHeader('Connection', 'keep-alive');\n+ res.setHeader('Age', '0');\n+ res.setHeader('cache-control', 'no-store');\n+\n if (urlParse.pathname === '/reset') {\n console.log('Image Server Reset');\n console.log('---------------------------');\n requestedUrls.length = 0;\n start = Date.now();\n- res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Content-Type', 'text/plain');\n res.end('reset');\n return;\n }\n@@ -48,9 +54,8 @@ task('test.imageserver', () => {\n \n setTimeout(() => {\n res.setHeader('Content-Type', 'image/svg+xml');\n- res.setHeader('Access-Control-Allow-Origin', '*');\n res.end(`<svg xmlns=\"http://www.w3.org/2000/svg\" xmlns:xlink=\"http://www.w3.org/1999/xlink\"\n- style=\"background-color: ${color}; width: ${width}px; height: ${height}px;\">\n+ viewBox=\"0 0 ${width} ${height}\" style=\"background-color: ${color};\">\n <text x=\"5\" y=\"22\" style=\"font-family: Courier; font-size: 24px\">${id}</text>\n </svg>`);\n }, delay);\n"]
2
["3f2eec37f76c1ad9408e423e49fe5bfe3e17d943", "32b76173a259ea1993298289b436cf10c1e800bf"]
["ci", "test"]
add documentation to use react-native-paper with CRA (#874),set name for topology module
["diff --git a/docs/pages/4.react-native-web.md b/docs/pages/4.react-native-web.md\nindex 69e4e52..8d6ae2a 100644\n--- a/docs/pages/4.react-native-web.md\n+++ b/docs/pages/4.react-native-web.md\n@@ -16,6 +16,63 @@ To install `react-native-web`, run:\n yarn add react-native-web react-dom react-art\n ```\n \n+### Using CRA ([Create React App](https://github.com/facebook/create-react-app))\n+\n+Install [`react-app-rewired`](https://github.com/timarney/react-app-rewired) to override `webpack` configuration:\n+\n+```sh\n+yarn add --dev react-app-rewired\n+```\n+\n+[Configure `babel-loader`](#2-configure-babel-loader) using a new file called `config-overrides.js`:\n+\n+```js\n+module.exports = function override(config, env) {\n+ config.module.rules.push({\n+ test: /\\.js$/,\n+ exclude: /node_modules[/\\\\](?!react-native-paper|react-native-vector-icons|react-native-safe-area-view)/,\n+ use: {\n+ loader: \"babel-loader\",\n+ options: {\n+ // Disable reading babel configuration\n+ babelrc: false,\n+ configFile: false,\n+\n+ // The configration for compilation\n+ presets: [\n+ [\"@babel/preset-env\", { useBuiltIns: \"usage\" }],\n+ \"@babel/preset-react\",\n+ \"@babel/preset-flow\"\n+ ],\n+ plugins: [\n+ \"@babel/plugin-proposal-class-properties\",\n+ \"@babel/plugin-proposal-object-rest-spread\"\n+ ]\n+ }\n+ }\n+ });\n+\n+ return config;\n+};\n+```\n+\n+Change your script in `package.json`:\n+\n+```diff\n+/* package.json */\n+\n+ \"scripts\": {\n+- \"start\": \"react-scripts start\",\n++ \"start\": \"react-app-rewired start\",\n+- \"build\": \"react-scripts build\",\n++ \"build\": \"react-app-rewired build\",\n+- \"test\": \"react-scripts test --env=jsdom\",\n++ \"test\": \"react-app-rewired test --env=jsdom\"\n+}\n+```\n+\n+### Custom webpack setup\n+\n To install `webpack`, run:\n \n ```sh\n", "diff --git a/topology/pom.xml b/topology/pom.xml\nindex 389508e..ee6239a 100644\n--- a/topology/pom.xml\n+++ b/topology/pom.xml\n@@ -16,6 +16,7 @@\n </parent>\n \n <artifactId>zeebe-cluster-topology</artifactId>\n+ <name>Zeebe Cluster Topology</name>\n \n <properties>\n <proto.dir>${maven.multiModuleProjectDirectory}/topology/src/main/resources/proto</proto.dir>\n"]
2
["ee7cc5d5a940fba774e715b1f029c6361110b108", "8911a972222dc80a242f3f1d9b3596321b3fdeaa"]
["docs", "build"]
fix pagination spacing,offset tests for min and max read cursors
["diff --git a/website/layouts/Base.tsx b/website/layouts/Base.tsx\nindex 22d36a2..40f7130 100644\n--- a/website/layouts/Base.tsx\n+++ b/website/layouts/Base.tsx\n@@ -399,7 +399,7 @@ export function Base({ children, headings }: BaseProps) {\n >\n <a className=\"flex items-center space-x-4 group\">\n <ArrowLeftIcon className=\"h-4 transition-transform duration-100 ease-in-out transform group-hover:-translate-x-1\" />\n- <div className=\"flex flex-col space-x-1\">\n+ <div className=\"flex flex-col space-y-1\">\n <span className=\"text-sm text-gray-500 transition-colors duration-100 ease-in-out group-hover:text-gray-700\">\n Previous\n </span>\n@@ -418,7 +418,7 @@ export function Base({ children, headings }: BaseProps) {\n aria-label={`Go to ${next.resource?.label}`}\n >\n <a className=\"flex items-center space-x-4 group\">\n- <div className=\"flex flex-col space-x-1\">\n+ <div className=\"flex flex-col space-y-1\">\n <span className=\"text-sm text-gray-500 transition-colors duration-100 ease-in-out group-hover:text-gray-700\">\n Next\n </span>\n", "diff --git a/storage/reads/array_cursor_test.go b/storage/reads/array_cursor_test.go\nindex 7c7ad0c..c1e6ff9 100644\n--- a/storage/reads/array_cursor_test.go\n+++ b/storage/reads/array_cursor_test.go\n@@ -1541,6 +1541,34 @@ func TestWindowMinArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := (i % 4) * 15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:30:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:30:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:30:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:30:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{0, 30, 130, 230, 330},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window desc values\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n@@ -1560,6 +1588,34 @@ func TestWindowMinArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset desc values\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := 60 - (i%4)*15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:45:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:45:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:45:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:45:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{45, 15, 115, 215, 315},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window min int\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n@@ -1693,6 +1749,34 @@ func TestWindowMaxArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := (i % 4) * 15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:15:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:45:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{15, 115, 215, 315, 345},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window desc values\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n@@ -1712,6 +1796,34 @@ func TestWindowMaxArrayCursor(t *testing.T) {\n \t\t\t},\n \t\t},\n \t\t{\n+\t\t\tname: \"window offset desc values\",\n+\t\t\tevery: time.Hour,\n+\t\t\toffset: 30 * time.Minute,\n+\t\t\tinputArrays: []*cursors.IntegerArray{\n+\t\t\t\tmakeIntegerArray(\n+\t\t\t\t\t16,\n+\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\"), 15*time.Minute,\n+\t\t\t\t\tfunc(i int64) int64 {\n+\t\t\t\t\t\tbase := (i / 4) * 100\n+\t\t\t\t\t\tm := 60 - (i%4)*15\n+\t\t\t\t\t\treturn base + m\n+\t\t\t\t\t},\n+\t\t\t\t),\n+\t\t\t},\n+\t\t\twant: []*cursors.IntegerArray{\n+\t\t\t\t{\n+\t\t\t\t\tTimestamps: []int64{\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T00:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T01:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T02:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:00:00Z\").UnixNano(),\n+\t\t\t\t\t\tmustParseTime(\"2010-01-01T03:30:00Z\").UnixNano(),\n+\t\t\t\t\t},\n+\t\t\t\t\tValues: []int64{60, 160, 260, 360, 330},\n+\t\t\t\t},\n+\t\t\t},\n+\t\t},\n+\t\t{\n \t\t\tname: \"window min int\",\n \t\t\tevery: time.Hour,\n \t\t\tinputArrays: []*cursors.IntegerArray{\n"]
2
["1e05a24486f15889ddf6bf1c711ea2bbffc1a88e", "b7e2330fa3a8d7b8a9bff01b707c44e64b845c7b"]
["fix", "test"]
add instruction for finding version,include `token_version` value in jwt payload use a generic function to populate jwt token to avoid repetition re #2361 Signed-off-by: Pranav C <[email protected]>
["diff --git a/.github/ISSUE_TEMPLATE/_bug_report_chs.md b/.github/ISSUE_TEMPLATE/_bug_report_chs.md\nindex 42a2e0f..44a33db 100644\n--- a/.github/ISSUE_TEMPLATE/_bug_report_chs.md\n+++ b/.github/ISSUE_TEMPLATE/_bug_report_chs.md\n@@ -36,7 +36,7 @@ assignees: ''\n ## \u8bbe\u5907\u4fe1\u606f\n - \u64cd\u4f5c\u7cfb\u7edf: [] <!-- \u5982 [Window10] -->\n - \u6d4f\u89c8\u5668\u7248\u672c: [] <!-- \u5982 [Chrome77] -->\n-- \u6c99\u62c9\u67e5\u8bcd\u7248\u672c: [] <!-- \u5982 [v7.0.0] -->\n+- \u6c99\u62c9\u67e5\u8bcd\u7248\u672c: [] <!-- \u5982 [v7.0.0] \uff08\u5728\u6269\u5c55\u7ba1\u7406\u9875\u9762\u67e5\u770b\uff09 -->\n \n <!-- \u8bf7\u5728\u4e0b\u65b9 ## \u5f00\u5934\u884c\u4e4b\u95f4\u7684\u7a7a\u767d\u5904\u586b\u5199 -->\n \n", "diff --git a/packages/nocodb/src/lib/meta/api/sync/importApis.ts b/packages/nocodb/src/lib/meta/api/sync/importApis.ts\nindex 2facc67..1f8d137 100644\n--- a/packages/nocodb/src/lib/meta/api/sync/importApis.ts\n+++ b/packages/nocodb/src/lib/meta/api/sync/importApis.ts\n@@ -7,7 +7,7 @@ import NocoJobs from '../../../jobs/NocoJobs';\n import job, { AirtableSyncConfig } from './helpers/job';\n import SyncSource from '../../../models/SyncSource';\n import Noco from '../../../Noco';\n-import * as jwt from 'jsonwebtoken';\n+import { genJwt } from '../userApi/helpers';\n const AIRTABLE_IMPORT_JOB = 'AIRTABLE_IMPORT_JOB';\n const AIRTABLE_PROGRESS_JOB = 'AIRTABLE_PROGRESS_JOB';\n \n@@ -76,18 +76,7 @@ export default (router: Router, clients: { [id: string]: Socket }) => {\n const syncSource = await SyncSource.get(req.params.syncId);\n \n const user = await syncSource.getUser();\n- const token = jwt.sign(\n- {\n- email: user.email,\n- firstname: user.firstname,\n- lastname: user.lastname,\n- id: user.id,\n- roles: user.roles\n- },\n-\n- Noco.getConfig().auth.jwt.secret,\n- Noco.getConfig().auth.jwt.options\n- );\n+ const token = genJwt(user, Noco.getConfig());\n \n // Treat default baseUrl as siteUrl from req object\n let baseURL = (req as any).ncSiteUrl;\ndiff --git a/packages/nocodb/src/lib/meta/api/userApi/helpers.ts b/packages/nocodb/src/lib/meta/api/userApi/helpers.ts\nnew file mode 100644\nindex 0000000..48f7c11\n--- /dev/null\n+++ b/packages/nocodb/src/lib/meta/api/userApi/helpers.ts\n@@ -0,0 +1,18 @@\n+import * as jwt from 'jsonwebtoken';\n+import User from '../../../models/User';\n+import { NcConfig } from '../../../../interface/config';\n+\n+export function genJwt(user: User, config: NcConfig) {\n+ return jwt.sign(\n+ {\n+ email: user.email,\n+ firstname: user.firstname,\n+ lastname: user.lastname,\n+ id: user.id,\n+ roles: user.roles,\n+ token_version: user.token_version\n+ },\n+ config.auth.jwt.secret,\n+ config.auth.jwt.options\n+ );\n+}\ndiff --git a/packages/nocodb/src/lib/meta/api/userApi/userApis.ts b/packages/nocodb/src/lib/meta/api/userApi/userApis.ts\nindex 2b8384a..7a6fa31 100644\n--- a/packages/nocodb/src/lib/meta/api/userApi/userApis.ts\n+++ b/packages/nocodb/src/lib/meta/api/userApi/userApis.ts\n@@ -10,7 +10,6 @@ import User from '../../../models/User';\n import { Tele } from 'nc-help';\n \n const { v4: uuidv4 } = require('uuid');\n-import * as jwt from 'jsonwebtoken';\n import Audit from '../../../models/Audit';\n import crypto from 'crypto';\n import NcPluginMgrv2 from '../../helpers/NcPluginMgrv2';\n@@ -20,6 +19,7 @@ import extractProjectIdAndAuthenticate from '../../helpers/extractProjectIdAndAu\n import ncMetaAclMw from '../../helpers/ncMetaAclMw';\n import { MetaTable } from '../../../utils/globals';\n import Noco from '../../../Noco';\n+import { genJwt } from './helpers';\n \n export async function signup(req: Request, res: Response<TableType>) {\n const {\n@@ -147,18 +147,7 @@ export async function signup(req: Request, res: Response<TableType>) {\n });\n \n res.json({\n- token: jwt.sign(\n- {\n- email: user.email,\n- firstname: user.firstname,\n- lastname: user.lastname,\n- id: user.id,\n- roles: user.roles,\n- token_version: user.token_version\n- },\n- Noco.getConfig().auth.jwt.secret,\n- Noco.getConfig().auth.jwt.options\n- )\n+ token: genJwt(user, Noco.getConfig())\n } as any);\n }\n \n@@ -205,19 +194,7 @@ async function successfulSignIn({\n });\n \n res.json({\n- token: jwt.sign(\n- {\n- email: user.email,\n- firstname: user.firstname,\n- lastname: user.lastname,\n- id: user.id,\n- roles: user.roles,\n- token_version\n- },\n-\n- Noco.getConfig().auth.jwt.secret,\n- Noco.getConfig().auth.jwt.options\n- )\n+ token: genJwt(user, Noco.getConfig())\n } as any);\n } catch (e) {\n console.log(e);\n@@ -477,17 +454,7 @@ async function refreshToken(req, res): Promise<any> {\n setTokenCookie(res, refreshToken);\n \n res.json({\n- token: jwt.sign(\n- {\n- email: user.email,\n- firstname: user.firstname,\n- lastname: user.lastname,\n- id: user.id,\n- roles: user.roles\n- },\n- Noco.getConfig().auth.jwt.secret,\n- Noco.getConfig().auth.jwt.options\n- )\n+ token: genJwt(user, Noco.getConfig())\n } as any);\n } catch (e) {\n return res.status(400).json({ msg: e.message });\n"]
2
["af0a5f7ab9d71fe20aa0888f682368f32b26fe18", "5ac169d840351733849207e082d2a65abf629c92"]
["docs", "fix"]
correct width when --no-quotes is used,fix unstable MessageCorrelationTest
["diff --git a/src/output/grid.rs b/src/output/grid.rs\nindex 37f6c57..ce989e5 100644\n--- a/src/output/grid.rs\n+++ b/src/output/grid.rs\n@@ -8,6 +8,8 @@ use crate::output::file_name::{Classify, Options as FileStyle};\n use crate::output::file_name::{EmbedHyperlinks, ShowIcons};\n use crate::theme::Theme;\n \n+use super::file_name::QuoteStyle;\n+\n #[derive(PartialEq, Eq, Debug, Copy, Clone)]\n pub struct Options {\n pub across: bool,\n@@ -55,27 +57,34 @@ impl<'a> Render<'a> {\n } else {\n 0\n };\n-\n- let space_filename_offset = if file.name.contains(' ') || file.name.contains('\\'') {\n- 2\n- } else {\n- 0\n+ let space_filename_offset = match self.file_style.quote_style {\n+ QuoteStyle::QuoteSpaces if file.name.contains(' ') => 2,\n+ QuoteStyle::NoQuotes => 0,\n+ _ => 0, // Default case\n };\n-\n let contents = filename.paint();\n- #[rustfmt::skip]\n let width = match (\n filename.options.embed_hyperlinks,\n filename.options.show_icons,\n ) {\n- ( EmbedHyperlinks::On, ShowIcons::Always(spacing) | ShowIcons::Automatic(spacing) )\n- => filename.bare_width() + classification_width + 1 + (spacing as usize) + space_filename_offset,\n- ( EmbedHyperlinks::On, ShowIcons::Never )\n- => filename.bare_width() + classification_width + space_filename_offset,\n- ( EmbedHyperlinks::Off, ShowIcons::Always(spacing) | ShowIcons::Automatic(spacing) )\n- => filename.bare_width() + 1 + (spacing as usize) + space_filename_offset,\n- ( EmbedHyperlinks::Off, _ )\n- => *contents.width(),\n+ (\n+ EmbedHyperlinks::On,\n+ ShowIcons::Always(spacing) | ShowIcons::Automatic(spacing),\n+ ) => {\n+ filename.bare_width()\n+ + classification_width\n+ + 1\n+ + (spacing as usize)\n+ + space_filename_offset\n+ }\n+ (EmbedHyperlinks::On, ShowIcons::Never) => {\n+ filename.bare_width() + classification_width + space_filename_offset\n+ }\n+ (\n+ EmbedHyperlinks::Off,\n+ ShowIcons::Always(spacing) | ShowIcons::Automatic(spacing),\n+ ) => filename.bare_width() + 1 + (spacing as usize) + space_filename_offset,\n+ (EmbedHyperlinks::Off, _) => *contents.width(),\n };\n \n grid.add(tg::Cell {\n", "diff --git a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\nindex 0f5fed9..796393c 100644\n--- a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\n+++ b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\n@@ -27,7 +27,6 @@ import static io.zeebe.test.util.MsgPackUtil.asMsgPack;\n import static org.assertj.core.api.Assertions.assertThat;\n import static org.assertj.core.api.Assertions.entry;\n \n-import io.zeebe.UnstableTest;\n import io.zeebe.broker.test.EmbeddedBrokerRule;\n import io.zeebe.model.bpmn.Bpmn;\n import io.zeebe.model.bpmn.BpmnModelInstance;\n@@ -50,7 +49,6 @@ import org.agrona.DirectBuffer;\n import org.junit.Before;\n import org.junit.Rule;\n import org.junit.Test;\n-import org.junit.experimental.categories.Category;\n import org.junit.rules.RuleChain;\n import org.junit.runner.RunWith;\n import org.junit.runners.Parameterized;\n@@ -165,7 +163,7 @@ public class MessageCorrelationTest {\n \"receive-message\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\n \n final SubscribedRecord messageSubscription =\n- findMessageSubscription(testClient, MessageSubscriptionIntent.OPENED);\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\n assertThat(messageSubscription.valueType()).isEqualTo(ValueType.MESSAGE_SUBSCRIPTION);\n assertThat(messageSubscription.recordType()).isEqualTo(RecordType.EVENT);\n assertThat(messageSubscription.value())\n@@ -244,7 +242,7 @@ public class MessageCorrelationTest {\n final long workflowInstanceKey =\n testClient.createWorkflowInstance(\"wf\", asMsgPack(\"orderId\", \"order-123\"));\n \n- testClient.receiveFirstWorkflowInstanceEvent(WorkflowInstanceIntent.ELEMENT_ACTIVATED);\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\n \n // when\n testClient.publishMessage(\"order canceled\", \"order-123\", asMsgPack(\"foo\", \"bar\"));\n@@ -308,13 +306,12 @@ public class MessageCorrelationTest {\n }\n \n @Test\n- @Category(UnstableTest.class) // => https://github.com/zeebe-io/zeebe/issues/1234\n public void shouldCorrelateMessageWithZeroTTL() throws Exception {\n // given\n final long workflowInstanceKey =\n testClient.createWorkflowInstance(\"wf\", asMsgPack(\"orderId\", \"order-123\"));\n \n- testClient.receiveElementInState(\"receive-message\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\n \n // when\n testClient.publishMessage(\"order canceled\", \"order-123\", asMsgPack(\"foo\", \"bar\"), 0);\n@@ -499,10 +496,9 @@ public class MessageCorrelationTest {\n .containsEntry(\"activityInstanceKey\", catchEventEntered.key());\n }\n \n- private SubscribedRecord findMessageSubscription(\n- final TestPartitionClient client, final MessageSubscriptionIntent intent)\n+ private SubscribedRecord findMessageSubscription(final MessageSubscriptionIntent intent)\n throws AssertionError {\n- return client\n+ return testClient\n .receiveEvents()\n .filter(intent(intent))\n .findFirst()\n"]
2
["61eaa2d0cca9bd27d6c5f0a8f9b34200b77fdbb0", "98bed2a8137930149559bc1cae9bd34a1a75e556"]
["fix", "test"]
remove sync ts check,simplyfy statement
["diff --git a/config/webpack.config.prod.js b/config/webpack.config.prod.js\nindex 8b23fba..58a4c17 100644\n--- a/config/webpack.config.prod.js\n+++ b/config/webpack.config.prod.js\n@@ -251,7 +251,7 @@ module.exports = {\n plugins: [\n argv.notypecheck\n ? null\n- : new ForkTsCheckerWebpackPlugin({tslint: true, async: false}),\n+ : new ForkTsCheckerWebpackPlugin({tslint: true}),\n // Makes some environment variables available in index.html.\n // The public URL is available as %PUBLIC_URL% in index.html, e.g.:\n // <link rel=\"shortcut icon\" href=\"%PUBLIC_URL%/favicon.ico\">\n", "diff --git a/src/Object/Merge.ts b/src/Object/Merge.ts\nindex 1f48efb..06caad1 100644\n--- a/src/Object/Merge.ts\n+++ b/src/Object/Merge.ts\n@@ -96,9 +96,11 @@ type ChooseMergeDeep<OK, O1K, K extends Key, OOK extends Key, style extends Merg\n @hidden\n */\n export type _MergeDeep<O, O1, K extends Key, OOK extends Key, style extends MergeStyle> =\n- Or<Extends<[O], [never]>, Extends<[O1], [never]>> extends 1 // filter never\n+ [O] extends [never]\n ? MergeProp<O, O1, K, OOK, style>\n- : LibStyle<ChooseMergeDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n+ : [O1] extends [never]\n+ ? MergeProp<O, O1, K, OOK, style>\n+ : LibStyle<ChooseMergeDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n \n /**\n @hidden\ndiff --git a/src/Object/Patch.ts b/src/Object/Patch.ts\nindex 2d73784..2c8bd42 100644\n--- a/src/Object/Patch.ts\n+++ b/src/Object/Patch.ts\n@@ -89,9 +89,11 @@ type ChoosePatchDeep<OK, O1K, K extends Key, OOK extends Key, style extends Merg\n @hidden\n */\n export type _PatchDeep<O, O1, K extends Key, OOK extends Key, style extends MergeStyle> =\n- Or<Extends<[O], [never]>, Extends<[O1], [never]>> extends 1 // filter never\n+ [O] extends [never]\n ? PatchProp<O, O1, K, OOK>\n- : LibStyle<ChoosePatchDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n+ : [O1] extends [never]\n+ ? PatchProp<O, O1, K, OOK>\n+ : LibStyle<ChoosePatchDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n \n /**\n @hidden\n"]
2
["411be831591b2ea15ca9138eaf8db81f51b5101e", "f86944ff00b970d7e2da48abbff43e58bdf29b99"]
["build", "refactor"]
fix node test cases run into infinite loop,remove writers from interface
["diff --git a/packages/designer/tests/document/node/node.test.ts b/packages/designer/tests/document/node/node.test.ts\nindex dd20bd3..113360d 100644\n--- a/packages/designer/tests/document/node/node.test.ts\n+++ b/packages/designer/tests/document/node/node.test.ts\n@@ -26,7 +26,7 @@ import rootHeaderMetadata from '../../fixtures/component-metadata/root-header';\n import rootContentMetadata from '../../fixtures/component-metadata/root-content';\n import rootFooterMetadata from '../../fixtures/component-metadata/root-footer';\n \n-describe.skip('Node \u65b9\u6cd5\u6d4b\u8bd5', () => {\n+describe('Node \u65b9\u6cd5\u6d4b\u8bd5', () => {\n let editor: Editor;\n let designer: Designer;\n let project: Project;\n@@ -474,15 +474,16 @@ describe.skip('Node \u65b9\u6cd5\u6d4b\u8bd5', () => {\n it('didDropIn / didDropOut', () => {\n const form = doc.getNode('node_k1ow3cbo');\n designer.createComponentMeta(divMetadata);\n+ designer.createComponentMeta(formMetadata);\n const callbacks = form.componentMeta.getMetadata().configure.advanced?.callbacks;\n const fn1 = callbacks.onNodeAdd = jest.fn();\n const fn2 = callbacks.onNodeRemove = jest.fn();\n const textField = doc.getNode('node_k1ow3cc9');\n form.didDropIn(textField);\n- expect(fn1).toHaveBeenCalledWith(textField, form);\n+ expect(fn1).toHaveBeenCalledWith(textField.internalToShellNode(), form.internalToShellNode());\n \n form.didDropOut(textField);\n- expect(fn2).toHaveBeenCalledWith(textField, form);\n+ expect(fn2).toHaveBeenCalledWith(textField.internalToShellNode(), form.internalToShellNode());\n });\n \n it('hover', () => {\n", "diff --git a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\nindex 91f1b41..eb4b9a8 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\n@@ -81,8 +81,6 @@ public class Engine implements RecordProcessor<EngineContext> {\n \n engineContext.setLifecycleListeners(typedRecordProcessors.getLifecycleListeners());\n recordProcessorMap = typedRecordProcessors.getRecordProcessorMap();\n-\n- engineContext.setWriters(writers);\n }\n \n @Override\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\nindex a8e5538..a27b6e6 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\n@@ -15,7 +15,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.StreamProcessorListene\n import io.camunda.zeebe.engine.processing.streamprocessor.TypedRecordProcessorFactory;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriter;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\n import io.camunda.zeebe.engine.state.EventApplier;\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\n import java.util.Collections;\n@@ -34,7 +33,6 @@ public final class EngineContext {\n private final TypedRecordProcessorFactory typedRecordProcessorFactory;\n private List<StreamProcessorLifecycleAware> lifecycleListeners = Collections.EMPTY_LIST;\n private StreamProcessorListener streamProcessorListener;\n- private Writers writers;\n \n public EngineContext(\n final int partitionId,\n@@ -102,12 +100,4 @@ public final class EngineContext {\n public void setStreamProcessorListener(final StreamProcessorListener streamProcessorListener) {\n this.streamProcessorListener = streamProcessorListener;\n }\n-\n- public Writers getWriters() {\n- return writers;\n- }\n-\n- public void setWriters(final Writers writers) {\n- this.writers = writers;\n- }\n }\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\nindex f30c7cc..834b421 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\n@@ -8,7 +8,6 @@\n package io.camunda.zeebe.engine.api;\n \n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\n import io.camunda.zeebe.logstreams.log.LogStream;\n \n@@ -27,11 +26,6 @@ public interface ReadonlyStreamProcessorContext {\n LegacyTypedStreamWriter getLogStreamWriter();\n \n /**\n- * @return the specific writers, like command, response, etc\n- */\n- Writers getWriters();\n-\n- /**\n * @return the state, where the data is stored during processing\n */\n MutableZeebeState getZeebeState();\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\nindex 844e487..49fd8e2 100755\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\n@@ -346,7 +346,6 @@ public class StreamProcessor extends Actor implements HealthMonitorable, LogReco\n if (listener != null) {\n streamProcessorContext.listener(engineContext.getStreamProcessorListener());\n }\n- streamProcessorContext.writers(engineContext.getWriters());\n }\n \n private long recoverFromSnapshot() {\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\nindex d02b273..b527d3c 100644\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\n@@ -17,7 +17,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.writers.CommandRespons\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriterImpl;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.NoopLegacyTypedStreamWriter;\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\n import io.camunda.zeebe.engine.state.EventApplier;\n import io.camunda.zeebe.engine.state.KeyGeneratorControls;\n import io.camunda.zeebe.engine.state.ZeebeDbState;\n@@ -55,7 +54,7 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\n private StreamProcessorMode streamProcessorMode = StreamProcessorMode.PROCESSING;\n private ProcessingScheduleService processingScheduleService;\n private MutableLastProcessedPositionState lastProcessedPositionState;\n- private Writers writers;\n+\n private LogStreamBatchWriter logStreamBatchWriter;\n private CommandResponseWriter commandResponseWriter;\n \n@@ -85,11 +84,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\n }\n \n @Override\n- public Writers getWriters() {\n- return writers;\n- }\n-\n- @Override\n public MutableZeebeState getZeebeState() {\n return zeebeState;\n }\n@@ -216,10 +210,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\n return streamProcessorMode;\n }\n \n- public void writers(final Writers writers) {\n- this.writers = writers;\n- }\n-\n public void logStreamBatchWriter(final LogStreamBatchWriter batchWriter) {\n logStreamBatchWriter = batchWriter;\n }\n"]
2
["d2c3f0ba6f85b659b76636a91ea9ab2b5a95a720", "d2709cab63295109dcd1a49f57da9418110e9044"]
["test", "refactor"]
updated riot to v6, fixed build,process CommandDistribution ACKNOWLEDGED event Adds an EventApplier for the CommandDistribution ACKNOWLEDGED event. This applier will be responsible to remove a pending distribution from the state. This will be used to mark the distribution to a specific partition as completed.
["diff --git a/components/riot/package.json b/components/riot/package.json\nindex c41743a..eb69756 100644\n--- a/components/riot/package.json\n+++ b/components/riot/package.json\n@@ -61,7 +61,7 @@\n },\n \"devDependencies\": {\n \"@babel/preset-typescript\": \"^7.14.5\",\n- \"@riotjs/cli\": \"^6.0.4\",\n+ \"@riotjs/cli\": \"^6.0.5\",\n \"@riotjs/compiler\": \"^6.0.0\",\n \"chai\": \"^4.3.4\",\n \"esm\": \"^3.2.25\",\n", "diff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java\nnew file mode 100644\nindex 0000000..4abf2e3\n--- /dev/null\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java\n@@ -0,0 +1,28 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.engine.state.appliers;\n+\n+import io.camunda.zeebe.engine.state.TypedEventApplier;\n+import io.camunda.zeebe.engine.state.mutable.MutableDistributionState;\n+import io.camunda.zeebe.protocol.impl.record.value.distribution.CommandDistributionRecord;\n+import io.camunda.zeebe.protocol.record.intent.CommandDistributionIntent;\n+\n+public final class CommandDistributionAcknowledgedApplier\n+ implements TypedEventApplier<CommandDistributionIntent, CommandDistributionRecord> {\n+\n+ private final MutableDistributionState distributionState;\n+\n+ public CommandDistributionAcknowledgedApplier(final MutableDistributionState distributionState) {\n+ this.distributionState = distributionState;\n+ }\n+\n+ @Override\n+ public void applyState(final long key, final CommandDistributionRecord value) {\n+ distributionState.removePendingDistribution(key, value.getPartitionId());\n+ }\n+}\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\nindex a72309b..4793315 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n@@ -284,6 +284,9 @@ public final class EventAppliers implements EventApplier {\n CommandDistributionIntent.DISTRIBUTING,\n new CommandDistributionDistributingApplier(distributionState));\n register(\n+ CommandDistributionIntent.ACKNOWLEDGED,\n+ new CommandDistributionAcknowledgedApplier(distributionState));\n+ register(\n CommandDistributionIntent.FINISHED,\n new CommandDistributionFinishedApplier(distributionState));\n }\n"]
2
["5d256f937f93e5a5ed003df86d38c44834095a11", "6f4c06076abff94f8bb5c634beaba55483a78b72"]
["build", "feat"]
fix deploy,add clean up test Add another clean up test, which verifies that the state is cleaned up after the timer (non-recurring) is triggered.
["diff --git a/.github/workflows/deploy.yaml b/.github/workflows/deploy.yaml\nindex 3830f4c..3b14ee5 100644\n--- a/.github/workflows/deploy.yaml\n+++ b/.github/workflows/deploy.yaml\n@@ -67,7 +67,7 @@ jobs:\n run: aws s3 cp .next/static s3://cdn.rs.school/_next/static/ --recursive --cache-control \"public,max-age=15552000,immutable\"\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n@@ -117,7 +117,7 @@ jobs:\n run: npm run build\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n@@ -167,7 +167,7 @@ jobs:\n run: npm run build\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n", "diff --git a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\nindex d36b4c9..ca5047f 100644\n--- a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n@@ -630,6 +630,40 @@ public final class ProcessExecutionCleanStateTest {\n }\n \n @Test\n+ public void testProcessWithTriggerTimerStartEvent() {\n+ // given\n+ final var deployment =\n+ engineRule\n+ .deployment()\n+ .withXmlResource(\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .timerWithDate(\"=now() + duration(\\\"PT15S\\\")\")\n+ .endEvent()\n+ .done())\n+ .deploy();\n+\n+ final var processDefinitionKey =\n+ deployment.getValue().getProcessesMetadata().get(0).getProcessDefinitionKey();\n+\n+ // when\n+ engineRule.awaitProcessingOf(\n+ RecordingExporter.timerRecords(TimerIntent.CREATED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .getFirst());\n+\n+ engineRule.increaseTime(Duration.ofSeconds(15));\n+\n+ RecordingExporter.processInstanceRecords(ProcessInstanceIntent.ELEMENT_COMPLETED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .withElementType(BpmnElementType.PROCESS)\n+ .await();\n+\n+ // then\n+ assertThatStateIsEmpty();\n+ }\n+\n+ @Test\n public void testProcessWithTimerStartEventRedeployment() {\n // given\n final var deployment =\n"]
2
["7785be09053049b30cf41b420c59f051cd0129fc", "aa746b764e6c54bbbd631210fce35df842d09b12"]
["ci", "test"]
use lambda to define backend operations,build improvements
["diff --git a/ibis/backends/duckdb/registry.py b/ibis/backends/duckdb/registry.py\nindex 20ffd6f..3f56f2a 100644\n--- a/ibis/backends/duckdb/registry.py\n+++ b/ibis/backends/duckdb/registry.py\n@@ -107,28 +107,13 @@ def _literal(_, op):\n return sa.cast(sa.literal(value), sqla_type)\n \n \n-def _array_column(t, op):\n- (arg,) = op.args\n- sqla_type = to_sqla_type(op.output_dtype)\n- return sa.cast(sa.func.list_value(*map(t.translate, arg)), sqla_type)\n-\n-\n def _neg_idx_to_pos(array, idx):\n if_ = getattr(sa.func, \"if\")\n arg_length = sa.func.array_length(array)\n return if_(idx < 0, arg_length + sa.func.greatest(idx, -arg_length), idx)\n \n \n-def _struct_field(t, op):\n- return sa.func.struct_extract(\n- t.translate(op.arg),\n- sa.text(repr(op.field)),\n- type_=to_sqla_type(op.output_dtype),\n- )\n-\n-\n-def _regex_extract(t, op):\n- string, pattern, index = map(t.translate, op.args)\n+def _regex_extract(string, pattern, index):\n result = sa.case(\n [\n (\n@@ -149,8 +134,7 @@ def _regex_extract(t, op):\n return result\n \n \n-def _json_get_item(t, op):\n- left, path = map(t.translate, op.args)\n+def _json_get_item(left, path):\n # Workaround for https://github.com/duckdb/duckdb/issues/5063\n # In some situations duckdb silently does the wrong thing if\n # the path is parametrized.\n@@ -197,7 +181,12 @@ def _struct_column(t, op):\n \n operation_registry.update(\n {\n- ops.ArrayColumn: _array_column,\n+ ops.ArrayColumn: (\n+ lambda t, op: sa.cast(\n+ sa.func.list_value(*map(t.translate, op.cols)),\n+ to_sqla_type(op.output_dtype),\n+ )\n+ ),\n ops.ArrayConcat: fixed_arity(sa.func.array_concat, 2),\n ops.ArrayRepeat: fixed_arity(\n lambda arg, times: sa.func.flatten(\n@@ -222,7 +211,13 @@ operation_registry.update(\n # TODO: map operations, but DuckDB's maps are multimaps\n ops.Modulus: fixed_arity(operator.mod, 2),\n ops.Round: _round,\n- ops.StructField: _struct_field,\n+ ops.StructField: (\n+ lambda t, op: sa.func.struct_extract(\n+ t.translate(op.arg),\n+ sa.text(repr(op.field)),\n+ type_=to_sqla_type(op.output_dtype),\n+ )\n+ ),\n ops.TableColumn: _table_column,\n ops.TimestampDiff: fixed_arity(sa.func.age, 2),\n ops.TimestampFromUNIX: _timestamp_from_unix,\n@@ -232,7 +227,7 @@ operation_registry.update(\n lambda *_: sa.cast(sa.func.now(), sa.TIMESTAMP),\n 0,\n ),\n- ops.RegexExtract: _regex_extract,\n+ ops.RegexExtract: fixed_arity(_regex_extract, 3),\n ops.RegexReplace: fixed_arity(\n lambda *args: sa.func.regexp_replace(*args, \"g\"), 3\n ),\n@@ -255,7 +250,7 @@ operation_registry.update(\n ops.ArgMin: reduction(sa.func.min_by),\n ops.ArgMax: reduction(sa.func.max_by),\n ops.BitwiseXor: fixed_arity(sa.func.xor, 2),\n- ops.JSONGetItem: _json_get_item,\n+ ops.JSONGetItem: fixed_arity(_json_get_item, 2),\n ops.RowID: lambda *_: sa.literal_column('rowid'),\n ops.StringToTimestamp: fixed_arity(sa.func.strptime, 2),\n }\n", "diff --git a/.travis.yml b/.travis.yml\nindex 9e1b926..3144244 100644\n--- a/.travis.yml\n+++ b/.travis.yml\n@@ -1,5 +1,6 @@\n language: node_js\n dist: trusty\n+sudo: required\n node_js:\n - '6.9.5'\n before_install:\ndiff --git a/e2e/schematics/command-line.test.ts b/e2e/schematics/command-line.test.ts\nindex 16d8b34..ea91494 100644\n--- a/e2e/schematics/command-line.test.ts\n+++ b/e2e/schematics/command-line.test.ts\n@@ -68,8 +68,6 @@ describe('Command line', () => {\n \n updateFile('apps/myapp/src/app/app.component.spec.ts', `import '@nrwl/mylib';`);\n \n- updateRunAffectedToWorkInE2ESetup();\n-\n const affectedApps = runCommand('npm run affected:apps -- --files=\"libs/mylib/index.ts\"');\n expect(affectedApps).toContain('myapp');\n expect(affectedApps).not.toContain('myapp2');\n@@ -147,11 +145,3 @@ describe('Command line', () => {\n 1000000\n );\n });\n-\n-function updateRunAffectedToWorkInE2ESetup() {\n- const runAffected = readFile('node_modules/@nrwl/schematics/src/command-line/affected.js');\n- const newRunAffected = runAffected\n- .replace('ng build', '../../node_modules/.bin/ng build')\n- .replace('ng e2e', '../../node_modules/.bin/ng e2e');\n- updateFile('node_modules/@nrwl/schematics/src/command-line/affected.js', newRunAffected);\n-}\ndiff --git a/e2e/schematics/workspace.test.ts b/e2e/schematics/workspace.test.ts\nindex 8a41070..8749926 100644\n--- a/e2e/schematics/workspace.test.ts\n+++ b/e2e/schematics/workspace.test.ts\n@@ -82,7 +82,7 @@ describe('Nrwl Convert to Nx Workspace', () => {\n \n it('should generate a workspace and not change dependencies or devDependencies if they already exist', () => {\n // create a new AngularCLI app\n- runNgNew('--skip-install');\n+ runNgNew();\n const nxVersion = '0.0.0';\n const schematicsVersion = '0.0.0';\n const ngrxVersion = '0.0.0';\ndiff --git a/e2e/utils.ts b/e2e/utils.ts\nindex 422d866..a03104f 100644\n--- a/e2e/utils.ts\n+++ b/e2e/utils.ts\n@@ -17,8 +17,7 @@ export function newProject(): void {\n copyMissingPackages();\n execSync('mv ./tmp/proj ./tmp/proj_backup');\n }\n- execSync('cp -r ./tmp/proj_backup ./tmp/proj');\n- setUpSynLink();\n+ execSync('cp -a ./tmp/proj_backup ./tmp/proj');\n }\n \n export function copyMissingPackages(): void {\n@@ -26,14 +25,9 @@ export function copyMissingPackages(): void {\n modulesToCopy.forEach(m => copyNodeModule(projectName, m));\n }\n \n-export function setUpSynLink(): void {\n- execSync(`ln -s ../@nrwl/schematics/src/command-line/nx.js tmp/${projectName}/node_modules/.bin/nx`);\n- execSync(`chmod +x tmp/${projectName}/node_modules/.bin/nx`);\n-}\n-\n function copyNodeModule(path: string, name: string) {\n execSync(`rm -rf tmp/${path}/node_modules/${name}`);\n- execSync(`cp -r node_modules/${name} tmp/${path}/node_modules/${name}`);\n+ execSync(`cp -a node_modules/${name} tmp/${path}/node_modules/${name}`);\n }\n \n export function runCLI(\n@@ -43,7 +37,7 @@ export function runCLI(\n }\n ): string {\n try {\n- return execSync(`../../node_modules/.bin/ng ${command}`, {\n+ return execSync(`./node_modules/.bin/ng ${command}`, {\n cwd: `./tmp/${projectName}`\n })\n .toString()\n@@ -67,7 +61,7 @@ export function newLib(name: string): string {\n }\n \n export function runSchematic(command: string): string {\n- return execSync(`../../node_modules/.bin/schematics ${command}`, {\n+ return execSync(`./node_modules/.bin/schematics ${command}`, {\n cwd: `./tmp/${projectName}`\n }).toString();\n }\ndiff --git a/package.json b/package.json\nindex bef54f8..9186a58 100644\n--- a/package.json\n+++ b/package.json\n@@ -6,7 +6,7 @@\n \"private\": true,\n \"scripts\": {\n \"build\": \"./scripts/build.sh\",\n- \"e2e\": \"yarn build && ./scripts/e2e.sh\",\n+ \"e2e\": \"./scripts/e2e.sh\",\n \"format\": \"./scripts/format.sh\",\n \"linknpm\": \"./scripts/link.sh\",\n \"package\": \"./scripts/package.sh\",\n@@ -14,7 +14,7 @@\n \"copy\": \"./scripts/copy.sh\",\n \"test:schematics\": \"yarn build && ./scripts/test_schematics.sh\",\n \"test:nx\": \"yarn build && ./scripts/test_nx.sh\",\n- \"test\": \"yarn build && ./scripts/test_nx.sh && ./scripts/test_schematics.sh\",\n+ \"test\": \"yarn linknpm && ./scripts/test_nx.sh && ./scripts/test_schematics.sh\",\n \"checkformat\": \"./scripts/check-format.sh\",\n \"publish_npm\": \"./scripts/publish.sh\"\n },\ndiff --git a/packages/schematics/src/collection/workspace/index.ts b/packages/schematics/src/collection/workspace/index.ts\nindex 8f8897f..c70d161 100644\n--- a/packages/schematics/src/collection/workspace/index.ts\n+++ b/packages/schematics/src/collection/workspace/index.ts\n@@ -254,20 +254,7 @@ function moveFiles(options: Schema) {\n \n function copyAngularCliTgz() {\n return (host: Tree) => {\n- copyFile(\n- path.join(\n- 'node_modules',\n- '@nrwl',\n- 'schematics',\n- 'src',\n- 'collection',\n- 'application',\n- 'files',\n- '__directory__',\n- '.angular_cli.tgz'\n- ),\n- '.'\n- );\n+ copyFile(path.join(__dirname, '..', 'application', 'files', '__directory__', '.angular_cli.tgz'), '.');\n return host;\n };\n }\ndiff --git a/packages/schematics/src/command-line/affected.ts b/packages/schematics/src/command-line/affected.ts\nindex b7f9173..89a4f72 100644\n--- a/packages/schematics/src/command-line/affected.ts\n+++ b/packages/schematics/src/command-line/affected.ts\n@@ -1,5 +1,7 @@\n import { execSync } from 'child_process';\n import { getAffectedApps, parseFiles } from './shared';\n+import * as path from 'path';\n+import * as resolve from 'resolve';\n \n export function affected(args: string[]): void {\n const command = args[0];\n@@ -39,7 +41,7 @@ function build(apps: string[], rest: string[]) {\n if (apps.length > 0) {\n console.log(`Building ${apps.join(', ')}`);\n apps.forEach(app => {\n- execSync(`ng build ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n+ execSync(`${ngPath()} build ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n });\n } else {\n console.log('No apps to build');\n@@ -50,9 +52,13 @@ function e2e(apps: string[], rest: string[]) {\n if (apps.length > 0) {\n console.log(`Testing ${apps.join(', ')}`);\n apps.forEach(app => {\n- execSync(`ng e2e ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n+ execSync(`${ngPath()} e2e ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\n });\n } else {\n- console.log('No apps to tst');\n+ console.log('No apps to test');\n }\n }\n+\n+function ngPath() {\n+ return `${path.dirname(path.dirname(path.dirname(resolve.sync('@angular/cli', { basedir: __dirname }))))}/bin/ng`;\n+}\ndiff --git a/scripts/build.sh b/scripts/build.sh\nindex ac533b5..9b8891b 100755\n--- a/scripts/build.sh\n+++ b/scripts/build.sh\n@@ -3,6 +3,8 @@\n rm -rf build\n ngc\n rsync -a --exclude=*.ts packages/ build/packages\n+chmod +x build/packages/schematics/bin/create-nx-workspace.js\n+chmod +x build/packages/schematics/src/command-line/nx.js\n rm -rf build/packages/install\n cp README.md build/packages/schematics\n cp README.md build/packages/nx\n\\ No newline at end of file\n"]
2
["5d14de6722eb34c6604a124f6f11cb711f16bd44", "e0a977b2d316e7612b5d72cb02cd7d78e75dbc55"]
["refactor", "build"]
add instruction for finding version,getBorderSize() missing "width" The correct property name to use is "borderWidth", not just "border". "border" works in Chrome but was breaking in Firefox. Also had to change .ui-grid-header's box-sizing to content-box so IE11 would include the border in height calcs. AND finally IE11 was returning fractional heights so Grid parseInt()s the returned values.
["diff --git a/.github/ISSUE_TEMPLATE/_bug_report_chs.md b/.github/ISSUE_TEMPLATE/_bug_report_chs.md\nindex 42a2e0f..44a33db 100644\n--- a/.github/ISSUE_TEMPLATE/_bug_report_chs.md\n+++ b/.github/ISSUE_TEMPLATE/_bug_report_chs.md\n@@ -36,7 +36,7 @@ assignees: ''\n ## \u8bbe\u5907\u4fe1\u606f\n - \u64cd\u4f5c\u7cfb\u7edf: [] <!-- \u5982 [Window10] -->\n - \u6d4f\u89c8\u5668\u7248\u672c: [] <!-- \u5982 [Chrome77] -->\n-- \u6c99\u62c9\u67e5\u8bcd\u7248\u672c: [] <!-- \u5982 [v7.0.0] -->\n+- \u6c99\u62c9\u67e5\u8bcd\u7248\u672c: [] <!-- \u5982 [v7.0.0] \uff08\u5728\u6269\u5c55\u7ba1\u7406\u9875\u9762\u67e5\u770b\uff09 -->\n \n <!-- \u8bf7\u5728\u4e0b\u65b9 ## \u5f00\u5934\u884c\u4e4b\u95f4\u7684\u7a7a\u767d\u5904\u586b\u5199 -->\n \n", "diff --git a/src/js/core/factories/Grid.js b/src/js/core/factories/Grid.js\nindex dcf10af..2be7842 100644\n--- a/src/js/core/factories/Grid.js\n+++ b/src/js/core/factories/Grid.js\n@@ -1525,7 +1525,7 @@ angular.module('ui.grid')\n var oldHeaderHeight = container.headerHeight;\n var headerHeight = gridUtil.outerElementHeight(container.header);\n \n- container.headerHeight = headerHeight;\n+ container.headerHeight = parseInt(headerHeight, 10);\n \n if (oldHeaderHeight !== headerHeight) {\n rebuildStyles = true;\n@@ -1534,7 +1534,9 @@ angular.module('ui.grid')\n // Get the \"inner\" header height, that is the height minus the top and bottom borders, if present. We'll use it to make sure all the headers have a consistent height\n var topBorder = gridUtil.getBorderSize(container.header, 'top');\n var bottomBorder = gridUtil.getBorderSize(container.header, 'bottom');\n- var innerHeaderHeight = headerHeight - topBorder - bottomBorder;\n+ var innerHeaderHeight = parseInt(headerHeight - topBorder - bottomBorder, 10);\n+\n+ innerHeaderHeight = innerHeaderHeight < 0 ? 0 : innerHeaderHeight;\n \n container.innerHeaderHeight = innerHeaderHeight;\n \ndiff --git a/src/js/core/services/ui-grid-util.js b/src/js/core/services/ui-grid-util.js\nindex 2c32cbe..cc7c36c 100644\n--- a/src/js/core/services/ui-grid-util.js\n+++ b/src/js/core/services/ui-grid-util.js\n@@ -757,6 +757,8 @@ module.service('gridUtil', ['$log', '$window', '$document', '$http', '$templateC\n borderType = 'border';\n }\n \n+ borderType += 'Width';\n+\n var val = parseInt(styles[borderType], 10);\n \n if (isNaN(val)) {\ndiff --git a/src/less/header.less b/src/less/header.less\nindex 5468a43..de8ff0b 100644\n--- a/src/less/header.less\n+++ b/src/less/header.less\n@@ -7,6 +7,7 @@\n \n .ui-grid-header {\n border-bottom: 1px solid @borderColor;\n+ box-sizing: content-box;;\n }\n \n .ui-grid-top-panel {\n"]
2
["af0a5f7ab9d71fe20aa0888f682368f32b26fe18", "174f25214caa10ec643db6c81aaa0f3511bf78f4"]
["docs", "fix"]
add ability to get all encoded values,adds test for exec with http proxy Signed-off-by: Sam Alba <[email protected]>
["diff --git a/delorean_mem_qe/src/column.rs b/delorean_mem_qe/src/column.rs\nindex bc89cb2..b3df18e 100644\n--- a/delorean_mem_qe/src/column.rs\n+++ b/delorean_mem_qe/src/column.rs\n@@ -537,6 +537,22 @@ impl Column {\n }\n }\n \n+ /// Materialise all of the encoded values.\n+ pub fn all_encoded_values(&self) -> Vector {\n+ match self {\n+ Column::String(c) => {\n+ let now = std::time::Instant::now();\n+ let v = c.all_encoded_values();\n+ log::debug!(\"time getting all encoded values {:?}\", now.elapsed());\n+\n+ log::debug!(\"dictionary {:?}\", c.data.dictionary());\n+ Vector::Integer(v)\n+ }\n+ Column::Float(c) => Vector::Float(c.all_encoded_values()),\n+ Column::Integer(c) => Vector::Integer(c.all_encoded_values()),\n+ }\n+ }\n+\n /// Given an encoded value for a row, materialise and return the decoded\n /// version.\n ///\n@@ -986,6 +1002,10 @@ impl String {\n self.data.encoded_values(row_ids)\n }\n \n+ pub fn all_encoded_values(&self) -> Vec<i64> {\n+ self.data.all_encoded_values()\n+ }\n+\n /// Return the decoded value for an encoded ID.\n ///\n /// Panics if there is no decoded value for the provided id\n@@ -1037,6 +1057,10 @@ impl Float {\n self.data.encoded_values(row_ids)\n }\n \n+ pub fn all_encoded_values(&self) -> Vec<f64> {\n+ self.data.all_encoded_values()\n+ }\n+\n pub fn scan_from(&self, row_id: usize) -> &[f64] {\n self.data.scan_from(row_id)\n }\n@@ -1106,6 +1130,10 @@ impl Integer {\n self.data.encoded_values(row_ids)\n }\n \n+ pub fn all_encoded_values(&self) -> Vec<i64> {\n+ self.data.all_encoded_values()\n+ }\n+\n pub fn scan_from(&self, row_id: usize) -> &[i64] {\n self.data.scan_from(row_id)\n }\ndiff --git a/delorean_mem_qe/src/encoding.rs b/delorean_mem_qe/src/encoding.rs\nindex d6a865a..4b057cf 100644\n--- a/delorean_mem_qe/src/encoding.rs\n+++ b/delorean_mem_qe/src/encoding.rs\n@@ -68,6 +68,12 @@ where\n self.values(row_ids)\n }\n \n+ /// Return all encoded values. For this encoding this is just the decoded\n+ /// values\n+ pub fn all_encoded_values(&self) -> Vec<T> {\n+ self.values.clone()\n+ }\n+\n // TODO(edd): fix this when added NULL support\n pub fn scan_from_until_some(&self, _row_id: usize) -> Option<T> {\n unreachable!(\"to remove\");\n@@ -485,6 +491,26 @@ impl DictionaryRLE {\n out\n }\n \n+ // values materialises a vector of references to all logical values in the\n+ // encoding.\n+ pub fn all_values(&mut self) -> Vec<Option<&String>> {\n+ let mut out: Vec<Option<&String>> = Vec::with_capacity(self.total as usize);\n+\n+ // build reverse mapping.\n+ let mut idx_value = BTreeMap::new();\n+ for (k, v) in &self.entry_index {\n+ idx_value.insert(v, k);\n+ }\n+ assert_eq!(idx_value.len(), self.entry_index.len());\n+\n+ for (idx, rl) in &self.run_lengths {\n+ // TODO(edd): fix unwrap - we know that the value exists in map...\n+ let v = idx_value.get(&idx).unwrap().as_ref();\n+ out.extend(iter::repeat(v).take(*rl as usize));\n+ }\n+ out\n+ }\n+\n /// Return the decoded value for an encoded ID.\n ///\n /// Panics if there is no decoded value for the provided id\n@@ -528,22 +554,13 @@ impl DictionaryRLE {\n out\n }\n \n- // values materialises a vector of references to all logical values in the\n- // encoding.\n- pub fn all_values(&mut self) -> Vec<Option<&String>> {\n- let mut out: Vec<Option<&String>> = Vec::with_capacity(self.total as usize);\n-\n- // build reverse mapping.\n- let mut idx_value = BTreeMap::new();\n- for (k, v) in &self.entry_index {\n- idx_value.insert(v, k);\n- }\n- assert_eq!(idx_value.len(), self.entry_index.len());\n+ // all_encoded_values materialises a vector of all encoded values for the\n+ // column.\n+ pub fn all_encoded_values(&self) -> Vec<i64> {\n+ let mut out: Vec<i64> = Vec::with_capacity(self.total as usize);\n \n for (idx, rl) in &self.run_lengths {\n- // TODO(edd): fix unwrap - we know that the value exists in map...\n- let v = idx_value.get(&idx).unwrap().as_ref();\n- out.extend(iter::repeat(v).take(*rl as usize));\n+ out.extend(iter::repeat(*idx as i64).take(*rl as usize));\n }\n out\n }\ndiff --git a/delorean_mem_qe/src/segment.rs b/delorean_mem_qe/src/segment.rs\nindex c058df0..f8c5005 100644\n--- a/delorean_mem_qe/src/segment.rs\n+++ b/delorean_mem_qe/src/segment.rs\n@@ -228,7 +228,7 @@ impl Segment {\n group_columns: &[String],\n aggregates: &[(String, AggregateType)],\n window: i64,\n- ) -> BTreeMap<Vec<String>, Vec<(String, Option<column::Aggregate>)>> {\n+ ) -> BTreeMap<Vec<i64>, Vec<(&String, &AggregateType, Option<column::Aggregate>)>> {\n // Build a hash table - essentially, scan columns for matching row ids,\n // emitting the encoded value for each column and track those value\n // combinations in a hashmap with running aggregates.\n@@ -242,6 +242,10 @@ impl Segment {\n assert_ne!(group_columns[group_columns.len() - 1], \"time\");\n }\n \n+ // TODO(edd): Perf - if there is no predicate and we want entire segment\n+ // then it will be a lot faster to not build filtered_row_ids and just\n+ // get all encoded values for each grouping column...\n+\n // filter on predicates and time\n let filtered_row_ids: croaring::Bitmap;\n if let Some(row_ids) = self.filter_by_predicates_eq(time_range, predicates) {\n@@ -263,7 +267,12 @@ impl Segment {\n let mut group_column_encoded_values = Vec::with_capacity(group_columns.len());\n for group_column in group_columns {\n if let Some(column) = self.column(&group_column) {\n- let encoded_values = column.encoded_values(&filtered_row_ids_vec);\n+ let encoded_values = if filtered_row_ids_vec.len() == self.meta.rows {\n+ column.all_encoded_values()\n+ } else {\n+ column.encoded_values(&filtered_row_ids_vec)\n+ };\n+\n assert_eq!(\n filtered_row_ids.cardinality() as usize,\n encoded_values.len()\n@@ -325,10 +334,10 @@ impl Segment {\n .collect::<Vec<_>>();\n \n // hashMap is about 20% faster than BTreeMap in this case\n- let mut hash_table: HashMap<\n+ let mut hash_table: BTreeMap<\n Vec<i64>,\n Vec<(&String, &AggregateType, Option<column::Aggregate>)>,\n- > = HashMap::new();\n+ > = BTreeMap::new();\n \n let mut aggregate_row: Vec<(&str, Option<column::Scalar>)> =\n std::iter::repeat_with(|| (\"\", None))\n@@ -406,8 +415,10 @@ impl Segment {\n }\n processed_rows += 1;\n }\n+ // println!(\"groups: {:?}\", hash_table.len());\n log::debug!(\"({:?} rows processed) {:?}\", processed_rows, hash_table);\n BTreeMap::new()\n+ // hash_table\n }\n \n pub fn aggregate_by_group_using_sort(\n@@ -451,7 +462,11 @@ impl Segment {\n let mut group_column_encoded_values = Vec::with_capacity(group_columns.len());\n for group_column in group_columns {\n if let Some(column) = self.column(&group_column) {\n- let encoded_values = column.encoded_values(&filtered_row_ids_vec);\n+ let encoded_values = if filtered_row_ids_vec.len() == self.meta.rows {\n+ column.all_encoded_values()\n+ } else {\n+ column.encoded_values(&filtered_row_ids_vec)\n+ };\n assert_eq!(\n filtered_row_ids.cardinality() as usize,\n encoded_values.len()\n@@ -557,6 +572,10 @@ impl Segment {\n assert_ne!(group_columns[group_columns.len() - 1], \"time\");\n }\n \n+ // TODO(edd): Perf - if there is no predicate and we want entire segment\n+ // then it will be a lot faster to not build filtered_row_ids and just\n+ // get all encoded values for each grouping column...\n+\n // filter on predicates and time\n let filtered_row_ids: croaring::Bitmap;\n if let Some(row_ids) = self.filter_by_predicates_eq(time_range, predicates) {\n@@ -577,7 +596,11 @@ impl Segment {\n let mut group_column_encoded_values = Vec::with_capacity(group_columns.len());\n for group_column in group_columns {\n if let Some(column) = self.column(&group_column) {\n- let encoded_values = column.encoded_values(&filtered_row_ids_vec);\n+ let encoded_values = if filtered_row_ids_vec.len() == self.meta.rows {\n+ column.all_encoded_values()\n+ } else {\n+ column.encoded_values(&filtered_row_ids_vec)\n+ };\n assert_eq!(\n filtered_row_ids.cardinality() as usize,\n encoded_values.len()\n@@ -709,6 +732,7 @@ impl Segment {\n aggregates: group_key_aggregates,\n });\n \n+ // println!(\"groups: {:?}\", results.len());\n log::debug!(\"({:?} rows processed) {:?}\", processed_rows, results);\n // results\n vec![]\n", "diff --git a/tests/tasks.bats b/tests/tasks.bats\nindex e3b95c4..4cfba15 100644\n--- a/tests/tasks.bats\n+++ b/tests/tasks.bats\n@@ -101,6 +101,14 @@ setup() {\n assert_line --partial --index 9 'actions.basicTest.stop'\n }\n \n+@test \"task: #Exec with HTTP proxy\" {\n+ cd ./tasks/exec\n+ export HTTPS_PROXY=\"https://localhost:4242/\"\n+ run \"$DAGGER\" \"do\" -p ./http_proxy.cue curlProxy\n+ assert_failure\n+ unset HTTP_PROXY\n+}\n+\n @test \"task: #Start #Stop params\" {\n cd ./tasks/exec\n \"$DAGGER\" \"do\" -p ./start_stop_exec.cue execParamsTest\n@@ -297,4 +305,4 @@ setup() {\n \n @test \"task: #Rm\" {\n \"$DAGGER\" \"do\" -p ./tasks/rm/rm.cue test\n-}\n\\ No newline at end of file\n+}\ndiff --git a/tests/tasks/exec/http_proxy.cue b/tests/tasks/exec/http_proxy.cue\nnew file mode 100644\nindex 0000000..05de4b9\n--- /dev/null\n+++ b/tests/tasks/exec/http_proxy.cue\n@@ -0,0 +1,25 @@\n+package main\n+\n+import (\n+\t\"dagger.io/dagger\"\n+\t\"dagger.io/dagger/core\"\n+)\n+\n+dagger.#Plan & {\n+\tactions: {\n+\t\timage: core.#Pull & {\n+\t\t\tsource: \"alpine:3.15.0@sha256:e7d88de73db3d3fd9b2d63aa7f447a10fd0220b7cbf39803c803f2af9ba256b3\"\n+\t\t}\n+\n+\t\tcurlProxy: core.#Exec & {\n+\t\t\tinput: image.output\n+\t\t\targs: [\n+\t\t\t\t\"sh\", \"-c\",\n+\t\t\t\t\"\"\"\n+\t\t\t\t\tapk add --no-cache curl\n+\t\t\t\t\tcurl -sfL -o /dev/null https://www.google.com/\n+\t\t\t\t\t\"\"\",\n+\t\t\t]\n+\t\t}\n+\t}\n+}\n"]
2
["cad5e45208346528ad02cd04dcac863f90faa037", "6c7398993bc567ec84e4573b6ededbf50b1ef606"]
["feat", "test"]
refactor generate_completion,add documentation to use react-native-paper with CRA (#874)
["diff --git a/src/lib.rs b/src/lib.rs\nindex dfd8014..15850f7 100644\n--- a/src/lib.rs\n+++ b/src/lib.rs\n@@ -1,11 +1,106 @@\n //! Generates [Nushell](https://github.com/nushell/nushell) completions for [`clap`](https://github.com/clap-rs/clap) based CLIs\n \n-use clap::Command;\n+use clap::{Arg, Command};\n use clap_complete::Generator;\n \n /// Generate Nushell complete file\n pub struct Nushell;\n \n+enum Argument {\n+ Short(char),\n+ Long(String),\n+ ShortAndLong(char, String),\n+ Positional(String, bool),\n+}\n+\n+struct ArgumentLine {\n+ arg: Argument,\n+ takes_values: bool,\n+ help: Option<String>,\n+}\n+\n+impl From<&Arg> for ArgumentLine {\n+ fn from(arg: &Arg) -> Self {\n+ let takes_values = arg\n+ .get_num_args()\n+ .map(|v| v.takes_values())\n+ .unwrap_or(false);\n+\n+ let help = arg.get_help().map(|s| s.to_string());\n+\n+ if arg.is_positional() {\n+ let id = arg.get_id().to_string();\n+ let required = arg.is_required_set();\n+ let arg = Argument::Positional(id, required);\n+\n+ return Self {\n+ arg,\n+ takes_values,\n+ help,\n+ };\n+ }\n+\n+ let short = arg.get_short();\n+ let long = arg.get_long();\n+\n+ match short {\n+ Some(short) => match long {\n+ Some(long) => Self {\n+ arg: Argument::ShortAndLong(short, long.into()),\n+ takes_values,\n+ help,\n+ },\n+ None => Self {\n+ arg: Argument::Short(short),\n+ takes_values,\n+ help,\n+ },\n+ },\n+ None => match long {\n+ Some(long) => Self {\n+ arg: Argument::Long(long.into()),\n+ takes_values,\n+ help,\n+ },\n+ None => unreachable!(\"No short or long option found\"),\n+ },\n+ }\n+ }\n+}\n+\n+impl ToString for ArgumentLine {\n+ fn to_string(&self) -> String {\n+ let mut s = String::new();\n+\n+ match &self.arg {\n+ Argument::Short(short) => s.push_str(format!(\" -{}\", short).as_str()),\n+ Argument::Long(long) => s.push_str(format!(\" --{}\", long).as_str()),\n+ Argument::ShortAndLong(short, long) => {\n+ s.push_str(format!(\" --{}(-{})\", long, short).as_str())\n+ }\n+ Argument::Positional(positional, required) => {\n+ s.push_str(format!(\" {}\", positional).as_str());\n+\n+ if !*required {\n+ s.push('?');\n+ }\n+ }\n+ }\n+\n+ if self.takes_values {\n+ s.push_str(\": string\");\n+ }\n+\n+ if let Some(help) = &self.help {\n+ s.push_str(format!(\"\\t# {}\", help).as_str());\n+ }\n+\n+ s.push('\\n');\n+\n+ s\n+ }\n+}\n+\n impl Generator for Nushell {\n fn file_name(&self, name: &str) -> String {\n format!(\"{}.nu\", name)\n@@ -37,51 +132,18 @@ fn generate_completion(completions: &mut String, cmd: &Command, is_subcommand: b\n \n let bin_name = cmd.get_bin_name().expect(\"Failed to get bin name\");\n \n- if is_subcommand {\n- completions.push_str(format!(\" export extern \\\"{}\\\" [\\n\", bin_name).as_str());\n+ let name = if is_subcommand {\n+ format!(r#\"\"{}\"\"#, bin_name)\n } else {\n- completions.push_str(format!(\" export extern {} [\\n\", bin_name).as_str());\n- }\n+ bin_name.into()\n+ };\n \n- let mut s = String::new();\n- for arg in cmd.get_arguments() {\n- if arg.is_positional() {\n- s.push_str(format!(\" {}\", arg.get_id()).as_str());\n- if !arg.is_required_set() {\n- s.push('?');\n- }\n- }\n-\n- let long = arg.get_long();\n- if let Some(opt) = long {\n- s.push_str(format!(\" --{}\", opt).as_str());\n- }\n+ completions.push_str(format!(\" export extern {} [\\n\", name).as_str());\n \n- let short = arg.get_short();\n- if let Some(opt) = short {\n- if long.is_some() {\n- s.push_str(format!(\"(-{})\", opt).as_str());\n- } else {\n- s.push_str(format!(\" -{}\", opt).as_str());\n- }\n- }\n-\n- if let Some(v) = arg.get_num_args() {\n- if v.takes_values() {\n- // TODO: add more types?\n- // TODO: add possible values?\n- s.push_str(\": string\");\n- }\n- }\n-\n- if let Some(msg) = arg.get_help() {\n- if arg.is_positional() || long.is_some() || short.is_some() {\n- s.push_str(format!(\"\\t# {}\", msg).as_str());\n- }\n- }\n-\n- s.push('\\n');\n- }\n+ let s: String = cmd\n+ .get_arguments()\n+ .map(|arg| ArgumentLine::from(arg).to_string())\n+ .collect();\n \n completions.push_str(&s);\n completions.push_str(\" ]\\n\\n\");\n", "diff --git a/docs/pages/4.react-native-web.md b/docs/pages/4.react-native-web.md\nindex 69e4e52..8d6ae2a 100644\n--- a/docs/pages/4.react-native-web.md\n+++ b/docs/pages/4.react-native-web.md\n@@ -16,6 +16,63 @@ To install `react-native-web`, run:\n yarn add react-native-web react-dom react-art\n ```\n \n+### Using CRA ([Create React App](https://github.com/facebook/create-react-app))\n+\n+Install [`react-app-rewired`](https://github.com/timarney/react-app-rewired) to override `webpack` configuration:\n+\n+```sh\n+yarn add --dev react-app-rewired\n+```\n+\n+[Configure `babel-loader`](#2-configure-babel-loader) using a new file called `config-overrides.js`:\n+\n+```js\n+module.exports = function override(config, env) {\n+ config.module.rules.push({\n+ test: /\\.js$/,\n+ exclude: /node_modules[/\\\\](?!react-native-paper|react-native-vector-icons|react-native-safe-area-view)/,\n+ use: {\n+ loader: \"babel-loader\",\n+ options: {\n+ // Disable reading babel configuration\n+ babelrc: false,\n+ configFile: false,\n+\n+ // The configration for compilation\n+ presets: [\n+ [\"@babel/preset-env\", { useBuiltIns: \"usage\" }],\n+ \"@babel/preset-react\",\n+ \"@babel/preset-flow\"\n+ ],\n+ plugins: [\n+ \"@babel/plugin-proposal-class-properties\",\n+ \"@babel/plugin-proposal-object-rest-spread\"\n+ ]\n+ }\n+ }\n+ });\n+\n+ return config;\n+};\n+```\n+\n+Change your script in `package.json`:\n+\n+```diff\n+/* package.json */\n+\n+ \"scripts\": {\n+- \"start\": \"react-scripts start\",\n++ \"start\": \"react-app-rewired start\",\n+- \"build\": \"react-scripts build\",\n++ \"build\": \"react-app-rewired build\",\n+- \"test\": \"react-scripts test --env=jsdom\",\n++ \"test\": \"react-app-rewired test --env=jsdom\"\n+}\n+```\n+\n+### Custom webpack setup\n+\n To install `webpack`, run:\n \n ```sh\n"]
2
["f1bc5a554af4e617c7d7508f7f16f8fd25c78c91", "ee7cc5d5a940fba774e715b1f029c6361110b108"]
["refactor", "docs"]
refactor generate_completion,fix "types" field in dist
["diff --git a/src/lib.rs b/src/lib.rs\nindex dfd8014..15850f7 100644\n--- a/src/lib.rs\n+++ b/src/lib.rs\n@@ -1,11 +1,106 @@\n //! Generates [Nushell](https://github.com/nushell/nushell) completions for [`clap`](https://github.com/clap-rs/clap) based CLIs\n \n-use clap::Command;\n+use clap::{Arg, Command};\n use clap_complete::Generator;\n \n /// Generate Nushell complete file\n pub struct Nushell;\n \n+enum Argument {\n+ Short(char),\n+ Long(String),\n+ ShortAndLong(char, String),\n+ Positional(String, bool),\n+}\n+\n+struct ArgumentLine {\n+ arg: Argument,\n+ takes_values: bool,\n+ help: Option<String>,\n+}\n+\n+impl From<&Arg> for ArgumentLine {\n+ fn from(arg: &Arg) -> Self {\n+ let takes_values = arg\n+ .get_num_args()\n+ .map(|v| v.takes_values())\n+ .unwrap_or(false);\n+\n+ let help = arg.get_help().map(|s| s.to_string());\n+\n+ if arg.is_positional() {\n+ let id = arg.get_id().to_string();\n+ let required = arg.is_required_set();\n+ let arg = Argument::Positional(id, required);\n+\n+ return Self {\n+ arg,\n+ takes_values,\n+ help,\n+ };\n+ }\n+\n+ let short = arg.get_short();\n+ let long = arg.get_long();\n+\n+ match short {\n+ Some(short) => match long {\n+ Some(long) => Self {\n+ arg: Argument::ShortAndLong(short, long.into()),\n+ takes_values,\n+ help,\n+ },\n+ None => Self {\n+ arg: Argument::Short(short),\n+ takes_values,\n+ help,\n+ },\n+ },\n+ None => match long {\n+ Some(long) => Self {\n+ arg: Argument::Long(long.into()),\n+ takes_values,\n+ help,\n+ },\n+ None => unreachable!(\"No short or long option found\"),\n+ },\n+ }\n+ }\n+}\n+\n+impl ToString for ArgumentLine {\n+ fn to_string(&self) -> String {\n+ let mut s = String::new();\n+\n+ match &self.arg {\n+ Argument::Short(short) => s.push_str(format!(\" -{}\", short).as_str()),\n+ Argument::Long(long) => s.push_str(format!(\" --{}\", long).as_str()),\n+ Argument::ShortAndLong(short, long) => {\n+ s.push_str(format!(\" --{}(-{})\", long, short).as_str())\n+ }\n+ Argument::Positional(positional, required) => {\n+ s.push_str(format!(\" {}\", positional).as_str());\n+\n+ if !*required {\n+ s.push('?');\n+ }\n+ }\n+ }\n+\n+ if self.takes_values {\n+ s.push_str(\": string\");\n+ }\n+\n+ if let Some(help) = &self.help {\n+ s.push_str(format!(\"\\t# {}\", help).as_str());\n+ }\n+\n+ s.push('\\n');\n+\n+ s\n+ }\n+}\n+\n impl Generator for Nushell {\n fn file_name(&self, name: &str) -> String {\n format!(\"{}.nu\", name)\n@@ -37,51 +132,18 @@ fn generate_completion(completions: &mut String, cmd: &Command, is_subcommand: b\n \n let bin_name = cmd.get_bin_name().expect(\"Failed to get bin name\");\n \n- if is_subcommand {\n- completions.push_str(format!(\" export extern \\\"{}\\\" [\\n\", bin_name).as_str());\n+ let name = if is_subcommand {\n+ format!(r#\"\"{}\"\"#, bin_name)\n } else {\n- completions.push_str(format!(\" export extern {} [\\n\", bin_name).as_str());\n- }\n+ bin_name.into()\n+ };\n \n- let mut s = String::new();\n- for arg in cmd.get_arguments() {\n- if arg.is_positional() {\n- s.push_str(format!(\" {}\", arg.get_id()).as_str());\n- if !arg.is_required_set() {\n- s.push('?');\n- }\n- }\n-\n- let long = arg.get_long();\n- if let Some(opt) = long {\n- s.push_str(format!(\" --{}\", opt).as_str());\n- }\n+ completions.push_str(format!(\" export extern {} [\\n\", name).as_str());\n \n- let short = arg.get_short();\n- if let Some(opt) = short {\n- if long.is_some() {\n- s.push_str(format!(\"(-{})\", opt).as_str());\n- } else {\n- s.push_str(format!(\" -{}\", opt).as_str());\n- }\n- }\n-\n- if let Some(v) = arg.get_num_args() {\n- if v.takes_values() {\n- // TODO: add more types?\n- // TODO: add possible values?\n- s.push_str(\": string\");\n- }\n- }\n-\n- if let Some(msg) = arg.get_help() {\n- if arg.is_positional() || long.is_some() || short.is_some() {\n- s.push_str(format!(\"\\t# {}\", msg).as_str());\n- }\n- }\n-\n- s.push('\\n');\n- }\n+ let s: String = cmd\n+ .get_arguments()\n+ .map(|arg| ArgumentLine::from(arg).to_string())\n+ .collect();\n \n completions.push_str(&s);\n completions.push_str(\" ]\\n\\n\");\n", "diff --git a/scripts/prepare.js b/scripts/prepare.js\nindex 4bab09b..55f459b 100644\n--- a/scripts/prepare.js\n+++ b/scripts/prepare.js\n@@ -96,7 +96,6 @@ async function prepare() {\n delete json.private\n delete json.scripts\n delete json.devDependencies\n- delete json.types\n \n // Add \"postinstall\" script for donations.\n if (/(native|core)$/.test(name))\n@@ -128,6 +127,7 @@ async function prepare() {\n else {\n json.main = json.main.replace(/^dist\\//, '')\n if (json.main.endsWith('.cjs.js')) {\n+ json.types = json.main.replace('.cjs.js', '.d.ts')\n json.module = json.main.replace('.cjs', '')\n }\n }\n"]
2
["f1bc5a554af4e617c7d7508f7f16f8fd25c78c91", "f14ef3809f456aadd73523e47cb16c5d15e9a9df"]
["refactor", "build"]
test,post installers compatiblity with Windows #2520
["diff --git a/tests/playwright/pages/Dashboard/Command/CmdKPage.ts b/tests/playwright/pages/Dashboard/Command/CmdKPage.ts\nindex 5ac62b2..0457243 100644\n--- a/tests/playwright/pages/Dashboard/Command/CmdKPage.ts\n+++ b/tests/playwright/pages/Dashboard/Command/CmdKPage.ts\n@@ -21,6 +21,7 @@ export class CmdK extends BasePage {\n async searchText(text: string) {\n await this.dashboardPage.rootPage.fill('.cmdk-input', text);\n await this.rootPage.keyboard.press('Enter');\n+ await this.rootPage.keyboard.press('Enter');\n }\n \n async isCmdKVisible() {\n", "diff --git a/packages/cubejs-databricks-jdbc-driver/package.json b/packages/cubejs-databricks-jdbc-driver/package.json\nindex cc164f0..fd7ad45 100644\n--- a/packages/cubejs-databricks-jdbc-driver/package.json\n+++ b/packages/cubejs-databricks-jdbc-driver/package.json\n@@ -14,13 +14,16 @@\n },\n \"main\": \"dist/src/index.js\",\n \"typings\": \"dist/src/index.d.ts\",\n+ \"bin\": {\n+ \"databricks-jdbc-installer\": \"bin/post-install\"\n+ },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n \"tsc\": \"tsc\",\n \"watch\": \"tsc -w\",\n \"lint\": \"eslint src/* --ext .ts\",\n \"lint:fix\": \"eslint --fix src/* --ext .ts\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"databricks-jdbc-installer\"\n },\n \"files\": [\n \"README.md\",\ndiff --git a/rust/package.json b/rust/package.json\nindex b139279..5bf6446 100644\n--- a/rust/package.json\n+++ b/rust/package.json\n@@ -8,7 +8,8 @@\n \"node\": \">=10.8.0\"\n },\n \"bin\": {\n- \"cubestore-dev\": \"bin/cubestore-dev\"\n+ \"cubestore-dev\": \"bin/cubestore-dev\",\n+ \"cubestore-installer\": \"bin/post-install\"\n },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n@@ -18,7 +19,7 @@\n \"lint:fix\": \"eslint --fix js-wrapper/* --ext .ts,js\",\n \"unit\": \"jest\",\n \"unit:debug\": \"jest --runInBand\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"cubestore-installer\"\n },\n \"files\": [\n \"dist\",\ndiff --git a/yarn.lock b/yarn.lock\nindex d2a4038..b59bb77 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -4036,9 +4036,9 @@\n integrity sha512-7btbphLrKvo5yl/5CC2OCxUSMx1wV1wvGT1qDXkSt7yi00/YW7E8k6qzXqJHsp+WU0eoG7r6MTQQXI9lIvd0qA==\n \n \"@types/fs-extra@^9.0.1\", \"@types/fs-extra@^9.0.2\", \"@types/fs-extra@^9.0.8\":\n- version \"9.0.10\"\n- resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.10.tgz#8023a72e3d06cf54929ea47ec7634e47f33f4046\"\n- integrity sha512-O9T2LLkRDiTlalOBdjEkcnT0MRdT2+wglCl7pJUJ3mkWkR8hX4K+5bg2raQNJcLv4V8zGuTXe7Ud3wSqkTyuyQ==\n+ version \"9.0.11\"\n+ resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.11.tgz#8cc99e103499eab9f347dbc6ca4e99fb8d2c2b87\"\n+ integrity sha512-mZsifGG4QeQ7hlkhO56u7zt/ycBgGxSVsFI/6lGTU34VtwkiqrrSDgw0+ygs8kFGWcXnFQWMrzF2h7TtDFNixA==\n dependencies:\n \"@types/node\" \"*\"\n \n@@ -5306,9 +5306,9 @@ acorn@^7.0.0, acorn@^7.1.0, acorn@^7.1.1, acorn@^7.4.0:\n integrity sha512-nQyp0o1/mNdbTO1PO6kHkwSrmgZ0MT/jCCpNiwbUjGoRN4dlBhqJtoQuCnEOKzgTVwg0ZWiCoQy6SxMebQVh8A==\n \n acorn@^8.1.0:\n- version \"8.1.0\"\n- resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.0.tgz#52311fd7037ae119cbb134309e901aa46295b3fe\"\n- integrity sha512-LWCF/Wn0nfHOmJ9rzQApGnxnvgfROzGilS8936rqN/lfcYkY9MYZzdMqN+2NJ4SlTc+m5HiSa+kNfDtI64dwUA==\n+ version \"8.1.1\"\n+ resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.1.tgz#fb0026885b9ac9f48bac1e185e4af472971149ff\"\n+ integrity sha512-xYiIVjNuqtKXMxlRMDc6mZUhXehod4a3gbZ1qRlM7icK4EbxUFNLhWoPblCvFtB2Y9CIqHP3CF/rdxLItaQv8g==\n \n adal-node@^0.1.28:\n version \"0.1.28\"\n@@ -5441,9 +5441,9 @@ ajv@^6.1.0, ajv@^6.10.0, ajv@^6.10.2, ajv@^6.12.2, ajv@^6.12.3, ajv@^6.12.4, ajv\n uri-js \"^4.2.2\"\n \n ajv@^8.0.1:\n- version \"8.0.5\"\n- resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.0.5.tgz#f07d6fdeffcdbb80485570ce3f1bc845fcc812b9\"\n- integrity sha512-RkiLa/AeJx7+9OvniQ/qeWu0w74A8DiPPBclQ6ji3ZQkv5KamO+QGpqmi7O4JIw3rHGUXZ6CoP9tsAkn3gyazg==\n+ version \"8.1.0\"\n+ resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.1.0.tgz#45d5d3d36c7cdd808930cc3e603cf6200dbeb736\"\n+ integrity sha512-B/Sk2Ix7A36fs/ZkuGLIR86EdjbgR6fsAcbx9lOP/QBSXujDNbVmIS/U4Itz5k8fPFDeVZl/zQ/gJW4Jrq6XjQ==\n dependencies:\n fast-deep-equal \"^3.1.1\"\n json-schema-traverse \"^1.0.0\"\n@@ -6828,15 +6828,15 @@ [email protected]:\n node-releases \"^1.1.61\"\n \n browserslist@^4.0.0, browserslist@^4.11.1, browserslist@^4.12.0, browserslist@^4.14.5, browserslist@^4.16.3, browserslist@^4.3.4, browserslist@^4.6.2, browserslist@^4.6.4, browserslist@^4.7.0, browserslist@^4.9.1:\n- version \"4.16.3\"\n- resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.3.tgz#340aa46940d7db878748567c5dea24a48ddf3717\"\n- integrity sha512-vIyhWmIkULaq04Gt93txdh+j02yX/JzlyhLYbV3YQCn/zvES3JnY7TifHHvvr1w5hTDluNKMkV05cs4vy8Q7sw==\n+ version \"4.16.4\"\n+ resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.4.tgz#7ebf913487f40caf4637b892b268069951c35d58\"\n+ integrity sha512-d7rCxYV8I9kj41RH8UKYnvDYCRENUlHRgyXy/Rhr/1BaeLGfiCptEdFE8MIrvGfWbBFNjVYx76SQWvNX1j+/cQ==\n dependencies:\n- caniuse-lite \"^1.0.30001181\"\n- colorette \"^1.2.1\"\n- electron-to-chromium \"^1.3.649\"\n+ caniuse-lite \"^1.0.30001208\"\n+ colorette \"^1.2.2\"\n+ electron-to-chromium \"^1.3.712\"\n escalade \"^3.1.1\"\n- node-releases \"^1.1.70\"\n+ node-releases \"^1.1.71\"\n \n [email protected]:\n version \"0.2.6\"\n@@ -7217,7 +7217,7 @@ caniuse-api@^3.0.0:\n lodash.memoize \"^4.1.2\"\n lodash.uniq \"^4.5.0\"\n \n-caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001181:\n+caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001208:\n version \"1.0.30001208\"\n resolved \"https://registry.yarnpkg.com/caniuse-lite/-/caniuse-lite-1.0.30001208.tgz#a999014a35cebd4f98c405930a057a0d75352eb9\"\n integrity sha512-OE5UE4+nBOro8Dyvv0lfx+SRtfVIOM9uhKqFmJeUbGriqhhStgp1A0OyBpgy3OUF8AhYCT+PVwPC1gMl2ZcQMA==\n@@ -9549,10 +9549,10 @@ ejs@^2.6.1:\n resolved \"https://registry.yarnpkg.com/ejs/-/ejs-2.7.4.tgz#48661287573dcc53e366c7a1ae52c3a120eec9ba\"\n integrity sha512-7vmuyh5+kuUyJKePhQfRQBhXV5Ce+RnaeeQArKu1EAMpL3WbgMt5WG6uQZpEVvYSSsxMXRKOewtDk9RaTKXRlA==\n \n-electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.649:\n- version \"1.3.711\"\n- resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.711.tgz#92c3caf7ffed5e18bf63f66b4b57b4db2409c450\"\n- integrity sha512-XbklBVCDiUeho0PZQCjC25Ha6uBwqqJeyDhPLwLwfWRAo4x+FZFsmu1pPPkXT+B4MQMQoQULfyaMltDopfeiHQ==\n+electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.712:\n+ version \"1.3.712\"\n+ resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.712.tgz#ae467ffe5f95961c6d41ceefe858fc36eb53b38f\"\n+ integrity sha512-3kRVibBeCM4vsgoHHGKHmPocLqtFAGTrebXxxtgKs87hNUzXrX2NuS3jnBys7IozCnw7viQlozxKkmty2KNfrw==\n \n elegant-spinner@^1.0.1:\n version \"1.0.1\"\n@@ -9945,9 +9945,9 @@ eslint-plugin-import@^2.16.0, eslint-plugin-import@^2.18.2, eslint-plugin-import\n tsconfig-paths \"^3.9.0\"\n \n eslint-plugin-jest@^24.1.0:\n- version \"24.3.4\"\n- resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.4.tgz#6d90c3554de0302e879603dd6405474c98849f19\"\n- integrity sha512-3n5oY1+fictanuFkTWPwSlehugBTAgwLnYLFsCllzE3Pl1BwywHl5fL0HFxmMjoQY8xhUDk8uAWc3S4JOHGh3A==\n+ version \"24.3.5\"\n+ resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.5.tgz#71f0b580f87915695c286c3f0eb88cf23664d044\"\n+ integrity sha512-XG4rtxYDuJykuqhsOqokYIR84/C8pRihRtEpVskYLbIIKGwPNW2ySxdctuVzETZE+MbF/e7wmsnbNVpzM0rDug==\n dependencies:\n \"@typescript-eslint/experimental-utils\" \"^4.0.1\"\n \n@@ -12140,12 +12140,11 @@ [email protected]:\n micromatch \"^3.1.10\"\n \n http-proxy-middleware@^1.0.0:\n- version \"1.1.0\"\n- resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.0.tgz#b896b2cc6836019af4a4f2d5f7b21b99c77ea13f\"\n- integrity sha512-OnjU5vyVgcZVe2AjLJyMrk8YLNOC2lspCHirB5ldM+B/dwEfZ5bgVTrFyzE9R7xRWAP/i/FXtvIqKjTNEZBhBg==\n+ version \"1.1.1\"\n+ resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.1.tgz#48900a68cd9d388c735d1dd97302c919b7e94a13\"\n+ integrity sha512-FIDg9zPvOwMhQ3XKB2+vdxK6WWbVAH7s5QpqQCif7a1TNL76GNAATWA1sy6q2gSfss8UJ/Nwza3N6QnFkKclpA==\n dependencies:\n \"@types/http-proxy\" \"^1.17.5\"\n- camelcase \"^6.2.0\"\n http-proxy \"^1.18.1\"\n is-glob \"^4.0.1\"\n is-plain-obj \"^3.0.0\"\n@@ -14341,9 +14340,9 @@ jsdom@^15.2.1:\n xml-name-validator \"^3.0.0\"\n \n jsdom@^16.4.0:\n- version \"16.5.2\"\n- resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.2.tgz#583fac89a0aea31dbf6237e7e4bedccd9beab472\"\n- integrity sha512-JxNtPt9C1ut85boCbJmffaQ06NBnzkQY/MWO3YxPW8IWS38A26z+B1oBvA9LwKrytewdfymnhi4UNH3/RAgZrg==\n+ version \"16.5.3\"\n+ resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.3.tgz#13a755b3950eb938b4482c407238ddf16f0d2136\"\n+ integrity sha512-Qj1H+PEvUsOtdPJ056ewXM4UJPCi4hhLA8wpiz9F2YvsRBhuFsXxtrIFAgGBDynQA9isAMGE91PfUYbdMPXuTA==\n dependencies:\n abab \"^2.0.5\"\n acorn \"^8.1.0\"\n@@ -15590,12 +15589,12 @@ micromatch@^3.1.10, micromatch@^3.1.4:\n to-regex \"^3.0.2\"\n \n micromatch@^4.0.2:\n- version \"4.0.3\"\n- resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.3.tgz#fdad8352bf0cbeb89b391b5d244bc22ff3dd4ec8\"\n- integrity sha512-ueuSaP4i67F/FAUac9zzZ0Dz/5KeKDkITYIS/k4fps+9qeh1SkeH6gbljcqz97mNBOsaWZ+iv2UobMKK/yD+aw==\n+ version \"4.0.4\"\n+ resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.4.tgz#896d519dfe9db25fce94ceb7a500919bf881ebf9\"\n+ integrity sha512-pRmzw/XUcwXGpD9aI9q/0XOwLNygjETJ8y0ao0wdqprrzDa4YnxLcz7fQRZr8voh8V10kGhABbNcHVk5wHgWwg==\n dependencies:\n braces \"^3.0.1\"\n- picomatch \"^2.2.1\"\n+ picomatch \"^2.2.3\"\n \n miller-rabin@^4.0.0:\n version \"4.0.1\"\n@@ -16356,7 +16355,7 @@ node-pre-gyp@^0.11.0:\n semver \"^5.3.0\"\n tar \"^4\"\n \n-node-releases@^1.1.61, node-releases@^1.1.70:\n+node-releases@^1.1.61, node-releases@^1.1.71:\n version \"1.1.71\"\n resolved \"https://registry.yarnpkg.com/node-releases/-/node-releases-1.1.71.tgz#cb1334b179896b1c89ecfdd4b725fb7bbdfc7dbb\"\n integrity sha512-zR6HoT6LrLCRBwukmrVbHv0EpEQjksO6GmFcZQQuCAy139BEsoVKPYnf3jongYW83fAa1torLGYwxxky/p28sg==\n@@ -17571,10 +17570,10 @@ [email protected]:\n dependencies:\n split2 \"^3.1.1\"\n \n-picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2:\n- version \"2.2.2\"\n- resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.2.tgz#21f333e9b6b8eaff02468f5146ea406d345f4dad\"\n- integrity sha512-q0M/9eZHzmr0AulXyPwNfZjtwZ/RBZlbN3K3CErVrk50T2ASYI7Bye0EvekFY3IP1Nt2DHu0re+V2ZHIpMkuWg==\n+picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2, picomatch@^2.2.3:\n+ version \"2.2.3\"\n+ resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.3.tgz#465547f359ccc206d3c48e46a1bcb89bf7ee619d\"\n+ integrity sha512-KpELjfwcCDUb9PeigTs2mBJzXUPzAuP2oPcA989He8Rte0+YUAjw1JVedDhuTKPkHjSYzMN3npC9luThGYEKdg==\n \n pify@^2.0.0, pify@^2.2.0, pify@^2.3.0:\n version \"2.3.0\"\n@@ -18446,9 +18445,9 @@ postcss@^7, postcss@^7.0.0, postcss@^7.0.1, postcss@^7.0.14, postcss@^7.0.17, po\n supports-color \"^6.1.0\"\n \n postcss@^8.1.0, postcss@^8.2.8:\n- version \"8.2.9\"\n- resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.9.tgz#fd95ff37b5cee55c409b3fdd237296ab4096fba3\"\n- integrity sha512-b+TmuIL4jGtCHtoLi+G/PisuIl9avxs8IZMSmlABRwNz5RLUUACrC+ws81dcomz1nRezm5YPdXiMEzBEKgYn+Q==\n+ version \"8.2.10\"\n+ resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.10.tgz#ca7a042aa8aff494b334d0ff3e9e77079f6f702b\"\n+ integrity sha512-b/h7CPV7QEdrqIxtAf2j31U5ef05uBDuvoXv6L51Q4rcS1jdlXAVKJv+atCFdUXYl9dyTHGyoMzIepwowRJjFw==\n dependencies:\n colorette \"^1.2.2\"\n nanoid \"^3.1.22\"\n@@ -19318,9 +19317,9 @@ rc-tree@^4.0.0, rc-tree@~4.1.0:\n rc-virtual-list \"^3.0.1\"\n \n rc-trigger@^5.0.0, rc-trigger@^5.0.4, rc-trigger@^5.1.2, rc-trigger@^5.2.1:\n- version \"5.2.3\"\n- resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.3.tgz#8c55046ab432d7b52d51c69afb57ebb5bbe37e17\"\n- integrity sha512-6Fokao07HUbqKIDkDRFEM0AGZvsvK0Fbp8A/KFgl1ngaqfO1nY037cISCG1Jm5fxImVsXp9awdkP7Vu5cxjjog==\n+ version \"5.2.4\"\n+ resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.4.tgz#f1cca4a6c1f378a5d6fadec010292250772069d3\"\n+ integrity sha512-nLZa4XYo3hOAVauQr7HsGrBtE8/pyoIWhHZnpr7x/H/dd6pPeRzH0//+1TzaBAXylbFgsY6hogKAMeJwaKeDFw==\n dependencies:\n \"@babel/runtime\" \"^7.11.2\"\n classnames \"^2.2.6\"\n@@ -20516,9 +20515,9 @@ rollup@^1.31.1:\n acorn \"^7.1.0\"\n \n rollup@^2.40.0, rollup@^2.8.0:\n- version \"2.45.0\"\n- resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.0.tgz#bfcce2347c96f15f5c78ac860bc38e3349ba27c9\"\n- integrity sha512-JJznbtGIsHZfKH0Sa9RpCAy5JarH8SWvBzRAGuRkgzAafb8e8D7VSMJ0O1Bsix1nn91koN/Ecvl2+ZWhljcuTw==\n+ version \"2.45.1\"\n+ resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.1.tgz#eae2b94dc2088b4e0a3b7197a5a1ee0bdd589d5c\"\n+ integrity sha512-vPD+JoDj3CY8k6m1bLcAFttXMe78P4CMxoau0iLVS60+S9kLsv2379xaGy4NgYWu+h2WTlucpoLPAoUoixFBag==\n optionalDependencies:\n fsevents \"~2.3.1\"\n \n@@ -22971,9 +22970,9 @@ typescript@~4.1.5:\n integrity sha512-6OSu9PTIzmn9TCDiovULTnET6BgXtDYL4Gg4szY+cGsc3JP1dQL8qvE8kShTRx1NIw4Q9IBHlwODjkjWEtMUyA==\n \n ua-parser-js@^0.7.18:\n- version \"0.7.27\"\n- resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.27.tgz#b54f8ce9eb6c7abf3584edeaf9a3d8b3bd92edba\"\n- integrity sha512-eXMaRYK2skomGocoX0x9sBXzx5A1ZVQgXfrW4mTc8dT0zS7olEcyfudAzRC5tIIRgLxQ69B6jut3DI+n5hslPA==\n+ version \"0.7.28\"\n+ resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.28.tgz#8ba04e653f35ce210239c64661685bf9121dec31\"\n+ integrity sha512-6Gurc1n//gjp9eQNXjD9O3M/sMwVtN5S8Lv9bvOYBfKfDNiIIhqiyi01vMBO45u4zkDE420w/e0se7Vs+sIg+g==\n \n [email protected]:\n version \"3.4.10\"\n@@ -22984,9 +22983,9 @@ [email protected]:\n source-map \"~0.6.1\"\n \n uglify-js@^3.1.4, uglify-js@^3.4.9:\n- version \"3.13.3\"\n- resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.3.tgz#ce72a1ad154348ea2af61f50933c76cc8802276e\"\n- integrity sha512-otIc7O9LyxpUcQoXzj2hL4LPWKklO6LJWoJUzNa8A17Xgi4fOeDC8FBDOLHnC/Slo1CQgsZMcM6as0M76BZaig==\n+ version \"3.13.4\"\n+ resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.4.tgz#592588bb9f47ae03b24916e2471218d914955574\"\n+ integrity sha512-kv7fCkIXyQIilD5/yQy8O+uagsYIOt5cZvs890W40/e/rvjMSzJw81o9Bg0tkURxzZBROtDQhW2LFjOGoK3RZw==\n \n [email protected]:\n version \"0.0.6\"\n"]
2
["990699ff4a84a5bac3abfecbec002f30e2714de9", "7e9bd7c86df1032d53e752654fe4a446951480bb"]
["test", "build"]
remove unnecessary import,add clean up test Add another clean up test, which verifies that the state is cleaned up after the timer (non-recurring) is triggered.
["diff --git a/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java b/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\nindex 14c6f30..ebaef60 100644\n--- a/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\n+++ b/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\n@@ -8,7 +8,6 @@\n package io.camunda.zeebe.transport.stream.impl;\n \n import io.camunda.zeebe.util.buffer.BufferUtil;\n-import org.agrona.BitUtil;\n import org.agrona.concurrent.UnsafeBuffer;\n \n /**\n", "diff --git a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\nindex d36b4c9..ca5047f 100644\n--- a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n@@ -630,6 +630,40 @@ public final class ProcessExecutionCleanStateTest {\n }\n \n @Test\n+ public void testProcessWithTriggerTimerStartEvent() {\n+ // given\n+ final var deployment =\n+ engineRule\n+ .deployment()\n+ .withXmlResource(\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .timerWithDate(\"=now() + duration(\\\"PT15S\\\")\")\n+ .endEvent()\n+ .done())\n+ .deploy();\n+\n+ final var processDefinitionKey =\n+ deployment.getValue().getProcessesMetadata().get(0).getProcessDefinitionKey();\n+\n+ // when\n+ engineRule.awaitProcessingOf(\n+ RecordingExporter.timerRecords(TimerIntent.CREATED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .getFirst());\n+\n+ engineRule.increaseTime(Duration.ofSeconds(15));\n+\n+ RecordingExporter.processInstanceRecords(ProcessInstanceIntent.ELEMENT_COMPLETED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .withElementType(BpmnElementType.PROCESS)\n+ .await();\n+\n+ // then\n+ assertThatStateIsEmpty();\n+ }\n+\n+ @Test\n public void testProcessWithTimerStartEventRedeployment() {\n // given\n final var deployment =\n"]
2
["84529bcb10c6fe02e2c0079d069ab6c6ac7683d6", "aa746b764e6c54bbbd631210fce35df842d09b12"]
["refactor", "test"]
increase timeout of multiregion failover test Due to the nature of the test, restarts and failovers can take long. If the recovery takes longer than 15m, then the test will fail unnecessarily. Since we are not really testing for how was it can recover, it is ok to increase the maxInstanceDuration.,rework RaftCommittedEntryListener Iterate over RaftCommittedEntryListener and refactor the listener such it serves the actual need. We have some services (to be specific the AsyncSnapshotDirector) which need the committed position, and want to listen to new updates. In raft we know which record we are committing and whether it was an application record so we can pass this information threw the listeners. This avoids to pass in the whole IndexedRecord object, and reduce the potential of going out of OOM because of keeping to much data in heap (when commit is not possible).
["diff --git a/.github/workflows/e2e-testbench.yaml b/.github/workflows/e2e-testbench.yaml\nindex 708f97f..fd0b918 100644\n--- a/.github/workflows/e2e-testbench.yaml\n+++ b/.github/workflows/e2e-testbench.yaml\n@@ -31,6 +31,11 @@ on:\n default: null\n required: false\n type: string\n+ maxInstanceDuration:\n+ description: 'If an instance takes longer than the given duration to complete, test will fail.'\n+ default: '15m'\n+ required: false\n+ type: string\n \n workflow_call:\n inputs:\n@@ -59,6 +64,11 @@ on:\n default: null\n required: false\n type: string\n+ maxInstanceDuration:\n+ description: 'If an instance takes longer than the given duration to complete, test will fail.'\n+ default: '15m'\n+ required: false\n+ type: string\n \n jobs:\n e2e:\n@@ -81,7 +91,7 @@ jobs:\n {\n \\\"maxTestDuration\\\": \\\"${{ inputs.maxTestDuration || 'P5D' }}\\\",\n \\\"starter\\\": [ {\\\"rate\\\": 50, \\\"processId\\\": \\\"one-task-one-timer\\\" } ],\n- \\\"verifier\\\" : { \\\"maxInstanceDuration\\\" : \\\"15m\\\" },\n+ \\\"verifier\\\" : { \\\"maxInstanceDuration\\\" : \\\"${{ inputs.maxInstanceDuration }}\\\" },\n \\\"fault\\\": ${{ inputs.fault || 'null' }}\n }\n }\ndiff --git a/.github/workflows/weekly-e2e.yml b/.github/workflows/weekly-e2e.yml\nindex 93aaeb5..4bd0afd 100644\n--- a/.github/workflows/weekly-e2e.yml\n+++ b/.github/workflows/weekly-e2e.yml\n@@ -31,4 +31,5 @@ jobs:\n maxTestDuration: P1D\n clusterPlan: Multiregion test simulation\n fault: \\\"2-region-dataloss-failover\\\"\n+ maxInstanceDuration: 40m\n secrets: inherit\n", "diff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\nnew file mode 100644\nindex 0000000..57c28a9\n--- /dev/null\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\n@@ -0,0 +1,31 @@\n+/*\n+ * Copyright 2016-present Open Networking Foundation\n+ * Copyright \u00a9 2020 camunda services GmbH ([email protected])\n+ *\n+ * Licensed under the Apache License, Version 2.0 (the \"License\");\n+ * you may not use this file except in compliance with the License.\n+ * You may obtain a copy of the License at\n+ *\n+ * http://www.apache.org/licenses/LICENSE-2.0\n+ *\n+ * Unless required by applicable law or agreed to in writing, software\n+ * distributed under the License is distributed on an \"AS IS\" BASIS,\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n+ * See the License for the specific language governing permissions and\n+ * limitations under the License.\n+ */\n+package io.atomix.raft;\n+\n+/**\n+ * This listener will only be called by the Leader, when it commits an application entry.\n+ *\n+ * <p>If RAFT is currently running in a follower role, it will not call this listener.\n+ */\n+@FunctionalInterface\n+public interface RaftApplicationEntryCommittedPositionListener {\n+\n+ /**\n+ * @param committedPosition the new committed position which is related to the application entries\n+ */\n+ void onCommit(long committedPosition);\n+}\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\ndeleted file mode 100644\nindex 3d11d75..0000000\n--- a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\n+++ /dev/null\n@@ -1,32 +0,0 @@\n-/*\n- * Copyright 2016-present Open Networking Foundation\n- * Copyright \u00a9 2020 camunda services GmbH ([email protected])\n- *\n- * Licensed under the Apache License, Version 2.0 (the \"License\");\n- * you may not use this file except in compliance with the License.\n- * You may obtain a copy of the License at\n- *\n- * http://www.apache.org/licenses/LICENSE-2.0\n- *\n- * Unless required by applicable law or agreed to in writing, software\n- * distributed under the License is distributed on an \"AS IS\" BASIS,\n- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n- * See the License for the specific language governing permissions and\n- * limitations under the License.\n- */\n-package io.atomix.raft;\n-\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\n-\n-/**\n- * This listener will only be called by the Leader, when it commits an entry. If RAFT is currently\n- * running in a follower role, it will not call this listener.\n- */\n-@FunctionalInterface\n-public interface RaftCommittedEntryListener {\n-\n- /**\n- * @param indexedRaftLogEntry the new committed entry\n- */\n- void onCommit(IndexedRaftLogEntry indexedRaftLogEntry);\n-}\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\nindex 1f4ee98..c177cb1 100644\n--- a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\n@@ -27,8 +27,8 @@ import io.atomix.cluster.MemberId;\n import io.atomix.cluster.messaging.MessagingException.NoRemoteHandler;\n import io.atomix.cluster.messaging.MessagingException.NoSuchMemberException;\n import io.atomix.raft.ElectionTimer;\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\n import io.atomix.raft.RaftCommitListener;\n-import io.atomix.raft.RaftCommittedEntryListener;\n import io.atomix.raft.RaftError;\n import io.atomix.raft.RaftException.ProtocolException;\n import io.atomix.raft.RaftRoleChangeListener;\n@@ -61,7 +61,6 @@ import io.atomix.raft.roles.PromotableRole;\n import io.atomix.raft.roles.RaftRole;\n import io.atomix.raft.storage.RaftStorage;\n import io.atomix.raft.storage.StorageException;\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\n import io.atomix.raft.storage.log.RaftLog;\n import io.atomix.raft.storage.system.MetaStore;\n import io.atomix.raft.utils.StateUtil;\n@@ -115,7 +114,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\n private final Set<Consumer<State>> stateChangeListeners = new CopyOnWriteArraySet<>();\n private final Set<Consumer<RaftMember>> electionListeners = new CopyOnWriteArraySet<>();\n private final Set<RaftCommitListener> commitListeners = new CopyOnWriteArraySet<>();\n- private final Set<RaftCommittedEntryListener> committedEntryListeners =\n+ private final Set<RaftApplicationEntryCommittedPositionListener> committedEntryListeners =\n new CopyOnWriteArraySet<>();\n private final Set<SnapshotReplicationListener> snapshotReplicationListeners =\n new CopyOnWriteArraySet<>();\n@@ -433,21 +432,23 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\n * <p>Note that it will be called on the Raft thread, and as such should not perform any heavy\n * computation.\n *\n- * @param raftCommittedEntryListener the listener to add\n+ * @param raftApplicationEntryCommittedPositionListener the listener to add\n */\n public void addCommittedEntryListener(\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\n- committedEntryListeners.add(raftCommittedEntryListener);\n+ final RaftApplicationEntryCommittedPositionListener\n+ raftApplicationEntryCommittedPositionListener) {\n+ committedEntryListeners.add(raftApplicationEntryCommittedPositionListener);\n }\n \n /**\n * Removes registered committedEntryListener\n *\n- * @param raftCommittedEntryListener the listener to remove\n+ * @param raftApplicationEntryCommittedPositionListener the listener to remove\n */\n public void removeCommittedEntryListener(\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\n- committedEntryListeners.remove(raftCommittedEntryListener);\n+ final RaftApplicationEntryCommittedPositionListener\n+ raftApplicationEntryCommittedPositionListener) {\n+ committedEntryListeners.remove(raftApplicationEntryCommittedPositionListener);\n }\n \n /**\n@@ -464,7 +465,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\n *\n * @param committedEntry the most recently committed entry\n */\n- public void notifyCommittedEntryListeners(final IndexedRaftLogEntry committedEntry) {\n+ public void notifyApplicationEntryCommittedPositionListeners(final long committedEntry) {\n committedEntryListeners.forEach(listener -> listener.onCommit(committedEntry));\n }\n \ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\nindex 56c7172..d075fca 100644\n--- a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\n@@ -21,8 +21,8 @@ import io.atomix.cluster.MemberId;\n import io.atomix.cluster.messaging.ClusterCommunicationService;\n import io.atomix.primitive.partition.Partition;\n import io.atomix.primitive.partition.PartitionMetadata;\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\n import io.atomix.raft.RaftCommitListener;\n-import io.atomix.raft.RaftCommittedEntryListener;\n import io.atomix.raft.RaftRoleChangeListener;\n import io.atomix.raft.RaftServer;\n import io.atomix.raft.RaftServer.Role;\n@@ -205,16 +205,20 @@ public class RaftPartitionServer implements HealthMonitorable {\n }\n \n /**\n- * @see io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftCommittedEntryListener)\n+ * @see\n+ * io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\n */\n- public void addCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\n+ public void addCommittedEntryListener(\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\n server.getContext().addCommittedEntryListener(commitListener);\n }\n \n /**\n- * @see io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftCommittedEntryListener)\n+ * @see\n+ * io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\n */\n- public void removeCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\n+ public void removeCommittedEntryListener(\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\n server.getContext().removeCommittedEntryListener(commitListener);\n }\n \ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\nindex e54df1a..fcfd177 100644\n--- a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\n@@ -630,27 +630,47 @@ public final class LeaderRole extends ActiveRole implements ZeebeLogAppender {\n \n private void replicate(final IndexedRaftLogEntry indexed, final AppendListener appendListener) {\n raft.checkThread();\n- appender\n- .appendEntries(indexed.index())\n- .whenCompleteAsync(\n- (commitIndex, commitError) -> {\n- if (!isRunning()) {\n- return;\n- }\n+ final var appendEntriesFuture = appender.appendEntries(indexed.index());\n+\n+ final boolean applicationEntryWasCommitted = indexed.isApplicationEntry();\n+ if (applicationEntryWasCommitted) {\n+ // We have some services which are waiting for the application records, especially position\n+ // to be committed. This is our glue code to notify them, instead of\n+ // passing the complete object (IndexedRaftLogEntry) threw the listeners and\n+ // keep them in heap until they are committed. This had the risk of going out of OOM\n+ // if records can't be committed, see https://github.com/camunda/zeebe/issues/14275\n+ final var committedPosition = indexed.getApplicationEntry().highestPosition();\n+ appendEntriesFuture.whenCompleteAsync(\n+ (commitIndex, commitError) -> {\n+ if (!isRunning()) {\n+ return;\n+ }\n+\n+ if (commitError == null) {\n+ raft.notifyApplicationEntryCommittedPositionListeners(committedPosition);\n+ }\n+ },\n+ raft.getThreadContext());\n+ }\n \n- // have the state machine apply the index which should do nothing but ensures it keeps\n- // up to date with the latest entries, so it can handle configuration and initial\n- // entries properly on fail over\n- if (commitError == null) {\n- appendListener.onCommit(indexed.index());\n- raft.notifyCommittedEntryListeners(indexed);\n- } else {\n- appendListener.onCommitError(indexed.index(), commitError);\n- // replicating the entry will be retried on the next append request\n- log.error(\"Failed to replicate entry: {}\", indexed, commitError);\n- }\n- },\n- raft.getThreadContext());\n+ appendEntriesFuture.whenCompleteAsync(\n+ (commitIndex, commitError) -> {\n+ if (!isRunning()) {\n+ return;\n+ }\n+\n+ // have the state machine apply the index which should do nothing but ensures it keeps\n+ // up to date with the latest entries, so it can handle configuration and initial\n+ // entries properly on fail over\n+ if (commitError == null) {\n+ appendListener.onCommit(indexed.index());\n+ } else {\n+ appendListener.onCommitError(indexed.index(), commitError);\n+ // replicating the entry will be retried on the next append request\n+ log.error(\"Failed to replicate entry: {}\", indexed, commitError);\n+ }\n+ },\n+ raft.getThreadContext());\n }\n \n public synchronized void onInitialEntriesCommitted(final Runnable runnable) {\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\nindex b217586..8029766 100644\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\n@@ -82,7 +82,7 @@ public class RaftAppendTest {\n @Test\n public void shouldNotifyCommittedEntryListenerOnLeaderOnly() throws Throwable {\n // given\n- final var committedEntryListener = mock(RaftCommittedEntryListener.class);\n+ final var committedEntryListener = mock(RaftApplicationEntryCommittedPositionListener.class);\n raftRule.addCommittedEntryListener(committedEntryListener);\n \n // when\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\nindex 8f73cba..193a176 100644\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\n@@ -644,9 +644,12 @@ public final class RaftRule extends ExternalResource {\n }\n \n public void addCommittedEntryListener(\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\n+ final RaftApplicationEntryCommittedPositionListener\n+ raftApplicationEntryCommittedPositionListener) {\n servers.forEach(\n- (id, raft) -> raft.getContext().addCommittedEntryListener(raftCommittedEntryListener));\n+ (id, raft) ->\n+ raft.getContext()\n+ .addCommittedEntryListener(raftApplicationEntryCommittedPositionListener));\n }\n \n public void partition(final RaftServer follower) {\ndiff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\nindex a61571f..6c082d7 100644\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n@@ -7,8 +7,7 @@\n */\n package io.camunda.zeebe.broker.system.partitions.impl;\n \n-import io.atomix.raft.RaftCommittedEntryListener;\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\n import io.camunda.zeebe.broker.system.partitions.NoEntryAtSnapshotPosition;\n import io.camunda.zeebe.broker.system.partitions.StateController;\n import io.camunda.zeebe.logstreams.impl.Loggers;\n@@ -36,7 +35,7 @@ import java.util.function.Consumer;\n import org.slf4j.Logger;\n \n public final class AsyncSnapshotDirector extends Actor\n- implements RaftCommittedEntryListener, HealthMonitorable {\n+ implements RaftApplicationEntryCommittedPositionListener, HealthMonitorable {\n \n public static final Duration MINIMUM_SNAPSHOT_PERIOD = Duration.ofMinutes(1);\n \n@@ -115,7 +114,7 @@ public final class AsyncSnapshotDirector extends Actor\n @Override\n protected void handleFailure(final Throwable failure) {\n LOG.error(\n- \"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}. {}\",\n+ \"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}.\",\n actorName,\n snapshotRate,\n failure);\n@@ -407,13 +406,8 @@ public final class AsyncSnapshotDirector extends Actor\n }\n \n @Override\n- public void onCommit(final IndexedRaftLogEntry indexedRaftLogEntry) {\n- // is called by the Leader Role and gives the last committed entry, where we\n- // can extract the highest position, which corresponds to the last committed position\n- if (indexedRaftLogEntry.isApplicationEntry()) {\n- final var committedPosition = indexedRaftLogEntry.getApplicationEntry().highestPosition();\n- newPositionCommitted(committedPosition);\n- }\n+ public void onCommit(final long committedPosition) {\n+ newPositionCommitted(committedPosition);\n }\n \n public void newPositionCommitted(final long currentCommitPosition) {\n"]
2
["ee824ddd71cbc4ccc26f7c6876d379c4927b79e6", "323cf81961cdd3748a7ba6ba470ecb13e5374e9f"]
["ci", "refactor"]
getBorderSize() missing "width" The correct property name to use is "borderWidth", not just "border". "border" works in Chrome but was breaking in Firefox. Also had to change .ui-grid-header's box-sizing to content-box so IE11 would include the border in height calcs. AND finally IE11 was returning fractional heights so Grid parseInt()s the returned values.,await job creation to ensure asserted event sequence
["diff --git a/src/js/core/factories/Grid.js b/src/js/core/factories/Grid.js\nindex dcf10af..2be7842 100644\n--- a/src/js/core/factories/Grid.js\n+++ b/src/js/core/factories/Grid.js\n@@ -1525,7 +1525,7 @@ angular.module('ui.grid')\n var oldHeaderHeight = container.headerHeight;\n var headerHeight = gridUtil.outerElementHeight(container.header);\n \n- container.headerHeight = headerHeight;\n+ container.headerHeight = parseInt(headerHeight, 10);\n \n if (oldHeaderHeight !== headerHeight) {\n rebuildStyles = true;\n@@ -1534,7 +1534,9 @@ angular.module('ui.grid')\n // Get the \"inner\" header height, that is the height minus the top and bottom borders, if present. We'll use it to make sure all the headers have a consistent height\n var topBorder = gridUtil.getBorderSize(container.header, 'top');\n var bottomBorder = gridUtil.getBorderSize(container.header, 'bottom');\n- var innerHeaderHeight = headerHeight - topBorder - bottomBorder;\n+ var innerHeaderHeight = parseInt(headerHeight - topBorder - bottomBorder, 10);\n+\n+ innerHeaderHeight = innerHeaderHeight < 0 ? 0 : innerHeaderHeight;\n \n container.innerHeaderHeight = innerHeaderHeight;\n \ndiff --git a/src/js/core/services/ui-grid-util.js b/src/js/core/services/ui-grid-util.js\nindex 2c32cbe..cc7c36c 100644\n--- a/src/js/core/services/ui-grid-util.js\n+++ b/src/js/core/services/ui-grid-util.js\n@@ -757,6 +757,8 @@ module.service('gridUtil', ['$log', '$window', '$document', '$http', '$templateC\n borderType = 'border';\n }\n \n+ borderType += 'Width';\n+\n var val = parseInt(styles[borderType], 10);\n \n if (isNaN(val)) {\ndiff --git a/src/less/header.less b/src/less/header.less\nindex 5468a43..de8ff0b 100644\n--- a/src/less/header.less\n+++ b/src/less/header.less\n@@ -7,6 +7,7 @@\n \n .ui-grid-header {\n border-bottom: 1px solid @borderColor;\n+ box-sizing: content-box;;\n }\n \n .ui-grid-top-panel {\n", "diff --git a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\nindex 9ffa1fa..4333db0 100644\n--- a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\n+++ b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\n@@ -114,12 +114,18 @@ public class BoundaryEventTest {\n ENGINE.deployment().withXmlResource(MULTIPLE_SEQUENCE_FLOWS).deploy();\n final long workflowInstanceKey = ENGINE.workflowInstance().ofBpmnProcessId(PROCESS_ID).create();\n \n- // when\n RecordingExporter.timerRecords()\n .withHandlerNodeId(\"timer\")\n .withIntent(TimerIntent.CREATED)\n .withWorkflowInstanceKey(workflowInstanceKey)\n .getFirst();\n+\n+ RecordingExporter.jobRecords(JobIntent.CREATED)\n+ .withType(\"type\")\n+ .withWorkflowInstanceKey(workflowInstanceKey)\n+ .getFirst();\n+\n+ // when\n ENGINE.increaseTime(Duration.ofMinutes(1));\n \n // then\n"]
2
["174f25214caa10ec643db6c81aaa0f3511bf78f4", "a8d1a60fd48d3fbd76d4271987a1b0f538d498f1"]
["fix", "test"]
pin version of actionlint used,add unit test for query API
["diff --git a/.github/workflows/introspect.yml b/.github/workflows/introspect.yml\nindex b6d9125..82d22a5 100644\n--- a/.github/workflows/introspect.yml\n+++ b/.github/workflows/introspect.yml\n@@ -25,5 +25,5 @@ jobs:\n # From https://github.com/rhysd/actionlint/blob/main/docs/usage.md#use-actionlint-on-github-actions\n - name: Check workflow files\n run: |\n- bash <(curl https://raw.githubusercontent.com/rhysd/actionlint/590d3bd9dde0c91f7a66071d40eb84716526e5a6/scripts/download-actionlint.bash)\n+ bash <(curl https://raw.githubusercontent.com/rhysd/actionlint/590d3bd9dde0c91f7a66071d40eb84716526e5a6/scripts/download-actionlint.bash) 1.6.25\n ./actionlint -color -shellcheck=\"\"\n", "diff --git a/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java b/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java\nindex 2d2d084..38261ad 100644\n--- a/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java\n+++ b/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java\n@@ -25,6 +25,7 @@ import java.util.HashMap;\n import java.util.List;\n import java.util.Map;\n import java.util.concurrent.CompletableFuture;\n+import java.util.concurrent.TimeUnit;\n import java.util.function.Consumer;\n \n public final class StubbedBrokerClient implements BrokerClient {\n@@ -67,7 +68,15 @@ public final class StubbedBrokerClient implements BrokerClient {\n @Override\n public <T> CompletableFuture<BrokerResponse<T>> sendRequestWithRetry(\n final BrokerRequest<T> request, final Duration requestTimeout) {\n- throw new UnsupportedOperationException(\"not implemented\");\n+ final CompletableFuture<BrokerResponse<T>> result = new CompletableFuture<>();\n+\n+ sendRequestWithRetry(\n+ request,\n+ (key, response) ->\n+ result.complete(new BrokerResponse<>(response, Protocol.decodePartitionId(key), key)),\n+ result::completeExceptionally);\n+\n+ return result.orTimeout(requestTimeout.toNanos(), TimeUnit.NANOSECONDS);\n }\n \n @Override\ndiff --git a/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryApiTest.java b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryApiTest.java\nnew file mode 100644\nindex 0000000..ec9ec80\n--- /dev/null\n+++ b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryApiTest.java\n@@ -0,0 +1,91 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.gateway.query;\n+\n+import static org.assertj.core.api.Assertions.assertThat;\n+\n+import io.camunda.zeebe.gateway.api.util.GatewayTest;\n+import io.camunda.zeebe.gateway.cmd.BrokerErrorException;\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerError;\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerErrorResponse;\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerResponse;\n+import io.camunda.zeebe.gateway.query.impl.QueryApiImpl;\n+import io.camunda.zeebe.protocol.Protocol;\n+import io.camunda.zeebe.protocol.record.ErrorCode;\n+import java.time.Duration;\n+import java.util.concurrent.CompletionStage;\n+import java.util.concurrent.ExecutionException;\n+import org.junit.Test;\n+import org.junit.runner.RunWith;\n+import org.junit.runners.Parameterized;\n+import org.junit.runners.Parameterized.Parameter;\n+import org.junit.runners.Parameterized.Parameters;\n+\n+@RunWith(Parameterized.class)\n+public final class QueryApiTest extends GatewayTest {\n+ @Parameter(0)\n+ public String name;\n+\n+ @Parameter(1)\n+ public Querier querier;\n+\n+ @Parameters(name = \"{index}: {0}\")\n+ public static Object[][] queries() {\n+ return new Object[][] {\n+ new Object[] {\"getBpmnProcessIdForProcess\", (Querier) QueryApi::getBpmnProcessIdFromProcess},\n+ new Object[] {\n+ \"getBpmnProcessIdForProcessInstance\",\n+ (Querier) QueryApi::getBpmnProcessIdFromProcessInstance\n+ },\n+ new Object[] {\"getBpmnProcessIdForProcessJob\", (Querier) QueryApi::getBpmnProcessIdFromJob},\n+ };\n+ }\n+\n+ @Test\n+ public void shouldGetBpmnProcessId() {\n+ // given\n+ final var key = Protocol.encodePartitionId(1, 1);\n+ final var api = new QueryApiImpl(brokerClient);\n+ final var timeout = Duration.ofSeconds(5);\n+ final var stub = new QueryStub(new BrokerResponse<>(\"myProcess\", 1, 1));\n+ stub.registerWith(brokerClient);\n+\n+ // when\n+ final var result = querier.query(api, key, timeout);\n+\n+ // then\n+ assertThat(result).succeedsWithin(timeout).isEqualTo(\"myProcess\");\n+ }\n+\n+ @Test\n+ public void shouldCompleteExceptionallyOnError() {\n+ // given\n+ final var key = Protocol.encodePartitionId(1, 1);\n+ final var api = new QueryApiImpl(brokerClient);\n+ final var timeout = Duration.ofSeconds(5);\n+ final var stub =\n+ new QueryStub(\n+ new BrokerErrorResponse<>(\n+ new BrokerError(ErrorCode.PARTITION_LEADER_MISMATCH, \"Leader mismatch\")));\n+ stub.registerWith(brokerClient);\n+\n+ // when\n+ final var result = querier.query(api, key, timeout);\n+\n+ // then\n+ assertThat(result)\n+ .failsWithin(timeout)\n+ .withThrowableOfType(ExecutionException.class)\n+ .havingRootCause()\n+ .isInstanceOf(BrokerErrorException.class);\n+ }\n+\n+ private interface Querier {\n+ CompletionStage<String> query(final QueryApi api, final long key, final Duration timeout);\n+ }\n+}\ndiff --git a/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryStub.java b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryStub.java\nnew file mode 100644\nindex 0000000..2f8334e\n--- /dev/null\n+++ b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryStub.java\n@@ -0,0 +1,31 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.gateway.query;\n+\n+import io.camunda.zeebe.gateway.api.util.StubbedBrokerClient;\n+import io.camunda.zeebe.gateway.api.util.StubbedBrokerClient.RequestStub;\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerResponse;\n+import io.camunda.zeebe.gateway.query.impl.BrokerExecuteQuery;\n+\n+final class QueryStub implements RequestStub<BrokerExecuteQuery, BrokerResponse<String>> {\n+ private final BrokerResponse<String> response;\n+\n+ public QueryStub(final BrokerResponse<String> response) {\n+ this.response = response;\n+ }\n+\n+ @Override\n+ public void registerWith(final StubbedBrokerClient gateway) {\n+ gateway.registerHandler(BrokerExecuteQuery.class, this);\n+ }\n+\n+ @Override\n+ public BrokerResponse<String> handle(final BrokerExecuteQuery request) throws Exception {\n+ return response;\n+ }\n+}\n"]
2
["b702adc245f679ae20d84de39f0d63b14aabed5d", "bed86aeae8dad2dd6371635cd24bf8ef3db80361"]
["ci", "test"]
generate terminate end event compatible execution steps part 1 The random execution tests don't know the concept of flow scopes. This makes it challenging to generate a correct execution path for terminate end events, as they terminate a specific flow scope. Processing should continue as normal once the flow scope has been terminated. Whilst we don't have flow scopes, we do have execution path segments. These segments don't map 1 to 1 to flow scopes. However, since every flow scope starts a new segment we can use these segments to get the desired behavior. Each segment must keep track whether is has reached a terminate end event. If this is the case that means that we don't expect any further execution steps. We can isolate this behavior in a single location, during the appending of one segment to another segment. In order to differentiate between flow scopes a new append method has been added which takes the boolean `changesFlowScope` as a parameter. Blockbuilder where the flow scope changes (e.g. SubProcessBlockBuilder) can use this to indicate that even though a terminate end event has been reached. Execution steps after this specific segment still need to added to complete the process. When a segment is appended to a different segment and the flow scope does not change we can use the segment that should be appended to identify whether new segment can still be added to the current segment. If passed segment has reached a terminate end event and the flow scope has not been changed it is guaranteed that the current segment is in the same flow scope has the previous segment and thus has also reached the terminate end event.,use connect instead of begin to avoid nesting transactions
["diff --git a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\nindex da33c23..23c43be 100644\n--- a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\n+++ b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\n@@ -29,6 +29,10 @@ import org.apache.commons.lang3.builder.ToStringStyle;\n */\n public final class ExecutionPathSegment {\n \n+ // If we have reached a terminate end event we want to stop generating execution steps for a\n+ // specific flow scope. By setting this flag to true no new execution steps will be added for the\n+ // flow scope this segment is in.\n+ private boolean reachedTerminateEndEvent = false;\n private final List<ScheduledExecutionStep> scheduledSteps = new ArrayList<>();\n private final Map<String, Object> variableDefaults = new HashMap<>();\n \n@@ -87,10 +91,28 @@ public final class ExecutionPathSegment {\n new ScheduledExecutionStep(logicalPredecessor, executionPredecessor, executionStep));\n }\n \n+ /**\n+ * Appends the steps of the passed execution path segment to the current segment.\n+ *\n+ * @param pathToAdd execution path segment to append to this segment\n+ */\n public void append(final ExecutionPathSegment pathToAdd) {\n+ append(pathToAdd, false);\n+ }\n+\n+ /**\n+ * Appends the step of the passed execution path segment to the current segment if the current\n+ *\n+ * @param pathToAdd\n+ * @param changesFlowScope\n+ */\n+ public void append(final ExecutionPathSegment pathToAdd, final boolean changesFlowScope) {\n mergeVariableDefaults(pathToAdd);\n \n- pathToAdd.getScheduledSteps().forEach(this::append);\n+ if (!hasReachedTerminateEndEvent() || changesFlowScope) {\n+ pathToAdd.getScheduledSteps().forEach(this::append);\n+ }\n+ reachedTerminateEndEvent = pathToAdd.hasReachedTerminateEndEvent() && !changesFlowScope;\n }\n \n public void append(final ScheduledExecutionStep scheduledExecutionStep) {\n@@ -259,6 +281,14 @@ public final class ExecutionPathSegment {\n return ToStringBuilder.reflectionToString(this, ToStringStyle.SHORT_PREFIX_STYLE);\n }\n \n+ public boolean hasReachedTerminateEndEvent() {\n+ return reachedTerminateEndEvent;\n+ }\n+\n+ public void setReachedTerminateEndEvent(final boolean reachedTerminateEndEvent) {\n+ this.reachedTerminateEndEvent = reachedTerminateEndEvent;\n+ }\n+\n /**\n * An execution boundary is the point where automatic and non-automatic {@link\n * ScheduledExecutionStep}'s meet each other. This class contains information about the existing\n", "diff --git a/ibis/backends/duckdb/__init__.py b/ibis/backends/duckdb/__init__.py\nindex 2006f59..bb2028e 100644\n--- a/ibis/backends/duckdb/__init__.py\n+++ b/ibis/backends/duckdb/__init__.py\n@@ -1180,7 +1180,7 @@ WHERE catalog_name = :database\"\"\"\n def _register_udfs(self, expr: ir.Expr) -> None:\n import ibis.expr.operations as ops\n \n- with self.begin() as con:\n+ with self.con.connect() as con:\n for udf_node in expr.op().find(ops.ScalarUDF):\n compile_func = getattr(\n self, f\"_compile_{udf_node.__input_type__.name.lower()}_udf\"\n"]
2
["40597fb4de41c7194eb99479a914db70da7909ea", "6889543bec720e7e7da66535e1012cb66edfe081"]
["feat", "fix"]
verify process can start at supported element types Verifies a PI can be started at specific element types. The test will deploy the process, start an instance at the desired start element and verify that it has been activated succesfully.,fix typos (#90)
["diff --git a/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\nnew file mode 100644\nindex 0000000..a505307\n--- /dev/null\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\n@@ -0,0 +1,233 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.engine.processing.processinstance;\n+\n+import static org.assertj.core.api.Assertions.assertThat;\n+import static org.assertj.core.groups.Tuple.tuple;\n+\n+import io.camunda.zeebe.engine.util.EngineRule;\n+import io.camunda.zeebe.model.bpmn.Bpmn;\n+import io.camunda.zeebe.model.bpmn.BpmnModelInstance;\n+import io.camunda.zeebe.protocol.record.Record;\n+import io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent;\n+import io.camunda.zeebe.protocol.record.value.BpmnElementType;\n+import io.camunda.zeebe.test.util.record.RecordingExporter;\n+import io.camunda.zeebe.test.util.record.RecordingExporterTestWatcher;\n+import java.util.Collection;\n+import java.util.Collections;\n+import java.util.List;\n+import java.util.Map;\n+import org.junit.ClassRule;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.runner.RunWith;\n+import org.junit.runners.Parameterized;\n+import org.junit.runners.Parameterized.Parameters;\n+\n+@RunWith(Parameterized.class)\n+public class CreateProcessInstanceSupportedElementTest {\n+\n+ @ClassRule public static final EngineRule ENGINE = EngineRule.singlePartition();\n+ private static final String PROCESS_ID = \"processId\";\n+ private static final String CHILD_PROCESS_ID = \"childProcessId\";\n+ private static final String START_ELEMENT_ID = \"startElement\";\n+ private static final String MESSAGE = \"message\";\n+ private static final String JOBTYPE = \"jobtype\";\n+\n+ @Rule\n+ public final RecordingExporterTestWatcher recordingExporterTestWatcher =\n+ new RecordingExporterTestWatcher();\n+\n+ private final Scenario scenario;\n+\n+ public CreateProcessInstanceSupportedElementTest(final Scenario scenario) {\n+ this.scenario = scenario;\n+ }\n+\n+ @Parameters(name = \"{0}\")\n+ public static Collection<Object> scenarios() {\n+ return List.of(\n+ new Scenario(\n+ BpmnElementType.SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .subProcess(START_ELEMENT_ID)\n+ .embeddedSubProcess()\n+ .startEvent()\n+ .subProcessDone()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .eventSubProcess(\n+ START_ELEMENT_ID, e -> e.startEvent().timerWithDuration(\"PT1H\").endEvent())\n+ .startEvent()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_CATCH_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateCatchEvent(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_THROW_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateThrowEvent(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.END_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().endEvent(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SERVICE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.RECEIVE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .receiveTask(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.USER_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().userTask(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.MANUAL_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .manualTask(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EXCLUSIVE_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .exclusiveGateway(START_ELEMENT_ID)\n+ .defaultFlow()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.PARALLEL_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .parallelGateway(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_BASED_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .eventBasedGateway(START_ELEMENT_ID)\n+ .intermediateCatchEvent()\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .moveToLastGateway()\n+ .intermediateCatchEvent()\n+ .timerWithDuration(\"PT1H\")\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.MULTI_INSTANCE_BODY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(\n+ START_ELEMENT_ID,\n+ t ->\n+ t.zeebeJobType(JOBTYPE)\n+ .multiInstance(m -> m.parallel().zeebeInputCollectionExpression(\"[1]\")))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.CALL_ACTIVITY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .callActivity(START_ELEMENT_ID, c -> c.zeebeProcessId(CHILD_PROCESS_ID))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.BUSINESS_RULE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .businessRuleTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SCRIPT_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .scriptTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SEND_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .sendTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()));\n+ }\n+\n+ @Test\n+ public void testProcessInstanceCanStartAtElementType() {\n+ // given\n+ ENGINE.deployment().withXmlResource(scenario.modelInstance).deploy();\n+ if (scenario.type == BpmnElementType.CALL_ACTIVITY) {\n+ ENGINE.deployment().withXmlResource(getChildProcess()).deploy();\n+ }\n+\n+ // when\n+ final long instanceKey =\n+ ENGINE\n+ .processInstance()\n+ .ofBpmnProcessId(PROCESS_ID)\n+ .withStartInstruction(START_ELEMENT_ID)\n+ .withVariables(scenario.variables)\n+ .create();\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.processInstanceRecords()\n+ .withProcessInstanceKey(instanceKey)\n+ .onlyEvents()\n+ .limit(\n+ r ->\n+ r.getValue().getBpmnElementType() == scenario.type\n+ && r.getIntent() == ProcessInstanceIntent.ELEMENT_ACTIVATED))\n+ .extracting(record -> record.getValue().getBpmnElementType(), Record::getIntent)\n+ .containsSequence(\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATED),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATED));\n+ }\n+\n+ private BpmnModelInstance getChildProcess() {\n+ return Bpmn.createExecutableProcess(CHILD_PROCESS_ID).startEvent().endEvent().done();\n+ }\n+\n+ record Scenario(\n+ BpmnElementType type, BpmnModelInstance modelInstance, Map<String, Object> variables) {}\n+}\n", "diff --git a/README.md b/README.md\nindex de15ac5..5ad8b47 100755\n--- a/README.md\n+++ b/README.md\n@@ -16,13 +16,13 @@ content that will be loaded, similar to Facebook cards loaders.\n \n ## Features\n \n-* :gear: **Complety customizable:** you can change the colors, speed and sizes;\n+* :gear: **Completely customizable:** you can change the colors, speed and sizes;\n * :pencil2: **Create your own loading:** use the\n [create-react-content-loader](https://danilowoz.github.io/create-react-content-loader/) to create\n- your customs loadings easily;\n+ your custom loadings easily;\n * :ok_hand: **You can use right now:** there are a lot of presets to use the loader, see the\n [options](#options);\n-* :rocket: **Perfomance:** react-content-loader uses pure SVG to work, so it's works without any extra scritpt,\n+* :rocket: **Performance:** react-content-loader uses pure SVG to work, so it works without any extra scripts,\n canvas, etc;\n \n ## Usage\n"]
2
["a5ecfdf49b0d4c43fbbbf7947be7c0327ccb3415", "88257ee720ed8ba136d49087c0d31373e8397dd5"]
["test", "docs"]
expose the means by which we process each reward cycle's affirmation maps at reward cycle boundaries,Fix readme Signed-off-by: Ben Johnson <[email protected]>
["diff --git a/src/burnchains/burnchain.rs b/src/burnchains/burnchain.rs\nindex 92105d6..60c608a 100644\n--- a/src/burnchains/burnchain.rs\n+++ b/src/burnchains/burnchain.rs\n@@ -851,8 +851,26 @@ impl Burnchain {\n );\n \n burnchain_db.store_new_burnchain_block(burnchain, indexer, &block)?;\n- let block_height = block.block_height();\n+ Burnchain::process_affirmation_maps(\n+ burnchain,\n+ burnchain_db,\n+ indexer,\n+ block.block_height(),\n+ )?;\n+\n+ let header = block.header();\n+ Ok(header)\n+ }\n \n+ /// Update the affirmation maps for the previous reward cycle's commits.\n+ /// This is a no-op unless the given burnchain block height falls on a reward cycle boundary. In that\n+ /// case, the previous reward cycle's block commits' affirmation maps are all re-calculated.\n+ pub fn process_affirmation_maps<B: BurnchainHeaderReader>(\n+ burnchain: &Burnchain,\n+ burnchain_db: &mut BurnchainDB,\n+ indexer: &B,\n+ block_height: u64,\n+ ) -> Result<(), burnchain_error> {\n let this_reward_cycle = burnchain\n .block_height_to_reward_cycle(block_height)\n .unwrap_or(0);\n@@ -872,10 +890,7 @@ impl Burnchain {\n );\n update_pox_affirmation_maps(burnchain_db, indexer, prev_reward_cycle, burnchain)?;\n }\n-\n- let header = block.header();\n-\n- Ok(header)\n+ Ok(())\n }\n \n /// Hand off the block to the ChainsCoordinator _and_ process the sortition\n", "diff --git a/README.md b/README.md\nindex 587d655..da746bb 100644\n--- a/README.md\n+++ b/README.md\n@@ -38,15 +38,20 @@ simple and unified.\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\n * [**Configuration**][docs.configuration]\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\n \n-#### [Components](https://vector.dev/components)\n+#### Reference\n \n-* [**Sources**][docs.sources] - \n-* [**Transforms**][docs.transforms]\n-* [**Sinks**][docs.sinks]\n+* [**Sources**][docs.sources] - [docker][docs.sources.docker], [file][docs.sources.file], [journald][docs.sources.journald], [kafka][docs.sources.kafka]\n+* [**Transforms**][docs.transforms] - [json_parser][docs.transforms.json_parser], [log_to_metric][docs.transforms.log_to_metric], [lua][docs.transforms.lua], [regex_parser][docs.transforms.regex_parser]\n+* [**Sinks**][docs.sinks] - [aws_cloudwatch_logs][docs.sinks.aws_cloudwatch_logs], [aws_cloudwatch_metrics][docs.sinks.aws_cloudwatch_metrics], [aws_kinesis_streams][docs.sinks.aws_kinesis_streams], [aws_s3][docs.sinks.aws_s3], [clickhouse][docs.sinks.clickhouse], [elasticsearch][docs.sinks.elasticsearch], and [15 more][docs.sinks]\n \n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\n-* [**Guides**][docs.guides]\n+#### Administration\n+\n+* [**Process management**][docs.process-management]\n+* [**Monitoring**][docs.monitoring]\n+* [**Updating**][docs.updating]\n+* [**Validating**][docs.validating]\n \n #### Resources\n \n@@ -105,88 +110,6 @@ curl --proto '=https' --tlsv1.2 -sSf https://sh.vector.dev | sh\n \n Or view [platform specific installation instructions][docs.installation].\n \n-\n-## Sources\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`docker`**][docs.sources.docker] | Ingests data through the docker engine daemon and outputs [`log`][docs.data-model#log] events. |\n-| [**`file`**][docs.sources.file] | Ingests data through one or more local files and outputs [`log`][docs.data-model#log] events. |\n-| [**`journald`**][docs.sources.journald] | Ingests data through log records from journald and outputs [`log`][docs.data-model#log] events. |\n-| [**`kafka`**][docs.sources.kafka] | Ingests data through Kafka 0.9 or later and outputs [`log`][docs.data-model#log] events. |\n-| [**`statsd`**][docs.sources.statsd] | Ingests data through the StatsD UDP protocol and outputs [`metric`][docs.data-model#metric] events. |\n-| [**`stdin`**][docs.sources.stdin] | Ingests data through standard input (STDIN) and outputs [`log`][docs.data-model#log] events. |\n-| [**`syslog`**][docs.sources.syslog] | Ingests data through the Syslog 5424 protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`tcp`**][docs.sources.tcp] | Ingests data through the TCP protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`udp`**][docs.sources.udp] | Ingests data through the UDP protocol and outputs [`log`][docs.data-model#log] events. |\n-| [**`vector`**][docs.sources.vector] | Ingests data through another upstream [`vector` sink][docs.sinks.vector] and outputs [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events. |\n-\n-[+ request a new source][urls.new_source]\n-\n-\n-## Transforms\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`add_fields`**][docs.transforms.add_fields] | Accepts [`log`][docs.data-model#log] events and allows you to add one or more log fields. |\n-| [**`add_tags`**][docs.transforms.add_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to add one or more metric tags. |\n-| [**`coercer`**][docs.transforms.coercer] | Accepts [`log`][docs.data-model#log] events and allows you to coerce log fields into fixed types. |\n-| [**`field_filter`**][docs.transforms.field_filter] | Accepts [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events and allows you to filter events by a log field's value. |\n-| [**`grok_parser`**][docs.transforms.grok_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value with [Grok][urls.grok]. |\n-| [**`json_parser`**][docs.transforms.json_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value as JSON. |\n-| [**`log_to_metric`**][docs.transforms.log_to_metric] | Accepts [`log`][docs.data-model#log] events and allows you to convert logs into one or more metrics. |\n-| [**`lua`**][docs.transforms.lua] | Accepts [`log`][docs.data-model#log] events and allows you to transform events with a full embedded [Lua][urls.lua] engine. |\n-| [**`regex_parser`**][docs.transforms.regex_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field's value with a [Regular Expression][urls.regex]. |\n-| [**`remove_fields`**][docs.transforms.remove_fields] | Accepts [`log`][docs.data-model#log] events and allows you to remove one or more log fields. |\n-| [**`remove_tags`**][docs.transforms.remove_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to remove one or more metric tags. |\n-| [**`sampler`**][docs.transforms.sampler] | Accepts [`log`][docs.data-model#log] events and allows you to sample events with a configurable rate. |\n-| [**`split`**][docs.transforms.split] | Accepts [`log`][docs.data-model#log] events and allows you to split a field's value on a given separator and zip the tokens into ordered field names. |\n-| [**`tokenizer`**][docs.transforms.tokenizer] | Accepts [`log`][docs.data-model#log] events and allows you to tokenize a field's value by splitting on white space, ignoring special wrapping characters, and zip the tokens into ordered field names. |\n-\n-[+ request a new transform][urls.new_transform]\n-\n-\n-## Sinks\n-\n-| Name | Description |\n-|:------|:------------|\n-| [**`aws_cloudwatch_logs`**][docs.sinks.aws_cloudwatch_logs] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS CloudWatch Logs][urls.aws_cw_logs] via the [`PutLogEvents` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutLogEvents.html). |\n-| [**`aws_cloudwatch_metrics`**][docs.sinks.aws_cloudwatch_metrics] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [AWS CloudWatch Metrics][urls.aws_cw_metrics] via the [`PutMetricData` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatch/latest/APIReference/API_PutMetricData.html). |\n-| [**`aws_kinesis_streams`**][docs.sinks.aws_kinesis_streams] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS Kinesis Data Stream][urls.aws_kinesis_data_streams] via the [`PutRecords` API endpoint](https://docs.aws.amazon.com/kinesis/latest/APIReference/API_PutRecords.html). |\n-| [**`aws_s3`**][docs.sinks.aws_s3] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS S3][urls.aws_s3] via the [`PutObject` API endpoint](https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectPUT.html). |\n-| [**`blackhole`**][docs.sinks.blackhole] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to a blackhole that simply discards data, designed for testing and benchmarking purposes. |\n-| [**`clickhouse`**][docs.sinks.clickhouse] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Clickhouse][urls.clickhouse] via the [`HTTP` Interface][urls.clickhouse_http]. |\n-| [**`console`**][docs.sinks.console] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to [standard output streams][urls.standard_streams], such as `STDOUT` and `STDERR`. |\n-| [**`datadog_metrics`**][docs.sinks.datadog_metrics] | [Batches](#buffers-and-batches) [`metric`][docs.data-model#metric] events to [Datadog][urls.datadog] metrics service using [HTTP API](https://docs.datadoghq.com/api/?lang=bash#metrics). |\n-| [**`elasticsearch`**][docs.sinks.elasticsearch] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Elasticsearch][urls.elasticsearch] via the [`_bulk` API endpoint](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-bulk.html). |\n-| [**`file`**][docs.sinks.file] | [Streams](#streaming) [`log`][docs.data-model#log] events to a file. |\n-| [**`http`**][docs.sinks.http] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a generic HTTP endpoint. |\n-| [**`kafka`**][docs.sinks.kafka] | [Streams](#streaming) [`log`][docs.data-model#log] events to [Apache Kafka][urls.kafka] via the [Kafka protocol][urls.kafka_protocol]. |\n-| [**`prometheus`**][docs.sinks.prometheus] | [Exposes](#exposing-and-scraping) [`metric`][docs.data-model#metric] events to [Prometheus][urls.prometheus] metrics service. |\n-| [**`splunk_hec`**][docs.sinks.splunk_hec] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a [Splunk HTTP Event Collector][urls.splunk_hec]. |\n-| [**`statsd`**][docs.sinks.statsd] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [StatsD][urls.statsd] metrics service. |\n-| [**`tcp`**][docs.sinks.tcp] | [Streams](#streaming) [`log`][docs.data-model#log] events to a TCP connection. |\n-| [**`vector`**][docs.sinks.vector] | [Streams](#streaming) [`log`][docs.data-model#log] events to another downstream [`vector` source][docs.sources.vector]. |\n-\n-[+ request a new sink][urls.new_sink]\n-\n-\n-## License\n-\n-Copyright 2019, Vector Authors. All rights reserved.\n-\n-Licensed under the Apache License, Version 2.0 (the \"License\"); you may not\n-use these files except in compliance with the License. You may obtain a copy\n-of the License at\n-\n-http://www.apache.org/licenses/LICENSE-2.0\n-\n-Unless required by applicable law or agreed to in writing, software\n-distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n-License for the specific language governing permissions and limitations under\n-the License.\n-\n ---\n \n <p align=\"center\">\n@@ -200,8 +123,6 @@ the License.\n [docs.configuration]: https://vector.dev/docs/setup/configuration\n [docs.containers]: https://vector.dev/docs/setup/installation/containers\n [docs.correctness]: https://vector.dev/docs/about/correctness\n-[docs.data-model#log]: https://vector.dev/docs/about/data-model#log\n-[docs.data-model#metric]: https://vector.dev/docs/about/data-model#metric\n [docs.data-model.log]: https://vector.dev/docs/about/data-model/log\n [docs.data-model.metric]: https://vector.dev/docs/about/data-model/metric\n [docs.data_model]: https://vector.dev/docs/about/data-model\n@@ -209,6 +130,7 @@ the License.\n [docs.from-archives]: https://vector.dev/docs/setup/installation/manual/from-archives\n [docs.from-source]: https://vector.dev/docs/setup/installation/manual/from-source\n [docs.guarantees]: https://vector.dev/docs/about/guarantees\n+[docs.guides.getting_started]: https://vector.dev/docs/setup/guides/getting-started\n [docs.guides]: https://vector.dev/docs/setup/guides\n [docs.installation]: https://vector.dev/docs/setup/installation\n [docs.monitoring]: https://vector.dev/docs/administration/monitoring\n@@ -224,72 +146,25 @@ the License.\n [docs.sinks.aws_cloudwatch_metrics]: https://vector.dev/docs/reference/sinks/aws_cloudwatch_metrics\n [docs.sinks.aws_kinesis_streams]: https://vector.dev/docs/reference/sinks/aws_kinesis_streams\n [docs.sinks.aws_s3]: https://vector.dev/docs/reference/sinks/aws_s3\n-[docs.sinks.blackhole]: https://vector.dev/docs/reference/sinks/blackhole\n [docs.sinks.clickhouse]: https://vector.dev/docs/reference/sinks/clickhouse\n-[docs.sinks.console]: https://vector.dev/docs/reference/sinks/console\n-[docs.sinks.datadog_metrics]: https://vector.dev/docs/reference/sinks/datadog_metrics\n [docs.sinks.elasticsearch]: https://vector.dev/docs/reference/sinks/elasticsearch\n-[docs.sinks.file]: https://vector.dev/docs/reference/sinks/file\n-[docs.sinks.http]: https://vector.dev/docs/reference/sinks/http\n-[docs.sinks.kafka]: https://vector.dev/docs/reference/sinks/kafka\n-[docs.sinks.prometheus]: https://vector.dev/docs/reference/sinks/prometheus\n-[docs.sinks.splunk_hec]: https://vector.dev/docs/reference/sinks/splunk_hec\n-[docs.sinks.statsd]: https://vector.dev/docs/reference/sinks/statsd\n-[docs.sinks.tcp]: https://vector.dev/docs/reference/sinks/tcp\n-[docs.sinks.vector]: https://vector.dev/docs/reference/sinks/vector\n [docs.sinks]: https://vector.dev/docs/reference/sinks\n [docs.sources.docker]: https://vector.dev/docs/reference/sources/docker\n [docs.sources.file]: https://vector.dev/docs/reference/sources/file\n [docs.sources.journald]: https://vector.dev/docs/reference/sources/journald\n [docs.sources.kafka]: https://vector.dev/docs/reference/sources/kafka\n-[docs.sources.statsd]: https://vector.dev/docs/reference/sources/statsd\n-[docs.sources.stdin]: https://vector.dev/docs/reference/sources/stdin\n-[docs.sources.syslog]: https://vector.dev/docs/reference/sources/syslog\n-[docs.sources.tcp]: https://vector.dev/docs/reference/sources/tcp\n-[docs.sources.udp]: https://vector.dev/docs/reference/sources/udp\n-[docs.sources.vector]: https://vector.dev/docs/reference/sources/vector\n [docs.sources]: https://vector.dev/docs/reference/sources\n [docs.topologies]: https://vector.dev/docs/setup/deployment/topologies\n-[docs.transforms.add_fields]: https://vector.dev/docs/reference/transforms/add_fields\n-[docs.transforms.add_tags]: https://vector.dev/docs/reference/transforms/add_tags\n-[docs.transforms.coercer]: https://vector.dev/docs/reference/transforms/coercer\n-[docs.transforms.field_filter]: https://vector.dev/docs/reference/transforms/field_filter\n-[docs.transforms.grok_parser]: https://vector.dev/docs/reference/transforms/grok_parser\n [docs.transforms.json_parser]: https://vector.dev/docs/reference/transforms/json_parser\n [docs.transforms.log_to_metric]: https://vector.dev/docs/reference/transforms/log_to_metric\n [docs.transforms.lua]: https://vector.dev/docs/reference/transforms/lua\n [docs.transforms.regex_parser]: https://vector.dev/docs/reference/transforms/regex_parser\n-[docs.transforms.remove_fields]: https://vector.dev/docs/reference/transforms/remove_fields\n-[docs.transforms.remove_tags]: https://vector.dev/docs/reference/transforms/remove_tags\n-[docs.transforms.sampler]: https://vector.dev/docs/reference/transforms/sampler\n-[docs.transforms.split]: https://vector.dev/docs/reference/transforms/split\n-[docs.transforms.tokenizer]: https://vector.dev/docs/reference/transforms/tokenizer\n [docs.transforms]: https://vector.dev/docs/reference/transforms\n [docs.updating]: https://vector.dev/docs/administration/updating\n [docs.use_cases]: https://vector.dev/docs/use_cases\n [docs.validating]: https://vector.dev/docs/administration/validating\n-[urls.aws_cw_logs]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/WhatIsCloudWatchLogs.html\n-[urls.aws_cw_metrics]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/working_with_metrics.html\n-[urls.aws_kinesis_data_streams]: https://aws.amazon.com/kinesis/data-streams/\n-[urls.aws_s3]: https://aws.amazon.com/s3/\n-[urls.clickhouse]: https://clickhouse.yandex/\n-[urls.clickhouse_http]: https://clickhouse.yandex/docs/en/interfaces/http/\n-[urls.datadog]: https://www.datadoghq.com\n-[urls.elasticsearch]: https://www.elastic.co/products/elasticsearch\n-[urls.grok]: http://grokdebug.herokuapp.com/\n-[urls.kafka]: https://kafka.apache.org/\n-[urls.kafka_protocol]: https://kafka.apache.org/protocol\n-[urls.lua]: https://www.lua.org/\n [urls.mailing_list]: https://vector.dev/mailing_list/\n-[urls.new_sink]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.new_source]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.new_transform]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\n-[urls.prometheus]: https://prometheus.io/\n-[urls.regex]: https://en.wikipedia.org/wiki/Regular_expression\n [urls.rust]: https://www.rust-lang.org/\n-[urls.splunk_hec]: http://dev.splunk.com/view/event-collector/SP-CAAAE6M\n-[urls.standard_streams]: https://en.wikipedia.org/wiki/Standard_streams\n-[urls.statsd]: https://github.com/statsd/statsd\n [urls.test_harness]: https://github.com/timberio/vector-test-harness/\n [urls.v0.5.0]: https://github.com/timberio/vector/releases/tag/v0.5.0\n [urls.vector_changelog]: https://github.com/timberio/vector/blob/master/CHANGELOG.md\ndiff --git a/README.md.erb b/README.md.erb\nindex 3b14aa0..cc241eb 100644\n--- a/README.md.erb\n+++ b/README.md.erb\n@@ -38,15 +38,20 @@ simple and unified.\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\n * [**Configuration**][docs.configuration]\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\n \n-#### [Components](https://vector.dev/components)\n+#### Reference\n \n-* [**Sources**][docs.sources] - \n-* [**Transforms**][docs.transforms]\n-* [**Sinks**][docs.sinks]\n+* [**Sources**][docs.sources] - <%= common_component_links(:source) %>\n+* [**Transforms**][docs.transforms] - <%= common_component_links(:transform) %>\n+* [**Sinks**][docs.sinks] - <%= common_component_links(:sink) %>\n \n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\n-* [**Guides**][docs.guides]\n+#### Administration\n+\n+* [**Process management**][docs.process-management]\n+* [**Monitoring**][docs.monitoring]\n+* [**Updating**][docs.updating]\n+* [**Validating**][docs.validating]\n \n #### Resources\n \n@@ -105,44 +110,6 @@ Run the following in your terminal, then follow the on-screen instructions.\n \n Or view [platform specific installation instructions][docs.installation].\n \n-\n-## Sources\n-\n-<%= components_table(metadata.sources.to_h.values.sort) %>\n-\n-[+ request a new source][urls.new_source]\n-\n-\n-## Transforms\n-\n-<%= components_table(metadata.transforms.to_h.values.sort) %>\n-\n-[+ request a new transform][urls.new_transform]\n-\n-\n-## Sinks\n-\n-<%= components_table(metadata.sinks.to_h.values.sort) %>\n-\n-[+ request a new sink][urls.new_sink]\n-\n-\n-## License\n-\n-Copyright <%= Time.now.year %>, Vector Authors. All rights reserved.\n-\n-Licensed under the Apache License, Version 2.0 (the \"License\"); you may not\n-use these files except in compliance with the License. You may obtain a copy\n-of the License at\n-\n-http://www.apache.org/licenses/LICENSE-2.0\n-\n-Unless required by applicable law or agreed to in writing, software\n-distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n-License for the specific language governing permissions and limitations under\n-the License.\n-\n ---\n \n <p align=\"center\">\ndiff --git a/scripts/generate/templates.rb b/scripts/generate/templates.rb\nindex e5e7ce7..c793ae0 100644\n--- a/scripts/generate/templates.rb\n+++ b/scripts/generate/templates.rb\n@@ -89,6 +89,23 @@ class Templates\n render(\"#{partials_path}/_commit_type_toc_item.md\", binding).gsub(/,$/, \"\")\n end\n \n+ def common_component_links(type, limit = 5)\n+ common = metadata.send(\"#{type.to_s.pluralize}_list\").select(&:common?)\n+\n+ links =\n+ common[0..limit].collect do |component|\n+ \"[#{component.name}][docs.#{type.to_s.pluralize}.#{component.name}]\"\n+ end\n+\n+ num_leftover = common.size - links.size\n+\n+ if num_leftover > 0\n+ links << \"and [15 more][docs.#{type.to_s.pluralize}]\"\n+ end\n+\n+ links.join(\", \")\n+ end\n+\n def component_config_example(component)\n render(\"#{partials_path}/_component_config_example.md\", binding).strip\n end\ndiff --git a/scripts/util/metadata/component.rb b/scripts/util/metadata/component.rb\nindex 0873b2e..4dc5650 100644\n--- a/scripts/util/metadata/component.rb\n+++ b/scripts/util/metadata/component.rb\n@@ -9,6 +9,7 @@ class Component\n include Comparable\n \n attr_reader :beta,\n+ :common,\n :function_category,\n :id,\n :name,\n@@ -18,6 +19,7 @@ class Component\n \n def initialize(hash)\n @beta = hash[\"beta\"] == true\n+ @common = hash[\"common\"] == true\n @function_category = hash.fetch(\"function_category\")\n @name = hash.fetch(\"name\")\n @type ||= self.class.name.downcase\n@@ -71,6 +73,10 @@ class Component\n beta == true\n end\n \n+ def common?\n+ common == true\n+ end\n+\n def context_options\n options_list.select(&:context?)\n end\ndiff --git a/website/src/components/VectorComponents/index.js b/website/src/components/VectorComponents/index.js\nindex b6c5c13..d3c9adf 100644\n--- a/website/src/components/VectorComponents/index.js\n+++ b/website/src/components/VectorComponents/index.js\n@@ -154,7 +154,7 @@ function VectorComponents(props) {\n //\n \n const [onlyAtLeastOnce, setOnlyAtLeastOnce] = useState(queryObj['at-least-once'] == 'true');\n- const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['providers']));\n+ const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['functions']));\n const [onlyLog, setOnlyLog] = useState(queryObj['log'] == 'true');\n const [onlyMetric, setOnlyMetric] = useState(queryObj['metric'] == 'true');\n const [onlyProductionReady, setOnlyProductionReady] = useState(queryObj['prod-ready'] == 'true');\n"]
2
["d7972da833257c073403dec3c2ac3a7f297e328a", "662c5d1346ea2b01c0bc3c11c648cbdf92035fe2"]
["refactor", "docs"]
make sure root is being watched and setRoot called when it changes,add react ecosystem
["diff --git a/packages/core/src/components/nav/nav.tsx b/packages/core/src/components/nav/nav.tsx\nindex 5aaacb6..27241ee 100644\n--- a/packages/core/src/components/nav/nav.tsx\n+++ b/packages/core/src/components/nav/nav.tsx\n@@ -1,4 +1,4 @@\n-import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n+import { Component, Element, Event, EventEmitter, Listen, Method, Prop, Watch } from '@stencil/core';\n import {\n Animation,\n AnimationController,\n@@ -103,10 +103,19 @@ export class Nav implements PublicNav, NavContainer {\n }\n this.init = true;\n if (!this.useRouter) {\n+ console.log('componentDidLoadImpl: ', this.root);\n componentDidLoadImpl(this);\n }\n }\n \n+ @Watch('root')\n+ updateRootComponent(): any {\n+ console.log('updateRootComponent: ', this.root);\n+ if (this.init) {\n+ return this.setRoot(this.root);\n+ }\n+ }\n+\n getViews(): PublicViewController[] {\n return getViews(this);\n }\ndiff --git a/packages/core/src/components/nav/test/set-root/index.html b/packages/core/src/components/nav/test/set-root/index.html\nnew file mode 100644\nindex 0000000..823c9ed\n--- /dev/null\n+++ b/packages/core/src/components/nav/test/set-root/index.html\n@@ -0,0 +1,110 @@\n+<!DOCTYPE html>\n+<html dir=\"ltr\">\n+<head>\n+ <meta charset=\"UTF-8\">\n+ <title>Nav</title>\n+ <meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0, minimum-scale=1.0, maximum-scale=1.0, user-scalable=no\">\n+ <script src=\"/dist/ionic.js\"></script>\n+</head>\n+<body onload=\"initiaize()\">\n+ <ion-app>\n+ <ion-nav root=\"page-one\"></ion-nav>\n+ </ion-app>\n+</body>\n+\n+<script>\n+\n+ class PageOne extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page One</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page One</h1>\n+ <ion-button class=\"next\">Go to Page Two</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const button = this.querySelector('ion-button');\n+ button.addEventListener('click', async () => {\n+ this.closest('ion-nav').push('page-two');\n+ });\n+ }\n+ }\n+\n+ class PageTwo extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page Two</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page Two</h1>\n+ <ion-button class=\"next\">Go to Page Three</ion-button>\n+ <ion-button class=\"previous\">Go Back</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const previousButton = this.querySelector('ion-button.previous');\n+ previousButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').pop();\n+ });\n+\n+ const nextButton = this.querySelector('ion-button.next');\n+ nextButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').push('page-three');\n+ });\n+ }\n+ }\n+\n+ class PageThree extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page Three</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page Three</h1>\n+ <ion-button class=\"previous\">Go Back</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const previousButton = this.querySelector('ion-button.previous');\n+ previousButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').pop();\n+ });\n+ }\n+ }\n+\n+ customElements.define('page-one', PageOne);\n+ customElements.define('page-two', PageTwo);\n+ customElements.define('page-three', PageThree);\n+\n+ async function initiaize() {\n+ const nav = document.querySelector('ion-nav');\n+ await nav.componentOnReady();\n+ nav.root = 'page-one';\n+\n+ setInterval(() => {\n+ if (nav.root === 'page-one') {\n+ nav.root = 'page-two';\n+ } else if ( nav.root === 'page-two') {\n+ nav.root = 'page-three';\n+ } else {\n+ nav.root = 'page-one';\n+ }\n+ }, 1000);\n+ }\n+\n+</script>\n+</html>\n", "diff --git a/package.json b/package.json\nindex 1ba8c4f..d1de9a0 100644\n--- a/package.json\n+++ b/package.json\n@@ -36,14 +36,19 @@\n \"@types/node\": \"^9.3.0\",\n \"@types/react\": \"^16.0.34\",\n \"@types/react-dom\": \"^16.0.3\",\n+ \"@types/react-motion\": \"^0.0.25\",\n \"bootstrap-sass\": \"^3.3.7\",\n \"highcharts\": \"^6.0.4\",\n \"html2canvas\": \"^1.0.0-alpha.9\",\n+ \"immer\": \"^1.2.1\",\n \"lodash\": \"^4.17.4\",\n \"moment\": \"^2.20.1\",\n \"normalize.css\": \"^8.0.0\",\n- \"react\": \"^16.2.0\",\n- \"react-dom\": \"^16.2.0\",\n+ \"react\": \"^16.3.1\",\n+ \"react-dom\": \"^16.3.1\",\n+ \"react-motion\": \"^0.5.2\",\n+ \"react-redux\": \"^5.0.7\",\n+ \"redux\": \"^3.7.2\",\n \"rxjs\": \"^5.5.6\",\n \"vue\": \"^2.5.13\",\n \"vue-plugin-webextension-i18n\": \"^0.1.0\",\ndiff --git a/yarn.lock b/yarn.lock\nindex c8898d8..5d0fc9f 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -187,6 +187,12 @@\n \"@types/node\" \"*\"\n \"@types/react\" \"*\"\n \n+\"@types/react-motion@^0.0.25\":\n+ version \"0.0.25\"\n+ resolved \"https://registry.npmjs.org/@types/react-motion/-/react-motion-0.0.25.tgz#2445745ee8e8e6149faa47a36ff6b0d4c21dbf94\"\n+ dependencies:\n+ \"@types/react\" \"*\"\n+\n \"@types/react@*\", \"@types/react@^16.0.34\":\n version \"16.0.40\"\n resolved \"https://registry.npmjs.org/@types/react/-/react-16.0.40.tgz#caabc2296886f40b67f6fc80f0f3464476461df9\"\n@@ -3837,6 +3843,10 @@ [email protected]:\n version \"4.2.1\"\n resolved \"https://registry.npmjs.org/hoek/-/hoek-4.2.1.tgz#9634502aa12c445dd5a7c5734b572bb8738aacbb\"\n \n+hoist-non-react-statics@^2.5.0:\n+ version \"2.5.0\"\n+ resolved \"https://registry.npmjs.org/hoist-non-react-statics/-/hoist-non-react-statics-2.5.0.tgz#d2ca2dfc19c5a91c5a6615ce8e564ef0347e2a40\"\n+\n home-or-tmp@^2.0.0:\n version \"2.0.0\"\n resolved \"https://registry.npmjs.org/home-or-tmp/-/home-or-tmp-2.0.0.tgz#e36c3f2d2cae7d746a857e38d18d5f32a7882db8\"\n@@ -4004,6 +4014,10 @@ ignore@^3.3.5:\n version \"3.3.7\"\n resolved \"https://registry.npmjs.org/ignore/-/ignore-3.3.7.tgz#612289bfb3c220e186a58118618d5be8c1bab021\"\n \n+immer@^1.2.1:\n+ version \"1.2.1\"\n+ resolved \"https://registry.npmjs.org/immer/-/immer-1.2.1.tgz#96e2ae29cdfc428f28120b832701931b92fa597c\"\n+\n import-local@^1.0.0:\n version \"1.0.0\"\n resolved \"https://registry.npmjs.org/import-local/-/import-local-1.0.0.tgz#5e4ffdc03f4fe6c009c6729beb29631c2f8227bc\"\n@@ -4104,7 +4118,7 @@ interpret@^1.0.0:\n version \"1.1.0\"\n resolved \"https://registry.npmjs.org/interpret/-/interpret-1.1.0.tgz#7ed1b1410c6a0e0f78cf95d3b8440c63f78b8614\"\n \n-invariant@^2.2.2:\n+invariant@^2.0.0, invariant@^2.2.2:\n version \"2.2.4\"\n resolved \"https://registry.npmjs.org/invariant/-/invariant-2.2.4.tgz#610f3c92c9359ce1db616e538008d23ff35158e6\"\n dependencies:\n@@ -5040,6 +5054,10 @@ locate-path@^2.0.0:\n p-locate \"^2.0.0\"\n path-exists \"^3.0.0\"\n \n+lodash-es@^4.17.5, lodash-es@^4.2.1:\n+ version \"4.17.8\"\n+ resolved \"https://registry.npmjs.org/lodash-es/-/lodash-es-4.17.8.tgz#6fa8c8c5d337481df0bdf1c0d899d42473121e45\"\n+\n lodash._reinterpolate@~3.0.0:\n version \"3.0.0\"\n resolved \"https://registry.npmjs.org/lodash._reinterpolate/-/lodash._reinterpolate-3.0.0.tgz#0ccf2d89166af03b3663c796538b75ac6e114d9d\"\n@@ -5149,7 +5167,7 @@ [email protected]:\n version \"4.17.2\"\n resolved \"https://registry.npmjs.org/lodash/-/lodash-4.17.2.tgz#34a3055babe04ce42467b607d700072c7ff6bf42\"\n \[email protected], lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\[email protected], lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.17.5, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\n version \"4.17.5\"\n resolved \"https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz#99a92d65c0272debe8c96b6057bc8fbfa3bed511\"\n \n@@ -6467,7 +6485,7 @@ promise@^7.1.1:\n dependencies:\n asap \"~2.0.3\"\n \n-prop-types@^15.6.0:\n+prop-types@^15.5.8, prop-types@^15.6.0:\n version \"15.6.1\"\n resolved \"https://registry.npmjs.org/prop-types/-/prop-types-15.6.1.tgz#36644453564255ddda391191fb3a125cbdf654ca\"\n dependencies:\n@@ -6574,7 +6592,7 @@ quick-lru@^1.0.0:\n version \"1.1.0\"\n resolved \"https://registry.npmjs.org/quick-lru/-/quick-lru-1.1.0.tgz#4360b17c61136ad38078397ff11416e186dcfbb8\"\n \[email protected]:\[email protected], raf@^3.1.0:\n version \"3.4.0\"\n resolved \"https://registry.npmjs.org/raf/-/raf-3.4.0.tgz#a28876881b4bc2ca9117d4138163ddb80f781575\"\n dependencies:\n@@ -6645,9 +6663,9 @@ react-dev-utils@^5.0.0:\n strip-ansi \"3.0.1\"\n text-table \"0.2.0\"\n \n-react-dom@^16.2.0:\n- version \"16.2.0\"\n- resolved \"https://registry.npmjs.org/react-dom/-/react-dom-16.2.0.tgz#69003178601c0ca19b709b33a83369fe6124c044\"\n+react-dom@^16.3.1:\n+ version \"16.3.1\"\n+ resolved \"https://registry.npmjs.org/react-dom/-/react-dom-16.3.1.tgz#6a3c90a4fb62f915bdbcf6204422d93a7d4ca573\"\n dependencies:\n fbjs \"^0.8.16\"\n loose-envify \"^1.1.0\"\n@@ -6658,9 +6676,28 @@ react-error-overlay@^4.0.0:\n version \"4.0.0\"\n resolved \"https://registry.npmjs.org/react-error-overlay/-/react-error-overlay-4.0.0.tgz#d198408a85b4070937a98667f500c832f86bd5d4\"\n \n-react@^16.2.0:\n- version \"16.2.0\"\n- resolved \"https://registry.npmjs.org/react/-/react-16.2.0.tgz#a31bd2dab89bff65d42134fa187f24d054c273ba\"\n+react-motion@^0.5.2:\n+ version \"0.5.2\"\n+ resolved \"https://registry.npmjs.org/react-motion/-/react-motion-0.5.2.tgz#0dd3a69e411316567927917c6626551ba0607316\"\n+ dependencies:\n+ performance-now \"^0.2.0\"\n+ prop-types \"^15.5.8\"\n+ raf \"^3.1.0\"\n+\n+react-redux@^5.0.7:\n+ version \"5.0.7\"\n+ resolved \"https://registry.npmjs.org/react-redux/-/react-redux-5.0.7.tgz#0dc1076d9afb4670f993ffaef44b8f8c1155a4c8\"\n+ dependencies:\n+ hoist-non-react-statics \"^2.5.0\"\n+ invariant \"^2.0.0\"\n+ lodash \"^4.17.5\"\n+ lodash-es \"^4.17.5\"\n+ loose-envify \"^1.1.0\"\n+ prop-types \"^15.6.0\"\n+\n+react@^16.3.1:\n+ version \"16.3.1\"\n+ resolved \"https://registry.npmjs.org/react/-/react-16.3.1.tgz#4a2da433d471251c69b6033ada30e2ed1202cfd8\"\n dependencies:\n fbjs \"^0.8.16\"\n loose-envify \"^1.1.0\"\n@@ -6788,6 +6825,15 @@ reduce-function-call@^1.0.1:\n dependencies:\n balanced-match \"^0.4.2\"\n \n+redux@^3.7.2:\n+ version \"3.7.2\"\n+ resolved \"https://registry.npmjs.org/redux/-/redux-3.7.2.tgz#06b73123215901d25d065be342eb026bc1c8537b\"\n+ dependencies:\n+ lodash \"^4.2.1\"\n+ lodash-es \"^4.2.1\"\n+ loose-envify \"^1.1.0\"\n+ symbol-observable \"^1.0.3\"\n+\n regenerate@^1.2.1:\n version \"1.3.3\"\n resolved \"https://registry.npmjs.org/regenerate/-/regenerate-1.3.3.tgz#0c336d3980553d755c39b586ae3b20aa49c82b7f\"\n@@ -7811,6 +7857,10 @@ [email protected]:\n version \"1.0.1\"\n resolved \"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.0.1.tgz#8340fc4702c3122df5d22288f88283f513d3fdd4\"\n \n+symbol-observable@^1.0.3:\n+ version \"1.2.0\"\n+ resolved \"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.2.0.tgz#c22688aed4eab3cdc2dfeacbb561660560a00804\"\n+\n symbol-tree@^3.2.2:\n version \"3.2.2\"\n resolved \"https://registry.npmjs.org/symbol-tree/-/symbol-tree-3.2.2.tgz#ae27db38f660a7ae2e1c3b7d1bc290819b8519e6\"\n"]
2
["4be836f5655fb5356fde5ddd7437125f8574705d", "7e04a5e829d7416e312ac342a00a11787745753b"]
["refactor", "build"]
rework RaftCommittedEntryListener Iterate over RaftCommittedEntryListener and refactor the listener such it serves the actual need. We have some services (to be specific the AsyncSnapshotDirector) which need the committed position, and want to listen to new updates. In raft we know which record we are committing and whether it was an application record so we can pass this information threw the listeners. This avoids to pass in the whole IndexedRecord object, and reduce the potential of going out of OOM because of keeping to much data in heap (when commit is not possible).,apply element migrated events This is a very straightforward event applier. All it needs to do is update the persisted data for the element instance using the data in the event.
["diff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\nnew file mode 100644\nindex 0000000..57c28a9\n--- /dev/null\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\n@@ -0,0 +1,31 @@\n+/*\n+ * Copyright 2016-present Open Networking Foundation\n+ * Copyright \u00a9 2020 camunda services GmbH ([email protected])\n+ *\n+ * Licensed under the Apache License, Version 2.0 (the \"License\");\n+ * you may not use this file except in compliance with the License.\n+ * You may obtain a copy of the License at\n+ *\n+ * http://www.apache.org/licenses/LICENSE-2.0\n+ *\n+ * Unless required by applicable law or agreed to in writing, software\n+ * distributed under the License is distributed on an \"AS IS\" BASIS,\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n+ * See the License for the specific language governing permissions and\n+ * limitations under the License.\n+ */\n+package io.atomix.raft;\n+\n+/**\n+ * This listener will only be called by the Leader, when it commits an application entry.\n+ *\n+ * <p>If RAFT is currently running in a follower role, it will not call this listener.\n+ */\n+@FunctionalInterface\n+public interface RaftApplicationEntryCommittedPositionListener {\n+\n+ /**\n+ * @param committedPosition the new committed position which is related to the application entries\n+ */\n+ void onCommit(long committedPosition);\n+}\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\ndeleted file mode 100644\nindex 3d11d75..0000000\n--- a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\n+++ /dev/null\n@@ -1,32 +0,0 @@\n-/*\n- * Copyright 2016-present Open Networking Foundation\n- * Copyright \u00a9 2020 camunda services GmbH ([email protected])\n- *\n- * Licensed under the Apache License, Version 2.0 (the \"License\");\n- * you may not use this file except in compliance with the License.\n- * You may obtain a copy of the License at\n- *\n- * http://www.apache.org/licenses/LICENSE-2.0\n- *\n- * Unless required by applicable law or agreed to in writing, software\n- * distributed under the License is distributed on an \"AS IS\" BASIS,\n- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n- * See the License for the specific language governing permissions and\n- * limitations under the License.\n- */\n-package io.atomix.raft;\n-\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\n-\n-/**\n- * This listener will only be called by the Leader, when it commits an entry. If RAFT is currently\n- * running in a follower role, it will not call this listener.\n- */\n-@FunctionalInterface\n-public interface RaftCommittedEntryListener {\n-\n- /**\n- * @param indexedRaftLogEntry the new committed entry\n- */\n- void onCommit(IndexedRaftLogEntry indexedRaftLogEntry);\n-}\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\nindex 1f4ee98..c177cb1 100644\n--- a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\n@@ -27,8 +27,8 @@ import io.atomix.cluster.MemberId;\n import io.atomix.cluster.messaging.MessagingException.NoRemoteHandler;\n import io.atomix.cluster.messaging.MessagingException.NoSuchMemberException;\n import io.atomix.raft.ElectionTimer;\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\n import io.atomix.raft.RaftCommitListener;\n-import io.atomix.raft.RaftCommittedEntryListener;\n import io.atomix.raft.RaftError;\n import io.atomix.raft.RaftException.ProtocolException;\n import io.atomix.raft.RaftRoleChangeListener;\n@@ -61,7 +61,6 @@ import io.atomix.raft.roles.PromotableRole;\n import io.atomix.raft.roles.RaftRole;\n import io.atomix.raft.storage.RaftStorage;\n import io.atomix.raft.storage.StorageException;\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\n import io.atomix.raft.storage.log.RaftLog;\n import io.atomix.raft.storage.system.MetaStore;\n import io.atomix.raft.utils.StateUtil;\n@@ -115,7 +114,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\n private final Set<Consumer<State>> stateChangeListeners = new CopyOnWriteArraySet<>();\n private final Set<Consumer<RaftMember>> electionListeners = new CopyOnWriteArraySet<>();\n private final Set<RaftCommitListener> commitListeners = new CopyOnWriteArraySet<>();\n- private final Set<RaftCommittedEntryListener> committedEntryListeners =\n+ private final Set<RaftApplicationEntryCommittedPositionListener> committedEntryListeners =\n new CopyOnWriteArraySet<>();\n private final Set<SnapshotReplicationListener> snapshotReplicationListeners =\n new CopyOnWriteArraySet<>();\n@@ -433,21 +432,23 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\n * <p>Note that it will be called on the Raft thread, and as such should not perform any heavy\n * computation.\n *\n- * @param raftCommittedEntryListener the listener to add\n+ * @param raftApplicationEntryCommittedPositionListener the listener to add\n */\n public void addCommittedEntryListener(\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\n- committedEntryListeners.add(raftCommittedEntryListener);\n+ final RaftApplicationEntryCommittedPositionListener\n+ raftApplicationEntryCommittedPositionListener) {\n+ committedEntryListeners.add(raftApplicationEntryCommittedPositionListener);\n }\n \n /**\n * Removes registered committedEntryListener\n *\n- * @param raftCommittedEntryListener the listener to remove\n+ * @param raftApplicationEntryCommittedPositionListener the listener to remove\n */\n public void removeCommittedEntryListener(\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\n- committedEntryListeners.remove(raftCommittedEntryListener);\n+ final RaftApplicationEntryCommittedPositionListener\n+ raftApplicationEntryCommittedPositionListener) {\n+ committedEntryListeners.remove(raftApplicationEntryCommittedPositionListener);\n }\n \n /**\n@@ -464,7 +465,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\n *\n * @param committedEntry the most recently committed entry\n */\n- public void notifyCommittedEntryListeners(final IndexedRaftLogEntry committedEntry) {\n+ public void notifyApplicationEntryCommittedPositionListeners(final long committedEntry) {\n committedEntryListeners.forEach(listener -> listener.onCommit(committedEntry));\n }\n \ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\nindex 56c7172..d075fca 100644\n--- a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\n@@ -21,8 +21,8 @@ import io.atomix.cluster.MemberId;\n import io.atomix.cluster.messaging.ClusterCommunicationService;\n import io.atomix.primitive.partition.Partition;\n import io.atomix.primitive.partition.PartitionMetadata;\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\n import io.atomix.raft.RaftCommitListener;\n-import io.atomix.raft.RaftCommittedEntryListener;\n import io.atomix.raft.RaftRoleChangeListener;\n import io.atomix.raft.RaftServer;\n import io.atomix.raft.RaftServer.Role;\n@@ -205,16 +205,20 @@ public class RaftPartitionServer implements HealthMonitorable {\n }\n \n /**\n- * @see io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftCommittedEntryListener)\n+ * @see\n+ * io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\n */\n- public void addCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\n+ public void addCommittedEntryListener(\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\n server.getContext().addCommittedEntryListener(commitListener);\n }\n \n /**\n- * @see io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftCommittedEntryListener)\n+ * @see\n+ * io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\n */\n- public void removeCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\n+ public void removeCommittedEntryListener(\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\n server.getContext().removeCommittedEntryListener(commitListener);\n }\n \ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\nindex e54df1a..fcfd177 100644\n--- a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\n@@ -630,27 +630,47 @@ public final class LeaderRole extends ActiveRole implements ZeebeLogAppender {\n \n private void replicate(final IndexedRaftLogEntry indexed, final AppendListener appendListener) {\n raft.checkThread();\n- appender\n- .appendEntries(indexed.index())\n- .whenCompleteAsync(\n- (commitIndex, commitError) -> {\n- if (!isRunning()) {\n- return;\n- }\n+ final var appendEntriesFuture = appender.appendEntries(indexed.index());\n+\n+ final boolean applicationEntryWasCommitted = indexed.isApplicationEntry();\n+ if (applicationEntryWasCommitted) {\n+ // We have some services which are waiting for the application records, especially position\n+ // to be committed. This is our glue code to notify them, instead of\n+ // passing the complete object (IndexedRaftLogEntry) threw the listeners and\n+ // keep them in heap until they are committed. This had the risk of going out of OOM\n+ // if records can't be committed, see https://github.com/camunda/zeebe/issues/14275\n+ final var committedPosition = indexed.getApplicationEntry().highestPosition();\n+ appendEntriesFuture.whenCompleteAsync(\n+ (commitIndex, commitError) -> {\n+ if (!isRunning()) {\n+ return;\n+ }\n+\n+ if (commitError == null) {\n+ raft.notifyApplicationEntryCommittedPositionListeners(committedPosition);\n+ }\n+ },\n+ raft.getThreadContext());\n+ }\n \n- // have the state machine apply the index which should do nothing but ensures it keeps\n- // up to date with the latest entries, so it can handle configuration and initial\n- // entries properly on fail over\n- if (commitError == null) {\n- appendListener.onCommit(indexed.index());\n- raft.notifyCommittedEntryListeners(indexed);\n- } else {\n- appendListener.onCommitError(indexed.index(), commitError);\n- // replicating the entry will be retried on the next append request\n- log.error(\"Failed to replicate entry: {}\", indexed, commitError);\n- }\n- },\n- raft.getThreadContext());\n+ appendEntriesFuture.whenCompleteAsync(\n+ (commitIndex, commitError) -> {\n+ if (!isRunning()) {\n+ return;\n+ }\n+\n+ // have the state machine apply the index which should do nothing but ensures it keeps\n+ // up to date with the latest entries, so it can handle configuration and initial\n+ // entries properly on fail over\n+ if (commitError == null) {\n+ appendListener.onCommit(indexed.index());\n+ } else {\n+ appendListener.onCommitError(indexed.index(), commitError);\n+ // replicating the entry will be retried on the next append request\n+ log.error(\"Failed to replicate entry: {}\", indexed, commitError);\n+ }\n+ },\n+ raft.getThreadContext());\n }\n \n public synchronized void onInitialEntriesCommitted(final Runnable runnable) {\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\nindex b217586..8029766 100644\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\n@@ -82,7 +82,7 @@ public class RaftAppendTest {\n @Test\n public void shouldNotifyCommittedEntryListenerOnLeaderOnly() throws Throwable {\n // given\n- final var committedEntryListener = mock(RaftCommittedEntryListener.class);\n+ final var committedEntryListener = mock(RaftApplicationEntryCommittedPositionListener.class);\n raftRule.addCommittedEntryListener(committedEntryListener);\n \n // when\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\nindex 8f73cba..193a176 100644\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\n@@ -644,9 +644,12 @@ public final class RaftRule extends ExternalResource {\n }\n \n public void addCommittedEntryListener(\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\n+ final RaftApplicationEntryCommittedPositionListener\n+ raftApplicationEntryCommittedPositionListener) {\n servers.forEach(\n- (id, raft) -> raft.getContext().addCommittedEntryListener(raftCommittedEntryListener));\n+ (id, raft) ->\n+ raft.getContext()\n+ .addCommittedEntryListener(raftApplicationEntryCommittedPositionListener));\n }\n \n public void partition(final RaftServer follower) {\ndiff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\nindex a61571f..6c082d7 100644\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n@@ -7,8 +7,7 @@\n */\n package io.camunda.zeebe.broker.system.partitions.impl;\n \n-import io.atomix.raft.RaftCommittedEntryListener;\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\n import io.camunda.zeebe.broker.system.partitions.NoEntryAtSnapshotPosition;\n import io.camunda.zeebe.broker.system.partitions.StateController;\n import io.camunda.zeebe.logstreams.impl.Loggers;\n@@ -36,7 +35,7 @@ import java.util.function.Consumer;\n import org.slf4j.Logger;\n \n public final class AsyncSnapshotDirector extends Actor\n- implements RaftCommittedEntryListener, HealthMonitorable {\n+ implements RaftApplicationEntryCommittedPositionListener, HealthMonitorable {\n \n public static final Duration MINIMUM_SNAPSHOT_PERIOD = Duration.ofMinutes(1);\n \n@@ -115,7 +114,7 @@ public final class AsyncSnapshotDirector extends Actor\n @Override\n protected void handleFailure(final Throwable failure) {\n LOG.error(\n- \"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}. {}\",\n+ \"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}.\",\n actorName,\n snapshotRate,\n failure);\n@@ -407,13 +406,8 @@ public final class AsyncSnapshotDirector extends Actor\n }\n \n @Override\n- public void onCommit(final IndexedRaftLogEntry indexedRaftLogEntry) {\n- // is called by the Leader Role and gives the last committed entry, where we\n- // can extract the highest position, which corresponds to the last committed position\n- if (indexedRaftLogEntry.isApplicationEntry()) {\n- final var committedPosition = indexedRaftLogEntry.getApplicationEntry().highestPosition();\n- newPositionCommitted(committedPosition);\n- }\n+ public void onCommit(final long committedPosition) {\n+ newPositionCommitted(committedPosition);\n }\n \n public void newPositionCommitted(final long currentCommitPosition) {\n", "diff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\nindex da05e13..9231df3 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n@@ -154,6 +154,9 @@ public final class EventAppliers implements EventApplier {\n register(\n ProcessInstanceIntent.SEQUENCE_FLOW_TAKEN,\n new ProcessInstanceSequenceFlowTakenApplier(elementInstanceState, processState));\n+ register(\n+ ProcessInstanceIntent.ELEMENT_MIGRATED,\n+ new ProcessInstanceElementMigratedApplier(elementInstanceState));\n }\n \n private void registerProcessInstanceCreationAppliers(final MutableProcessingState state) {\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\nindex e5a0f3a..d38358f 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\n@@ -24,5 +24,16 @@ final class ProcessInstanceElementMigratedApplier\n }\n \n @Override\n- public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {}\n+ public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {\n+ elementInstanceState.updateInstance(\n+ elementInstanceKey,\n+ elementInstance ->\n+ elementInstance\n+ .getValue()\n+ .setProcessDefinitionKey(value.getProcessDefinitionKey())\n+ .setBpmnProcessId(value.getBpmnProcessId())\n+ .setVersion(value.getVersion())\n+ .setElementId(value.getElementId())\n+ .setFlowScopeKey(value.getFlowScopeKey()));\n+ }\n }\n"]
2
["323cf81961cdd3748a7ba6ba470ecb13e5374e9f", "39d5d1cfe8d2210305df2c8fab4a4ae430732cf7"]
["refactor", "feat"]
fix deploy,set name for topology module
["diff --git a/.github/workflows/deploy.yaml b/.github/workflows/deploy.yaml\nindex 3830f4c..3b14ee5 100644\n--- a/.github/workflows/deploy.yaml\n+++ b/.github/workflows/deploy.yaml\n@@ -67,7 +67,7 @@ jobs:\n run: aws s3 cp .next/static s3://cdn.rs.school/_next/static/ --recursive --cache-control \"public,max-age=15552000,immutable\"\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n@@ -117,7 +117,7 @@ jobs:\n run: npm run build\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n@@ -167,7 +167,7 @@ jobs:\n run: npm run build\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n", "diff --git a/topology/pom.xml b/topology/pom.xml\nindex 389508e..ee6239a 100644\n--- a/topology/pom.xml\n+++ b/topology/pom.xml\n@@ -16,6 +16,7 @@\n </parent>\n \n <artifactId>zeebe-cluster-topology</artifactId>\n+ <name>Zeebe Cluster Topology</name>\n \n <properties>\n <proto.dir>${maven.multiModuleProjectDirectory}/topology/src/main/resources/proto</proto.dir>\n"]
2
["7785be09053049b30cf41b420c59f051cd0129fc", "8911a972222dc80a242f3f1d9b3596321b3fdeaa"]
["ci", "build"]
add Expr.equals benchmark,remove duplicated code
["diff --git a/ibis/tests/benchmarks/test_benchmarks.py b/ibis/tests/benchmarks/test_benchmarks.py\nindex 78305bb..9c7e6d7 100644\n--- a/ibis/tests/benchmarks/test_benchmarks.py\n+++ b/ibis/tests/benchmarks/test_benchmarks.py\n@@ -1,3 +1,4 @@\n+import copy\n import functools\n import itertools\n import string\n@@ -340,8 +341,9 @@ def test_execute(benchmark, expression_fn, pt):\n benchmark(expr.execute)\n \n \n-def test_repr_tpc_h02(benchmark):\n- part = ibis.table(\[email protected]\n+def part():\n+ return ibis.table(\n dict(\n p_partkey=\"int64\",\n p_size=\"int64\",\n@@ -350,7 +352,11 @@ def test_repr_tpc_h02(benchmark):\n ),\n name=\"part\",\n )\n- supplier = ibis.table(\n+\n+\[email protected]\n+def supplier():\n+ return ibis.table(\n dict(\n s_suppkey=\"int64\",\n s_nationkey=\"int64\",\n@@ -362,7 +368,11 @@ def test_repr_tpc_h02(benchmark):\n ),\n name=\"supplier\",\n )\n- partsupp = ibis.table(\n+\n+\[email protected]\n+def partsupp():\n+ return ibis.table(\n dict(\n ps_partkey=\"int64\",\n ps_suppkey=\"int64\",\n@@ -370,14 +380,25 @@ def test_repr_tpc_h02(benchmark):\n ),\n name=\"partsupp\",\n )\n- nation = ibis.table(\n+\n+\[email protected]\n+def nation():\n+ return ibis.table(\n dict(n_nationkey=\"int64\", n_regionkey=\"int64\", n_name=\"string\"),\n name=\"nation\",\n )\n- region = ibis.table(\n+\n+\[email protected]\n+def region():\n+ return ibis.table(\n dict(r_regionkey=\"int64\", r_name=\"string\"), name=\"region\"\n )\n \n+\[email protected]\n+def tpc_h02(part, supplier, partsupp, nation, region):\n REGION = \"EUROPE\"\n SIZE = 25\n TYPE = \"BRASS\"\n@@ -420,7 +441,7 @@ def test_repr_tpc_h02(benchmark):\n ]\n )\n \n- expr = q.sort_by(\n+ return q.sort_by(\n [\n ibis.desc(q.s_acctbal),\n q.n_name,\n@@ -429,7 +450,9 @@ def test_repr_tpc_h02(benchmark):\n ]\n ).limit(100)\n \n- benchmark(repr, expr)\n+\n+def test_repr_tpc_h02(benchmark, tpc_h02):\n+ benchmark(repr, tpc_h02)\n \n \n def test_repr_huge_union(benchmark):\n@@ -478,3 +501,7 @@ def test_complex_datatype_builtins(benchmark, func):\n )\n )\n benchmark(func, datatype)\n+\n+\n+def test_large_expr_equals(benchmark, tpc_h02):\n+ benchmark(ir.Expr.equals, tpc_h02, copy.deepcopy(tpc_h02))\n", "diff --git a/packages/core/src/components/action-sheet/action-sheet.tsx b/packages/core/src/components/action-sheet/action-sheet.tsx\nindex 7166508..dad7daf 100644\n--- a/packages/core/src/components/action-sheet/action-sheet.tsx\n+++ b/packages/core/src/components/action-sheet/action-sheet.tsx\n@@ -1,9 +1,9 @@\n import { Component, CssClassMap, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n \n-import { domControllerAsync, isDef, playAnimationAsync } from '../../utils/helpers';\n+import { domControllerAsync } from '../../utils/helpers';\n import { createThemedClasses, getClassMap } from '../../utils/theme';\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\n \n import iosEnterAnimation from './animations/ios.enter';\n import iosLeaveAnimation from './animations/ios.leave';\n@@ -23,15 +23,15 @@ import mdLeaveAnimation from './animations/md.leave';\n })\n export class ActionSheet implements OverlayInterface {\n \n+ private presented = false;\n+\n mode: string;\n color: string;\n-\n- private presented = false;\n- private animation: Animation | null = null;\n+ animation: Animation;\n \n @Element() private el: HTMLElement;\n \n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\n @Prop({ context: 'config' }) config: Config;\n @Prop({ context: 'dom' }) dom: DomController;\n @Prop() overlayId: number;\n@@ -178,25 +178,8 @@ export class ActionSheet implements OverlayInterface {\n });\n }\n \n- private playAnimation(animationBuilder: AnimationBuilder) {\n- if (this.animation) {\n- this.animation.destroy();\n- this.animation = null;\n- }\n-\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\n- this.animation = animation;\n- // Check if prop animate is false or if the config for animate is defined/false\n- if (!this.willAnimate || (isDef(this.config.get('willAnimate')) && this.config.get('willAnimate') === false)) {\n- // if the duration is 0, it won't actually animate I don't think\n- // TODO - validate this\n- this.animation = animation.duration(0);\n- }\n- return playAnimationAsync(animation);\n- }).then((animation) => {\n- animation.destroy();\n- this.animation = null;\n- });\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise<void> {\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\n }\n \n protected buttonClick(button: ActionSheetButton) {\ndiff --git a/packages/core/src/components/alert/alert.tsx b/packages/core/src/components/alert/alert.tsx\nindex 800b77b..bdf4fc5 100644\n--- a/packages/core/src/components/alert/alert.tsx\n+++ b/packages/core/src/components/alert/alert.tsx\n@@ -1,8 +1,8 @@\n import { Component, CssClassMap, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n-import { domControllerAsync, playAnimationAsync, autoFocus } from '../../utils/helpers';\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n+import { domControllerAsync, autoFocus } from '../../utils/helpers';\n import { createThemedClasses, getClassMap } from '../../utils/theme';\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\n \n import iosEnterAnimation from './animations/ios.enter';\n import iosLeaveAnimation from './animations/ios.leave';\n@@ -21,18 +21,19 @@ import mdLeaveAnimation from './animations/md.leave';\n }\n })\n export class Alert implements OverlayInterface {\n- mode: string;\n- color: string;\n \n private presented = false;\n- private animation: Animation | null = null;\n private activeId: string;\n private inputType: string | null = null;\n private hdrId: string;\n \n+ animation: Animation;\n+ mode: string;\n+ color: string;\n+\n @Element() private el: HTMLElement;\n \n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\n @Prop({ context: 'config' }) config: Config;\n @Prop({ context: 'dom' }) dom: DomController;\n @Prop() overlayId: number;\n@@ -264,25 +265,10 @@ export class Alert implements OverlayInterface {\n return values;\n }\n \n- private playAnimation(animationBuilder: AnimationBuilder) {\n- if (this.animation) {\n- this.animation.destroy();\n- this.animation = null;\n- }\n-\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\n- this.animation = animation;\n- if (!this.willAnimate) {\n- animation.duration(0);\n- }\n- return playAnimationAsync(animation);\n- }).then(animation => {\n- animation.destroy();\n- this.animation = null;\n- });\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise<void> {\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\n }\n \n-\n private renderCheckbox(inputs: AlertInput[]) {\n if (inputs.length === 0) return null;\n \ndiff --git a/packages/core/src/components/loading/loading.tsx b/packages/core/src/components/loading/loading.tsx\nindex f45eaf1..cc4f511 100644\n--- a/packages/core/src/components/loading/loading.tsx\n+++ b/packages/core/src/components/loading/loading.tsx\n@@ -1,13 +1,13 @@\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n+import { domControllerAsync } from '../../utils/helpers';\n import { createThemedClasses, getClassMap } from '../../utils/theme';\n \n import iosEnterAnimation from './animations/ios.enter';\n import iosLeaveAnimation from './animations/ios.leave';\n import mdEnterAnimation from './animations/md.enter';\n import mdLeaveAnimation from './animations/md.leave';\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\n \n @Component({\n tag: 'ion-loading',\n@@ -21,16 +21,17 @@ import { OverlayInterface, BACKDROP } from '../../utils/overlays';\n })\n \n export class Loading implements OverlayInterface {\n- color: string;\n- mode: string;\n \n private presented = false;\n- private animation: Animation;\n private durationTimeout: any;\n \n+ animation: Animation;\n+ color: string;\n+ mode: string;\n+\n @Element() private el: HTMLElement;\n \n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\n @Prop({ context: 'config' }) config: Config;\n @Prop({ context: 'dom' }) dom: DomController;\n @Prop() overlayId: number;\n@@ -199,24 +200,8 @@ export class Loading implements OverlayInterface {\n });\n }\n \n- private playAnimation(animationBuilder: AnimationBuilder) {\n- if (this.animation) {\n- this.animation.destroy();\n- this.animation = null;\n- }\n-\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\n- this.animation = animation;\n- if (!this.willAnimate) {\n- // if the duration is 0, it won't actually animate I don't think\n- // TODO - validate this\n- animation.duration(0);\n- }\n- return playAnimationAsync(animation);\n- }).then(animation => {\n- animation.destroy();\n- this.animation = null;\n- });\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise<void> {\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\n }\n \n hostData() {\ndiff --git a/packages/core/src/components/modal/modal.tsx b/packages/core/src/components/modal/modal.tsx\nindex af50d63..2b7510c 100644\n--- a/packages/core/src/components/modal/modal.tsx\n+++ b/packages/core/src/components/modal/modal.tsx\n@@ -1,10 +1,10 @@\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n+import { Animation, AnimationBuilder, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n \n import { DomFrameworkDelegate } from '../../utils/dom-framework-delegate';\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\n+import { domControllerAsync } from '../../utils/helpers';\n import { createThemedClasses } from '../../utils/theme';\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\n \n import iosEnterAnimation from './animations/ios.enter';\n import iosLeaveAnimation from './animations/ios.leave';\n@@ -25,14 +25,16 @@ import mdLeaveAnimation from './animations/md.leave';\n export class Modal implements OverlayInterface {\n \n private presented = false;\n- private animation: Animation;\n private usersComponentElement: HTMLElement;\n \n+ animation: Animation;\n+\n @Element() private el: HTMLElement;\n \n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\n @Prop({ context: 'config' }) config: Config;\n @Prop({ context: 'dom' }) dom: DomController;\n+\n @Prop() overlayId: number;\n @Prop({ mutable: true }) delegate: FrameworkDelegate;\n \n@@ -208,22 +210,8 @@ export class Modal implements OverlayInterface {\n });\n }\n \n- private playAnimation(animationBuilder: AnimationBuilder) {\n- if (this.animation) {\n- this.animation.destroy();\n- this.animation = null;\n- }\n-\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\n- this.animation = animation;\n- if (!this.willAnimate) {\n- animation.duration(0);\n- }\n- return playAnimationAsync(animation);\n- }).then((animation) => {\n- animation.destroy();\n- this.animation = null;\n- });\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise<void> {\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\n }\n \n @Method()\ndiff --git a/packages/core/src/components/picker/picker.tsx b/packages/core/src/components/picker/picker.tsx\nindex 13faa3e..d70381e 100644\n--- a/packages/core/src/components/picker/picker.tsx\n+++ b/packages/core/src/components/picker/picker.tsx\n@@ -1,9 +1,9 @@\n import { Component, CssClassMap, Element, Event, EventEmitter, Listen, Method, Prop, State } from '@stencil/core';\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n \n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\n+import { domControllerAsync } from '../../utils/helpers';\n import { getClassMap } from '../../utils/theme';\n-import { OverlayInterface } from '../../utils/overlays';\n+import { OverlayInterface, overlayAnimation } from '../../utils/overlays';\n \n import iosEnterAnimation from './animations/ios.enter';\n import iosLeaveAnimation from './animations/ios.leave';\n@@ -21,16 +21,17 @@ import iosLeaveAnimation from './animations/ios.leave';\n export class Picker implements OverlayInterface {\n \n private presented = false;\n- private animation: Animation;\n private durationTimeout: any;\n private mode: string;\n \n+ animation: Animation;\n+\n @Element() private el: HTMLElement;\n \n @State() private showSpinner: boolean = null;\n @State() private spinner: string;\n \n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\n @Prop({ context: 'config' }) config: Config;\n @Prop({ context: 'dom' }) dom: DomController;\n @Prop() overlayId: number;\n@@ -231,22 +232,8 @@ export class Picker implements OverlayInterface {\n return this.columns;\n }\n \n- private playAnimation(animationBuilder: AnimationBuilder) {\n- if (this.animation) {\n- this.animation.destroy();\n- this.animation = null;\n- }\n-\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\n- this.animation = animation;\n- if (!this.willAnimate) {\n- animation.duration(0);\n- }\n- return playAnimationAsync(animation);\n- }).then(animation => {\n- animation.destroy();\n- this.animation = null;\n- })\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise<void> {\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\n }\n \n private buttonClick(button: PickerButton) {\ndiff --git a/packages/core/src/components/popover/popover.tsx b/packages/core/src/components/popover/popover.tsx\nindex 65031ff..6a47bf6 100644\n--- a/packages/core/src/components/popover/popover.tsx\n+++ b/packages/core/src/components/popover/popover.tsx\n@@ -1,10 +1,10 @@\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n+import { Animation, AnimationBuilder, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n \n import { DomFrameworkDelegate } from '../../utils/dom-framework-delegate';\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\n+import { domControllerAsync } from '../../utils/helpers';\n import { createThemedClasses } from '../../utils/theme';\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\n \n import iosEnterAnimation from './animations/ios.enter';\n import iosLeaveAnimation from './animations/ios.leave';\n@@ -24,12 +24,13 @@ import mdLeaveAnimation from './animations/md.leave';\n export class Popover implements OverlayInterface {\n \n private presented = false;\n- private animation: Animation;\n private usersComponentElement: HTMLElement;\n \n+ animation: Animation;\n+\n @Element() private el: HTMLElement;\n \n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\n @Prop({ context: 'config' }) config: Config;\n @Prop({ context: 'dom' }) dom: DomController;\n @Prop({ mutable: true }) delegate: FrameworkDelegate;\n@@ -224,22 +225,8 @@ export class Popover implements OverlayInterface {\n });\n }\n \n- private playAnimation(animationBuilder: AnimationBuilder) {\n- if (this.animation) {\n- this.animation.destroy();\n- this.animation = null;\n- }\n-\n- return this.animationCtrl.create(animationBuilder, this.el, this.ev).then((animation) => {\n- this.animation = animation;\n- if (!this.willAnimate) {\n- animation.duration(0);\n- }\n- return playAnimationAsync(animation);\n- }).then(animation => {\n- animation.destroy();\n- this.animation = null;\n- })\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise<void> {\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, this.ev);\n }\n \n hostData() {\ndiff --git a/packages/core/src/components/toast/toast.tsx b/packages/core/src/components/toast/toast.tsx\nindex 1afa318..372070a 100644\n--- a/packages/core/src/components/toast/toast.tsx\n+++ b/packages/core/src/components/toast/toast.tsx\n@@ -1,9 +1,9 @@\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n-import { Animation, AnimationBuilder, AnimationController, Config, CssClassMap, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n+import { Animation, AnimationBuilder, Config, CssClassMap, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\n \n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\n+import { domControllerAsync } from '../../utils/helpers';\n import { createThemedClasses, getClassMap } from '../../utils/theme';\n-import { OverlayInterface } from '../../utils/overlays';\n+import { OverlayInterface, overlayAnimation } from '../../utils/overlays';\n \n import iosEnterAnimation from './animations/ios.enter';\n import iosLeaveAnimation from './animations/ios.leave';\n@@ -24,14 +24,14 @@ import mdLeaveAnimation from './animations/md.leave';\n export class Toast implements OverlayInterface {\n \n private presented = false;\n- private animation: Animation | null;\n \n @Element() private el: HTMLElement;\n \n mode: string;\n color: string;\n+ animation: Animation | null;\n \n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\n @Prop({ context: 'config' }) config: Config;\n @Prop({ context: 'dom' }) dom: DomController;\n @Prop() overlayId: number;\n@@ -123,6 +123,22 @@ export class Toast implements OverlayInterface {\n */\n @Event() ionToastDidUnload: EventEmitter<ToastEventDetail>;\n \n+ componentDidLoad() {\n+ this.ionToastDidLoad.emit();\n+ }\n+\n+ componentDidUnload() {\n+ this.ionToastDidUnload.emit();\n+ }\n+\n+ @Listen('ionDismiss')\n+ protected onDismiss(ev: UIEvent) {\n+ ev.stopPropagation();\n+ ev.preventDefault();\n+\n+ this.dismiss();\n+ }\n+\n /**\n * Present the toast overlay after it has been created.\n */\n@@ -169,38 +185,8 @@ export class Toast implements OverlayInterface {\n });\n }\n \n- playAnimation(animationBuilder: AnimationBuilder) {\n- if (this.animation) {\n- this.animation.destroy();\n- this.animation = null;\n- }\n-\n- return this.animationCtrl.create(animationBuilder, this.el, this.position).then(animation => {\n- this.animation = animation;\n- if (!this.willAnimate) {\n- animation.duration(0);\n- }\n- return playAnimationAsync(animation);\n- }).then((animation) => {\n- animation.destroy();\n- this.animation = null;\n- });\n- }\n-\n- componentDidLoad() {\n- this.ionToastDidLoad.emit();\n- }\n-\n- componentDidUnload() {\n- this.ionToastDidUnload.emit();\n- }\n-\n- @Listen('ionDismiss')\n- protected onDismiss(ev: UIEvent) {\n- ev.stopPropagation();\n- ev.preventDefault();\n-\n- this.dismiss();\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise<void> {\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, this.position);\n }\n \n private wrapperClass(): CssClassMap {\ndiff --git a/packages/core/src/utils/overlays.ts b/packages/core/src/utils/overlays.ts\nindex 8926544..634df43 100644\n--- a/packages/core/src/utils/overlays.ts\n+++ b/packages/core/src/utils/overlays.ts\n@@ -1,3 +1,5 @@\n+import { AnimationBuilder, Animation } from \"..\";\n+import { playAnimationAsync } from \"./helpers\";\n \n let lastId = 1;\n \n@@ -56,8 +58,33 @@ export function removeLastOverlay(overlays: OverlayMap) {\n return toRemove ? toRemove.dismiss() : Promise.resolve();\n }\n \n+export function overlayAnimation(\n+ overlay: OverlayInterface,\n+ animationBuilder: AnimationBuilder,\n+ animate: boolean,\n+ baseEl: HTMLElement,\n+ opts: any\n+): Promise<void> {\n+ if (overlay.animation) {\n+ overlay.animation.destroy();\n+ overlay.animation = null;\n+ }\n+ return overlay.animationCtrl.create(animationBuilder, baseEl, opts).then(animation => {\n+ overlay.animation = animation;\n+ if (!animate) {\n+ animation.duration(0);\n+ }\n+ return playAnimationAsync(animation);\n+ }).then((animation) => {\n+ animation.destroy();\n+ overlay.animation = null;\n+ });\n+}\n+\n export interface OverlayInterface {\n overlayId: number;\n+ animation: Animation;\n+ animationCtrl: HTMLIonAnimationControllerElement;\n \n present(): Promise<void>;\n dismiss(data?: any, role?: string): Promise<void>;\n"]
2
["b700285c1f27588922d9c56527cee721bb884682", "9e3f295bbfd4098ffda1ae6656699f60b86c1f92"]
["test", "refactor"]
new ShowDebug parameter calculate each segment timing new parameter to show/hide segment debug information set-poshprompt updated with the new showDebug parameter Force disabled segment to be visible for debug purpose,typo
["diff --git a/engine.go b/engine.go\nindex 6cc1ff3..4617ceb 100644\n--- a/engine.go\n+++ b/engine.go\n@@ -67,6 +67,9 @@ func (e *engine) renderText(text string) {\n \tprefix := e.activeSegment.getValue(Prefix, \" \")\n \tpostfix := e.activeSegment.getValue(Postfix, \" \")\n \te.renderer.write(e.activeSegment.Background, e.activeSegment.Foreground, fmt.Sprintf(\"%s%s%s\", prefix, text, postfix))\n+\tif *e.env.getArgs().Debug {\n+\t\te.renderer.write(e.activeSegment.Background, e.activeSegment.Foreground, fmt.Sprintf(\"(%s:%s)\", e.activeSegment.Type, e.activeSegment.timing))\n+\t}\n }\n \n func (e *engine) renderSegmentText(text string) {\n@@ -107,13 +110,11 @@ func (e *engine) setStringValues(segments []*Segment) {\n \twg.Add(len(segments))\n \tdefer wg.Wait()\n \tcwd := e.env.getcwd()\n+\tdebug := *e.env.getArgs().Debug\n \tfor _, segment := range segments {\n \t\tgo func(s *Segment) {\n \t\t\tdefer wg.Done()\n-\t\t\terr := s.mapSegmentWithWriter(e.env)\n-\t\t\tif err == nil && !s.hasValue(IgnoreFolders, cwd) && s.enabled() {\n-\t\t\t\ts.stringValue = s.string()\n-\t\t\t}\n+\t\t\ts.setStringValue(e.env, cwd, debug)\n \t\t}(segment)\n \t}\n }\ndiff --git a/main.go b/main.go\nindex 56ae8a5..d67a640 100644\n--- a/main.go\n+++ b/main.go\n@@ -14,6 +14,7 @@ type args struct {\n \tConfig *string\n \tShell *string\n \tPWD *string\n+\tDebug *bool\n }\n \n func main() {\n@@ -42,6 +43,10 @@ func main() {\n \t\t\t\"pwd\",\n \t\t\t\"\",\n \t\t\t\"the path you are working in\"),\n+\t\tDebug: flag.Bool(\n+\t\t\t\"debug\",\n+\t\t\tfalse,\n+\t\t\t\"Print debug information\"),\n \t}\n \tflag.Parse()\n \tenv := &environment{\ndiff --git a/packages/powershell/oh-my-posh/oh-my-posh.psm1 b/packages/powershell/oh-my-posh/oh-my-posh.psm1\nindex 9234fc6..1450eb3 100644\n--- a/packages/powershell/oh-my-posh/oh-my-posh.psm1\n+++ b/packages/powershell/oh-my-posh/oh-my-posh.psm1\n@@ -5,6 +5,7 @@\n \n $global:PoshSettings = New-Object -TypeName PSObject -Property @{\n Theme = \"$PSScriptRoot\\themes\\jandedobbeleer.json\";\n+ ShowDebug = $false\n }\n \n function Get-PoshCommand {\n@@ -36,9 +37,14 @@ function Set-PoshPrompt {\n param(\n [Parameter(Mandatory = $false)]\n [string]\n- $Theme\n+ $Theme,\n+ [Parameter(Mandatory = $false)]\n+ [bool]\n+ $ShowDebug = $false\n )\n \n+ $global:PoshSettings.ShowDebug = $ShowDebug\n+\n if (Test-Path \"$PSScriptRoot/themes/$Theme.json\") {\n $global:PoshSettings.Theme = \"$PSScriptRoot/themes/$Theme.json\"\n }\n@@ -68,8 +74,9 @@ function Set-PoshPrompt {\n $startInfo = New-Object System.Diagnostics.ProcessStartInfo\n $startInfo.FileName = Get-PoshCommand\n $config = $global:PoshSettings.Theme\n+ $showDebug = $global:PoshSettings.ShowDebug\n $cleanPWD = $PWD.ProviderPath.TrimEnd(\"\\\")\n- $startInfo.Arguments = \"-config=\"\"$config\"\" -error=$errorCode -pwd=\"\"$cleanPWD\"\"\"\n+ $startInfo.Arguments = \"-debug=\"\"$showDebug\"\" -config=\"\"$config\"\" -error=$errorCode -pwd=\"\"$cleanPWD\"\"\"\n $startInfo.Environment[\"TERM\"] = \"xterm-256color\"\n $startInfo.CreateNoWindow = $true\n $startInfo.StandardOutputEncoding = [System.Text.Encoding]::UTF8\ndiff --git a/segment.go b/segment.go\nindex 27dd416..4015dac 100644\n--- a/segment.go\n+++ b/segment.go\n@@ -1,6 +1,9 @@\n package main\n \n-import \"errors\"\n+import (\n+\t\"errors\"\n+\t\"time\"\n+)\n \n // Segment represent a single segment and it's configuration\n type Segment struct {\n@@ -17,6 +20,7 @@ type Segment struct {\n \twriter SegmentWriter\n \tstringValue string\n \tactive bool\n+\ttiming time.Duration\n }\n \n // SegmentWriter is the interface used to define what and if to write to the prompt\n@@ -149,3 +153,26 @@ func (segment *Segment) mapSegmentWithWriter(env environmentInfo) error {\n \t}\n \treturn errors.New(\"unable to map writer\")\n }\n+\n+func (segment *Segment) setStringValue(env environmentInfo, cwd string, debug bool) {\n+\terr := segment.mapSegmentWithWriter(env)\n+\tif err != nil || segment.hasValue(IgnoreFolders, cwd) {\n+\t\treturn\n+\t}\n+\t// add timing only in debug\n+\tif debug {\n+\t\tstart := time.Now()\n+\t\tdefer (func() {\n+\t\t\t// force segment rendering to display the time it took\n+\t\t\t// to check if the segment is enabled or not\n+\t\t\t// depending on the segement, calling enabled()\n+\t\t\t// can be time consuming\n+\t\t\tsegment.active = true\n+\t\t\telapsed := time.Since(start)\n+\t\t\tsegment.timing = elapsed\n+\t\t})()\n+\t}\n+\tif segment.enabled() {\n+\t\tsegment.stringValue = segment.string()\n+\t}\n+}\n", "diff --git a/README.md b/README.md\nindex b823d75..53f265d 100644\n--- a/README.md\n+++ b/README.md\n@@ -74,7 +74,7 @@ With oclif you can create 2 different CLI types, single and multi.\n \n Single CLIs are like `ls` or `cat`. They can accept arguments and flags. Single CLIs can optionally be just be a single file.\n \n-Multi CLIs are like `git` or `heroku`. They have subcommands that are themselves single CLIs commands. In the `package.json` there is a field `oclif.commands` that points to a directory. This directory contains all the subcommands for the CLI. For example, if you wanted a CLI called `mycli` with the commands `mycli create` and `mycli destroy`, you would have a project like the following:\n+Multi CLIs are like `git` or `heroku`. They have subcommands that are themselves single CLI commands. In the `package.json` there is a field `oclif.commands` that points to a directory. This directory contains all the subcommands for the CLI. For example, if you wanted a CLI called `mycli` with the commands `mycli create` and `mycli destroy`, you would have a project like the following:\n \n ```\n package.json\n"]
2
["bea32587586ca08f390c901a95e9b9c25263f4df", "06c12fb603e3a38eca0340a92719ee59d34a9f47"]
["feat", "docs"]
Template using kube api version Signed-off-by: rjshrjndrn <[email protected]>,support react@17 in peer deps resolves #1478
["diff --git a/.github/workflows/api-ee.yaml b/.github/workflows/api-ee.yaml\nindex c014f34..2a12e0d 100644\n--- a/.github/workflows/api-ee.yaml\n+++ b/.github/workflows/api-ee.yaml\n@@ -8,7 +8,7 @@ on:\n default: 'false'\n push:\n branches:\n- - dev\n+ - test_ci\n paths:\n - ee/api/**\n - api/**\n@@ -112,7 +112,8 @@ jobs:\n # Deploy command\n kubectl config set-context --namespace=app --current\n kubectl config get-contexts\n- helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks | kubectl apply -f -\n+ k_version=$(kubectl version --short 2>/dev/null | awk '/Server/{print $NF}')\n+ helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks --kube-version=$k_version | kubectl apply -f -\n env:\n DOCKER_REPO: ${{ secrets.EE_REGISTRY_URL }}\n # We're not passing -ee flag, because helm will add that.\n", "diff --git a/packages/animated/package.json b/packages/animated/package.json\nindex 2249a2f..e35a1fd 100644\n--- a/packages/animated/package.json\n+++ b/packages/animated/package.json\n@@ -33,6 +33,6 @@\n \"react-layout-effect\": \"^1.0.1\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n }\n }\ndiff --git a/packages/core/package.json b/packages/core/package.json\nindex 584bbc2..c934253 100644\n--- a/packages/core/package.json\n+++ b/packages/core/package.json\n@@ -36,7 +36,7 @@\n \"react-layout-effect\": \"^1.0.1\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n },\n \"devDependencies\": {\n \"rafz\": \"^0.1.13\"\ndiff --git a/packages/parallax/package.json b/packages/parallax/package.json\nindex 49f8391..5a181fe 100644\n--- a/packages/parallax/package.json\n+++ b/packages/parallax/package.json\n@@ -31,6 +31,6 @@\n \"@react-spring/web\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n }\n }\ndiff --git a/packages/shared/package.json b/packages/shared/package.json\nindex 67d286c..12f7db3 100644\n--- a/packages/shared/package.json\n+++ b/packages/shared/package.json\n@@ -33,6 +33,6 @@\n \"rafz\": \"^0.1.13\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n }\n }\ndiff --git a/targets/konva/package.json b/targets/konva/package.json\nindex 17675ac..271d58c 100644\n--- a/targets/konva/package.json\n+++ b/targets/konva/package.json\n@@ -34,7 +34,7 @@\n },\n \"peerDependencies\": {\n \"konva\": \">=2.6\",\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-konva\": \">=16.8\"\n },\n \"devDependencies\": {\ndiff --git a/targets/native/package.json b/targets/native/package.json\nindex e97aa97..802a66c 100644\n--- a/targets/native/package.json\n+++ b/targets/native/package.json\n@@ -33,7 +33,7 @@\n \"@react-spring/types\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-native\": \">=0.58\"\n },\n \"devDependencies\": {\ndiff --git a/targets/web/package.json b/targets/web/package.json\nindex d74c25c..f7ac000 100644\n--- a/targets/web/package.json\n+++ b/targets/web/package.json\n@@ -33,7 +33,7 @@\n \"@react-spring/types\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-dom\": \">=16.8\"\n }\n }\ndiff --git a/targets/zdog/package.json b/targets/zdog/package.json\nindex aa57890..f65945a 100644\n--- a/targets/zdog/package.json\n+++ b/targets/zdog/package.json\n@@ -33,7 +33,7 @@\n \"@react-spring/types\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-dom\": \">=16.8\",\n \"react-zdog\": \">=1.0\",\n \"zdog\": \">=1.0\"\n"]
2
["c3531347fe5a4cc82d426db195026a5bdad15e7a", "27169897c0e58bc4fbca724f290ad54fa39abec7"]
["ci", "build"]
fixing deploying to kubernetes Signed-off-by: Rajesh Rajendran <[email protected]>,bump version Signed-off-by: rjshrjndrn <[email protected]>
["diff --git a/.github/workflows/frontend.yaml b/.github/workflows/frontend.yaml\nindex 7e42967..77e4abf 100644\n--- a/.github/workflows/frontend.yaml\n+++ b/.github/workflows/frontend.yaml\n@@ -22,26 +22,22 @@ jobs:\n ${{ runner.OS }}-build-\n ${{ runner.OS }}-\n \n+ - uses: azure/k8s-set-context@v1\n+ with:\n+ method: kubeconfig\n+ kubeconfig: ${{ secrets.OSS_KUBECONFIG }} # Use content of kubeconfig in secret.\n+ id: setcontext\n - name: Install\n run: npm install\n \n- - name: Build\n- run: npm run build:staging\n- env:\n- ENVIRONMENT: staging\n-\n- - name: Deploy\n- env:\n- AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }}\n- AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }}\n- AWS_REGION: eu-central-1\n- AWS_S3_BUCKET_NAME: ${{ secrets.AWS_S3_BUCKET_NAME }}\n+ - name: Build and deploy\n run: |\n- aws configure set default.s3.signature_version s3v4\n- aws --endpoint-url https://${{secrets.DOMAIN_NAME}}/frontend/ s3 cp \\\n- --recursive \\\n- --region \"$AWS_REGION\" \\\n- public s3://$AWS_S3_BUCKET_NAME\n+ cd frontend\n+ bash build.sh\n+ cp -arl public frontend\n+ minio_pod=$(kubectl get po -n db -l app.kubernetes.io/name=minio -n db --output custom-columns=name:.metadata.name | tail -n+2)\n+ kubectl -n db cp frontend $minio_pod:/data/\n+ rm -rf frontend\n \n # - name: Debug Job\n # if: ${{ failure() }}\n", "diff --git a/scripts/helmcharts/init.sh b/scripts/helmcharts/init.sh\nindex 5a2b4b0..69a6944 100644\n--- a/scripts/helmcharts/init.sh\n+++ b/scripts/helmcharts/init.sh\n@@ -26,7 +26,7 @@ usr=$(whoami)\n \n # Installing k3s\n function install_k8s() {\n- curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\"644\" INSTALL_K3S_VERSION='v1.22.8+k3s1' INSTALL_K3S_EXEC=\"--no-deploy=traefik\" sh -\n+ curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\"644\" INSTALL_K3S_VERSION='v1.25.6+k3s1' INSTALL_K3S_EXEC=\"--disable=traefik\" sh -\n [[ -d ~/.kube ]] || mkdir ~/.kube\n sudo cp /etc/rancher/k3s/k3s.yaml ~/.kube/config\n sudo chmod 0644 ~/.kube/config\n"]
2
["3f2eec37f76c1ad9408e423e49fe5bfe3e17d943", "9a25fe59dfb63d32505afcea3a164ff0b8ea4c71"]
["ci", "build"]
coordinator accepts a request transformer instead of a list of operations The request transformer can generate the operations from the current topology. This helps to - ensure that the operations are generated based on the latest topology. When concurrent changes happens, coordinator can detect it. Previously it was unclear because by the time handle apply operations, the cluster topology might have changed. - return the simulated final topology as part of the result,fixa few issues
["diff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\nindex 8bb5c3d..f8f5e24 100644\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\n@@ -10,6 +10,7 @@ package io.camunda.zeebe.topology.changes;\n import io.camunda.zeebe.scheduler.future.ActorFuture;\n import io.camunda.zeebe.topology.state.ClusterTopology;\n import io.camunda.zeebe.topology.state.TopologyChangeOperation;\n+import io.camunda.zeebe.util.Either;\n import java.util.List;\n \n public interface TopologyChangeCoordinator {\n@@ -39,4 +40,16 @@ public interface TopologyChangeCoordinator {\n ActorFuture<Boolean> hasCompletedChanges(final long version);\n \n ActorFuture<ClusterTopology> getCurrentTopology();\n+\n+ ActorFuture<TopologyChangeResult> applyOperations(TopologyChangeRequest request);\n+\n+ record TopologyChangeResult(\n+ ClusterTopology currentTopology,\n+ ClusterTopology finalTopology,\n+ List<TopologyChangeOperation> operations) {}\n+\n+ interface TopologyChangeRequest {\n+ Either<Exception, List<TopologyChangeOperation>> operations(\n+ final ClusterTopology currentTopology);\n+ }\n }\ndiff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\nindex 13ec754..877fc3c 100644\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\n@@ -103,6 +103,62 @@ public class TopologyChangeCoordinatorImpl implements TopologyChangeCoordinator \n return clusterTopologyManager.getClusterTopology();\n }\n \n+ @Override\n+ public ActorFuture<TopologyChangeResult> applyOperations(final TopologyChangeRequest request) {\n+ final ActorFuture<TopologyChangeResult> future = executor.createFuture();\n+ clusterTopologyManager\n+ .getClusterTopology()\n+ .onComplete(\n+ (currentClusterTopology, errorOnGettingTopology) -> {\n+ if (errorOnGettingTopology != null) {\n+ future.completeExceptionally(errorOnGettingTopology);\n+ return;\n+ }\n+\n+ final var operationsEither = request.operations(currentClusterTopology);\n+ if (operationsEither.isLeft()) {\n+ future.completeExceptionally(operationsEither.getLeft());\n+ return;\n+ }\n+ final var operations = operationsEither.get();\n+ if (operations.isEmpty()) {\n+ // No operations to apply\n+ future.complete(\n+ new TopologyChangeResult(\n+ currentClusterTopology, currentClusterTopology, operations));\n+ return;\n+ }\n+\n+ final ActorFuture<ClusterTopology> validation =\n+ validateTopologyChangeRequest(currentClusterTopology, operations);\n+\n+ validation.onComplete(\n+ (simulatedFinalTopology, validationError) -> {\n+ if (validationError != null) {\n+ future.completeExceptionally(validationError);\n+ return;\n+ }\n+\n+ // if the validation was successful, apply the changes\n+ final ActorFuture<ClusterTopology> applyFuture = executor.createFuture();\n+ applyTopologyChange(\n+ operations, currentClusterTopology, simulatedFinalTopology, applyFuture);\n+\n+ applyFuture.onComplete(\n+ (ignore, error) -> {\n+ if (error == null) {\n+ future.complete(\n+ new TopologyChangeResult(\n+ currentClusterTopology, simulatedFinalTopology, operations));\n+ } else {\n+ future.completeExceptionally(error);\n+ }\n+ });\n+ });\n+ });\n+ return future;\n+ }\n+\n private ActorFuture<ClusterTopology> validateTopologyChangeRequest(\n final ClusterTopology currentClusterTopology,\n final List<TopologyChangeOperation> operations) {\n", "diff --git a/README.md b/README.md\nindex d944d22..5099f03 100644\n--- a/README.md\n+++ b/README.md\n@@ -10,9 +10,8 @@ React state management with a minimal API. Made with :heart: and ES6 Proxies.\n \n <!-- toc -->\n \n-* [Motivation](#motivation)\n+* [Introduction](#introduction)\n * [Installation](#installation)\n- + [Setting up a quick project](#setting-up-a-quick-project)\n * [Usage](#usage)\n + [Creating stores](#creating-stores)\n + [Creating reactive views](#creating-reactive-views)\n@@ -35,12 +34,14 @@ React state management with a minimal API. Made with :heart: and ES6 Proxies.\n Easy State consists of two wrapper functions only. `store` creates state stores and `view` creates reactive components, which re-render whenever state stores are mutated. The rest is just plain JavaScript.\n \n ```js\n-import React, from 'react'\n+import React from 'react'\n import { store, view } from 'react-easy-state'\n \n+// stores are normal objects\n const clock = store({ time: new Date() })\n setInterval(() => clock.time = new Date(), 1000)\n \n+// reactive components re-render on store mutations\n function ClockComp () {\n return <div>{clock.time}</div>\n }\n"]
2
["dec860436916ef216998f80f8b2f9c39d00c064d", "b8a664c1b10f4e30a3e221a14211a3cdaf90b7f4"]
["feat", "docs"]
updated riot to v6, fixed build,add test for clickhouse-specific `create_table` parameters
["diff --git a/components/riot/package.json b/components/riot/package.json\nindex c41743a..eb69756 100644\n--- a/components/riot/package.json\n+++ b/components/riot/package.json\n@@ -61,7 +61,7 @@\n },\n \"devDependencies\": {\n \"@babel/preset-typescript\": \"^7.14.5\",\n- \"@riotjs/cli\": \"^6.0.4\",\n+ \"@riotjs/cli\": \"^6.0.5\",\n \"@riotjs/compiler\": \"^6.0.0\",\n \"chai\": \"^4.3.4\",\n \"esm\": \"^3.2.25\",\n", "diff --git a/ibis/backends/clickhouse/tests/test_client.py b/ibis/backends/clickhouse/tests/test_client.py\nindex 678683d..c4e2aec 100644\n--- a/ibis/backends/clickhouse/tests/test_client.py\n+++ b/ibis/backends/clickhouse/tests/test_client.py\n@@ -224,6 +224,21 @@ def test_create_table_data(con, data, engine, temp_table):\n assert len(t.execute()) == 3\n \n \n+def test_create_table_with_properties(con, temp_table):\n+ data = pd.DataFrame({\"a\": list(\"abcde\" * 20), \"b\": [1, 2, 3, 4, 5] * 20})\n+ n = len(data)\n+ t = con.create_table(\n+ temp_table,\n+ data,\n+ schema=ibis.schema(dict(a=\"string\", b=\"!uint32\")),\n+ order_by=[\"a\", \"b\"],\n+ partition_by=[\"a\"],\n+ sample_by=[\"b\"],\n+ settings={\"allow_nullable_key\": \"1\"},\n+ )\n+ assert t.count().execute() == n\n+\n+\n @pytest.mark.parametrize(\n \"engine\",\n [\n"]
2
["5d256f937f93e5a5ed003df86d38c44834095a11", "7e1ece7d3fd41d1e3ee38e479c119494bb269966"]
["build", "test"]
do not check mkdocs for older versions used in deployments,use a closure
["diff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\nindex 57d94a4..04de03b 100644\n--- a/.github/workflows/ibis-docs-lint.yml\n+++ b/.github/workflows/ibis-docs-lint.yml\n@@ -206,7 +206,7 @@ jobs:\n - name: build and push dev docs\n run: |\n nix develop --ignore-environment -c \\\n- mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}'\n+ mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}' --ignore-version\n \n simulate_release:\n runs-on: ubuntu-latest\n", "diff --git a/ibis/expr/analysis.py b/ibis/expr/analysis.py\nindex bb17a7a..975c658 100644\n--- a/ibis/expr/analysis.py\n+++ b/ibis/expr/analysis.py\n@@ -39,7 +39,9 @@ def sub_for(expr, substitutions):\n An Ibis expression\n \"\"\"\n \n- def fn(node, mapping={k.op(): v for k, v in substitutions}):\n+ mapping = {k.op(): v for k, v in substitutions}\n+\n+ def fn(node):\n try:\n return mapping[node]\n except KeyError:\n"]
2
["21228c55b7045d9b2225f65e6231184ff332b071", "ad52e1d67fd77f0b6a73fbf989b33f9abf395ecc"]
["ci", "refactor"]
fixing deploying to kubernetes Signed-off-by: Rajesh Rajendran <[email protected]>,don't consider cases where there are no txids
["diff --git a/.github/workflows/frontend.yaml b/.github/workflows/frontend.yaml\nindex 7e42967..77e4abf 100644\n--- a/.github/workflows/frontend.yaml\n+++ b/.github/workflows/frontend.yaml\n@@ -22,26 +22,22 @@ jobs:\n ${{ runner.OS }}-build-\n ${{ runner.OS }}-\n \n+ - uses: azure/k8s-set-context@v1\n+ with:\n+ method: kubeconfig\n+ kubeconfig: ${{ secrets.OSS_KUBECONFIG }} # Use content of kubeconfig in secret.\n+ id: setcontext\n - name: Install\n run: npm install\n \n- - name: Build\n- run: npm run build:staging\n- env:\n- ENVIRONMENT: staging\n-\n- - name: Deploy\n- env:\n- AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }}\n- AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }}\n- AWS_REGION: eu-central-1\n- AWS_S3_BUCKET_NAME: ${{ secrets.AWS_S3_BUCKET_NAME }}\n+ - name: Build and deploy\n run: |\n- aws configure set default.s3.signature_version s3v4\n- aws --endpoint-url https://${{secrets.DOMAIN_NAME}}/frontend/ s3 cp \\\n- --recursive \\\n- --region \"$AWS_REGION\" \\\n- public s3://$AWS_S3_BUCKET_NAME\n+ cd frontend\n+ bash build.sh\n+ cp -arl public frontend\n+ minio_pod=$(kubectl get po -n db -l app.kubernetes.io/name=minio -n db --output custom-columns=name:.metadata.name | tail -n+2)\n+ kubectl -n db cp frontend $minio_pod:/data/\n+ rm -rf frontend\n \n # - name: Debug Job\n # if: ${{ failure() }}\n", "diff --git a/src/main.rs b/src/main.rs\nindex 25d9580..9ba4e38 100644\n--- a/src/main.rs\n+++ b/src/main.rs\n@@ -441,6 +441,9 @@ fn main() {\n let mut delta_tx_fees = vec![];\n let empty_txids = vec![];\n let txids = tx_mined_deltas.get(&delta).unwrap_or(&empty_txids);\n+ if txids.len() == 0 {\n+ continue;\n+ }\n for txid in txids.iter() {\n delta_tx_fees.push(*tx_fees.get(txid).unwrap_or(&0));\n }\n"]
2
["3f2eec37f76c1ad9408e423e49fe5bfe3e17d943", "37a1b5bbb5270befcee5d9b9621af196c787a61f"]
["ci", "fix"]
update the formatting for python integration example,simplyfy statement
["diff --git a/website/docs/integration/python.md b/website/docs/integration/python.md\nindex 064cae3..b6b720d 100644\n--- a/website/docs/integration/python.md\n+++ b/website/docs/integration/python.md\n@@ -13,6 +13,7 @@ header = \"All notable changes to this project will be documented in this file.\"\n body = \"...\"\n footer = \"<!-- generated by git-cliff -->\"\n # see [changelog] section for more keys\n+\n [tool.git-cliff.git]\n conventional_commits = true\n commit_parsers = []\n", "diff --git a/src/Object/Merge.ts b/src/Object/Merge.ts\nindex 1f48efb..06caad1 100644\n--- a/src/Object/Merge.ts\n+++ b/src/Object/Merge.ts\n@@ -96,9 +96,11 @@ type ChooseMergeDeep<OK, O1K, K extends Key, OOK extends Key, style extends Merg\n @hidden\n */\n export type _MergeDeep<O, O1, K extends Key, OOK extends Key, style extends MergeStyle> =\n- Or<Extends<[O], [never]>, Extends<[O1], [never]>> extends 1 // filter never\n+ [O] extends [never]\n ? MergeProp<O, O1, K, OOK, style>\n- : LibStyle<ChooseMergeDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n+ : [O1] extends [never]\n+ ? MergeProp<O, O1, K, OOK, style>\n+ : LibStyle<ChooseMergeDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n \n /**\n @hidden\ndiff --git a/src/Object/Patch.ts b/src/Object/Patch.ts\nindex 2d73784..2c8bd42 100644\n--- a/src/Object/Patch.ts\n+++ b/src/Object/Patch.ts\n@@ -89,9 +89,11 @@ type ChoosePatchDeep<OK, O1K, K extends Key, OOK extends Key, style extends Merg\n @hidden\n */\n export type _PatchDeep<O, O1, K extends Key, OOK extends Key, style extends MergeStyle> =\n- Or<Extends<[O], [never]>, Extends<[O1], [never]>> extends 1 // filter never\n+ [O] extends [never]\n ? PatchProp<O, O1, K, OOK>\n- : LibStyle<ChoosePatchDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n+ : [O1] extends [never]\n+ ? PatchProp<O, O1, K, OOK>\n+ : LibStyle<ChoosePatchDeep<NoList<O>, NoList<O1>, K, OOK, style>, O, O1, style>\n \n /**\n @hidden\n"]
2
["3ee672483790ec71c700907a6e93af4698492026", "f86944ff00b970d7e2da48abbff43e58bdf29b99"]
["docs", "refactor"]
handle default_branch_monthly_cost having no cost Closes https://github.com/infracost/infracost-gh-action/issues/17,remove sync ts check
["diff --git a/scripts/ci/diff.sh b/scripts/ci/diff.sh\nindex 7472273..fa48e4b 100755\n--- a/scripts/ci/diff.sh\n+++ b/scripts/ci/diff.sh\n@@ -112,7 +112,12 @@ echo \"$default_branch_output\" > default_branch_infracost.txt\n default_branch_monthly_cost=$(cat default_branch_infracost.txt | awk '/OVERALL TOTAL/ { gsub(\",\",\"\"); printf(\"%.2f\",$NF) }')\n echo \"::set-output name=default_branch_monthly_cost::$default_branch_monthly_cost\"\n \n-percent_diff=$(echo \"scale=4; $current_branch_monthly_cost / $default_branch_monthly_cost * 100 - 100\" | bc)\n+if [ $(echo \"$default_branch_monthly_cost > 0\" | bc -l) = 1 ]; then\n+ percent_diff=$(echo \"scale=4; $current_branch_monthly_cost / $default_branch_monthly_cost * 100 - 100\" | bc)\n+else\n+ echo \"Default branch has no cost, setting percent_diff=100 to force a comment\"\n+ percent_diff=100\n+fi\n absolute_percent_diff=$(echo $percent_diff | tr -d -)\n \n if [ $(echo \"$absolute_percent_diff > $percentage_threshold\" | bc -l) = 1 ]; then\n", "diff --git a/config/webpack.config.prod.js b/config/webpack.config.prod.js\nindex 8b23fba..58a4c17 100644\n--- a/config/webpack.config.prod.js\n+++ b/config/webpack.config.prod.js\n@@ -251,7 +251,7 @@ module.exports = {\n plugins: [\n argv.notypecheck\n ? null\n- : new ForkTsCheckerWebpackPlugin({tslint: true, async: false}),\n+ : new ForkTsCheckerWebpackPlugin({tslint: true}),\n // Makes some environment variables available in index.html.\n // The public URL is available as %PUBLIC_URL% in index.html, e.g.:\n // <link rel=\"shortcut icon\" href=\"%PUBLIC_URL%/favicon.ico\">\n"]
2
["9474f58b44a35321e9157ca9890c589a7b3729b2", "411be831591b2ea15ca9138eaf8db81f51b5101e"]
["fix", "build"]
return Animated nodes passed to "getAnimated" ...instead of undefined. Also, stop using instanceof in "isAnimated" for perf.,add link to roadmap
["diff --git a/packages/animated/src/Animated.ts b/packages/animated/src/Animated.ts\nindex 00daa96..05ff7f9 100644\n--- a/packages/animated/src/Animated.ts\n+++ b/packages/animated/src/Animated.ts\n@@ -4,7 +4,7 @@ import { AnimatedValue } from './AnimatedValue'\n const $node: any = Symbol.for('Animated:node')\n \n export const isAnimated = (value: any): value is Animated =>\n- value instanceof Animated\n+ !!value && value[$node] === value\n \n /** Get the owner's `Animated` node. */\n export const getAnimated = (owner: any): Animated | undefined =>\n@@ -23,6 +23,10 @@ export abstract class Animated<T = any> {\n protected payload?: Payload\n \n /** Returns every value of the node. Pass true for only the animated values. */\n+ constructor() {\n+ setAnimated(this, this)\n+ }\n+\n abstract getValue(animated?: boolean): T\n \n abstract setValue(value: T): void\n", "diff --git a/packages/plugin-core/README.md b/packages/plugin-core/README.md\nindex 3c25c9b..c7506d4 100644\n--- a/packages/plugin-core/README.md\n+++ b/packages/plugin-core/README.md\n@@ -187,6 +187,10 @@ When the workspace opens, it will show dialogue to install the recommended exten\n \n See [[FAQ]] to answers for common questions.\n \n+# Roadmap\n+\n+Check out our [public roadmap](https://github.com/orgs/dendronhq/projects/1) to see the features we're working on and to vote for what you want to see next. \n+\n \n # Contributing\n \n"]
2
["eb513f7eeea7865f15e5bd561a471d1f4381ea70", "94202f01e44c58bee4419044f8a18ac5f1a50dff"]
["fix", "docs"]
simplify loadFiles code,post installers compatiblity with Windows #2520
["diff --git a/frontend/app/player/web/network/loadFiles.ts b/frontend/app/player/web/network/loadFiles.ts\nindex ec174fc..d164333 100644\n--- a/frontend/app/player/web/network/loadFiles.ts\n+++ b/frontend/app/player/web/network/loadFiles.ts\n@@ -1,43 +1,33 @@\n import APIClient from 'App/api_client';\n \n-const NO_NTH_FILE = \"nnf\"\n-const NO_UNPROCESSED_FILES = \"nuf\"\n+const NO_FILE_OK = \"No-file-but-this-is-ok\"\n+const NO_BACKUP_FILE = \"No-efs-file\"\n \n export const loadFiles = (\n urls: string[],\n onData: (data: Uint8Array) => void,\n ): Promise<void> => {\n- const firstFileURL = urls[0]\n- urls = urls.slice(1)\n- if (!firstFileURL) {\n+ if (!urls.length) {\n return Promise.reject(\"No urls provided\")\n }\n- return window.fetch(firstFileURL)\n- .then(r => {\n- return processAPIStreamResponse(r, true)\n- })\n- .then(onData)\n- .then(() =>\n- urls.reduce((p, url) =>\n- p.then(() =>\n- window.fetch(url)\n- .then(r => {\n- return processAPIStreamResponse(r, false)\n- })\n- .then(onData)\n- ),\n- Promise.resolve(),\n- )\n+ return urls.reduce((p, url, index) =>\n+ p.then(() =>\n+ window.fetch(url)\n+ .then(r => {\n+ return processAPIStreamResponse(r, index===0)\n+ })\n+ .then(onData)\n+ ),\n+ Promise.resolve(),\n )\n .catch(e => {\n- if (e === NO_NTH_FILE) {\n+ if (e === NO_FILE_OK) {\n return\n }\n throw e\n })\n }\n \n-\n export async function requestEFSDom(sessionId: string) {\n return await requestEFSMobFile(sessionId + \"/dom.mob\")\n }\n@@ -50,21 +40,18 @@ async function requestEFSMobFile(filename: string) {\n const api = new APIClient()\n const res = await api.fetch('/unprocessed/' + filename)\n if (res.status >= 400) {\n- throw NO_UNPROCESSED_FILES\n+ throw NO_BACKUP_FILE\n }\n return await processAPIStreamResponse(res, false)\n }\n \n-const processAPIStreamResponse = (response: Response, isFirstFile: boolean) => {\n+const processAPIStreamResponse = (response: Response, canBeMissed: boolean) => {\n return new Promise<ArrayBuffer>((res, rej) => {\n- if (response.status === 404 && !isFirstFile) {\n- return rej(NO_NTH_FILE)\n+ if (response.status === 404 && canBeMissed) {\n+ return rej(NO_FILE_OK)\n }\n if (response.status >= 400) {\n- return rej(\n- isFirstFile ? `no start file. status code ${ response.status }`\n- : `Bad endfile status code ${response.status}`\n- )\n+ return rej(`Bad file status code ${response.status}. Url: ${response.url}`)\n }\n res(response.arrayBuffer())\n }).then(buffer => new Uint8Array(buffer))\n", "diff --git a/packages/cubejs-databricks-jdbc-driver/package.json b/packages/cubejs-databricks-jdbc-driver/package.json\nindex cc164f0..fd7ad45 100644\n--- a/packages/cubejs-databricks-jdbc-driver/package.json\n+++ b/packages/cubejs-databricks-jdbc-driver/package.json\n@@ -14,13 +14,16 @@\n },\n \"main\": \"dist/src/index.js\",\n \"typings\": \"dist/src/index.d.ts\",\n+ \"bin\": {\n+ \"databricks-jdbc-installer\": \"bin/post-install\"\n+ },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n \"tsc\": \"tsc\",\n \"watch\": \"tsc -w\",\n \"lint\": \"eslint src/* --ext .ts\",\n \"lint:fix\": \"eslint --fix src/* --ext .ts\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"databricks-jdbc-installer\"\n },\n \"files\": [\n \"README.md\",\ndiff --git a/rust/package.json b/rust/package.json\nindex b139279..5bf6446 100644\n--- a/rust/package.json\n+++ b/rust/package.json\n@@ -8,7 +8,8 @@\n \"node\": \">=10.8.0\"\n },\n \"bin\": {\n- \"cubestore-dev\": \"bin/cubestore-dev\"\n+ \"cubestore-dev\": \"bin/cubestore-dev\",\n+ \"cubestore-installer\": \"bin/post-install\"\n },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n@@ -18,7 +19,7 @@\n \"lint:fix\": \"eslint --fix js-wrapper/* --ext .ts,js\",\n \"unit\": \"jest\",\n \"unit:debug\": \"jest --runInBand\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"cubestore-installer\"\n },\n \"files\": [\n \"dist\",\ndiff --git a/yarn.lock b/yarn.lock\nindex d2a4038..b59bb77 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -4036,9 +4036,9 @@\n integrity sha512-7btbphLrKvo5yl/5CC2OCxUSMx1wV1wvGT1qDXkSt7yi00/YW7E8k6qzXqJHsp+WU0eoG7r6MTQQXI9lIvd0qA==\n \n \"@types/fs-extra@^9.0.1\", \"@types/fs-extra@^9.0.2\", \"@types/fs-extra@^9.0.8\":\n- version \"9.0.10\"\n- resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.10.tgz#8023a72e3d06cf54929ea47ec7634e47f33f4046\"\n- integrity sha512-O9T2LLkRDiTlalOBdjEkcnT0MRdT2+wglCl7pJUJ3mkWkR8hX4K+5bg2raQNJcLv4V8zGuTXe7Ud3wSqkTyuyQ==\n+ version \"9.0.11\"\n+ resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.11.tgz#8cc99e103499eab9f347dbc6ca4e99fb8d2c2b87\"\n+ integrity sha512-mZsifGG4QeQ7hlkhO56u7zt/ycBgGxSVsFI/6lGTU34VtwkiqrrSDgw0+ygs8kFGWcXnFQWMrzF2h7TtDFNixA==\n dependencies:\n \"@types/node\" \"*\"\n \n@@ -5306,9 +5306,9 @@ acorn@^7.0.0, acorn@^7.1.0, acorn@^7.1.1, acorn@^7.4.0:\n integrity sha512-nQyp0o1/mNdbTO1PO6kHkwSrmgZ0MT/jCCpNiwbUjGoRN4dlBhqJtoQuCnEOKzgTVwg0ZWiCoQy6SxMebQVh8A==\n \n acorn@^8.1.0:\n- version \"8.1.0\"\n- resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.0.tgz#52311fd7037ae119cbb134309e901aa46295b3fe\"\n- integrity sha512-LWCF/Wn0nfHOmJ9rzQApGnxnvgfROzGilS8936rqN/lfcYkY9MYZzdMqN+2NJ4SlTc+m5HiSa+kNfDtI64dwUA==\n+ version \"8.1.1\"\n+ resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.1.tgz#fb0026885b9ac9f48bac1e185e4af472971149ff\"\n+ integrity sha512-xYiIVjNuqtKXMxlRMDc6mZUhXehod4a3gbZ1qRlM7icK4EbxUFNLhWoPblCvFtB2Y9CIqHP3CF/rdxLItaQv8g==\n \n adal-node@^0.1.28:\n version \"0.1.28\"\n@@ -5441,9 +5441,9 @@ ajv@^6.1.0, ajv@^6.10.0, ajv@^6.10.2, ajv@^6.12.2, ajv@^6.12.3, ajv@^6.12.4, ajv\n uri-js \"^4.2.2\"\n \n ajv@^8.0.1:\n- version \"8.0.5\"\n- resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.0.5.tgz#f07d6fdeffcdbb80485570ce3f1bc845fcc812b9\"\n- integrity sha512-RkiLa/AeJx7+9OvniQ/qeWu0w74A8DiPPBclQ6ji3ZQkv5KamO+QGpqmi7O4JIw3rHGUXZ6CoP9tsAkn3gyazg==\n+ version \"8.1.0\"\n+ resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.1.0.tgz#45d5d3d36c7cdd808930cc3e603cf6200dbeb736\"\n+ integrity sha512-B/Sk2Ix7A36fs/ZkuGLIR86EdjbgR6fsAcbx9lOP/QBSXujDNbVmIS/U4Itz5k8fPFDeVZl/zQ/gJW4Jrq6XjQ==\n dependencies:\n fast-deep-equal \"^3.1.1\"\n json-schema-traverse \"^1.0.0\"\n@@ -6828,15 +6828,15 @@ [email protected]:\n node-releases \"^1.1.61\"\n \n browserslist@^4.0.0, browserslist@^4.11.1, browserslist@^4.12.0, browserslist@^4.14.5, browserslist@^4.16.3, browserslist@^4.3.4, browserslist@^4.6.2, browserslist@^4.6.4, browserslist@^4.7.0, browserslist@^4.9.1:\n- version \"4.16.3\"\n- resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.3.tgz#340aa46940d7db878748567c5dea24a48ddf3717\"\n- integrity sha512-vIyhWmIkULaq04Gt93txdh+j02yX/JzlyhLYbV3YQCn/zvES3JnY7TifHHvvr1w5hTDluNKMkV05cs4vy8Q7sw==\n+ version \"4.16.4\"\n+ resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.4.tgz#7ebf913487f40caf4637b892b268069951c35d58\"\n+ integrity sha512-d7rCxYV8I9kj41RH8UKYnvDYCRENUlHRgyXy/Rhr/1BaeLGfiCptEdFE8MIrvGfWbBFNjVYx76SQWvNX1j+/cQ==\n dependencies:\n- caniuse-lite \"^1.0.30001181\"\n- colorette \"^1.2.1\"\n- electron-to-chromium \"^1.3.649\"\n+ caniuse-lite \"^1.0.30001208\"\n+ colorette \"^1.2.2\"\n+ electron-to-chromium \"^1.3.712\"\n escalade \"^3.1.1\"\n- node-releases \"^1.1.70\"\n+ node-releases \"^1.1.71\"\n \n [email protected]:\n version \"0.2.6\"\n@@ -7217,7 +7217,7 @@ caniuse-api@^3.0.0:\n lodash.memoize \"^4.1.2\"\n lodash.uniq \"^4.5.0\"\n \n-caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001181:\n+caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001208:\n version \"1.0.30001208\"\n resolved \"https://registry.yarnpkg.com/caniuse-lite/-/caniuse-lite-1.0.30001208.tgz#a999014a35cebd4f98c405930a057a0d75352eb9\"\n integrity sha512-OE5UE4+nBOro8Dyvv0lfx+SRtfVIOM9uhKqFmJeUbGriqhhStgp1A0OyBpgy3OUF8AhYCT+PVwPC1gMl2ZcQMA==\n@@ -9549,10 +9549,10 @@ ejs@^2.6.1:\n resolved \"https://registry.yarnpkg.com/ejs/-/ejs-2.7.4.tgz#48661287573dcc53e366c7a1ae52c3a120eec9ba\"\n integrity sha512-7vmuyh5+kuUyJKePhQfRQBhXV5Ce+RnaeeQArKu1EAMpL3WbgMt5WG6uQZpEVvYSSsxMXRKOewtDk9RaTKXRlA==\n \n-electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.649:\n- version \"1.3.711\"\n- resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.711.tgz#92c3caf7ffed5e18bf63f66b4b57b4db2409c450\"\n- integrity sha512-XbklBVCDiUeho0PZQCjC25Ha6uBwqqJeyDhPLwLwfWRAo4x+FZFsmu1pPPkXT+B4MQMQoQULfyaMltDopfeiHQ==\n+electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.712:\n+ version \"1.3.712\"\n+ resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.712.tgz#ae467ffe5f95961c6d41ceefe858fc36eb53b38f\"\n+ integrity sha512-3kRVibBeCM4vsgoHHGKHmPocLqtFAGTrebXxxtgKs87hNUzXrX2NuS3jnBys7IozCnw7viQlozxKkmty2KNfrw==\n \n elegant-spinner@^1.0.1:\n version \"1.0.1\"\n@@ -9945,9 +9945,9 @@ eslint-plugin-import@^2.16.0, eslint-plugin-import@^2.18.2, eslint-plugin-import\n tsconfig-paths \"^3.9.0\"\n \n eslint-plugin-jest@^24.1.0:\n- version \"24.3.4\"\n- resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.4.tgz#6d90c3554de0302e879603dd6405474c98849f19\"\n- integrity sha512-3n5oY1+fictanuFkTWPwSlehugBTAgwLnYLFsCllzE3Pl1BwywHl5fL0HFxmMjoQY8xhUDk8uAWc3S4JOHGh3A==\n+ version \"24.3.5\"\n+ resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.5.tgz#71f0b580f87915695c286c3f0eb88cf23664d044\"\n+ integrity sha512-XG4rtxYDuJykuqhsOqokYIR84/C8pRihRtEpVskYLbIIKGwPNW2ySxdctuVzETZE+MbF/e7wmsnbNVpzM0rDug==\n dependencies:\n \"@typescript-eslint/experimental-utils\" \"^4.0.1\"\n \n@@ -12140,12 +12140,11 @@ [email protected]:\n micromatch \"^3.1.10\"\n \n http-proxy-middleware@^1.0.0:\n- version \"1.1.0\"\n- resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.0.tgz#b896b2cc6836019af4a4f2d5f7b21b99c77ea13f\"\n- integrity sha512-OnjU5vyVgcZVe2AjLJyMrk8YLNOC2lspCHirB5ldM+B/dwEfZ5bgVTrFyzE9R7xRWAP/i/FXtvIqKjTNEZBhBg==\n+ version \"1.1.1\"\n+ resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.1.tgz#48900a68cd9d388c735d1dd97302c919b7e94a13\"\n+ integrity sha512-FIDg9zPvOwMhQ3XKB2+vdxK6WWbVAH7s5QpqQCif7a1TNL76GNAATWA1sy6q2gSfss8UJ/Nwza3N6QnFkKclpA==\n dependencies:\n \"@types/http-proxy\" \"^1.17.5\"\n- camelcase \"^6.2.0\"\n http-proxy \"^1.18.1\"\n is-glob \"^4.0.1\"\n is-plain-obj \"^3.0.0\"\n@@ -14341,9 +14340,9 @@ jsdom@^15.2.1:\n xml-name-validator \"^3.0.0\"\n \n jsdom@^16.4.0:\n- version \"16.5.2\"\n- resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.2.tgz#583fac89a0aea31dbf6237e7e4bedccd9beab472\"\n- integrity sha512-JxNtPt9C1ut85boCbJmffaQ06NBnzkQY/MWO3YxPW8IWS38A26z+B1oBvA9LwKrytewdfymnhi4UNH3/RAgZrg==\n+ version \"16.5.3\"\n+ resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.3.tgz#13a755b3950eb938b4482c407238ddf16f0d2136\"\n+ integrity sha512-Qj1H+PEvUsOtdPJ056ewXM4UJPCi4hhLA8wpiz9F2YvsRBhuFsXxtrIFAgGBDynQA9isAMGE91PfUYbdMPXuTA==\n dependencies:\n abab \"^2.0.5\"\n acorn \"^8.1.0\"\n@@ -15590,12 +15589,12 @@ micromatch@^3.1.10, micromatch@^3.1.4:\n to-regex \"^3.0.2\"\n \n micromatch@^4.0.2:\n- version \"4.0.3\"\n- resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.3.tgz#fdad8352bf0cbeb89b391b5d244bc22ff3dd4ec8\"\n- integrity sha512-ueuSaP4i67F/FAUac9zzZ0Dz/5KeKDkITYIS/k4fps+9qeh1SkeH6gbljcqz97mNBOsaWZ+iv2UobMKK/yD+aw==\n+ version \"4.0.4\"\n+ resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.4.tgz#896d519dfe9db25fce94ceb7a500919bf881ebf9\"\n+ integrity sha512-pRmzw/XUcwXGpD9aI9q/0XOwLNygjETJ8y0ao0wdqprrzDa4YnxLcz7fQRZr8voh8V10kGhABbNcHVk5wHgWwg==\n dependencies:\n braces \"^3.0.1\"\n- picomatch \"^2.2.1\"\n+ picomatch \"^2.2.3\"\n \n miller-rabin@^4.0.0:\n version \"4.0.1\"\n@@ -16356,7 +16355,7 @@ node-pre-gyp@^0.11.0:\n semver \"^5.3.0\"\n tar \"^4\"\n \n-node-releases@^1.1.61, node-releases@^1.1.70:\n+node-releases@^1.1.61, node-releases@^1.1.71:\n version \"1.1.71\"\n resolved \"https://registry.yarnpkg.com/node-releases/-/node-releases-1.1.71.tgz#cb1334b179896b1c89ecfdd4b725fb7bbdfc7dbb\"\n integrity sha512-zR6HoT6LrLCRBwukmrVbHv0EpEQjksO6GmFcZQQuCAy139BEsoVKPYnf3jongYW83fAa1torLGYwxxky/p28sg==\n@@ -17571,10 +17570,10 @@ [email protected]:\n dependencies:\n split2 \"^3.1.1\"\n \n-picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2:\n- version \"2.2.2\"\n- resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.2.tgz#21f333e9b6b8eaff02468f5146ea406d345f4dad\"\n- integrity sha512-q0M/9eZHzmr0AulXyPwNfZjtwZ/RBZlbN3K3CErVrk50T2ASYI7Bye0EvekFY3IP1Nt2DHu0re+V2ZHIpMkuWg==\n+picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2, picomatch@^2.2.3:\n+ version \"2.2.3\"\n+ resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.3.tgz#465547f359ccc206d3c48e46a1bcb89bf7ee619d\"\n+ integrity sha512-KpELjfwcCDUb9PeigTs2mBJzXUPzAuP2oPcA989He8Rte0+YUAjw1JVedDhuTKPkHjSYzMN3npC9luThGYEKdg==\n \n pify@^2.0.0, pify@^2.2.0, pify@^2.3.0:\n version \"2.3.0\"\n@@ -18446,9 +18445,9 @@ postcss@^7, postcss@^7.0.0, postcss@^7.0.1, postcss@^7.0.14, postcss@^7.0.17, po\n supports-color \"^6.1.0\"\n \n postcss@^8.1.0, postcss@^8.2.8:\n- version \"8.2.9\"\n- resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.9.tgz#fd95ff37b5cee55c409b3fdd237296ab4096fba3\"\n- integrity sha512-b+TmuIL4jGtCHtoLi+G/PisuIl9avxs8IZMSmlABRwNz5RLUUACrC+ws81dcomz1nRezm5YPdXiMEzBEKgYn+Q==\n+ version \"8.2.10\"\n+ resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.10.tgz#ca7a042aa8aff494b334d0ff3e9e77079f6f702b\"\n+ integrity sha512-b/h7CPV7QEdrqIxtAf2j31U5ef05uBDuvoXv6L51Q4rcS1jdlXAVKJv+atCFdUXYl9dyTHGyoMzIepwowRJjFw==\n dependencies:\n colorette \"^1.2.2\"\n nanoid \"^3.1.22\"\n@@ -19318,9 +19317,9 @@ rc-tree@^4.0.0, rc-tree@~4.1.0:\n rc-virtual-list \"^3.0.1\"\n \n rc-trigger@^5.0.0, rc-trigger@^5.0.4, rc-trigger@^5.1.2, rc-trigger@^5.2.1:\n- version \"5.2.3\"\n- resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.3.tgz#8c55046ab432d7b52d51c69afb57ebb5bbe37e17\"\n- integrity sha512-6Fokao07HUbqKIDkDRFEM0AGZvsvK0Fbp8A/KFgl1ngaqfO1nY037cISCG1Jm5fxImVsXp9awdkP7Vu5cxjjog==\n+ version \"5.2.4\"\n+ resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.4.tgz#f1cca4a6c1f378a5d6fadec010292250772069d3\"\n+ integrity sha512-nLZa4XYo3hOAVauQr7HsGrBtE8/pyoIWhHZnpr7x/H/dd6pPeRzH0//+1TzaBAXylbFgsY6hogKAMeJwaKeDFw==\n dependencies:\n \"@babel/runtime\" \"^7.11.2\"\n classnames \"^2.2.6\"\n@@ -20516,9 +20515,9 @@ rollup@^1.31.1:\n acorn \"^7.1.0\"\n \n rollup@^2.40.0, rollup@^2.8.0:\n- version \"2.45.0\"\n- resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.0.tgz#bfcce2347c96f15f5c78ac860bc38e3349ba27c9\"\n- integrity sha512-JJznbtGIsHZfKH0Sa9RpCAy5JarH8SWvBzRAGuRkgzAafb8e8D7VSMJ0O1Bsix1nn91koN/Ecvl2+ZWhljcuTw==\n+ version \"2.45.1\"\n+ resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.1.tgz#eae2b94dc2088b4e0a3b7197a5a1ee0bdd589d5c\"\n+ integrity sha512-vPD+JoDj3CY8k6m1bLcAFttXMe78P4CMxoau0iLVS60+S9kLsv2379xaGy4NgYWu+h2WTlucpoLPAoUoixFBag==\n optionalDependencies:\n fsevents \"~2.3.1\"\n \n@@ -22971,9 +22970,9 @@ typescript@~4.1.5:\n integrity sha512-6OSu9PTIzmn9TCDiovULTnET6BgXtDYL4Gg4szY+cGsc3JP1dQL8qvE8kShTRx1NIw4Q9IBHlwODjkjWEtMUyA==\n \n ua-parser-js@^0.7.18:\n- version \"0.7.27\"\n- resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.27.tgz#b54f8ce9eb6c7abf3584edeaf9a3d8b3bd92edba\"\n- integrity sha512-eXMaRYK2skomGocoX0x9sBXzx5A1ZVQgXfrW4mTc8dT0zS7olEcyfudAzRC5tIIRgLxQ69B6jut3DI+n5hslPA==\n+ version \"0.7.28\"\n+ resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.28.tgz#8ba04e653f35ce210239c64661685bf9121dec31\"\n+ integrity sha512-6Gurc1n//gjp9eQNXjD9O3M/sMwVtN5S8Lv9bvOYBfKfDNiIIhqiyi01vMBO45u4zkDE420w/e0se7Vs+sIg+g==\n \n [email protected]:\n version \"3.4.10\"\n@@ -22984,9 +22983,9 @@ [email protected]:\n source-map \"~0.6.1\"\n \n uglify-js@^3.1.4, uglify-js@^3.4.9:\n- version \"3.13.3\"\n- resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.3.tgz#ce72a1ad154348ea2af61f50933c76cc8802276e\"\n- integrity sha512-otIc7O9LyxpUcQoXzj2hL4LPWKklO6LJWoJUzNa8A17Xgi4fOeDC8FBDOLHnC/Slo1CQgsZMcM6as0M76BZaig==\n+ version \"3.13.4\"\n+ resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.4.tgz#592588bb9f47ae03b24916e2471218d914955574\"\n+ integrity sha512-kv7fCkIXyQIilD5/yQy8O+uagsYIOt5cZvs890W40/e/rvjMSzJw81o9Bg0tkURxzZBROtDQhW2LFjOGoK3RZw==\n \n [email protected]:\n version \"0.0.6\"\n"]
2
["983fef55ef08ca2ca25349bb2d5bdff10ecf89f4", "7e9bd7c86df1032d53e752654fe4a446951480bb"]
["refactor", "build"]
skip ruff format in pre-commit ci runner,ecma 7 ready
["diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml\nindex 6193d96..4ba39d6 100644\n--- a/.pre-commit-config.yaml\n+++ b/.pre-commit-config.yaml\n@@ -9,6 +9,7 @@ ci:\n - nixpkgs-fmt\n - prettier\n - ruff\n+ - ruff-format\n - shellcheck\n - shfmt\n - statix\n", "diff --git a/config/webpack.config.prod.js b/config/webpack.config.prod.js\nindex f7c6b23..4a00c65 100644\n--- a/config/webpack.config.prod.js\n+++ b/config/webpack.config.prod.js\n@@ -266,7 +266,7 @@ module.exports = {\n : new UglifyJsPlugin({\n uglifyOptions: {\n ie8: false,\n- ecma: 6,\n+ ecma: 7,\n compress: {\n warnings: false,\n // Disabled because of an issue with Uglify breaking seemingly valid code:\n"]
2
["9117fdedb9b5ce0345c31b3e1fa22ae8554944d4", "6aa63c9b8d4dcdbb401743adc3c9a1020d943250"]
["ci", "build"]
lint source on ci,ecma 7 ready
["diff --git a/.travis.yml b/.travis.yml\nindex d56185e..96510cb 100644\n--- a/.travis.yml\n+++ b/.travis.yml\n@@ -2,5 +2,6 @@ language: node_js\n node_js:\n - 'stable'\n script:\n+ - yarn lint\n - yarn build\n - yarn test\n", "diff --git a/config/webpack.config.prod.js b/config/webpack.config.prod.js\nindex f7c6b23..4a00c65 100644\n--- a/config/webpack.config.prod.js\n+++ b/config/webpack.config.prod.js\n@@ -266,7 +266,7 @@ module.exports = {\n : new UglifyJsPlugin({\n uglifyOptions: {\n ie8: false,\n- ecma: 6,\n+ ecma: 7,\n compress: {\n warnings: false,\n // Disabled because of an issue with Uglify breaking seemingly valid code:\n"]
2
["2ac99c0a66a1adc18ee4ef660608f814823dd198", "6aa63c9b8d4dcdbb401743adc3c9a1020d943250"]
["ci", "build"]
fix sonar integration,tests should pass now Make the code more safe
["diff --git a/.ci/scripts/distribution/analyse-java.sh b/.ci/scripts/distribution/analyse-java.sh\nindex a0122f7..0e965df 100755\n--- a/.ci/scripts/distribution/analyse-java.sh\n+++ b/.ci/scripts/distribution/analyse-java.sh\n@@ -23,12 +23,12 @@ else\n fi\n \n if [ \"${GIT_BRANCH}\" == \"master\" ] || [ \"${GIT_BRANCH}\" == \"develop\" ]; then\n- TARGET_BRANCH=\"master\"\n+ TARGET_BRANCH=\"${GIT_BRANCH}\"\n else\n TARGET_BRANCH=\"develop\"\n+ PROPERTIES+=(\"-Dsonar.branch.target=${TARGET_BRANCH}\")\n fi\n \n- PROPERTIES+=(\"-Dsonar.branch.target=${TARGET_BRANCH}\")\n git fetch --no-tags \"${GIT_URL}\" \"+refs/heads/${TARGET_BRANCH}:refs/remotes/origin/${TARGET_BRANCH}\"\n fi\n \ndiff --git a/parent/pom.xml b/parent/pom.xml\nindex f4c3160..d34b41f 100644\n--- a/parent/pom.xml\n+++ b/parent/pom.xml\n@@ -1570,7 +1570,7 @@\n <!-- sonarscanner integration -->\n <!-- sonar.login token must be passed at runtime to avoid sharing token -->\n <sonar.host.url>https://sonarcloud.io</sonar.host.url>\n- <sonar.organization>zeebe-io</sonar.organization>\n+ <sonar.organization>camunda-cloud</sonar.organization>\n <sonar.login>${env.SONARCLOUD_TOKEN}</sonar.login>\n <sonar.links.issue>${project.scm.url}/issues</sonar.links.issue>\n <sonar.cpd.exclusions>\n", "diff --git a/goreleaserlib/goreleaser.go b/goreleaserlib/goreleaser.go\nindex 28ba3f4..67ba95d 100644\n--- a/goreleaserlib/goreleaser.go\n+++ b/goreleaserlib/goreleaser.go\n@@ -27,6 +27,15 @@ import (\n \tyaml \"gopkg.in/yaml.v2\"\n )\n \n+var (\n+\tnormalPadding = cli.Default.Padding\n+\tincreasedPadding = normalPadding * 2\n+)\n+\n+func init() {\n+\tlog.SetHandler(cli.Default)\n+}\n+\n var pipes = []pipeline.Piper{\n \tdefaults.Pipe{}, // load default configs\n \tgit.Pipe{}, // get and validate git repo state\n@@ -89,17 +98,15 @@ func Release(flags Flags) error {\n \t\tctx.Publish = false\n \t}\n \tctx.RmDist = flags.Bool(\"rm-dist\")\n-\tlogger, _ := log.Log.(*log.Logger)\n-\thandler, _ := logger.Handler.(*cli.Handler)\n \tfor _, pipe := range pipes {\n-\t\thandler.Padding = 3\n+\t\tcli.Default.Padding = normalPadding\n \t\tlog.Infof(\"\\033[1m%s\\033[0m\", strings.ToUpper(pipe.String()))\n-\t\thandler.Padding = 6\n+\t\tcli.Default.Padding = increasedPadding\n \t\tif err := handle(pipe.Run(ctx)); err != nil {\n \t\t\treturn err\n \t\t}\n \t}\n-\thandler.Padding = 3\n+\tcli.Default.Padding = normalPadding\n \treturn nil\n }\n \ndiff --git a/main.go b/main.go\nindex b9b961d..7ced1dd 100644\n--- a/main.go\n+++ b/main.go\n@@ -18,7 +18,7 @@ var (\n )\n \n func init() {\n-\tlog.SetHandler(lcli.New(os.Stdout))\n+\tlog.SetHandler(lcli.Default)\n }\n \n func main() {\n"]
2
["6cbbd98dfe6c768dbe49f8d6d2448856a9a86089", "5636313d7c9cfbd9f48578fd104771d65eae9720"]
["build", "fix"]
add react ecosystem,apply permissions to profile request
["diff --git a/package.json b/package.json\nindex 1ba8c4f..d1de9a0 100644\n--- a/package.json\n+++ b/package.json\n@@ -36,14 +36,19 @@\n \"@types/node\": \"^9.3.0\",\n \"@types/react\": \"^16.0.34\",\n \"@types/react-dom\": \"^16.0.3\",\n+ \"@types/react-motion\": \"^0.0.25\",\n \"bootstrap-sass\": \"^3.3.7\",\n \"highcharts\": \"^6.0.4\",\n \"html2canvas\": \"^1.0.0-alpha.9\",\n+ \"immer\": \"^1.2.1\",\n \"lodash\": \"^4.17.4\",\n \"moment\": \"^2.20.1\",\n \"normalize.css\": \"^8.0.0\",\n- \"react\": \"^16.2.0\",\n- \"react-dom\": \"^16.2.0\",\n+ \"react\": \"^16.3.1\",\n+ \"react-dom\": \"^16.3.1\",\n+ \"react-motion\": \"^0.5.2\",\n+ \"react-redux\": \"^5.0.7\",\n+ \"redux\": \"^3.7.2\",\n \"rxjs\": \"^5.5.6\",\n \"vue\": \"^2.5.13\",\n \"vue-plugin-webextension-i18n\": \"^0.1.0\",\ndiff --git a/yarn.lock b/yarn.lock\nindex c8898d8..5d0fc9f 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -187,6 +187,12 @@\n \"@types/node\" \"*\"\n \"@types/react\" \"*\"\n \n+\"@types/react-motion@^0.0.25\":\n+ version \"0.0.25\"\n+ resolved \"https://registry.npmjs.org/@types/react-motion/-/react-motion-0.0.25.tgz#2445745ee8e8e6149faa47a36ff6b0d4c21dbf94\"\n+ dependencies:\n+ \"@types/react\" \"*\"\n+\n \"@types/react@*\", \"@types/react@^16.0.34\":\n version \"16.0.40\"\n resolved \"https://registry.npmjs.org/@types/react/-/react-16.0.40.tgz#caabc2296886f40b67f6fc80f0f3464476461df9\"\n@@ -3837,6 +3843,10 @@ [email protected]:\n version \"4.2.1\"\n resolved \"https://registry.npmjs.org/hoek/-/hoek-4.2.1.tgz#9634502aa12c445dd5a7c5734b572bb8738aacbb\"\n \n+hoist-non-react-statics@^2.5.0:\n+ version \"2.5.0\"\n+ resolved \"https://registry.npmjs.org/hoist-non-react-statics/-/hoist-non-react-statics-2.5.0.tgz#d2ca2dfc19c5a91c5a6615ce8e564ef0347e2a40\"\n+\n home-or-tmp@^2.0.0:\n version \"2.0.0\"\n resolved \"https://registry.npmjs.org/home-or-tmp/-/home-or-tmp-2.0.0.tgz#e36c3f2d2cae7d746a857e38d18d5f32a7882db8\"\n@@ -4004,6 +4014,10 @@ ignore@^3.3.5:\n version \"3.3.7\"\n resolved \"https://registry.npmjs.org/ignore/-/ignore-3.3.7.tgz#612289bfb3c220e186a58118618d5be8c1bab021\"\n \n+immer@^1.2.1:\n+ version \"1.2.1\"\n+ resolved \"https://registry.npmjs.org/immer/-/immer-1.2.1.tgz#96e2ae29cdfc428f28120b832701931b92fa597c\"\n+\n import-local@^1.0.0:\n version \"1.0.0\"\n resolved \"https://registry.npmjs.org/import-local/-/import-local-1.0.0.tgz#5e4ffdc03f4fe6c009c6729beb29631c2f8227bc\"\n@@ -4104,7 +4118,7 @@ interpret@^1.0.0:\n version \"1.1.0\"\n resolved \"https://registry.npmjs.org/interpret/-/interpret-1.1.0.tgz#7ed1b1410c6a0e0f78cf95d3b8440c63f78b8614\"\n \n-invariant@^2.2.2:\n+invariant@^2.0.0, invariant@^2.2.2:\n version \"2.2.4\"\n resolved \"https://registry.npmjs.org/invariant/-/invariant-2.2.4.tgz#610f3c92c9359ce1db616e538008d23ff35158e6\"\n dependencies:\n@@ -5040,6 +5054,10 @@ locate-path@^2.0.0:\n p-locate \"^2.0.0\"\n path-exists \"^3.0.0\"\n \n+lodash-es@^4.17.5, lodash-es@^4.2.1:\n+ version \"4.17.8\"\n+ resolved \"https://registry.npmjs.org/lodash-es/-/lodash-es-4.17.8.tgz#6fa8c8c5d337481df0bdf1c0d899d42473121e45\"\n+\n lodash._reinterpolate@~3.0.0:\n version \"3.0.0\"\n resolved \"https://registry.npmjs.org/lodash._reinterpolate/-/lodash._reinterpolate-3.0.0.tgz#0ccf2d89166af03b3663c796538b75ac6e114d9d\"\n@@ -5149,7 +5167,7 @@ [email protected]:\n version \"4.17.2\"\n resolved \"https://registry.npmjs.org/lodash/-/lodash-4.17.2.tgz#34a3055babe04ce42467b607d700072c7ff6bf42\"\n \[email protected], lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\[email protected], lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.17.5, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\n version \"4.17.5\"\n resolved \"https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz#99a92d65c0272debe8c96b6057bc8fbfa3bed511\"\n \n@@ -6467,7 +6485,7 @@ promise@^7.1.1:\n dependencies:\n asap \"~2.0.3\"\n \n-prop-types@^15.6.0:\n+prop-types@^15.5.8, prop-types@^15.6.0:\n version \"15.6.1\"\n resolved \"https://registry.npmjs.org/prop-types/-/prop-types-15.6.1.tgz#36644453564255ddda391191fb3a125cbdf654ca\"\n dependencies:\n@@ -6574,7 +6592,7 @@ quick-lru@^1.0.0:\n version \"1.1.0\"\n resolved \"https://registry.npmjs.org/quick-lru/-/quick-lru-1.1.0.tgz#4360b17c61136ad38078397ff11416e186dcfbb8\"\n \[email protected]:\[email protected], raf@^3.1.0:\n version \"3.4.0\"\n resolved \"https://registry.npmjs.org/raf/-/raf-3.4.0.tgz#a28876881b4bc2ca9117d4138163ddb80f781575\"\n dependencies:\n@@ -6645,9 +6663,9 @@ react-dev-utils@^5.0.0:\n strip-ansi \"3.0.1\"\n text-table \"0.2.0\"\n \n-react-dom@^16.2.0:\n- version \"16.2.0\"\n- resolved \"https://registry.npmjs.org/react-dom/-/react-dom-16.2.0.tgz#69003178601c0ca19b709b33a83369fe6124c044\"\n+react-dom@^16.3.1:\n+ version \"16.3.1\"\n+ resolved \"https://registry.npmjs.org/react-dom/-/react-dom-16.3.1.tgz#6a3c90a4fb62f915bdbcf6204422d93a7d4ca573\"\n dependencies:\n fbjs \"^0.8.16\"\n loose-envify \"^1.1.0\"\n@@ -6658,9 +6676,28 @@ react-error-overlay@^4.0.0:\n version \"4.0.0\"\n resolved \"https://registry.npmjs.org/react-error-overlay/-/react-error-overlay-4.0.0.tgz#d198408a85b4070937a98667f500c832f86bd5d4\"\n \n-react@^16.2.0:\n- version \"16.2.0\"\n- resolved \"https://registry.npmjs.org/react/-/react-16.2.0.tgz#a31bd2dab89bff65d42134fa187f24d054c273ba\"\n+react-motion@^0.5.2:\n+ version \"0.5.2\"\n+ resolved \"https://registry.npmjs.org/react-motion/-/react-motion-0.5.2.tgz#0dd3a69e411316567927917c6626551ba0607316\"\n+ dependencies:\n+ performance-now \"^0.2.0\"\n+ prop-types \"^15.5.8\"\n+ raf \"^3.1.0\"\n+\n+react-redux@^5.0.7:\n+ version \"5.0.7\"\n+ resolved \"https://registry.npmjs.org/react-redux/-/react-redux-5.0.7.tgz#0dc1076d9afb4670f993ffaef44b8f8c1155a4c8\"\n+ dependencies:\n+ hoist-non-react-statics \"^2.5.0\"\n+ invariant \"^2.0.0\"\n+ lodash \"^4.17.5\"\n+ lodash-es \"^4.17.5\"\n+ loose-envify \"^1.1.0\"\n+ prop-types \"^15.6.0\"\n+\n+react@^16.3.1:\n+ version \"16.3.1\"\n+ resolved \"https://registry.npmjs.org/react/-/react-16.3.1.tgz#4a2da433d471251c69b6033ada30e2ed1202cfd8\"\n dependencies:\n fbjs \"^0.8.16\"\n loose-envify \"^1.1.0\"\n@@ -6788,6 +6825,15 @@ reduce-function-call@^1.0.1:\n dependencies:\n balanced-match \"^0.4.2\"\n \n+redux@^3.7.2:\n+ version \"3.7.2\"\n+ resolved \"https://registry.npmjs.org/redux/-/redux-3.7.2.tgz#06b73123215901d25d065be342eb026bc1c8537b\"\n+ dependencies:\n+ lodash \"^4.2.1\"\n+ lodash-es \"^4.2.1\"\n+ loose-envify \"^1.1.0\"\n+ symbol-observable \"^1.0.3\"\n+\n regenerate@^1.2.1:\n version \"1.3.3\"\n resolved \"https://registry.npmjs.org/regenerate/-/regenerate-1.3.3.tgz#0c336d3980553d755c39b586ae3b20aa49c82b7f\"\n@@ -7811,6 +7857,10 @@ [email protected]:\n version \"1.0.1\"\n resolved \"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.0.1.tgz#8340fc4702c3122df5d22288f88283f513d3fdd4\"\n \n+symbol-observable@^1.0.3:\n+ version \"1.2.0\"\n+ resolved \"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.2.0.tgz#c22688aed4eab3cdc2dfeacbb561660560a00804\"\n+\n symbol-tree@^3.2.2:\n version \"3.2.2\"\n resolved \"https://registry.npmjs.org/symbol-tree/-/symbol-tree-3.2.2.tgz#ae27db38f660a7ae2e1c3b7d1bc290819b8519e6\"\n", "diff --git a/client/src/components/Profile/AboutCard.tsx b/client/src/components/Profile/AboutCard.tsx\nindex 3bd6e9a..e07ddb6 100644\n--- a/client/src/components/Profile/AboutCard.tsx\n+++ b/client/src/components/Profile/AboutCard.tsx\n@@ -11,6 +11,7 @@ import { InfoCircleOutlined } from '@ant-design/icons';\n \n type Props = {\n data: GeneralInfo;\n+ isEditingModeEnabled: boolean;\n };\n \n class AboutCard extends React.Component<Props> {\ndiff --git a/client/src/components/Profile/ContactsCard.tsx b/client/src/components/Profile/ContactsCard.tsx\nindex 6fe80a3..3a35c9f 100644\n--- a/client/src/components/Profile/ContactsCard.tsx\n+++ b/client/src/components/Profile/ContactsCard.tsx\n@@ -12,8 +12,11 @@ import { ContactsOutlined } from '@ant-design/icons';\n \n type Props = {\n data: Contacts;\n+ isEditingModeEnabled: boolean;\n };\n \n+type Contact = { name: string, value?: string };\n+\n class ContactsCard extends React.Component<Props> {\n render() {\n const { email, telegram, phone, skype, notes } = this.props.data;\n@@ -32,7 +35,7 @@ class ContactsCard extends React.Component<Props> {\n }, {\n name: 'Notes',\n value: notes,\n- }].filter(({ value }: { name: string, value: string | null }) => value);\n+ }].filter(({ value }: Contact) => value);\n \n return (\n <CommonCard\n@@ -42,7 +45,7 @@ class ContactsCard extends React.Component<Props> {\n <List\n itemLayout=\"horizontal\"\n dataSource={contacts}\n- renderItem={({ name, value }: { name: string, value: string }) => (\n+ renderItem={({ name, value }: Contact) => (\n <List.Item>\n <Text strong>{name}:</Text> {value}\n </List.Item>\ndiff --git a/client/src/components/Profile/EducationCard.tsx b/client/src/components/Profile/EducationCard.tsx\nindex 4279c9f..b409c29 100644\n--- a/client/src/components/Profile/EducationCard.tsx\n+++ b/client/src/components/Profile/EducationCard.tsx\n@@ -12,6 +12,7 @@ import { ReadOutlined } from '@ant-design/icons';\n \n type Props = {\n data: GeneralInfo;\n+ isEditingModeEnabled: boolean;\n };\n \n class EducationCard extends React.Component<Props> {\ndiff --git a/client/src/components/Profile/EnglishCard.tsx b/client/src/components/Profile/EnglishCard.tsx\nindex d8f8ab4..2d5efa0 100644\n--- a/client/src/components/Profile/EnglishCard.tsx\n+++ b/client/src/components/Profile/EnglishCard.tsx\n@@ -11,6 +11,7 @@ import { TagOutlined } from '@ant-design/icons';\n \n type Props = {\n data: GeneralInfo;\n+ isEditingModeEnabled: boolean;\n };\n \n class EnglishCard extends React.Component<Props> {\ndiff --git a/client/src/components/Profile/MainCard.tsx b/client/src/components/Profile/MainCard.tsx\nindex cbfb71b..c0d49cc 100644\n--- a/client/src/components/Profile/MainCard.tsx\n+++ b/client/src/components/Profile/MainCard.tsx\n@@ -4,6 +4,8 @@ import { GithubAvatar } from 'components';\n import {\n Card,\n Typography,\n+ Drawer,\n+ Checkbox,\n } from 'antd';\n \n const { Title, Paragraph } = Typography;\n@@ -11,30 +13,70 @@ const { Title, Paragraph } = Typography;\n import {\n GithubFilled,\n EnvironmentFilled,\n+ EditOutlined,\n+ SettingOutlined,\n } from '@ant-design/icons';\n \n type Props = {\n data: GeneralInfo;\n+ isEditingModeEnabled: boolean;\n };\n \n-class MainCard extends React.Component<Props> {\n+type State = {\n+ isSettingsVisible: boolean;\n+}\n+\n+class MainCard extends React.Component<Props, State> {\n+ state = {\n+ isSettingsVisible: false,\n+ }\n+\n+ private showSettings = () => {\n+ this.setState({ isSettingsVisible: true });\n+ }\n+\n+ private hideSettings = () => {\n+ this.setState({ isSettingsVisible: false });\n+ }\n+\n render() {\n const { githubId, name, locationName } = this.props.data;\n+ const { isSettingsVisible } = this.state;\n+\n return (\n- <Card>\n- <GithubAvatar size={96} githubId={githubId} style={{ margin: '0 auto 10px', display: 'block' }} />\n- <Title level={1} style={{ fontSize: 24, textAlign: 'center', margin: 0 }}>{name}</Title>\n- <Paragraph style={{ textAlign: 'center', marginBottom: 20 }}>\n- <a target=\"_blank\" href={`https://github.com/${githubId}`} style={{ marginLeft: '-14px', fontSize: 16 }}>\n- <GithubFilled /> {githubId}\n- </a>\n- </Paragraph>\n- <Paragraph style={{ textAlign: 'center', margin: 0 }}>\n- <span style={{ marginLeft: '-14px' }}>\n- <EnvironmentFilled /> {locationName}\n- </span>\n- </Paragraph>\n- </Card>\n+ <>\n+\n+ <Card\n+ actions={[\n+ <EditOutlined key=\"main-card-actions-edit\"/>,\n+ <SettingOutlined key=\"main-card-actions-settings\" onClick={this.showSettings} />,\n+ ]}\n+ >\n+ <GithubAvatar size={96} githubId={githubId} style={{ margin: '0 auto 10px', display: 'block' }} />\n+ <Title level={1} style={{ fontSize: 24, textAlign: 'center', margin: 0 }}>{name}</Title>\n+ <Paragraph style={{ textAlign: 'center', marginBottom: 20 }}>\n+ <a target=\"_blank\" href={`https://github.com/${githubId}`} style={{ marginLeft: '-14px', fontSize: 16 }}>\n+ <GithubFilled /> {githubId}\n+ </a>\n+ </Paragraph>\n+ <Paragraph style={{ textAlign: 'center', margin: 0 }}>\n+ <span style={{ marginLeft: '-14px' }}>\n+ <EnvironmentFilled /> {locationName}\n+ </span>\n+ </Paragraph>\n+ <Drawer\n+ title=\"Who can see my profile?\"\n+ placement=\"top\"\n+ closable={true}\n+ onClose={this.hideSettings}\n+ visible={isSettingsVisible}\n+ getContainer={false}\n+ style={{ position: 'absolute', display: isSettingsVisible ? 'block' : 'none' }}\n+ >\n+ <Checkbox>Nobody</Checkbox>\n+ </Drawer>\n+ </Card>\n+ </>\n );\n }\n }\ndiff --git a/client/src/components/Profile/MentorStatsCard.tsx b/client/src/components/Profile/MentorStatsCard.tsx\nindex ca54480..1ec3b9c 100644\n--- a/client/src/components/Profile/MentorStatsCard.tsx\n+++ b/client/src/components/Profile/MentorStatsCard.tsx\n@@ -18,6 +18,7 @@ import {\n \n type Props = {\n data: MentorStats[];\n+ isEditingModeEnabled: boolean;\n };\n \n type State = {\n@@ -80,7 +81,7 @@ class MentorStatsCard extends React.Component<Props, State> {\n <Text strong>{courseName}{locationName && ` / ${locationName}`}</Text>\n </p>\n {\n- idx === 0 && (\n+ students ? idx === 0 && (\n <List\n itemLayout=\"horizontal\"\n dataSource={students}\n@@ -116,12 +117,14 @@ class MentorStatsCard extends React.Component<Props, State> {\n </List.Item>\n )}\n />\n- )\n+ ) : <p>Doesn't have students at this course yet</p>\n }\n </div>\n- <Button type=\"dashed\" onClick={this.showMentorStatsModal.bind(null, idx)}>\n- <FullscreenOutlined/>\n- </Button>\n+ {\n+ students && <Button type=\"dashed\" onClick={this.showMentorStatsModal.bind(null, idx)}>\n+ <FullscreenOutlined/>\n+ </Button>\n+ }\n </List.Item>\n )}\n />\ndiff --git a/client/src/components/Profile/MentorStatsModal.tsx b/client/src/components/Profile/MentorStatsModal.tsx\nindex 47b5f2a..0e94cc1 100644\n--- a/client/src/components/Profile/MentorStatsModal.tsx\n+++ b/client/src/components/Profile/MentorStatsModal.tsx\n@@ -38,7 +38,7 @@ class MentorStatsModal extends React.Component<Props> {\n >\n <Row gutter={[16, 16]}>\n {\n- students.map(({ name, githubId, isExpelled, totalScore }) => {\n+ students?.map(({ name, githubId, isExpelled, totalScore }) => {\n const profile = `/profile?githubId=${githubId}`;\n const guithubLink = `https://github.com/${githubId}`;\n const privateRepoLink = `https://github.com/rolling-scopes-school/${githubId}-${courseYearPostfix}`;\ndiff --git a/client/src/components/Profile/PublicFeedbackCard.tsx b/client/src/components/Profile/PublicFeedbackCard.tsx\nindex 2f8a999..6ce1862 100644\n--- a/client/src/components/Profile/PublicFeedbackCard.tsx\n+++ b/client/src/components/Profile/PublicFeedbackCard.tsx\n@@ -22,6 +22,7 @@ import {\n \n type Props = {\n data: PublicFeedback[];\n+ isEditingModeEnabled: boolean;\n };\n \n interface State {\ndiff --git a/client/src/components/Profile/StudentStatsCard.tsx b/client/src/components/Profile/StudentStatsCard.tsx\nindex c811640..b472e49 100644\n--- a/client/src/components/Profile/StudentStatsCard.tsx\n+++ b/client/src/components/Profile/StudentStatsCard.tsx\n@@ -18,6 +18,7 @@ import {\n \n type Props = {\n data: StudentStats[];\n+ isEditingModeEnabled: boolean;\n };\n \n type State = {\ndiff --git a/client/src/pages/profile/index.tsx b/client/src/pages/profile/index.tsx\nindex 68b2a70..b6ffb1a 100644\n--- a/client/src/pages/profile/index.tsx\n+++ b/client/src/pages/profile/index.tsx\n@@ -1,6 +1,7 @@\n import * as React from 'react';\n import {\n Result,\n+ Button,\n } from 'antd';\n import css from 'styled-jsx/css';\n import Masonry from 'react-masonry-css';\n@@ -23,18 +24,25 @@ import CoreJsIviewsCard from 'components/Profile/CoreJsIviewsCard';\n import { CoreJsInterviewData } from 'components/Profile/CoreJsIviewsCard';\n import PreScreeningIviewCard from 'components/Profile/PreScreeningIviewCard';\n \n+import {\n+ EditOutlined,\n+ EyeOutlined,\n+} from '@ant-design/icons';\n+\n type Props = {\n router: NextRouter;\n session: Session;\n };\n \n type State = {\n+ isEditingModeEnabled: boolean;\n profile: ProfileInfo | null;\n isLoading: boolean;\n };\n \n class ProfilePage extends React.Component<Props, State> {\n state: State = {\n+ isEditingModeEnabled: false,\n isLoading: true,\n profile: null,\n };\n@@ -79,6 +87,12 @@ class ProfilePage extends React.Component<Props, State> {\n }\n };\n \n+ private toggleEditViewProfileButton = () => {\n+ const { isEditingModeEnabled } = this.state;\n+\n+ this.setState({ isEditingModeEnabled: !isEditingModeEnabled });\n+ }\n+\n async componentDidMount() {\n await this.fetchData();\n }\n@@ -90,21 +104,29 @@ class ProfilePage extends React.Component<Props, State> {\n }\n \n render() {\n- const { profile } = this.state;\n+ const { profile, isEditingModeEnabled } = this.state;\n \n const cards = [\n- profile?.generalInfo && <MainCard data={profile.generalInfo}/>,\n- profile?.generalInfo?.aboutMyself && <AboutCard data={profile.generalInfo}/>,\n- profile?.generalInfo?.englishLevel && <EnglishCard data={profile.generalInfo}/>,\n- profile?.generalInfo?.educationHistory.length && <EducationCard data={profile.generalInfo}/>,\n- profile?.contacts && <ContactsCard data={profile.contacts}/>,\n- profile?.publicFeedback.length && <PublicFeedbackCard data={profile.publicFeedback}/>,\n- profile?.studentStats.length && <StudentStatsCard data={profile.studentStats}/>,\n- profile?.mentorStats.length && <MentorStatsCard data={profile.mentorStats}/>,\n- profile?.studentStats.length &&\n- this.hadStudentCoreJSInterview(profile.studentStats) &&\n+ profile?.generalInfo &&\n+ <MainCard data={profile.generalInfo} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.generalInfo?.aboutMyself &&\n+ <AboutCard data={profile.generalInfo} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.generalInfo?.englishLevel &&\n+ <EnglishCard data={profile.generalInfo} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.generalInfo?.educationHistory?.length &&\n+ <EducationCard data={profile.generalInfo} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.contacts &&\n+ <ContactsCard data={profile.contacts} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.publicFeedback?.length &&\n+ <PublicFeedbackCard data={profile.publicFeedback} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.studentStats?.length &&\n+ <StudentStatsCard data={profile.studentStats} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.mentorStats?.length &&\n+ <MentorStatsCard data={profile.mentorStats} isEditingModeEnabled={isEditingModeEnabled}/>,\n+ profile?.studentStats?.length && this.hadStudentCoreJSInterview(profile.studentStats) &&\n <CoreJsIviewsCard data={this.getStudentCoreJSInterviews(profile.studentStats)}/>,\n- profile?.stageInterviewFeedback.length && <PreScreeningIviewCard data={profile.stageInterviewFeedback}/>,\n+ profile?.stageInterviewFeedback.length &&\n+ <PreScreeningIviewCard data={profile.stageInterviewFeedback}/>,\n ].filter(Boolean) as JSX.Element[];\n \n return (\n@@ -114,6 +136,17 @@ class ProfilePage extends React.Component<Props, State> {\n {\n this.state.profile\n ? <div style={{ padding: 10 }}>\n+ <Button\n+ type=\"ghost\"\n+ style={{ position: 'fixed', width: 80, right: 10, zIndex: 1 }}\n+ onClick={this.toggleEditViewProfileButton}\n+ >\n+ {\n+ isEditingModeEnabled ?\n+ <span><EditOutlined/> Edit</span> :\n+ <span><EyeOutlined /> View</span>\n+ }\n+ </Button>\n <Masonry\n breakpointCols={{\n default: 4,\ndiff --git a/common/models/profile.ts b/common/models/profile.ts\nindex 6a06fd1..ce7abc2 100644\n--- a/common/models/profile.ts\n+++ b/common/models/profile.ts\n@@ -3,26 +3,25 @@ import { EnglishLevel } from './';\n export interface GeneralInfo {\n name: string;\n githubId: string;\n- aboutMyself: string;\n+ aboutMyself?: string;\n locationName: string;\n- educationHistory: any;\n- employmentHistory: any;\n- englishLevel: EnglishLevel;\n+ educationHistory?: any;\n+ englishLevel?: EnglishLevel;\n }\n \n export interface Contacts {\n- phone: string;\n- email: string;\n- skype: string;\n- telegram: string;\n- notes: string;\n+ phone?: string;\n+ email?: string;\n+ skype?: string;\n+ telegram?: string;\n+ notes?: string;\n }\n \n export interface MentorStats {\n courseName: string;\n locationName: string;\n courseFullName: string;\n- students: {\n+ students?: {\n githubId: string;\n name: string;\n isExpelled: boolean;\n@@ -102,14 +101,14 @@ export interface StageInterviewDetailedFeedback {\n \n export interface UserInfo {\n generalInfo: GeneralInfo;\n- contacts: Contacts;\n+ contacts?: Contacts;\n };\n \n export interface ProfileInfo {\n generalInfo?: GeneralInfo;\n contacts?: Contacts;\n- mentorStats: MentorStats[];\n- studentStats: StudentStats[];\n- publicFeedback: PublicFeedback[];\n+ mentorStats?: MentorStats[];\n+ studentStats?: StudentStats[];\n+ publicFeedback?: PublicFeedback[];\n stageInterviewFeedback: StageInterviewDetailedFeedback[];\n };\ndiff --git a/server/package.json b/server/package.json\nindex 1bd6de1..bf2d5f0 100755\n--- a/server/package.json\n+++ b/server/package.json\n@@ -4,7 +4,7 @@\n \"private\": true,\n \"scripts\": {\n \"build\": \"tsc\",\n- \"start\": \"nodemon --inspect --watch 'src/**/*' -e ts --exec node -r ts-node/register -r dotenv/config ./index.ts | pino-pretty -i time,hostname,pid,host,method,remoteAddress\",\n+ \"start\": \"nodemon --inspect --watch \\\"src/**/*\\\" -e ts --exec node -r ts-node/register -r dotenv/config ./index.ts | pino-pretty -i time,hostname,pid,host,method,remoteAddress\",\n \"lint\": \"tslint -c tslint.json -p tsconfig.json\",\n \"swagger\": \"swagger-jsdoc -d swaggerDef.js -o ./public/swagger.yml ./src/routes/**/*.ts ./src/routes/**.ts\"\n },\ndiff --git a/server/src/models/profilePermissions.ts b/server/src/models/profilePermissions.ts\nindex 1b2a79a..fd06900 100644\n--- a/server/src/models/profilePermissions.ts\n+++ b/server/src/models/profilePermissions.ts\n@@ -1,20 +1,20 @@\n import { Entity, Column, CreateDateColumn, UpdateDateColumn, PrimaryGeneratedColumn, OneToOne } from 'typeorm';\n import { User } from './user';\n \n-interface PublicVisibilitySettings {\n+export interface PublicVisibilitySettings {\n all: boolean;\n }\n \n-interface VisibilitySettings extends PublicVisibilitySettings {\n+export interface VisibilitySettings extends PublicVisibilitySettings {\n mentor: boolean;\n student: boolean;\n }\n \n-const defaultPublicVisibilitySettings = {\n+export const defaultPublicVisibilitySettings = {\n all: false,\n };\n \n-const defaultVisibilitySettings = {\n+export const defaultVisibilitySettings = {\n mentor: false,\n student: false,\n all: false,\ndiff --git a/server/src/routes/profile/info.ts b/server/src/routes/profile/info.ts\nindex f5d249d..22a8132 100644\n--- a/server/src/routes/profile/info.ts\n+++ b/server/src/routes/profile/info.ts\n@@ -1,4 +1,4 @@\n-import { NOT_FOUND, OK } from 'http-status-codes';\n+import { NOT_FOUND, OK, FORBIDDEN } from 'http-status-codes';\n import Router from 'koa-router';\n import { ILogger } from '../../logger';\n import { setResponse } from '../utils';\n@@ -9,7 +9,7 @@ import { getPublicFeedback } from './public-feedback';\n import { getStageInterviewFeedback } from './stage-interview-feedback';\n import { getStudentStats } from './student-stats';\n import { getUserInfo } from './user-info';\n-import { getPermissions } from './permissions';\n+import { getPermissions, getOwnerPermissions } from './permissions';\n \n /*\n WHO CAN SEE\n@@ -60,13 +60,9 @@ import { getPermissions } from './permissions';\n */\n \n export const getProfileInfo = (_: ILogger) => async (ctx: Router.RouterContext) => {\n- const {\n- // id: userId,\n- githubId: userGithubId,\n- } = ctx.state!.user as IUserSession;\n+ const { githubId: userGithubId } = ctx.state!.user as IUserSession;\n // const { isAdmin, roles } = ctx.state!.user as IUserSession;\n- const { githubId } = ctx.query as { githubId: string | undefined };\n-\n+ const { githubId = userGithubId } = ctx.query as { githubId: string | undefined };\n // console.log('GITHUB =>', githubId);\n // console.log('ADMIN =>', isAdmin);\n // console.log('ROLES =>', roles);\n@@ -75,16 +71,28 @@ export const getProfileInfo = (_: ILogger) => async (ctx: Router.RouterContext) \n return setResponse(ctx, NOT_FOUND);\n }\n \n+ const isProfileOwner = githubId === userGithubId;\n+ console.log('isProfileOwner', isProfileOwner);\n // await getRepository(ProfilePermissions).save({ userId });\n \n- const permissions = await getPermissions(userGithubId, githubId);\n+ const permissions = await getPermissions(userGithubId, githubId, { isProfileOwner });\n \n- console.log(JSON.stringify(permissions, null, 2));\n+ const { isProfileVisible, isPublicFeedbackVisible, isMentorStatsVisible, isStudentStatsVisible } = permissions;\n+\n+ if (!isProfileVisible && !isProfileOwner) {\n+ return setResponse(ctx, FORBIDDEN);\n+ }\n+\n+ if (isProfileOwner) {\n+ const ownerPermissions = await getOwnerPermissions(userGithubId);\n+\n+ console.log('OWN =>', ownerPermissions);\n+ }\n \n const { generalInfo, contacts } = await getUserInfo(githubId, permissions);\n- const publicFeedback = await getPublicFeedback(githubId);\n- const mentorStats = await getMentorStats(githubId);\n- const studentStats = await getStudentStats(githubId);\n+ const publicFeedback = isPublicFeedbackVisible ? await getPublicFeedback(githubId) : undefined;\n+ const mentorStats = isMentorStatsVisible ? await getMentorStats(githubId) : undefined;\n+ const studentStats = isStudentStatsVisible ? await getStudentStats(githubId) : undefined;\n const stageInterviewFeedback = await getStageInterviewFeedback(githubId);\n \n const profileInfo: ProfileInfo = {\n@@ -96,7 +104,8 @@ export const getProfileInfo = (_: ILogger) => async (ctx: Router.RouterContext) \n studentStats,\n };\n \n- // console.log(JSON.stringify(profileInfo, null, 2));\n+ console.log(JSON.stringify(permissions, null, 2));\n+ console.log(JSON.stringify(profileInfo, null, 2));\n \n setResponse(ctx, OK, profileInfo);\n };\ndiff --git a/server/src/routes/profile/mentor-stats.ts b/server/src/routes/profile/mentor-stats.ts\nindex 843a2f7..72e6b30 100644\n--- a/server/src/routes/profile/mentor-stats.ts\n+++ b/server/src/routes/profile/mentor-stats.ts\n@@ -36,11 +36,11 @@ export const getMentorStats = async (githubId: string): Promise<MentorStats[]> =\n studentIsExpelledStatuses,\n studentTotalScores,\n }: any) => {\n- const students = studentGithubIds.map((githubId: string, idx: number) => ({\n+ const students = studentGithubIds[0] ? studentGithubIds.map((githubId: string, idx: number) => ({\n githubId,\n name: getFullName(studentFirstNames[idx], studentLastNames[idx], githubId),\n isExpelled: studentIsExpelledStatuses[idx],\n totalScore: studentTotalScores[idx],\n- }));\n+ })) : undefined;\n return { courseName, locationName, courseFullName, students };\n });\ndiff --git a/server/src/routes/profile/permissions.ts b/server/src/routes/profile/permissions.ts\nindex 61924a8..b40121c 100644\n--- a/server/src/routes/profile/permissions.ts\n+++ b/server/src/routes/profile/permissions.ts\n@@ -1,3 +1,4 @@\n+import { get, mapValues } from 'lodash';\n import { getRepository } from 'typeorm';\n import {\n User,\n@@ -8,6 +9,12 @@ import {\n TaskInterviewResult,\n StageInterview,\n } from '../../models';\n+import {\n+ PublicVisibilitySettings,\n+ VisibilitySettings,\n+ defaultPublicVisibilitySettings,\n+ defaultVisibilitySettings,\n+} from '../../models/profilePermissions';\n \n interface Relations {\n student: string;\n@@ -19,7 +26,43 @@ interface Relations {\n \n type RelationRole = 'student' | 'mentor' | 'all';\n \n-const getAllProfilePermissions = async (githubId: string): Promise<any> => (\n+interface SuperAccessRights {\n+ isProfileOwner: boolean;\n+}\n+\n+interface ConfigurableProfilePermissions {\n+ isProfileVisible: PublicVisibilitySettings;\n+ isAboutVisible: VisibilitySettings;\n+ isEducationVisible: VisibilitySettings;\n+ isEnglishVisible: VisibilitySettings;\n+ isEmailVisible: VisibilitySettings;\n+ isTelegramVisible: VisibilitySettings;\n+ isSkypeVisible: VisibilitySettings;\n+ isPhoneVisible: VisibilitySettings;\n+ isContactsNotesVisible: VisibilitySettings;\n+ isLinkedInVisible: VisibilitySettings;\n+ isPublicFeedbackVisible: VisibilitySettings;\n+ isMentorStatsVisible: VisibilitySettings;\n+ isStudentStatsVisible: VisibilitySettings;\n+}\n+\n+export interface Permissions {\n+ isProfileVisible: boolean;\n+ isAboutVisible: boolean;\n+ isEducationVisible: boolean;\n+ isEnglishVisible: boolean;\n+ isEmailVisible: boolean;\n+ isTelegramVisible: boolean;\n+ isSkypeVisible: boolean;\n+ isPhoneVisible: boolean;\n+ isContactsNotesVisible: boolean;\n+ isLinkedInVisible: boolean;\n+ isPublicFeedbackVisible: boolean;\n+ isMentorStatsVisible: boolean;\n+ isStudentStatsVisible: boolean;\n+}\n+\n+const getConfigurableProfilePermissions = async (githubId: string): Promise<ConfigurableProfilePermissions> => (\n (await getRepository(ProfilePermissions)\n .createQueryBuilder('pp')\n .select('\"pp\".\"isProfileVisible\" AS \"isProfileVisible\"')\n@@ -85,16 +128,67 @@ const getRelationRole = async (userGithubId: string, requestedGithubId: string):\n return 'all';\n };\n \n-const matchPermissions = (permissions: any, role: RelationRole) => {\n- const obj: any = {};\n- Object.keys(permissions).forEach((key) => {\n- obj[key] = permissions[key].all || permissions[key][role];\n- });\n- return obj;\n+const matchPermissions = (\n+ permissions: ConfigurableProfilePermissions,\n+ role: RelationRole,\n+ { isProfileOwner }: SuperAccessRights,\n+): Permissions => {\n+ const p: Permissions = {\n+ isProfileVisible: false,\n+ isAboutVisible: false,\n+ isEducationVisible: false,\n+ isEnglishVisible: false,\n+ isEmailVisible: false,\n+ isTelegramVisible: false,\n+ isSkypeVisible: false,\n+ isPhoneVisible: false,\n+ isContactsNotesVisible: false,\n+ isLinkedInVisible: false,\n+ isPublicFeedbackVisible: false,\n+ isMentorStatsVisible: false,\n+ isStudentStatsVisible: false,\n+ };\n+\n+ // (Object.keys(p) as (keyof Permissions)[]).forEach((key) => {\n+ // p[key] = isProfileOwner || permissions[key].all || permissions[key][role];\n+ // });\n+\n+ // return p;\n+\n+ return mapValues(p, (_, key) => isProfileOwner ||\n+ get(permissions, `${key}.all`) ||\n+ get(permissions, `${key}.${role}`) ||\n+ false,\n+ );\n };\n \n-export const getPermissions = async (userGithubId: string, requestedGithubId: string) => {\n- const permissions = await getAllProfilePermissions(requestedGithubId);\n+export const getPermissions = async (\n+ userGithubId: string,\n+ requestedGithubId: string,\n+ superAccessRights: SuperAccessRights,\n+) => {\n+ const permissions = await getConfigurableProfilePermissions(requestedGithubId);\n const role = await getRelationRole(userGithubId, requestedGithubId);\n- return matchPermissions(permissions, role);\n+ return matchPermissions(permissions, role, superAccessRights);\n+};\n+\n+export const getOwnerPermissions = async (githubId: string) => {\n+ const permissions = await getConfigurableProfilePermissions(githubId);\n+ const p: ConfigurableProfilePermissions = {\n+ isProfileVisible: defaultPublicVisibilitySettings,\n+ isAboutVisible: defaultVisibilitySettings,\n+ isEducationVisible: defaultVisibilitySettings,\n+ isEnglishVisible: defaultVisibilitySettings,\n+ isEmailVisible: defaultVisibilitySettings,\n+ isTelegramVisible: defaultVisibilitySettings,\n+ isSkypeVisible: defaultVisibilitySettings,\n+ isPhoneVisible: defaultVisibilitySettings,\n+ isContactsNotesVisible: defaultVisibilitySettings,\n+ isLinkedInVisible: defaultVisibilitySettings,\n+ isPublicFeedbackVisible: defaultVisibilitySettings,\n+ isMentorStatsVisible: defaultVisibilitySettings,\n+ isStudentStatsVisible: defaultVisibilitySettings,\n+ };\n+\n+ return mapValues(p, (value, key) => get(permissions, key, value));\n };\ndiff --git a/server/src/routes/profile/user-info.ts b/server/src/routes/profile/user-info.ts\nindex 5b871e0..1998ed0 100644\n--- a/server/src/routes/profile/user-info.ts\n+++ b/server/src/routes/profile/user-info.ts\n@@ -2,23 +2,53 @@ import { getRepository } from 'typeorm';\n import { UserInfo } from '../../../../common/models/profile';\n import { getFullName } from '../../lib/utils';\n import { User } from '../../models';\n+import { Permissions } from './permissions';\n \n-export const getUserInfo = async (githubId: string, permissions: any): Promise<UserInfo> => {\n- const { isAboutVisible } = permissions;\n+export const getUserInfo = async (githubId: string, permissions: Permissions): Promise<UserInfo> => {\n+ const {\n+ isAboutVisible,\n+ isEducationVisible,\n+ isEnglishVisible,\n+ isPhoneVisible,\n+ isEmailVisible,\n+ isTelegramVisible,\n+ isSkypeVisible,\n+ isContactsNotesVisible,\n+ } = permissions;\n \n const query = await getRepository(User)\n .createQueryBuilder('user')\n .select('\"user\".\"firstName\" AS \"firstName\", \"user\".\"lastName\" AS \"lastName\"')\n .addSelect('\"user\".\"githubId\" AS \"githubId\"')\n- .addSelect('\"user\".\"locationName\" AS \"locationName\"')\n- .addSelect('\"user\".\"educationHistory\" AS \"educationHistory\"')\n- .addSelect('\"user\".\"employmentHistory\" AS \"employmentHistory\"')\n- .addSelect('\"user\".\"englishLevel\" AS \"englishLevel\"')\n- .addSelect('\"user\".\"contactsPhone\" AS \"contactsPhone\"')\n- .addSelect('\"user\".\"contactsEmail\" AS \"contactsEmail\"')\n- .addSelect('\"user\".\"contactsTelegram\" AS \"contactsTelegram\"')\n- .addSelect('\"user\".\"contactsSkype\" AS \"contactsSkype\"')\n- .addSelect('\"user\".\"contactsNotes\" AS \"contactsNotes\"');\n+ .addSelect('\"user\".\"locationName\" AS \"locationName\"');\n+\n+ if (isEducationVisible) {\n+ query.addSelect('\"user\".\"educationHistory\" AS \"educationHistory\"');\n+ }\n+\n+ if (isEnglishVisible) {\n+ query.addSelect('\"user\".\"englishLevel\" AS \"englishLevel\"');\n+ }\n+\n+ if (isPhoneVisible) {\n+ query.addSelect('\"user\".\"contactsPhone\" AS \"contactsPhone\"');\n+ }\n+\n+ if (isEmailVisible) {\n+ query.addSelect('\"user\".\"contactsEmail\" AS \"contactsEmail\"');\n+ }\n+\n+ if (isTelegramVisible) {\n+ query.addSelect('\"user\".\"contactsTelegram\" AS \"contactsTelegram\"');\n+ }\n+\n+ if (isSkypeVisible) {\n+ query.addSelect('\"user\".\"contactsSkype\" AS \"contactsSkype\"');\n+ }\n+\n+ if (isContactsNotesVisible) {\n+ query.addSelect('\"user\".\"contactsNotes\" AS \"contactsNotes\"');\n+ }\n \n if (isAboutVisible) {\n query.addSelect('\"user\".\"aboutMyself\" AS \"aboutMyself\"');\n@@ -33,7 +63,6 @@ export const getUserInfo = async (githubId: string, permissions: any): Promise<U\n lastName,\n locationName,\n educationHistory,\n- employmentHistory,\n englishLevel,\n contactsPhone,\n contactsEmail,\n@@ -49,16 +78,15 @@ export const getUserInfo = async (githubId: string, permissions: any): Promise<U\n aboutMyself,\n locationName,\n educationHistory,\n- employmentHistory,\n englishLevel,\n name: getFullName(firstName, lastName, githubId),\n },\n- contacts: {\n+ contacts: contactsPhone || contactsEmail || contactsSkype || contactsTelegram || contactsNotes ? {\n phone: contactsPhone,\n email: contactsEmail,\n skype: contactsSkype,\n telegram: contactsTelegram,\n notes: contactsNotes,\n- },\n+ } : undefined,\n };\n };\n"]
2
["7e04a5e829d7416e312ac342a00a11787745753b", "1f15f71e415ba49b21684c7a3a51c8e3faaa7cf3"]
["build", "feat"]
import flux-lsp v0.5.21,run nix macos jobs on macos-13 to try and avoid SIP
["diff --git a/ui/package.json b/ui/package.json\nindex 7a44aad..a36fc3d 100644\n--- a/ui/package.json\n+++ b/ui/package.json\n@@ -134,7 +134,7 @@\n \"dependencies\": {\n \"@influxdata/clockface\": \"2.3.4\",\n \"@influxdata/flux\": \"^0.5.1\",\n- \"@influxdata/flux-lsp-browser\": \"0.5.20\",\n+ \"@influxdata/flux-lsp-browser\": \"0.5.21\",\n \"@influxdata/giraffe\": \"0.29.0\",\n \"@influxdata/influx\": \"0.5.5\",\n \"@influxdata/influxdb-templates\": \"0.9.0\",\ndiff --git a/ui/yarn.lock b/ui/yarn.lock\nindex 99ae766..e6e2a47 100644\n--- a/ui/yarn.lock\n+++ b/ui/yarn.lock\n@@ -752,10 +752,10 @@\n resolved \"https://registry.yarnpkg.com/@influxdata/clockface/-/clockface-2.3.4.tgz#9c496601253e1d49cbeae29a7b9cfb54862785f6\"\n integrity sha512-mmz3YElK8Ho+1onEafuas6sVhIT638JA4NbDTO3bVJgK1TG7AnU4rQP+c6fj7vZSfvrIwtOwGaMONJTaww5o6w==\n \n-\"@influxdata/[email protected]\":\n- version \"0.5.20\"\n- resolved \"https://registry.yarnpkg.com/@influxdata/flux-lsp-browser/-/flux-lsp-browser-0.5.20.tgz#150d261bab869e130f6d00ee73ea4e859e8969e4\"\n- integrity sha512-gUy19t/QndkJPmyv7Lb56zXxaW5v7R9TslTHt0hB0GJjo7lmYkRfkD7DELdFHrD2e/CLtcNQBnczIMIGkII8Bw==\n+\"@influxdata/[email protected]\":\n+ version \"0.5.21\"\n+ resolved \"https://registry.yarnpkg.com/@influxdata/flux-lsp-browser/-/flux-lsp-browser-0.5.21.tgz#d5632f45e925c09bae9501a00fbef2ed55567f9e\"\n+ integrity sha512-lcUwKX1yj0QqGiusQFOVi7UPsvp6+qNX7Cwf9qqS5/dRwoh7c++nFVRdGNrSWlsbyRrPaAWBoZWEnghSnIf6DQ==\n \n \"@influxdata/flux@^0.5.1\":\n version \"0.5.1\"\n", "diff --git a/.github/actionlint.yaml b/.github/actionlint.yaml\nnew file mode 100644\nindex 0000000..5be7d17\n--- /dev/null\n+++ b/.github/actionlint.yaml\n@@ -0,0 +1,7 @@\n+self-hosted-runner:\n+ # Labels of self-hosted runner in array of strings.\n+ labels: [macos-13]\n+# Configuration variables in array of strings defined in your repository or\n+# organization. `null` means disabling configuration variables check.\n+# Empty array means no configuration variable is allowed.\n+config-variables: null\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\nindex e37346c..dce77e1 100644\n--- a/.github/workflows/nix.yml\n+++ b/.github/workflows/nix.yml\n@@ -37,7 +37,7 @@ jobs:\n - \"3.10\"\n - \"3.11\"\n include:\n- - os: macos-latest\n+ - os: macos-13\n python-version: \"3.10\"\n steps:\n - name: checkout\ndiff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml\nindex 005a850..8db22e2 100644\n--- a/.pre-commit-config.yaml\n+++ b/.pre-commit-config.yaml\n@@ -3,7 +3,7 @@ ci:\n autofix_prs: false\n autoupdate_commit_msg: \"chore(deps): pre-commit.ci autoupdate\"\n skip:\n- - actionlint\n+ - actionlint-system\n - deadnix\n - just\n - nixpkgs-fmt\n@@ -17,9 +17,9 @@ default_stages:\n - commit\n repos:\n - repo: https://github.com/rhysd/actionlint\n- rev: v1.6.24\n+ rev: v1.6.25\n hooks:\n- - id: actionlint\n+ - id: actionlint-system\n - repo: https://github.com/psf/black\n rev: 23.3.0\n hooks:\n@@ -30,7 +30,7 @@ repos:\n - id: nbstripout\n exclude: .+/rendered/.+\n - repo: https://github.com/codespell-project/codespell\n- rev: v2.2.4\n+ rev: v2.2.5\n hooks:\n - id: codespell\n additional_dependencies:\n"]
2
["bfe32bf10e9b6d699f694fbd095af0b3f2e6275f", "54cb6d4643b4a072ff997592a7fa14a69a6c068d"]
["build", "ci"]
generate terminate end event compatible execution steps part 1 The random execution tests don't know the concept of flow scopes. This makes it challenging to generate a correct execution path for terminate end events, as they terminate a specific flow scope. Processing should continue as normal once the flow scope has been terminated. Whilst we don't have flow scopes, we do have execution path segments. These segments don't map 1 to 1 to flow scopes. However, since every flow scope starts a new segment we can use these segments to get the desired behavior. Each segment must keep track whether is has reached a terminate end event. If this is the case that means that we don't expect any further execution steps. We can isolate this behavior in a single location, during the appending of one segment to another segment. In order to differentiate between flow scopes a new append method has been added which takes the boolean `changesFlowScope` as a parameter. Blockbuilder where the flow scope changes (e.g. SubProcessBlockBuilder) can use this to indicate that even though a terminate end event has been reached. Execution steps after this specific segment still need to added to complete the process. When a segment is appended to a different segment and the flow scope does not change we can use the segment that should be appended to identify whether new segment can still be added to the current segment. If passed segment has reached a terminate end event and the flow scope has not been changed it is guaranteed that the current segment is in the same flow scope has the previous segment and thus has also reached the terminate end event.,remove unnecessary start argument from `range`
["diff --git a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\nindex da33c23..23c43be 100644\n--- a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\n+++ b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\n@@ -29,6 +29,10 @@ import org.apache.commons.lang3.builder.ToStringStyle;\n */\n public final class ExecutionPathSegment {\n \n+ // If we have reached a terminate end event we want to stop generating execution steps for a\n+ // specific flow scope. By setting this flag to true no new execution steps will be added for the\n+ // flow scope this segment is in.\n+ private boolean reachedTerminateEndEvent = false;\n private final List<ScheduledExecutionStep> scheduledSteps = new ArrayList<>();\n private final Map<String, Object> variableDefaults = new HashMap<>();\n \n@@ -87,10 +91,28 @@ public final class ExecutionPathSegment {\n new ScheduledExecutionStep(logicalPredecessor, executionPredecessor, executionStep));\n }\n \n+ /**\n+ * Appends the steps of the passed execution path segment to the current segment.\n+ *\n+ * @param pathToAdd execution path segment to append to this segment\n+ */\n public void append(final ExecutionPathSegment pathToAdd) {\n+ append(pathToAdd, false);\n+ }\n+\n+ /**\n+ * Appends the step of the passed execution path segment to the current segment if the current\n+ *\n+ * @param pathToAdd\n+ * @param changesFlowScope\n+ */\n+ public void append(final ExecutionPathSegment pathToAdd, final boolean changesFlowScope) {\n mergeVariableDefaults(pathToAdd);\n \n- pathToAdd.getScheduledSteps().forEach(this::append);\n+ if (!hasReachedTerminateEndEvent() || changesFlowScope) {\n+ pathToAdd.getScheduledSteps().forEach(this::append);\n+ }\n+ reachedTerminateEndEvent = pathToAdd.hasReachedTerminateEndEvent() && !changesFlowScope;\n }\n \n public void append(final ScheduledExecutionStep scheduledExecutionStep) {\n@@ -259,6 +281,14 @@ public final class ExecutionPathSegment {\n return ToStringBuilder.reflectionToString(this, ToStringStyle.SHORT_PREFIX_STYLE);\n }\n \n+ public boolean hasReachedTerminateEndEvent() {\n+ return reachedTerminateEndEvent;\n+ }\n+\n+ public void setReachedTerminateEndEvent(final boolean reachedTerminateEndEvent) {\n+ this.reachedTerminateEndEvent = reachedTerminateEndEvent;\n+ }\n+\n /**\n * An execution boundary is the point where automatic and non-automatic {@link\n * ScheduledExecutionStep}'s meet each other. This class contains information about the existing\n", "diff --git a/ibis/backends/dask/tests/execution/test_window.py b/ibis/backends/dask/tests/execution/test_window.py\nindex 75a7331..6bfc5e3 100644\n--- a/ibis/backends/dask/tests/execution/test_window.py\n+++ b/ibis/backends/dask/tests/execution/test_window.py\n@@ -489,7 +489,7 @@ def test_project_list_scalar(npartitions):\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\n result = expr.execute()\n \n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\"res\")\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\"res\")\n tm.assert_series_equal(result.res, expected)\n \n \ndiff --git a/ibis/backends/pandas/tests/execution/test_window.py b/ibis/backends/pandas/tests/execution/test_window.py\nindex 8f292b3..effa372 100644\n--- a/ibis/backends/pandas/tests/execution/test_window.py\n+++ b/ibis/backends/pandas/tests/execution/test_window.py\n@@ -436,7 +436,7 @@ def test_project_list_scalar():\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\n result = expr.execute()\n \n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\"res\")\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\"res\")\n tm.assert_series_equal(result.res, expected)\n \n \ndiff --git a/ibis/backends/pyspark/tests/test_basic.py b/ibis/backends/pyspark/tests/test_basic.py\nindex 3850919..14fe677 100644\n--- a/ibis/backends/pyspark/tests/test_basic.py\n+++ b/ibis/backends/pyspark/tests/test_basic.py\n@@ -19,7 +19,7 @@ from ibis.backends.pyspark.compiler import _can_be_replaced_by_column_name # no\n def test_basic(con):\n table = con.table(\"basic_table\")\n result = table.compile().toPandas()\n- expected = pd.DataFrame({\"id\": range(0, 10), \"str_col\": \"value\"})\n+ expected = pd.DataFrame({\"id\": range(10), \"str_col\": \"value\"})\n \n tm.assert_frame_equal(result, expected)\n \n@@ -28,9 +28,7 @@ def test_projection(con):\n table = con.table(\"basic_table\")\n result1 = table.mutate(v=table[\"id\"]).compile().toPandas()\n \n- expected1 = pd.DataFrame(\n- {\"id\": range(0, 10), \"str_col\": \"value\", \"v\": range(0, 10)}\n- )\n+ expected1 = pd.DataFrame({\"id\": range(10), \"str_col\": \"value\", \"v\": range(10)})\n \n result2 = (\n table.mutate(v=table[\"id\"])\n@@ -44,8 +42,8 @@ def test_projection(con):\n {\n \"id\": range(0, 20, 2),\n \"str_col\": \"value\",\n- \"v\": range(0, 10),\n- \"v2\": range(0, 10),\n+ \"v\": range(10),\n+ \"v2\": range(10),\n }\n )\n \n"]
2
["40597fb4de41c7194eb99479a914db70da7909ea", "15f8d95754a0b6865ea475ca9e515272a07bf6ba"]
["feat", "refactor"]
Template using kube api version Signed-off-by: rjshrjndrn <[email protected]>,verify process can start at supported element types Verifies a PI can be started at specific element types. The test will deploy the process, start an instance at the desired start element and verify that it has been activated succesfully.
["diff --git a/.github/workflows/api-ee.yaml b/.github/workflows/api-ee.yaml\nindex c014f34..2a12e0d 100644\n--- a/.github/workflows/api-ee.yaml\n+++ b/.github/workflows/api-ee.yaml\n@@ -8,7 +8,7 @@ on:\n default: 'false'\n push:\n branches:\n- - dev\n+ - test_ci\n paths:\n - ee/api/**\n - api/**\n@@ -112,7 +112,8 @@ jobs:\n # Deploy command\n kubectl config set-context --namespace=app --current\n kubectl config get-contexts\n- helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks | kubectl apply -f -\n+ k_version=$(kubectl version --short 2>/dev/null | awk '/Server/{print $NF}')\n+ helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks --kube-version=$k_version | kubectl apply -f -\n env:\n DOCKER_REPO: ${{ secrets.EE_REGISTRY_URL }}\n # We're not passing -ee flag, because helm will add that.\n", "diff --git a/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\nnew file mode 100644\nindex 0000000..a505307\n--- /dev/null\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceSupportedElementTest.java\n@@ -0,0 +1,233 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.engine.processing.processinstance;\n+\n+import static org.assertj.core.api.Assertions.assertThat;\n+import static org.assertj.core.groups.Tuple.tuple;\n+\n+import io.camunda.zeebe.engine.util.EngineRule;\n+import io.camunda.zeebe.model.bpmn.Bpmn;\n+import io.camunda.zeebe.model.bpmn.BpmnModelInstance;\n+import io.camunda.zeebe.protocol.record.Record;\n+import io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent;\n+import io.camunda.zeebe.protocol.record.value.BpmnElementType;\n+import io.camunda.zeebe.test.util.record.RecordingExporter;\n+import io.camunda.zeebe.test.util.record.RecordingExporterTestWatcher;\n+import java.util.Collection;\n+import java.util.Collections;\n+import java.util.List;\n+import java.util.Map;\n+import org.junit.ClassRule;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.runner.RunWith;\n+import org.junit.runners.Parameterized;\n+import org.junit.runners.Parameterized.Parameters;\n+\n+@RunWith(Parameterized.class)\n+public class CreateProcessInstanceSupportedElementTest {\n+\n+ @ClassRule public static final EngineRule ENGINE = EngineRule.singlePartition();\n+ private static final String PROCESS_ID = \"processId\";\n+ private static final String CHILD_PROCESS_ID = \"childProcessId\";\n+ private static final String START_ELEMENT_ID = \"startElement\";\n+ private static final String MESSAGE = \"message\";\n+ private static final String JOBTYPE = \"jobtype\";\n+\n+ @Rule\n+ public final RecordingExporterTestWatcher recordingExporterTestWatcher =\n+ new RecordingExporterTestWatcher();\n+\n+ private final Scenario scenario;\n+\n+ public CreateProcessInstanceSupportedElementTest(final Scenario scenario) {\n+ this.scenario = scenario;\n+ }\n+\n+ @Parameters(name = \"{0}\")\n+ public static Collection<Object> scenarios() {\n+ return List.of(\n+ new Scenario(\n+ BpmnElementType.SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .subProcess(START_ELEMENT_ID)\n+ .embeddedSubProcess()\n+ .startEvent()\n+ .subProcessDone()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_SUB_PROCESS,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .eventSubProcess(\n+ START_ELEMENT_ID, e -> e.startEvent().timerWithDuration(\"PT1H\").endEvent())\n+ .startEvent()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_CATCH_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateCatchEvent(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.INTERMEDIATE_THROW_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateThrowEvent(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.END_EVENT,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().endEvent(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SERVICE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.RECEIVE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .receiveTask(START_ELEMENT_ID)\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.USER_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID).startEvent().userTask(START_ELEMENT_ID).done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.MANUAL_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .manualTask(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EXCLUSIVE_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .exclusiveGateway(START_ELEMENT_ID)\n+ .defaultFlow()\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.PARALLEL_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .parallelGateway(START_ELEMENT_ID)\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.EVENT_BASED_GATEWAY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .eventBasedGateway(START_ELEMENT_ID)\n+ .intermediateCatchEvent()\n+ .message(b -> b.name(MESSAGE).zeebeCorrelationKeyExpression(\"correlationKey\"))\n+ .moveToLastGateway()\n+ .intermediateCatchEvent()\n+ .timerWithDuration(\"PT1H\")\n+ .done(),\n+ Map.of(\"correlationKey\", \"value\")),\n+ new Scenario(\n+ BpmnElementType.MULTI_INSTANCE_BODY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .serviceTask(\n+ START_ELEMENT_ID,\n+ t ->\n+ t.zeebeJobType(JOBTYPE)\n+ .multiInstance(m -> m.parallel().zeebeInputCollectionExpression(\"[1]\")))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.CALL_ACTIVITY,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .callActivity(START_ELEMENT_ID, c -> c.zeebeProcessId(CHILD_PROCESS_ID))\n+ .endEvent()\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.BUSINESS_RULE_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .businessRuleTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SCRIPT_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .scriptTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()),\n+ new Scenario(\n+ BpmnElementType.SEND_TASK,\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .sendTask(START_ELEMENT_ID, b -> b.zeebeJobType(JOBTYPE))\n+ .done(),\n+ Collections.emptyMap()));\n+ }\n+\n+ @Test\n+ public void testProcessInstanceCanStartAtElementType() {\n+ // given\n+ ENGINE.deployment().withXmlResource(scenario.modelInstance).deploy();\n+ if (scenario.type == BpmnElementType.CALL_ACTIVITY) {\n+ ENGINE.deployment().withXmlResource(getChildProcess()).deploy();\n+ }\n+\n+ // when\n+ final long instanceKey =\n+ ENGINE\n+ .processInstance()\n+ .ofBpmnProcessId(PROCESS_ID)\n+ .withStartInstruction(START_ELEMENT_ID)\n+ .withVariables(scenario.variables)\n+ .create();\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.processInstanceRecords()\n+ .withProcessInstanceKey(instanceKey)\n+ .onlyEvents()\n+ .limit(\n+ r ->\n+ r.getValue().getBpmnElementType() == scenario.type\n+ && r.getIntent() == ProcessInstanceIntent.ELEMENT_ACTIVATED))\n+ .extracting(record -> record.getValue().getBpmnElementType(), Record::getIntent)\n+ .containsSequence(\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(BpmnElementType.PROCESS, ProcessInstanceIntent.ELEMENT_ACTIVATED),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATING),\n+ tuple(scenario.type, ProcessInstanceIntent.ELEMENT_ACTIVATED));\n+ }\n+\n+ private BpmnModelInstance getChildProcess() {\n+ return Bpmn.createExecutableProcess(CHILD_PROCESS_ID).startEvent().endEvent().done();\n+ }\n+\n+ record Scenario(\n+ BpmnElementType type, BpmnModelInstance modelInstance, Map<String, Object> variables) {}\n+}\n"]
2
["c3531347fe5a4cc82d426db195026a5bdad15e7a", "a5ecfdf49b0d4c43fbbbf7947be7c0327ccb3415"]
["ci", "test"]
add ability to get all encoded values,101: fix import key cmd Signed-off-by: Sam Alba <[email protected]>
["diff --git a/delorean_mem_qe/src/column.rs b/delorean_mem_qe/src/column.rs\nindex bc89cb2..b3df18e 100644\n--- a/delorean_mem_qe/src/column.rs\n+++ b/delorean_mem_qe/src/column.rs\n@@ -537,6 +537,22 @@ impl Column {\n }\n }\n \n+ /// Materialise all of the encoded values.\n+ pub fn all_encoded_values(&self) -> Vector {\n+ match self {\n+ Column::String(c) => {\n+ let now = std::time::Instant::now();\n+ let v = c.all_encoded_values();\n+ log::debug!(\"time getting all encoded values {:?}\", now.elapsed());\n+\n+ log::debug!(\"dictionary {:?}\", c.data.dictionary());\n+ Vector::Integer(v)\n+ }\n+ Column::Float(c) => Vector::Float(c.all_encoded_values()),\n+ Column::Integer(c) => Vector::Integer(c.all_encoded_values()),\n+ }\n+ }\n+\n /// Given an encoded value for a row, materialise and return the decoded\n /// version.\n ///\n@@ -986,6 +1002,10 @@ impl String {\n self.data.encoded_values(row_ids)\n }\n \n+ pub fn all_encoded_values(&self) -> Vec<i64> {\n+ self.data.all_encoded_values()\n+ }\n+\n /// Return the decoded value for an encoded ID.\n ///\n /// Panics if there is no decoded value for the provided id\n@@ -1037,6 +1057,10 @@ impl Float {\n self.data.encoded_values(row_ids)\n }\n \n+ pub fn all_encoded_values(&self) -> Vec<f64> {\n+ self.data.all_encoded_values()\n+ }\n+\n pub fn scan_from(&self, row_id: usize) -> &[f64] {\n self.data.scan_from(row_id)\n }\n@@ -1106,6 +1130,10 @@ impl Integer {\n self.data.encoded_values(row_ids)\n }\n \n+ pub fn all_encoded_values(&self) -> Vec<i64> {\n+ self.data.all_encoded_values()\n+ }\n+\n pub fn scan_from(&self, row_id: usize) -> &[i64] {\n self.data.scan_from(row_id)\n }\ndiff --git a/delorean_mem_qe/src/encoding.rs b/delorean_mem_qe/src/encoding.rs\nindex d6a865a..4b057cf 100644\n--- a/delorean_mem_qe/src/encoding.rs\n+++ b/delorean_mem_qe/src/encoding.rs\n@@ -68,6 +68,12 @@ where\n self.values(row_ids)\n }\n \n+ /// Return all encoded values. For this encoding this is just the decoded\n+ /// values\n+ pub fn all_encoded_values(&self) -> Vec<T> {\n+ self.values.clone()\n+ }\n+\n // TODO(edd): fix this when added NULL support\n pub fn scan_from_until_some(&self, _row_id: usize) -> Option<T> {\n unreachable!(\"to remove\");\n@@ -485,6 +491,26 @@ impl DictionaryRLE {\n out\n }\n \n+ // values materialises a vector of references to all logical values in the\n+ // encoding.\n+ pub fn all_values(&mut self) -> Vec<Option<&String>> {\n+ let mut out: Vec<Option<&String>> = Vec::with_capacity(self.total as usize);\n+\n+ // build reverse mapping.\n+ let mut idx_value = BTreeMap::new();\n+ for (k, v) in &self.entry_index {\n+ idx_value.insert(v, k);\n+ }\n+ assert_eq!(idx_value.len(), self.entry_index.len());\n+\n+ for (idx, rl) in &self.run_lengths {\n+ // TODO(edd): fix unwrap - we know that the value exists in map...\n+ let v = idx_value.get(&idx).unwrap().as_ref();\n+ out.extend(iter::repeat(v).take(*rl as usize));\n+ }\n+ out\n+ }\n+\n /// Return the decoded value for an encoded ID.\n ///\n /// Panics if there is no decoded value for the provided id\n@@ -528,22 +554,13 @@ impl DictionaryRLE {\n out\n }\n \n- // values materialises a vector of references to all logical values in the\n- // encoding.\n- pub fn all_values(&mut self) -> Vec<Option<&String>> {\n- let mut out: Vec<Option<&String>> = Vec::with_capacity(self.total as usize);\n-\n- // build reverse mapping.\n- let mut idx_value = BTreeMap::new();\n- for (k, v) in &self.entry_index {\n- idx_value.insert(v, k);\n- }\n- assert_eq!(idx_value.len(), self.entry_index.len());\n+ // all_encoded_values materialises a vector of all encoded values for the\n+ // column.\n+ pub fn all_encoded_values(&self) -> Vec<i64> {\n+ let mut out: Vec<i64> = Vec::with_capacity(self.total as usize);\n \n for (idx, rl) in &self.run_lengths {\n- // TODO(edd): fix unwrap - we know that the value exists in map...\n- let v = idx_value.get(&idx).unwrap().as_ref();\n- out.extend(iter::repeat(v).take(*rl as usize));\n+ out.extend(iter::repeat(*idx as i64).take(*rl as usize));\n }\n out\n }\ndiff --git a/delorean_mem_qe/src/segment.rs b/delorean_mem_qe/src/segment.rs\nindex c058df0..f8c5005 100644\n--- a/delorean_mem_qe/src/segment.rs\n+++ b/delorean_mem_qe/src/segment.rs\n@@ -228,7 +228,7 @@ impl Segment {\n group_columns: &[String],\n aggregates: &[(String, AggregateType)],\n window: i64,\n- ) -> BTreeMap<Vec<String>, Vec<(String, Option<column::Aggregate>)>> {\n+ ) -> BTreeMap<Vec<i64>, Vec<(&String, &AggregateType, Option<column::Aggregate>)>> {\n // Build a hash table - essentially, scan columns for matching row ids,\n // emitting the encoded value for each column and track those value\n // combinations in a hashmap with running aggregates.\n@@ -242,6 +242,10 @@ impl Segment {\n assert_ne!(group_columns[group_columns.len() - 1], \"time\");\n }\n \n+ // TODO(edd): Perf - if there is no predicate and we want entire segment\n+ // then it will be a lot faster to not build filtered_row_ids and just\n+ // get all encoded values for each grouping column...\n+\n // filter on predicates and time\n let filtered_row_ids: croaring::Bitmap;\n if let Some(row_ids) = self.filter_by_predicates_eq(time_range, predicates) {\n@@ -263,7 +267,12 @@ impl Segment {\n let mut group_column_encoded_values = Vec::with_capacity(group_columns.len());\n for group_column in group_columns {\n if let Some(column) = self.column(&group_column) {\n- let encoded_values = column.encoded_values(&filtered_row_ids_vec);\n+ let encoded_values = if filtered_row_ids_vec.len() == self.meta.rows {\n+ column.all_encoded_values()\n+ } else {\n+ column.encoded_values(&filtered_row_ids_vec)\n+ };\n+\n assert_eq!(\n filtered_row_ids.cardinality() as usize,\n encoded_values.len()\n@@ -325,10 +334,10 @@ impl Segment {\n .collect::<Vec<_>>();\n \n // hashMap is about 20% faster than BTreeMap in this case\n- let mut hash_table: HashMap<\n+ let mut hash_table: BTreeMap<\n Vec<i64>,\n Vec<(&String, &AggregateType, Option<column::Aggregate>)>,\n- > = HashMap::new();\n+ > = BTreeMap::new();\n \n let mut aggregate_row: Vec<(&str, Option<column::Scalar>)> =\n std::iter::repeat_with(|| (\"\", None))\n@@ -406,8 +415,10 @@ impl Segment {\n }\n processed_rows += 1;\n }\n+ // println!(\"groups: {:?}\", hash_table.len());\n log::debug!(\"({:?} rows processed) {:?}\", processed_rows, hash_table);\n BTreeMap::new()\n+ // hash_table\n }\n \n pub fn aggregate_by_group_using_sort(\n@@ -451,7 +462,11 @@ impl Segment {\n let mut group_column_encoded_values = Vec::with_capacity(group_columns.len());\n for group_column in group_columns {\n if let Some(column) = self.column(&group_column) {\n- let encoded_values = column.encoded_values(&filtered_row_ids_vec);\n+ let encoded_values = if filtered_row_ids_vec.len() == self.meta.rows {\n+ column.all_encoded_values()\n+ } else {\n+ column.encoded_values(&filtered_row_ids_vec)\n+ };\n assert_eq!(\n filtered_row_ids.cardinality() as usize,\n encoded_values.len()\n@@ -557,6 +572,10 @@ impl Segment {\n assert_ne!(group_columns[group_columns.len() - 1], \"time\");\n }\n \n+ // TODO(edd): Perf - if there is no predicate and we want entire segment\n+ // then it will be a lot faster to not build filtered_row_ids and just\n+ // get all encoded values for each grouping column...\n+\n // filter on predicates and time\n let filtered_row_ids: croaring::Bitmap;\n if let Some(row_ids) = self.filter_by_predicates_eq(time_range, predicates) {\n@@ -577,7 +596,11 @@ impl Segment {\n let mut group_column_encoded_values = Vec::with_capacity(group_columns.len());\n for group_column in group_columns {\n if let Some(column) = self.column(&group_column) {\n- let encoded_values = column.encoded_values(&filtered_row_ids_vec);\n+ let encoded_values = if filtered_row_ids_vec.len() == self.meta.rows {\n+ column.all_encoded_values()\n+ } else {\n+ column.encoded_values(&filtered_row_ids_vec)\n+ };\n assert_eq!(\n filtered_row_ids.cardinality() as usize,\n encoded_values.len()\n@@ -709,6 +732,7 @@ impl Segment {\n aggregates: group_key_aggregates,\n });\n \n+ // println!(\"groups: {:?}\", results.len());\n log::debug!(\"({:?} rows processed) {:?}\", processed_rows, results);\n // results\n vec![]\n", "diff --git a/docs/learn/101-use.md b/docs/learn/101-use.md\nindex 283c1c1..2ec10f9 100644\n--- a/docs/learn/101-use.md\n+++ b/docs/learn/101-use.md\n@@ -41,8 +41,7 @@ cd ./examples/todoapp\n The example app contains encrypted secrets and other pre-configured inputs, here is how to decrypt them:\n \n ```sh\n-curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\n-dagger input list\n+dagger input list || curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\n ```\n \n **Step 4**: Deploy!\n"]
2
["cad5e45208346528ad02cd04dcac863f90faa037", "2b01808ec86fe9d8b4a93141a1b7f95e11fd6010"]
["feat", "docs"]
change tests to depend on BrokerContext,remove unnecessary spotless definition It receives this already from the parent pom.
["diff --git a/broker/src/main/java/io/camunda/zeebe/broker/Broker.java b/broker/src/main/java/io/camunda/zeebe/broker/Broker.java\nindex fe4e42d..37c7066 100644\n--- a/broker/src/main/java/io/camunda/zeebe/broker/Broker.java\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/Broker.java\n@@ -7,20 +7,14 @@\n */\n package io.camunda.zeebe.broker;\n \n-import io.atomix.cluster.AtomixCluster;\n import io.camunda.zeebe.broker.bootstrap.BrokerContext;\n import io.camunda.zeebe.broker.bootstrap.BrokerStartupContextImpl;\n import io.camunda.zeebe.broker.bootstrap.BrokerStartupProcess;\n-import io.camunda.zeebe.broker.clustering.ClusterServices;\n import io.camunda.zeebe.broker.exporter.repo.ExporterLoadException;\n import io.camunda.zeebe.broker.exporter.repo.ExporterRepository;\n-import io.camunda.zeebe.broker.partitioning.PartitionManager;\n-import io.camunda.zeebe.broker.system.EmbeddedGatewayService;\n import io.camunda.zeebe.broker.system.SystemContext;\n import io.camunda.zeebe.broker.system.configuration.BrokerCfg;\n-import io.camunda.zeebe.broker.system.management.BrokerAdminService;\n import io.camunda.zeebe.broker.system.monitoring.BrokerHealthCheckService;\n-import io.camunda.zeebe.broker.system.monitoring.DiskSpaceUsageMonitor;\n import io.camunda.zeebe.protocol.impl.encoding.BrokerInfo;\n import io.camunda.zeebe.util.LogUtil;\n import io.camunda.zeebe.util.VersionUtil;\n@@ -184,35 +178,15 @@ public final class Broker implements AutoCloseable {\n }\n \n // only used for tests\n- public EmbeddedGatewayService getEmbeddedGatewayService() {\n- return brokerContext.getEmbeddedGatewayService();\n- }\n-\n- public AtomixCluster getAtomixCluster() {\n- return brokerContext.getAtomixCluster();\n- }\n-\n- public ClusterServices getClusterServices() {\n- return brokerContext.getClusterServices();\n- }\n-\n- public DiskSpaceUsageMonitor getDiskSpaceUsageMonitor() {\n- return brokerContext.getDiskSpaceUsageMonitor();\n- }\n-\n- public BrokerAdminService getBrokerAdminService() {\n- return brokerContext.getBrokerAdminService();\n+ public BrokerContext getBrokerContext() {\n+ return brokerContext;\n }\n \n+ // only used for tests\n public SystemContext getSystemContext() {\n return systemContext;\n }\n \n- public PartitionManager getPartitionManager() {\n- return brokerContext.getPartitionManager();\n- }\n- // only used for tests\n-\n /**\n * Temporary helper object. This object is needed during the transition of broker startup/shutdown\n * steps to the new concept. Afterwards, the expectation is that this object will merge with the\ndiff --git a/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java b/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java\nindex bda5170..1accbc1 100644\n--- a/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java\n+++ b/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java\n@@ -45,11 +45,12 @@ public class BrokerSnapshotTest {\n (RaftPartition)\n brokerRule\n .getBroker()\n+ .getBrokerContext()\n .getPartitionManager()\n .getPartitionGroup()\n .getPartition(PartitionId.from(PartitionManagerImpl.GROUP_NAME, PARTITION_ID));\n journalReader = raftPartition.getServer().openReader();\n- brokerAdminService = brokerRule.getBroker().getBrokerAdminService();\n+ brokerAdminService = brokerRule.getBroker().getBrokerContext().getBrokerAdminService();\n \n final String contactPoint = NetUtil.toSocketAddressString(brokerRule.getGatewayAddress());\n final ZeebeClientBuilder zeebeClientBuilder =\ndiff --git a/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java b/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java\nindex e98e7d2..a831bfe 100644\n--- a/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java\n+++ b/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java\n@@ -173,11 +173,11 @@ public final class EmbeddedBrokerRule extends ExternalResource {\n }\n \n public ClusterServices getClusterServices() {\n- return broker.getClusterServices();\n+ return broker.getBrokerContext().getClusterServices();\n }\n \n public AtomixCluster getAtomixCluster() {\n- return broker.getAtomixCluster();\n+ return broker.getBrokerContext().getAtomixCluster();\n }\n \n public InetSocketAddress getGatewayAddress() {\n@@ -245,7 +245,8 @@ public final class EmbeddedBrokerRule extends ExternalResource {\n Thread.currentThread().interrupt();\n }\n \n- final EmbeddedGatewayService embeddedGatewayService = broker.getEmbeddedGatewayService();\n+ final EmbeddedGatewayService embeddedGatewayService =\n+ broker.getBrokerContext().getEmbeddedGatewayService();\n if (embeddedGatewayService != null) {\n final BrokerClient brokerClient = embeddedGatewayService.get().getBrokerClient();\n \ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java\nindex 890b596..8561cf1 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java\n@@ -29,6 +29,7 @@ import io.atomix.utils.net.Address;\n import io.camunda.zeebe.broker.Broker;\n import io.camunda.zeebe.broker.PartitionListener;\n import io.camunda.zeebe.broker.SpringBrokerBridge;\n+import io.camunda.zeebe.broker.bootstrap.BrokerContext;\n import io.camunda.zeebe.broker.exporter.stream.ExporterDirectorContext;\n import io.camunda.zeebe.broker.partitioning.PartitionManagerImpl;\n import io.camunda.zeebe.broker.system.SystemContext;\n@@ -602,11 +603,11 @@ public final class ClusteringRule extends ExternalResource {\n }\n \n public void stepDown(final Broker broker, final int partitionId) {\n- final var atomix = broker.getClusterServices();\n+ final var atomix = broker.getBrokerContext().getClusterServices();\n final MemberId nodeId = atomix.getMembershipService().getLocalMember().id();\n \n final var raftPartition =\n- broker.getPartitionManager().getPartitionGroup().getPartitions().stream()\n+ broker.getBrokerContext().getPartitionManager().getPartitionGroup().getPartitions().stream()\n .filter(partition -> partition.members().contains(nodeId))\n .filter(partition -> partition.id().id() == partitionId)\n .map(RaftPartition.class::cast)\n@@ -617,14 +618,14 @@ public final class ClusteringRule extends ExternalResource {\n }\n \n public void disconnect(final Broker broker) {\n- final var atomix = broker.getAtomixCluster();\n+ final var atomix = broker.getBrokerContext().getAtomixCluster();\n \n ((NettyUnicastService) atomix.getUnicastService()).stop().join();\n ((NettyMessagingService) atomix.getMessagingService()).stop().join();\n }\n \n public void connect(final Broker broker) {\n- final var atomix = broker.getAtomixCluster();\n+ final var atomix = broker.getBrokerContext().getAtomixCluster();\n \n ((NettyUnicastService) atomix.getUnicastService()).start().join();\n ((NettyMessagingService) atomix.getMessagingService()).start().join();\n@@ -666,11 +667,11 @@ public final class ClusteringRule extends ExternalResource {\n }\n \n final var broker = brokers.get(expectedLeader);\n- final var atomix = broker.getClusterServices();\n+ final var atomix = broker.getBrokerContext().getClusterServices();\n final MemberId nodeId = atomix.getMembershipService().getLocalMember().id();\n \n final var raftPartition =\n- broker.getPartitionManager().getPartitionGroup().getPartitions().stream()\n+ broker.getBrokerContext().getPartitionManager().getPartitionGroup().getPartitions().stream()\n .filter(partition -> partition.members().contains(nodeId))\n .filter(partition -> partition.id().id() == START_PARTITION_ID)\n .map(RaftPartition.class::cast)\n@@ -775,14 +776,15 @@ public final class ClusteringRule extends ExternalResource {\n }\n \n public void takeSnapshot(final Broker broker) {\n- broker.getBrokerAdminService().takeSnapshot();\n+ broker.getBrokerContext().getBrokerAdminService().takeSnapshot();\n }\n \n public void triggerAndWaitForSnapshots() {\n // Ensure that the exporter positions are distributed to the followers\n getClock().addTime(ExporterDirectorContext.DEFAULT_DISTRIBUTION_INTERVAL);\n getBrokers().stream()\n- .map(Broker::getBrokerAdminService)\n+ .map(Broker::getBrokerContext)\n+ .map(BrokerContext::getBrokerAdminService)\n .forEach(BrokerAdminService::takeSnapshot);\n \n getBrokers()\n@@ -794,7 +796,7 @@ public final class ClusteringRule extends ExternalResource {\n .until(\n () -> {\n // Trigger snapshot again in case snapshot is not already taken\n- broker.getBrokerAdminService().takeSnapshot();\n+ broker.getBrokerContext().getBrokerAdminService().takeSnapshot();\n return getSnapshot(broker);\n },\n Optional::isPresent));\n@@ -831,7 +833,7 @@ public final class ClusteringRule extends ExternalResource {\n \n private Optional<SnapshotId> getSnapshot(final Broker broker, final int partitionId) {\n \n- final var partitions = broker.getBrokerAdminService().getPartitionStatus();\n+ final var partitions = broker.getBrokerContext().getBrokerAdminService().getPartitionStatus();\n final var partitionStatus = partitions.get(partitionId);\n \n return Optional.ofNullable(partitionStatus)\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java\nindex f07961c..d46636b 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java\n@@ -88,6 +88,7 @@ public class DiskSpaceMonitoringFailOverTest {\n () ->\n clusteringRule\n .getBroker(newLeaderId)\n+ .getBrokerContext()\n .getBrokerAdminService()\n .getPartitionStatus()\n .get(1)\n@@ -96,7 +97,7 @@ public class DiskSpaceMonitoringFailOverTest {\n }\n \n private void waitUntilDiskSpaceNotAvailable(final Broker broker) throws InterruptedException {\n- final var diskSpaceMonitor = broker.getDiskSpaceUsageMonitor();\n+ final var diskSpaceMonitor = broker.getBrokerContext().getDiskSpaceUsageMonitor();\n \n final CountDownLatch diskSpaceNotAvailable = new CountDownLatch(1);\n diskSpaceMonitor.addDiskUsageListener(\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java\nindex 0a02a27..6e93cf9 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java\n@@ -165,7 +165,7 @@ public class DiskSpaceRecoveryClusteredTest {\n }\n \n private void waitUntilDiskSpaceNotAvailable(final Broker broker) throws InterruptedException {\n- final var diskSpaceMonitor = broker.getDiskSpaceUsageMonitor();\n+ final var diskSpaceMonitor = broker.getBrokerContext().getDiskSpaceUsageMonitor();\n \n final CountDownLatch diskSpaceNotAvailable = new CountDownLatch(1);\n diskSpaceMonitor.addDiskUsageListener(\n@@ -188,7 +188,7 @@ public class DiskSpaceRecoveryClusteredTest {\n }\n \n private void waitUntilDiskSpaceAvailable(final Broker broker) throws InterruptedException {\n- final var diskSpaceMonitor = broker.getDiskSpaceUsageMonitor();\n+ final var diskSpaceMonitor = broker.getBrokerContext().getDiskSpaceUsageMonitor();\n final CountDownLatch diskSpaceAvailableAgain = new CountDownLatch(1);\n diskSpaceMonitor.addDiskUsageListener(\n new DiskSpaceUsageListener() {\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java\nindex 9cef5a0..a487729 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java\n@@ -192,7 +192,8 @@ public class DiskSpaceRecoveryTest {\n }\n \n private void waitUntilDiskSpaceNotAvailable() throws InterruptedException {\n- final var diskSpaceMonitor = embeddedBrokerRule.getBroker().getDiskSpaceUsageMonitor();\n+ final var diskSpaceMonitor =\n+ embeddedBrokerRule.getBroker().getBrokerContext().getDiskSpaceUsageMonitor();\n \n final CountDownLatch diskSpaceNotAvailable = new CountDownLatch(1);\n diskSpaceMonitor.addDiskUsageListener(\n@@ -215,7 +216,8 @@ public class DiskSpaceRecoveryTest {\n }\n \n private void waitUntilDiskSpaceAvailable() throws InterruptedException {\n- final var diskSpaceMonitor = embeddedBrokerRule.getBroker().getDiskSpaceUsageMonitor();\n+ final var diskSpaceMonitor =\n+ embeddedBrokerRule.getBroker().getBrokerContext().getDiskSpaceUsageMonitor();\n final CountDownLatch diskSpaceAvailableAgain = new CountDownLatch(1);\n diskSpaceMonitor.addDiskUsageListener(\n new DiskSpaceUsageListener() {\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java\nindex 2d1e4f0..58f6f16 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java\n@@ -48,6 +48,7 @@ public class HealthMonitoringTest {\n final var raftPartition =\n (RaftPartition)\n leader\n+ .getBrokerContext()\n .getPartitionManager()\n .getPartitionGroup()\n .getPartition(\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java\nindex 468f83c..7ff03be 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java\n@@ -11,6 +11,7 @@ import static org.assertj.core.api.Assertions.assertThat;\n \n import io.atomix.raft.RaftServer.Role;\n import io.camunda.zeebe.broker.Broker;\n+import io.camunda.zeebe.broker.bootstrap.BrokerContext;\n import io.camunda.zeebe.broker.system.management.BrokerAdminService;\n import io.camunda.zeebe.engine.processing.streamprocessor.StreamProcessor.Phase;\n import io.camunda.zeebe.it.clustering.ClusteringRule;\n@@ -48,7 +49,7 @@ public class BrokerAdminServiceClusterTest {\n @Before\n public void before() {\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\n- leaderAdminService = leader.getBrokerAdminService();\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\n }\n \n @Test\n@@ -61,7 +62,8 @@ public class BrokerAdminServiceClusterTest {\n // when\n final var followerStatus =\n followers.stream()\n- .map(Broker::getBrokerAdminService)\n+ .map(Broker::getBrokerContext)\n+ .map(BrokerContext::getBrokerAdminService)\n .map(BrokerAdminService::getPartitionStatus)\n .map(status -> status.get(1));\n \n@@ -94,7 +96,8 @@ public class BrokerAdminServiceClusterTest {\n \n // then\n clusteringRule.getBrokers().stream()\n- .map(Broker::getBrokerAdminService)\n+ .map(Broker::getBrokerContext)\n+ .map(BrokerContext::getBrokerAdminService)\n .forEach(this::assertThatStatusContainsProcessedPositionInSnapshot);\n }\n \n@@ -102,7 +105,8 @@ public class BrokerAdminServiceClusterTest {\n public void shouldPauseAfterLeaderChange() {\n // given\n clusteringRule.getBrokers().stream()\n- .map(Broker::getBrokerAdminService)\n+ .map(Broker::getBrokerContext)\n+ .map(BrokerContext::getBrokerAdminService)\n .forEach(BrokerAdminService::pauseStreamProcessing);\n \n // when\n@@ -113,6 +117,7 @@ public class BrokerAdminServiceClusterTest {\n final var newLeaderAdminService =\n clusteringRule\n .getBroker(clusteringRule.getLeaderForPartition(1).getNodeId())\n+ .getBrokerContext()\n .getBrokerAdminService();\n assertStreamProcessorPhase(newLeaderAdminService, Phase.PAUSED);\n }\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java\nindex 5160b50..2185329 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java\n@@ -41,7 +41,7 @@ public class BrokerAdminServiceTest {\n @Before\n public void before() {\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\n- leaderAdminService = leader.getBrokerAdminService();\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\n }\n \n @Test\n@@ -144,7 +144,7 @@ public class BrokerAdminServiceTest {\n \n // then\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\n- leaderAdminService = leader.getBrokerAdminService();\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\n assertStreamProcessorPhase(leaderAdminService, Phase.PAUSED);\n }\n \n@@ -161,7 +161,7 @@ public class BrokerAdminServiceTest {\n \n // then\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\n- leaderAdminService = leader.getBrokerAdminService();\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\n assertStreamProcessorPhase(leaderAdminService, Phase.PROCESSING);\n }\n \n@@ -176,7 +176,7 @@ public class BrokerAdminServiceTest {\n \n // then\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\n- leaderAdminService = leader.getBrokerAdminService();\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\n assertExporterPhase(leaderAdminService, ExporterPhase.PAUSED);\n }\n \n@@ -193,7 +193,7 @@ public class BrokerAdminServiceTest {\n \n // then\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\n- leaderAdminService = leader.getBrokerAdminService();\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\n assertExporterPhase(leaderAdminService, ExporterPhase.EXPORTING);\n }\n \ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java\nindex d6c8ab3..4582ad2 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java\n@@ -30,7 +30,7 @@ public class BrokerAdminServiceWithOutExporterTest {\n // given\n final var leader =\n clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\n- final var leaderAdminService = leader.getBrokerAdminService();\n+ final var leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\n // when there are no exporters configured\n // then\n final var partitionStatus = leaderAdminService.getPartitionStatus().get(1);\ndiff --git a/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java b/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java\nindex 36bc0bf..d332201 100644\n--- a/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java\n+++ b/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java\n@@ -240,7 +240,8 @@ public class EmbeddedBrokerRule extends ExternalResource {\n Thread.currentThread().interrupt();\n }\n \n- final EmbeddedGatewayService embeddedGatewayService = broker.getEmbeddedGatewayService();\n+ final EmbeddedGatewayService embeddedGatewayService =\n+ broker.getBrokerContext().getEmbeddedGatewayService();\n if (embeddedGatewayService != null) {\n final BrokerClient brokerClient = embeddedGatewayService.get().getBrokerClient();\n \n", "diff --git a/benchmarks/project/pom.xml b/benchmarks/project/pom.xml\nindex 62030b6..ab87dea 100644\n--- a/benchmarks/project/pom.xml\n+++ b/benchmarks/project/pom.xml\n@@ -123,11 +123,6 @@\n </plugin>\n \n <plugin>\n- <groupId>com.diffplug.spotless</groupId>\n- <artifactId>spotless-maven-plugin</artifactId>\n- </plugin>\n-\n- <plugin>\n <groupId>org.apache.maven.plugins</groupId>\n <artifactId>maven-shade-plugin</artifactId>\n <executions>\n"]
2
["e52a6201093f273add4903dd5f4e55a63539386d", "7f9721dc9bbf66a3712d59352f64ca089da139f0"]
["refactor", "build"]
i18n for Time Picker,add page balckwhitelist and pdf
["diff --git a/packages/nc-gui/components/cell/TimePicker.vue b/packages/nc-gui/components/cell/TimePicker.vue\nindex 619ab45..7f66828 100644\n--- a/packages/nc-gui/components/cell/TimePicker.vue\n+++ b/packages/nc-gui/components/cell/TimePicker.vue\n@@ -38,6 +38,8 @@ const isTimeInvalid = ref(false)\n \n const dateFormat = isMysql(column.value.base_id) ? 'YYYY-MM-DD HH:mm:ss' : 'YYYY-MM-DD HH:mm:ssZ'\n \n+const { t } = useI18n()\n+\n const localState = computed({\n get() {\n if (!modelValue) {\n@@ -89,11 +91,11 @@ watch(\n \n const placeholder = computed(() => {\n if (isEditColumn.value && (modelValue === '' || modelValue === null)) {\n- return '(Optional)'\n+ return t('labels.optional')\n } else if (modelValue === null && showNull.value) {\n- return 'NULL'\n+ return t('general.null')\n } else if (isTimeInvalid.value) {\n- return 'Invalid time'\n+ return t('msg.invalidTime')\n } else {\n return ''\n }\n", "diff --git a/src/_locales/common/messages.json b/src/_locales/common/messages.json\nindex e8524ac..3a596d6 100644\n--- a/src/_locales/common/messages.json\n+++ b/src/_locales/common/messages.json\n@@ -4,11 +4,21 @@\n \"zh_CN\": \"\u6dfb\u52a0\",\n \"zh_TW\": \"\u65b0\u589e\"\n },\n+ \"blacklist\": {\n+ \"en\": \"Blacklist\",\n+ \"zh_CN\": \"\u9ed1\u540d\u5355\",\n+ \"zh_TW\": \"\u9ed1\u540d\u55ae\"\n+ },\n \"cancel\": {\n \"en\": \"Cancel\",\n \"zh_CN\": \"\u53d6\u6d88\",\n \"zh_TW\": \"\u53d6\u6d88\"\n },\n+ \"changes_confirm\": {\n+ \"en\": \"Changes not saved. Close anyway?\",\n+ \"zh_CN\": \"\u4fee\u6539\u672a\u4fdd\u5b58\u3002\u786e\u8ba4\u5173\u95ed\uff1f\",\n+ \"zh_TW\": \"\u4fee\u6539\u672a\u4fdd\u5b58\u3002\u78ba\u5b9a\u95dc\u9589\uff1f\"\n+ },\n \"confirm\": {\n \"en\": \"Confirm\",\n \"zh_CN\": \"\u786e\u8ba4\",\n@@ -93,5 +103,10 @@\n \"en\": \"words\",\n \"zh_CN\": \"\u4e2a\",\n \"zh_TW\": \"\u4e2a\"\n+ },\n+ \"whitelist\": {\n+ \"en\": \"Whitelist\",\n+ \"zh_CN\": \"\u767d\u540d\u5355\",\n+ \"zh_TW\": \"\u767d\u540d\u55ae\"\n }\n }\ndiff --git a/src/_locales/options/messages.json b/src/_locales/options/messages.json\nindex ada2488..e7d699a 100644\n--- a/src/_locales/options/messages.json\n+++ b/src/_locales/options/messages.json\n@@ -119,6 +119,11 @@\n \"zh_CN\": \"\u53cd\u9988\u95ee\u9898\",\n \"zh_TW\": \"\u8edf\u9ad4\u4f7f\u7528\u7591\u554f\u548c\u5efa\u8a00\"\n },\n+ \"match_pattern_description\": {\n+ \"en\": \"Specify URLs as match patterns. <a href=\\\"https://developer.mozilla.org/en-US/Add-ons/WebExtensions/Match_patterns#Examples\\\" target=\\\"_blank\\\">Examples</a>. Empty fields will be removed.\",\n+ \"zh_CN\": \"\u7f51\u5740\u652f\u6301\u5339\u914d\u6a21\u5f0f\uff08<a href=\\\"https://developer.mozilla.org/zh-CN/Add-ons/WebExtensions/Match_patterns#\u8303\u4f8b\\\" target=\\\"_blank\\\">\u4f8b\u5b50</a>\uff09\u3002\u7559\u7a7a\u4fdd\u5b58\u5373\u53ef\u6e05\u9664\u3002\",\n+ \"zh_TW\": \"\u7db2\u5740\u652f\u63f4\u5339\u914d\u6a21\u5f0f\uff08<a href=\\\"https://developer.mozilla.org/zh-CN/Add-ons/WebExtensions/Match_patterns#\u8303\u4f8b\\\" target=\\\"_blank\\\">\u4f8b\u5b50</a>\uff09\u3002\u7559\u7a7a\u5132\u5b58\u5373\u53ef\u6e05\u9664\u3002\"\n+ },\n \"msg_updated\": {\n \"en\": \"Successfully updated\",\n \"zh_CN\": \"\u8bbe\u7f6e\u5df2\u66f4\u65b0\",\n@@ -319,6 +324,21 @@\n \"zh_CN\": \"\u5f00\u542f\u540e\uff0c\u672c\u6269\u5c55\u4f1a\u81ea\u52a8\u8bc6\u522b\u8f93\u5165\u6846\u4ee5\u53ca\u5e38\u89c1\u7f16\u8f91\u5668\uff0c\u5982 CodeMirror\u3001ACE \u548c Monaco\u3002\",\n \"zh_TW\": \"\u958b\u555f\u540e\uff0c\u672c\u7a0b\u5f0f\u6703\u81ea\u52d5\u8b58\u5225\u8f38\u5165\u6846\u4ee5\u53ca\u5e38\u898b\u7de8\u8f2f\u5668\uff0c\u5982 CodeMirror\u3001ACE \u548c Monaco\u3002\"\n },\n+ \"opt_pdf_blackwhitelist_help\": {\n+ \"en\": \"Blacklisted PDF links will not jump to Saladict PDF Viewer.\",\n+ \"zh_CN\": \"\u9ed1\u540d\u5355\u5339\u914d\u7684 PDF \u94fe\u63a5\u5c06\u4e0d\u4f1a\u8df3\u8f6c\u5230 Saladict \u6253\u5f00\u3002\",\n+ \"zh_TW\": \"\u9ed1\u540d\u55ae\u5339\u914d\u7684 PDF \u9023\u7d50\u5c07\u4e0d\u6703\u8df3\u8f49\u5230 Saladict \u958b\u555f\u3002\"\n+ },\n+ \"opt_pdf_sniff\": {\n+ \"en\": \"Enable PDF Sniffer\",\n+ \"zh_CN\": \"\u9ed8\u8ba4\u7528\u672c\u6269\u5c55\u6d4f\u89c8 PDF\",\n+ \"zh_TW\": \"\u4f7f\u7528\u672c\u61c9\u7528\u7a0b\u5f0f\u700f\u89bd PDF\"\n+ },\n+ \"opt_pdf_sniff_help\": {\n+ \"en\": \"If turned on\uff0c PDF links will be automatically captured.\",\n+ \"zh_CN\": \"\u5f00\u542f\u540e\u6240\u6709 PDF \u94fe\u63a5\u5c06\u81ea\u52a8\u8df3\u8f6c\u5230\u672c\u6269\u5c55\u6253\u5f00\uff08\u5305\u62ec\u672c\u5730\uff0c\u5982\u679c\u5728\u6269\u5c55\u7ba1\u7406\u9875\u9762\u52fe\u9009\u4e86\u5141\u8bb8\uff09\u3002\",\n+ \"zh_TW\": \"\u958b\u555f\u5f8c\u6240\u6709 PDF \u9023\u7d50\u5c07\u81ea\u52d5\u8df3\u8f49\u5230\u672c\u64f4\u5145\u5957\u4ef6\u958b\u555f\uff08\u5305\u62ec\u672c\u5730\uff0c\u5982\u679c\u5728\u64f4\u5145\u5957\u4ef6\u7ba1\u7406\u9801\u9762\u52fe\u9078\u4e86\u5141\u8a31\uff09\u3002\"\n+ },\n \"opt_profile_change\": {\n \"en\": \"This option may change base on \\\"Profile\\\".\",\n \"zh_CN\": \"\u6b64\u9009\u9879\u4f1a\u56e0\u300c\u60c5\u666f\u6a21\u5f0f\u300d\u800c\u6539\u53d8\u3002\",\n@@ -329,6 +349,16 @@\n \"zh_CN\": \"\u8f93\u5165\u65f6\u663e\u793a\u5019\u9009\",\n \"zh_TW\": \"\u8f38\u5165\u6642\u986f\u793a\u5019\u9078\"\n },\n+ \"opt_sel_blackwhitelist\": {\n+ \"en\": \"Selection Black/White List\",\n+ \"zh_CN\": \"\u5212\u8bcd\u9ed1\u767d\u540d\u5355\",\n+ \"zh_TW\": \"\u9078\u8a5e\u9ed1\u767d\u540d\u55ae\"\n+ },\n+ \"opt_sel_blackwhitelist_help\": {\n+ \"en\": \"Saladict will not react to selection in blacklisted pages.\",\n+ \"zh_CN\": \"\u9ed1\u540d\u5355\u5339\u914d\u7684\u9875\u9762 Saladict \u5c06\u4e0d\u4f1a\u54cd\u5e94\u9f20\u6807\u5212\u8bcd\u3002\",\n+ \"zh_TW\": \"\u9ed1\u540d\u55ae\u5339\u914d\u7684\u9801\u9762 Saladict \u5c07\u4e0d\u6703\u97ff\u61c9\u6ed1\u9f20\u5283\u8a5e\u3002\"\n+ },\n \"opt_sel_lang\": {\n \"en\": \"Selection Languages\",\n \"zh_CN\": \"\u5212\u8bcd\u8bed\u8a00\",\ndiff --git a/src/options/components/options/BlackWhiteList/index.tsx b/src/options/components/options/BlackWhiteList/index.tsx\nnew file mode 100644\nindex 0000000..52708dd\n--- /dev/null\n+++ b/src/options/components/options/BlackWhiteList/index.tsx\n@@ -0,0 +1,69 @@\n+import React from 'react'\n+import { Props } from '../typings'\n+import { formItemLayout } from '../helpers'\n+import MatchPatternModal from '../../MatchPatternModal'\n+\n+import { FormComponentProps } from 'antd/lib/form'\n+import { Form, Button } from 'antd'\n+\n+export type BlackWhiteListProps = Props & FormComponentProps\n+\n+interface BlackWhiteListState {\n+ editingArea: '' | 'pdfWhitelist' | 'pdfBlacklist' | 'whitelist' | 'blacklist'\n+}\n+\n+export class BlackWhiteList extends React.Component<BlackWhiteListProps, BlackWhiteListState> {\n+ constructor (props: BlackWhiteListProps) {\n+ super(props)\n+ this.state = {\n+ editingArea: ''\n+ }\n+ }\n+\n+ closeModal = () => {\n+ this.setState({ editingArea: '' })\n+ }\n+\n+ render () {\n+ const { t, config } = this.props\n+\n+ return (\n+ <Form>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={t('opt_sel_blackwhitelist')}\n+ help={t('opt_sel_blackwhitelist_help')}\n+ >\n+ <Button\n+ style={{ marginRight: 10 }}\n+ onClick={() => this.setState({ editingArea: 'blacklist' })}\n+ >{t('common:blacklist')}</Button>\n+ <Button\n+ onClick={() => this.setState({ editingArea: 'whitelist' })}\n+ >{t('common:whitelist')}</Button>\n+ </Form.Item>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={`PDF ${t('nav_BlackWhiteList')}`}\n+ help={t('opt_pdf_blackwhitelist_help')}\n+ >\n+ <Button\n+ style={{ marginRight: 10 }}\n+ onClick={() => this.setState({ editingArea: 'pdfBlacklist' })}\n+ >PDF {t('common:blacklist')}</Button>\n+ <Button\n+ onClick={() => this.setState({ editingArea: 'pdfWhitelist' })}\n+ >PDF {t('common:whitelist')}</Button>\n+ </Form.Item>\n+ <MatchPatternModal\n+ t={t}\n+ config={config}\n+ area={this.state.editingArea}\n+ onClose={this.closeModal}\n+ />\n+ </Form>\n+ )\n+ }\n+}\n+\n+export default BlackWhiteList\ndiff --git a/src/options/components/options/PDF/index.tsx b/src/options/components/options/PDF/index.tsx\nnew file mode 100644\nindex 0000000..3e7772d\n--- /dev/null\n+++ b/src/options/components/options/PDF/index.tsx\n@@ -0,0 +1,72 @@\n+import React from 'react'\n+import { Props } from '../typings'\n+import { updateConfigOrProfile, formItemLayout } from '../helpers'\n+import MatchPatternModal from '../../MatchPatternModal'\n+\n+import { FormComponentProps } from 'antd/lib/form'\n+import { Form, Switch, Button } from 'antd'\n+\n+export type PDFProps = Props & FormComponentProps\n+\n+interface PDFState {\n+ editingArea: '' | 'pdfWhitelist' | 'pdfBlacklist'\n+}\n+\n+export class PDF extends React.Component<PDFProps, PDFState> {\n+ constructor (props: PDFProps) {\n+ super(props)\n+\n+ this.state = {\n+ editingArea: ''\n+ }\n+ }\n+\n+ closeModal = () => {\n+ this.setState({ editingArea: '' })\n+ }\n+\n+ render () {\n+ const { t, config } = this.props\n+ const { getFieldDecorator } = this.props.form\n+\n+ return (\n+ <Form>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={t('opt_pdf_sniff')}\n+ help={t('opt_pdf_sniff_help')}\n+ >{\n+ getFieldDecorator('config#pdfSniff', {\n+ initialValue: config.pdfSniff,\n+ valuePropName: 'checked',\n+ })(\n+ <Switch />\n+ )\n+ }</Form.Item>\n+ <Form.Item\n+ {...formItemLayout}\n+ label={t('nav_BlackWhiteList')}\n+ help={t('opt_pdf_blackwhitelist_help')}\n+ >\n+ <Button\n+ style={{ marginRight: 10 }}\n+ onClick={() => this.setState({ editingArea: 'pdfBlacklist' })}\n+ >PDF {t('common:blacklist')}</Button>\n+ <Button\n+ onClick={() => this.setState({ editingArea: 'pdfWhitelist' })}\n+ >PDF {t('common:whitelist')}</Button>\n+ </Form.Item>\n+ <MatchPatternModal\n+ t={t}\n+ config={config}\n+ area={this.state.editingArea}\n+ onClose={this.closeModal}\n+ />\n+ </Form>\n+ )\n+ }\n+}\n+\n+export default Form.create<PDFProps>({\n+ onValuesChange: updateConfigOrProfile\n+})(PDF)\n"]
2
["48806e3675c7b18327e7629827454d7c29be25a9", "9b1c0fc20b614513384a1e562317dbf076eb8ef0"]
["fix", "feat"]
remove unused,trigger build every hour for develop To better track stability of the develop branch the build should be triggered on commit and every hour. Other branches should not be effected. - add cron trigger to develop branch - extract variables to identify stable and develop branch
["diff --git a/src/content/redux/modules/dictionaries.ts b/src/content/redux/modules/dictionaries.ts\nindex 88f7215..570d397 100644\n--- a/src/content/redux/modules/dictionaries.ts\n+++ b/src/content/redux/modules/dictionaries.ts\n@@ -3,7 +3,6 @@ import { DictID, appConfigFactory, AppConfig } from '@/app-config'\n import isEqual from 'lodash/isEqual'\n import { saveWord } from '@/_helpers/record-manager'\n import { getDefaultSelectionInfo, SelectionInfo, isSameSelection } from '@/_helpers/selection'\n-import { createActiveConfigStream } from '@/_helpers/config-manager'\n import { isContainChinese, isContainEnglish, testerPunct, isContainMinor, testerChinese, testJapanese, testKorean } from '@/_helpers/lang-check'\n import { MsgType, MsgFetchDictResult } from '@/typings/message'\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\ndiff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\nindex 53ad550..68e0a3d 100644\n--- a/src/content/redux/modules/widget.ts\n+++ b/src/content/redux/modules/widget.ts\n@@ -1,9 +1,9 @@\n import * as recordManager from '@/_helpers/record-manager'\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\n-import appConfigFactory, { TCDirection, AppConfig, DictID } from '@/app-config'\n+import appConfigFactory, { TCDirection, DictID } from '@/app-config'\n import { message, storage } from '@/_helpers/browser-api'\n-import { createActiveConfigStream, createConfigIDListStream } from '@/_helpers/config-manager'\n-import { MsgSelection, MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\n+import { createConfigIDListStream } from '@/_helpers/config-manager'\n+import { MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\n import { searchText, restoreDicts } from '@/content/redux/modules/dictionaries'\n import { SelectionInfo, getDefaultSelectionInfo } from '@/_helpers/selection'\n import { Mutable } from '@/typings/helpers'\n", "diff --git a/Jenkinsfile b/Jenkinsfile\nindex 2c58f61..9daa38f 100644\n--- a/Jenkinsfile\n+++ b/Jenkinsfile\n@@ -4,9 +4,17 @@\n \n def buildName = \"${env.JOB_BASE_NAME.replaceAll(\"%2F\", \"-\").replaceAll(\"\\\\.\", \"-\").take(20)}-${env.BUILD_ID}\"\n \n+def masterBranchName = 'master'\n+def isMasterBranch = env.BRANCH_NAME == masterBranchName\n+def developBranchName = 'develop'\n+def isDevelopBranch = env.BRANCH_NAME == developBranchName\n+\n //for develop branch keep builds for 7 days to be able to analyse build errors, for all other branches, keep the last 10 builds\n-def daysToKeep = (env.BRANCH_NAME=='develop') ? '7' : '-1'\n-def numToKeep = (env.BRANCH_NAME=='develop') ? '-1' : '10'\n+def daysToKeep = isDevelopBranch ? '7' : '-1'\n+def numToKeep = isDevelopBranch ? '-1' : '10'\n+\n+//the develop branch should be run hourly to detect flaky tests and instability, other branches only on commit\n+def cronTrigger = isDevelopBranch ? '@hourly' : ''\n \n pipeline {\n agent {\n@@ -23,6 +31,10 @@ pipeline {\n SONARCLOUD_TOKEN = credentials('zeebe-sonarcloud-token')\n }\n \n+ triggers {\n+ cron(cronTrigger)\n+ }\n+\n options {\n buildDiscarder(logRotator(daysToKeepStr: daysToKeep, numToKeepStr: numToKeep))\n timestamps()\n@@ -201,7 +213,7 @@ pipeline {\n }\n \n stage('Upload') {\n- when { branch 'develop' }\n+ when { allOf { branch developBranchName ; not { triggeredBy 'TimerTrigger' } } }\n steps {\n retry(3) {\n container('maven') {\n@@ -214,9 +226,11 @@ pipeline {\n }\n \n stage('Post') {\n+ when { not { triggeredBy 'TimerTrigger' } }\n+\n parallel {\n stage('Docker') {\n- when { branch 'develop' }\n+ when { branch developBranchName }\n \n environment {\n VERSION = readMavenPom(file: 'parent/pom.xml').getVersion()\n@@ -227,20 +241,20 @@ pipeline {\n build job: 'zeebe-docker', parameters: [\n string(name: 'BRANCH', value: env.BRANCH_NAME),\n string(name: 'VERSION', value: env.VERSION),\n- booleanParam(name: 'IS_LATEST', value: env.BRANCH_NAME == 'master'),\n- booleanParam(name: 'PUSH', value: env.BRANCH_NAME == 'develop')\n+ booleanParam(name: 'IS_LATEST', value: isMasterBranch),\n+ booleanParam(name: 'PUSH', value: isDevelopBranch)\n ]\n }\n }\n }\n \n stage('Docs') {\n- when { anyOf { branch 'master'; branch 'develop' } }\n+ when { anyOf { branch masterBranchName; branch developBranchName } }\n steps {\n retry(3) {\n build job: 'zeebe-docs', parameters: [\n string(name: 'BRANCH', value: env.BRANCH_NAME),\n- booleanParam(name: 'LIVE', value: env.BRANCH_NAME == 'master')\n+ booleanParam(name: 'LIVE', value: isMasterBranch)\n ]\n }\n }\n"]
2
["a50b51999015e210918d9c8e95fd4cac347353be", "3bc1541d6c95ef8cb5ce5da741733f09c98e4b29"]
["refactor", "ci"]
updated webpack in react,add important to override paragraphs in items
["diff --git a/components/react/package.json b/components/react/package.json\nindex bbeb9ee..43ddebc 100644\n--- a/components/react/package.json\n+++ b/components/react/package.json\n@@ -114,7 +114,7 @@\n \"ts-loader\": \"^9.2.9\",\n \"ts-node\": \"^10.7.0\",\n \"typescript\": \"^4.7.3\",\n- \"webpack\": \"^5.72.0\",\n+ \"webpack\": \"^5.73.0\",\n \"webpack-bundle-analyzer\": \"^4.5.0\",\n \"webpack-cli\": \"^4.9.2\",\n \"webpack-node-externals\": \"^3.0.0\"\ndiff --git a/yarn.lock b/yarn.lock\nindex a3fdb26..19a0716 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -25212,7 +25212,7 @@ webpack@^4.38.0, webpack@^4.42.1:\n watchpack \"^1.7.4\"\n webpack-sources \"^1.4.1\"\n \n-webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0:\n+webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0, webpack@^5.73.0:\n version \"5.73.0\"\n resolved \"https://registry.yarnpkg.com/webpack/-/webpack-5.73.0.tgz#bbd17738f8a53ee5760ea2f59dce7f3431d35d38\"\n integrity sha512-svjudQRPPa0YiOYa2lM/Gacw0r6PvxptHj4FuEKQ2kX05ZLkjbVc5MnPs6its5j7IZljnIqSVo/OsY2X0IpHGA==\n", "diff --git a/packages/core/src/components/text/text.ios.scss b/packages/core/src/components/text/text.ios.scss\nindex a3c58e2..2a020ab 100644\n--- a/packages/core/src/components/text/text.ios.scss\n+++ b/packages/core/src/components/text/text.ios.scss\n@@ -9,8 +9,9 @@\n @each $color-name, $color-base, $color-contrast in get-colors($colors-ios) {\n \n .text-ios-#{$color-name},\n- .text-ios-#{$color-name} a {\n- color: $color-base;\n+ .text-ios-#{$color-name} a,\n+ .text-ios-#{$color-name} p {\n+ color: $color-base !important\n }\n \n }\ndiff --git a/packages/core/src/components/text/text.md.scss b/packages/core/src/components/text/text.md.scss\nindex b397acb..050af1a 100644\n--- a/packages/core/src/components/text/text.md.scss\n+++ b/packages/core/src/components/text/text.md.scss\n@@ -9,8 +9,9 @@\n @each $color-name, $color-base, $color-contrast in get-colors($colors-md) {\n \n .text-md-#{$color-name},\n- .text-md-#{$color-name} a {\n- color: $color-base;\n+ .text-md-#{$color-name} a,\n+ .text-md-#{$color-name} p {\n+ color: $color-base !important;\n }\n \n }\n"]
2
["78c446cbea61af2268b4c4da03a9ad4283f10049", "7ab363f7ba2807b3eb9895e47f4fcd058f43ae5e"]
["build", "test"]
pass absolute burnchain block height to pox sync watchdog so we correctly infer ibd status,enable performance test trigger This reverts commit 146c7b58154a5b3de957f87e3b193447e0576547.
["diff --git a/testnet/stacks-node/src/run_loop/neon.rs b/testnet/stacks-node/src/run_loop/neon.rs\nindex 677749b..dc4a7bd 100644\n--- a/testnet/stacks-node/src/run_loop/neon.rs\n+++ b/testnet/stacks-node/src/run_loop/neon.rs\n@@ -411,7 +411,6 @@ impl RunLoop {\n \n let mut burnchain_height = sortition_db_height;\n let mut num_sortitions_in_last_cycle = 1;\n- let mut learned_burnchain_height = false;\n \n // prepare to fetch the first reward cycle!\n target_burnchain_block_height = burnchain_height + pox_constants.reward_cycle_length as u64;\n@@ -439,18 +438,16 @@ impl RunLoop {\n break;\n }\n \n+ let remote_chain_height = burnchain.get_headers_height();\n+\n // wait for the p2p state-machine to do at least one pass\n- debug!(\"Wait until we reach steady-state before processing more burnchain blocks...\");\n+ debug!(\"Wait until we reach steady-state before processing more burnchain blocks (chain height is {}, we are at {})...\", remote_chain_height, burnchain_height);\n \n // wait until it's okay to process the next sortitions\n let ibd = match pox_watchdog.pox_sync_wait(\n &burnchain_config,\n &burnchain_tip,\n- if learned_burnchain_height {\n- Some(burnchain_height)\n- } else {\n- None\n- },\n+ Some(remote_chain_height),\n num_sortitions_in_last_cycle,\n ) {\n Ok(ibd) => ibd,\n@@ -478,7 +475,6 @@ impl RunLoop {\n };\n \n // *now* we know the burnchain height\n- learned_burnchain_height = true;\n burnchain_tip = next_burnchain_tip;\n burnchain_height = cmp::min(burnchain_height + 1, target_burnchain_block_height);\n \n", "diff --git a/Jenkinsfile b/Jenkinsfile\nindex 399f8b8..c3f8fde 100644\n--- a/Jenkinsfile\n+++ b/Jenkinsfile\n@@ -120,6 +120,12 @@ pipeline {\n }\n }\n \n+ stage('Trigger Performance Tests') {\n+ when { branch 'develop' }\n+ steps {\n+ build job: 'zeebe-cluster-performance-tests', wait: false\n+ }\n+ }\n }\n \n post {\n"]
2
["5b70e008c57efc89da4061f9adb7d0491b2ea644", "80944b7a513b442afcb2d0d6c7d71c0d79365dba"]
["fix", "ci"]
missing transformation for T,fix `memtable` docstrings
["diff --git a/src/Tuple/Merge.ts b/src/Tuple/Merge.ts\nindex dfa7ce5..5ba44b7 100644\n--- a/src/Tuple/Merge.ts\n+++ b/src/Tuple/Merge.ts\n@@ -30,7 +30,7 @@ type _MergeFlat<O extends object, O1P extends object> = {\n }\n \n type MergeDeep<T extends any[], T1 extends any[]> =\n- TupleOf<Compute<_MergeDeep<T, Omit<ObjectOf<T1>, keyof T>, ObjectOf<T1>>>>\n+ TupleOf<Compute<_MergeDeep<ObjectOf<T>, Omit<ObjectOf<T1>, keyof T>, ObjectOf<T1>>>>\n // same principle as above, but with a little tweak\n // we keep the original `O1` to know if we can merge\n // => if `O` and `O1` have `object` fields of same name\n", "diff --git a/ibis/expr/api.py b/ibis/expr/api.py\nindex 93fabaa..66a2ea9 100644\n--- a/ibis/expr/api.py\n+++ b/ibis/expr/api.py\n@@ -403,15 +403,21 @@ def memtable(\n >>> import ibis\n >>> t = ibis.memtable([{\"a\": 1}, {\"a\": 2}])\n >>> t\n+ PandasInMemoryTable\n+ data:\n+ DataFrameProxy:\n+ a\n+ 0 1\n+ 1 2\n \n >>> t = ibis.memtable([{\"a\": 1, \"b\": \"foo\"}, {\"a\": 2, \"b\": \"baz\"}])\n >>> t\n PandasInMemoryTable\n data:\n- ((1, 'foo'), (2, 'baz'))\n- schema:\n- a int8\n- b string\n+ DataFrameProxy:\n+ a b\n+ 0 1 foo\n+ 1 2 baz\n \n Create a table literal without column names embedded in the data and pass\n `columns`\n@@ -420,10 +426,22 @@ def memtable(\n >>> t\n PandasInMemoryTable\n data:\n- ((1, 'foo'), (2, 'baz'))\n- schema:\n- a int8\n- b string\n+ DataFrameProxy:\n+ a b\n+ 0 1 foo\n+ 1 2 baz\n+\n+ Create a table literal without column names embedded in the data. Ibis\n+ generates column names if none are provided.\n+\n+ >>> t = ibis.memtable([(1, \"foo\"), (2, \"baz\")])\n+ >>> t\n+ PandasInMemoryTable\n+ data:\n+ DataFrameProxy:\n+ col0 col1\n+ 0 1 foo\n+ 1 2 baz\n \"\"\"\n if columns is not None and schema is not None:\n raise NotImplementedError(\n"]
2
["c4d9e5023fa0f88ba283b37da27677ceda1cbfbb", "72bc0f5172c0a3d17bde29cfc00db4c60d2fee3a"]
["fix", "docs"]
remove unnecessary start argument from `range`,Improved Config Loading #423
["diff --git a/ibis/backends/dask/tests/execution/test_window.py b/ibis/backends/dask/tests/execution/test_window.py\nindex 75a7331..6bfc5e3 100644\n--- a/ibis/backends/dask/tests/execution/test_window.py\n+++ b/ibis/backends/dask/tests/execution/test_window.py\n@@ -489,7 +489,7 @@ def test_project_list_scalar(npartitions):\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\n result = expr.execute()\n \n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\"res\")\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\"res\")\n tm.assert_series_equal(result.res, expected)\n \n \ndiff --git a/ibis/backends/pandas/tests/execution/test_window.py b/ibis/backends/pandas/tests/execution/test_window.py\nindex 8f292b3..effa372 100644\n--- a/ibis/backends/pandas/tests/execution/test_window.py\n+++ b/ibis/backends/pandas/tests/execution/test_window.py\n@@ -436,7 +436,7 @@ def test_project_list_scalar():\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\n result = expr.execute()\n \n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\"res\")\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\"res\")\n tm.assert_series_equal(result.res, expected)\n \n \ndiff --git a/ibis/backends/pyspark/tests/test_basic.py b/ibis/backends/pyspark/tests/test_basic.py\nindex 3850919..14fe677 100644\n--- a/ibis/backends/pyspark/tests/test_basic.py\n+++ b/ibis/backends/pyspark/tests/test_basic.py\n@@ -19,7 +19,7 @@ from ibis.backends.pyspark.compiler import _can_be_replaced_by_column_name # no\n def test_basic(con):\n table = con.table(\"basic_table\")\n result = table.compile().toPandas()\n- expected = pd.DataFrame({\"id\": range(0, 10), \"str_col\": \"value\"})\n+ expected = pd.DataFrame({\"id\": range(10), \"str_col\": \"value\"})\n \n tm.assert_frame_equal(result, expected)\n \n@@ -28,9 +28,7 @@ def test_projection(con):\n table = con.table(\"basic_table\")\n result1 = table.mutate(v=table[\"id\"]).compile().toPandas()\n \n- expected1 = pd.DataFrame(\n- {\"id\": range(0, 10), \"str_col\": \"value\", \"v\": range(0, 10)}\n- )\n+ expected1 = pd.DataFrame({\"id\": range(10), \"str_col\": \"value\", \"v\": range(10)})\n \n result2 = (\n table.mutate(v=table[\"id\"])\n@@ -44,8 +42,8 @@ def test_projection(con):\n {\n \"id\": range(0, 20, 2),\n \"str_col\": \"value\",\n- \"v\": range(0, 10),\n- \"v2\": range(0, 10),\n+ \"v\": range(10),\n+ \"v2\": range(10),\n }\n )\n \n", "diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex 76dd749..2087803 100644\n--- a/CHANGELOG.md\n+++ b/CHANGELOG.md\n@@ -7,6 +7,7 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/)\n ## [Unreleased]\n * Fix the tab '(Sync)' suffix in named tabs (https://github.com/zellij-org/zellij/pull/410)\n * Improve performance when multiple panes are open (https://github.com/zellij-org/zellij/pull/318)\n+* Improve error reporting and tests of configuration (https://github.com/zellij-org/zellij/pull/423)\n \n ## [0.6.0] - 2021-04-29\n * Doesn't quit anymore on single `q` press while in tab mode (https://github.com/zellij-org/zellij/pull/342)\n"]
2
["15f8d95754a0b6865ea475ca9e515272a07bf6ba", "099861ff5b0f83773ca0af4c70e6e39be3b0336c"]
["refactor", "docs"]
updates the readme to improve the readability and contributing sections,reorder startup steps
["diff --git a/.github/CONTRIBUTING.md b/.github/CONTRIBUTING.md\nindex 3c4dd8d..f8b8514 100644\n--- a/.github/CONTRIBUTING.md\n+++ b/.github/CONTRIBUTING.md\n@@ -21,7 +21,8 @@ Contributions are always welcome! Please use the following guidelines when contr\n - `chore` - Catch all or things that have to do with the build system, etc\n - `examples` - Changes to existing example, or a new example\n * The `COMPONENT` is optional, and may be a single file, directory, or logical component. Can be omitted if commit applies globally\n-5. Run the tests (`cargo test --no-std-features && cargo test --features yaml`)\n+5. Run the tests (`cargo test --features \"yaml unstable\"`)\n+5. Run the lints (`cargo build --features lints`) (requires a nightly compiler)\n 6. `git rebase` into concise commits and remove `--fixup`s (`git rebase -i HEAD~NUM` where `NUM` is number of commits back)\n 7. Push your changes back to your fork (`git push origin $your-branch`)\n 8. Create a pull request! (You can also create the pull request first, and we'll merge when ready. This a good way to discuss proposed changes.)\ndiff --git a/README.md b/README.md\nindex 9e6efce..b74405d 100644\n--- a/README.md\n+++ b/README.md\n@@ -31,7 +31,9 @@ Table of Contents\n * [More Information](#more-information)\n * [Video Tutorials](#video-tutorials)\n * [How to Contribute](#how-to-contribute)\n- * [Running the tests](#running-the-tests)\n+ * [Testing Code](#testing-code)\n+ * [Linting Code](#linting-code)\n+ * [Debugging Code](#debugging-code)\n * [Goals](#goals)\n * [Compatibility Policy](#compatibility-policy)\n * [Minimum Version of Rust](#minimum-version-of-rust)\n@@ -43,288 +45,83 @@ Created by [gh-md-toc](https://github.com/ekalinin/github-markdown-toc)\n \n ## What's New\n \n-Here's what's new in v2.18.0\n+Here's the highlights from v2.0.0 to v2.18.0\n \n * **Completions:** Adds completion support for Microsoft PowerShell! (Thanks to @Arnavion)\n-\n-Here's what's new in v2.17.1\n-\n-* Fixes a bug where using low index multiples was propagated to subcommands\n-\n-Here's what's new in v2.17.0\n-\n * Allows specifying the second to last positional argument as `multiple(true)` (i.e. things such as `mv <files>... <target>`)\n * Adds an `App::get_name` and `App::get_bin_name`\n-\n-Here's what's new in v2.16.4\n-\n-* Fixes bug that caused panic on subcommands with aliases\n * Conflicting argument errors are now symetrical, meaning more consistent and better usage suggestions\n-* Fixes typo in example `13a_enum_values_automatic`\n-* Fixes failing yaml example (#715)\n-* Fixes the `debug` feature (#716)\n-\n-Here's the highlights for v2.16.3\n-\n-* Fixes a bug where the derived display order isn't propagated\n-* **yaml-example:** fixes some inconsistent args in the example\n-\n-Here's the highlights for v2.16.2\n-\n-* Fixes a bug where single quotes are not escaped\n-\n-Here's the highlights for v2.16.1\n-\n-* **Help Message:** fixes a regression bug where args with multiple(true) threw off alignment\n-\n-Here's the highlights for v2.16.0\n-\n * **Completions:** adds automatic ZSH completion script generation support! :tada: :tada:\n-\n-Here's a gif of them in action!\n-\n-![zsh-comppletions](http://i.imgur.com/rwlMbAv.gif)\n-\n-Here's the highlights for v2.15.0\n-\n * **AppSettings:** adds new setting `AppSettings::AllowNegativeNumbers` which functions like `AllowLeadingHyphen` except only allows undefined negative numbers to pass parsing.\n-* Improves some of the documentation of `AppSettings` by moving variants into roughly alphabetical order\n-\n-Here's the highlights for v2.14.1 (Huge thanks to all the contributors who put in a lot of work this cycle! Especially @tormol @nabijaczleweli and @wdv4758h)\n-\n * Stabilize `clap_app!` macro (i.e. no longer need to use `unstable` feature)\n-* Fixes a bug that made determining when to auto-wrap long help messages inconsistent\n-* Fixes fish completions for nested subcommands\n-* Improve documentation around features\n-* Reword docs for `ErrorKind` and `App::settings`\n-* Fix tests that fail when the `suggestions` feature is disabled\n-* Fix the `OsString`-using doc-tests\n-* Tag non-rust code blocks as such instead of ignoring them\n-* Improve some errors about subcommands\n-* Makes sure the doc-tests don't fail before \"missing file\" in YAML tests\n * Deprecate `App::with_defaults`\n-* Make lints not enable other nightly-requiring features\n-\n-Here's the highlights for v2.14.0\n-\n-* One can now alias arguments either visibly (whichc appears in the help text) or invisibly just like subcommands!\n+* One can now alias arguments either visibly (which appears in the help text) or invisibly just like subcommands!\n * The `from_usage` parser now correctly handles non-ascii names / options and help!\n-* Fixes a bug in the `require_delimiter` code which caused some incorrect parses\n-* Fixes various typos in the docs\n-* Various other small performance improvements and enhancements\n-\n-Here's the highlights for v2.13.0\n-\n * **Value Delimiters:** fixes the confusion around implicitly setting value delimiters. (The default is to *not* use a delimiter unless explicitly set)\n-* **Docs:** Updates README.md with new website information and updated video tutorials info\n-* **Docs:** Updates the docs about removing implicit `value_delimiter(true)`\n-* **Docs:** Adds better examples on using default values\n-\n-\n-Here's the highlights for v2.12.1\n-\n-* Fixes a regression-bug where the old `{n}` newline char stopped being replaced a properly re-aligned newline\n-\n-Here's the highlights for v2.12.0\n-\n * Changes the default value delimiter rules (i.e. the default is `use_delimiter(false)` *unless* a setting/method that implies multiple values was used) **[Bugfix that *may* \"break\" code]**\n * If code breaks, simply add `Arg::use_delimiter(true)` to the affected args\n-* Updates the docs for the `Arg::multiple` method WRT value delimiters and default settings\n * Adds ability to hide the possible values from the help text on a per argument basis, instead of command wide\n * Allows for limiting detected terminal width (i.e. wrap at `x` length, unless the terminal width is *smaller*)\n-* Removes some redundant `contains()` checks for minor performance improvements\n-* Fixes a bug where valid args aren't recognized with the `AppSettings::AllowLeadingHyphen` setting\n * `clap` now ignores hard newlines in help messages and properly re-aligns text, but still wraps if the term width is too small\n-* Makes some minor changes to when next line help is automatically used\n * Adds support for the setting `Arg::require_delimiter` from YAML\n-* Removes the verbage about using `'{n}'` to insert newlines in help text from the docs (the normal `\\n` can now be used)\n-* Documents `AppSetting::DisableVersion`\n-\n-Here's the highlights for v2.11.3\n-\n * `clap` no longer requires one to use `{n}` inside help text to insert a newline that is properly aligned. One can now use the normal `\\n`.\n * `clap` now ignores hard newlines in help messages and properly re-aligns text, but still wraps if the term width is too small\n-* Supports setting `Arg::require_delimiter` from YAML\n-\n-Here's the highlights for v2.11.2\n-\n-* Makes some minor changes to when next line help is automatically used for improved wrapping\n-\n-Here's the highlights for v2.11.1\n-\n-* Fixes an issue where settings weren't propogated down through grand-child subcommands\n * Errors can now have custom description\n * Uses `term_size` instead of home-grown solution on Windows\n-* Updates deps with some minor bug fixes\n-\n-\n-Here's the highlights for v2.11.0\n-\n * Adds the ability to wrap help text intelligently on Windows!\n-* Moves docs to [docs.rs!](https://docs.rs/clap/)\n-* Fixes some usage strings that contain both args in groups and ones that conflict with each other\n-* Uses standard conventions for bash completion files, namely `{bin}.bash-completion`\n+* Moves docs to [docs.rs!](https://docs.rs/clap/)!\n * Automatically moves help text to the next line and wraps when term width is determined to be too small, or help text is too long\n * Vastly improves *development* error messages when using YAML\n-* Adds `App::with_defaults` to automatically use `crate_authors!` and `crate_version!` macros\n-* Other minor improvements and bug fixes\n-\n-Here's the highlights for v2.10.4\n-\n-* Fixes a bug where help is wrapped incorrectly and causing a panic with some non-English characters\n-\n-Here's the highlights for v2.10.3\n-\n-* Fixes a bug with non-English characters in help text wrapping, where the character is stripped or causes a panic\n-* Fixes an issue with `strsim` which caused a panic in some scenarios\n * Adds a shorthand way to ignore help text wrapping and use source formatting (i.e. `App::set_term_width(0)`)\n-\n-Here's the highlights for v2.10.2\n-\n-* Fixes a critical bug where the help message is printed twice\n-\n-Here's the highlights for v2.10.1\n-\n * **Help Subcommand:** fixes misleading usage string when using multi-level subcommmands such as `myprog help subcmd1 subcmd2`\n * **YAML:** allows using lists or single values with certain arg declarations for increased ergonomics\n-\n-\n-Here's the highlights for v2.10.0\n-\n-\n * **Fish Shell Completions:** one can generate a basic fish completions script at compile time!\n-* **External SubCommands:** fixes a bug which now correctly preserves external subcommand name along with args to said command (Minor breaking change that breaks no known real world code)\n-* **YAML Documentation:** fixes example 17's incorrect reference to arg_groups instead of groups\n-\n-\n-Here's the highlights for v2.9.3\n-\n * Adds the ability to generate completions to an `io::Write` object\n * Adds an `App::unset_setting` and `App::unset_settings`\n-* Fixes bug where only first arg in list of `required_unless_one` is recognized\n-* Fixes a typo bug `SubcommandsRequired`->`SubcommandRequired`\n-\n-\n-Here's the highlights for v2.9.2\n-\n-\n-* fixes bug where --help and --version short weren't added to the completion list\n-* improves completions allowing multiple bins to have seperate completion files\n-\n-Here's the highlights for v2.9.0\n-\n * **Completions:** one can now [generate a bash completions](https://docs.rs/clap/2.9.0/clap/struct.App.html#method.gen_completions) script at compile time! These completions work with options using [possible values](https://docs.rs/clap/2.9.0/clap/struct.Arg.html#method.possible_values), [subcommand aliases](https://docs.rs/clap/2.9.0/clap/struct.App.html#method.aliases), and even multiple levels of subcommands\n-* Minor bug fixes when using `AppSettings::TrailingVarArg` and `AppSettings::AllowLeadingHyphen`\n-\n-Here's the highlights for v2.8.0\n-\n * **Arg:** adds new optional setting [`Arg::require_delimiter`](https://docs.rs/clap/2.8.0/clap/struct.Arg.html#method.require_delimiter) which requires val delimiter to parse multiple values\n * The terminal sizing portion has been factored out into a separate crate, [term_size](https://crates.io/crates/term_size)\n-* Minor bug fixes\n-\n-\n-Here's the highlights for v2.7.1\n-\n-* **Options:**\n- * options using multiple values and delimiters no longer parse additional values after a trailing space (i.e. `prog -o 1,2 file.txt` parses as `1,2` for `-o` and `file.txt` for a positional arg)\n- * using options using multiple values and with an `=` no longer parse args after the trailing space as values (i.e. `prog -o=1 file.txt` parses as `1` for `-o` and `file.txt` for a positional arg)\n-\n-Here's the highlights for v2.7.0\n-\n+* Options using multiple values and delimiters no longer parse additional values after a trailing space (i.e. `prog -o 1,2 file.txt` parses as `1,2` for `-o` and `file.txt` for a positional arg)\n+* Using options using multiple values and with an `=` no longer parse args after the trailing space as values (i.e. `prog -o=1 file.txt` parses as `1` for `-o` and `file.txt` for a positional arg)\n * **Usage Strings:** `[FLAGS]` and `[ARGS]` are no longer blindly added to usage strings, instead only when applicable\n * `arg_enum!`: allows using more than one meta item, or things like `#[repr(C)]` with `arg_enum!`s\n * `App::print_help`: now prints the same as would have been printed by `--help` or the like\n-* **Help Messages:**\n- * prevents invoking `<cmd> help help` and displaying incorrect help message\n- * subcommand help messages requested via `<cmd> help <sub>` now correctly match `<cmd> <sub> --help`\n-* **`ArgGroup`s:**\n- * one can now specify groups which require AT LEAST one of the args\n- * allows adding multiple ArgGroups per Arg\n- * **Documentation:** vastly improves `ArgGroup` docs by adding better examples\n-* **Documentation:** fixes a bunch of typos in the documentation\n-\n-Here's the highlights for v2.6.0\n-\n+* Prevents invoking `<cmd> help help` and displaying incorrect help message\n+* Subcommand help messages requested via `<cmd> help <sub>` now correctly match `<cmd> <sub> --help`\n+* One can now specify groups which require AT LEAST one of the args\n+* Allows adding multiple ArgGroups per Arg\n * **Global Settings:** One can now set an `AppSetting` which is propogated down through child subcommands\n * **Terminal Wrapping:** Allows wrapping at specified term width (Even on Windows!) (can now set an absolute width to \"smart\" wrap at)\n * **SubCommands/Aliases:** adds support for visible aliases for subcommands (i.e. aliases that are dipslayed in the help message)\n * **Subcommands/Aliases:** when viewing the help of an alias, it now display help of the aliased subcommand\n-* Improves the default usage string when only a single positional arg is present\n * Adds new setting to stop delimiting values with `--` or `AppSettings::TrailingVarArg`\n-* `App::before_help` and `App::after_help` now correctly wrap\n-* Fixes bug where positional args are printed out of order when using templates\n-* Fixes bug where one can't override the auto-generated version or help flags\n-* Fixes issue where `App::before_help` wasn't printed\n-* Fixes a failing windows build\n-* Fixes bug where new color settings couldn't be converted from strings\n-* Adds missing YAML methods for App and Arg\n-* Allows printing version to any io::Write object\n-* Removes extra newline from help and version output\n-\n-Here's what's new in v.2.5.2\n-\n-* Removes trailing newlines from help and version output\n-* Allows printing version to any io::Write object\n-* Inter-links all types and pages\n-* Makes all publicly available types viewable in docs\n-* Fixes bug where one can't override version or help flags\n-* Fixes bug where args are printed out of order when using templates\n-* Fixes issue where `App::before_help` wasn't printed properly\n-\n-Here's what's new in v.2.5.0\n-\n * Subcommands now support aliases - think of them as hidden subcommands that dispatch to said subcommand automatically\n-\n-Here's what's new in v2.4.3\n-\n-* Bug Fixes\n- * Usage strings get de-deuplicated when there are args which are also part ``ArgGroup`s`\n- * Fixed times when `ArgGroup`s are duplicated in usage strings\n-* Improvements\n- * Positional arguments which are part of a group are now formatted in a more readable way (fewer brackets)\n- * Positional arguments use the standard `<>` brackets to reduce confusion\n- * The default help string for the `help` subcommand has been shortened to fit in 80 columns\n-\n-Here's the highlights from v2.4.0\n-\n+* Fixed times when `ArgGroup`s are duplicated in usage strings\n * **Before Help:** adds support for displaying info before help message\n * **Required Unless:** adds support for allowing args that are required unless certain other args are present\n-* Bug fixes\n-\n-Here's the highlights from v2.3.0\n-\n * **New Help Template Engine!**: Now you have full control over the layout of your help message. Major thanks to @hgrecco\n * **Pull crate Authors from Cargo.toml**: One can now use the `crate_authors!` macro to automatically pull the crate authors from their Cargo.toml file\n * **Colored Help Messages**: Help messages can now be optionally colored (See the `AppSettings::ColoredHelp` setting). Screenshot below.\n-* A bunch of bug fixes\n-\n-Here's the highlights from v2.2.1\n-\n * **Help text auto wraps and aligns at for subcommands too!** - Long help strings of subcommands will now properly wrap and align to term width on Linux and OS X. This can be turned off as well.\n-* Bug fixes\n-\n-An example of the optional colored help:\n-\n-![screenshot](http://i.imgur.com/7fs2h5j.png)\n-\n-Here's the highlights from v2.2.0\n-\n * **Help text auto wraps and aligns at term width!** - Long help strings will now properly wrap and align to term width on Linux and OS X (and presumably Unix too). This can be turned off as well.\n * **Can customize the order of opts, flags, and subcommands in help messages** - Instead of using the default alphabetical order, you can now re-arrange the order of your args and subcommands in help message. This helps to emphasize more popular or important options.\n- * **Can auto-derive the order from declaration order** - Have a bunch of args or subcommmands to re-order? You can now just derive the order from the declaration order!\n+* **Can auto-derive the order from declaration order** - Have a bunch of args or subcommmands to re-order? You can now just derive the order from the declaration order!\n * **Help subcommand now accepts other subcommands as arguments!** - Similar to other CLI precedents, the `help` subcommand can now accept other subcommands as arguments to display their help message. i.e. `$ myprog help mysubcmd` (*Note* these can even be nested heavily such as `$ myprog help subcmd1 subcmd2 subcmd3` etc.)\n+* **Default Values**: Args can now specify default values\n+* **Next Line Help**: Args can have help strings on the line following the argument (useful for long arguments, or those with many values). This can be set command-wide or for individual args\n \n-* Other minor bug fixes\n+Here's a gif of them in action!\n+\n+![zsh-comppletions](http://i.imgur.com/rwlMbAv.gif)\n \n An example of the help text wrapping at term width:\n \n ![screenshot](http://i.imgur.com/PAJzJJG.png)\n \n-In v2.1.2\n+An example of the optional colored help:\n+\n+![screenshot](http://i.imgur.com/7fs2h5j.png)\n \n- * **Default Values**: Args can now specify default values\n- * **Next Line Help**: Args can have help strings on the line following the argument (useful for long arguments, or those with many values). This can be set command-wide or for individual args\n- * **Documentation Examples**: The examples in the documentation have been vastly improved\n \n For full details, see [CHANGELOG.md](https://github.com/kbknapp/clap-rs/blob/master/CHANGELOG.md)\n \n@@ -697,6 +494,7 @@ features = [ \"suggestions\", \"color\" ]\n #### Opt-in features\n \n * **\"yaml\"**: Enables building CLIs from YAML documents. (builds dependency `yaml-rust`)\n+* **\"unstable\"**: Enables unstable `clap` features that may change from release to release\n \n ### Dependencies Tree\n \n@@ -707,6 +505,7 @@ The following graphic depicts `clap`s dependency graph (generated using [cargo-g\n * **Blue** Color: Dev dependency, only used while developing.\n \n ![clap dependencies](clap_dep_graph.png)\n+\n ### More Information\n \n You can find complete documentation on the [docs.rs](https://docs.rs/clap/) for this project.\n@@ -727,20 +526,65 @@ Another really great way to help is if you find an interesting, or helpful way i\n \n Please read [CONTRIBUTING.md](.github/CONTRIBUTING.md) before you start contributing.\n \n+\n+### Testing Code\n+\n To test with all features both enabled and disabled, you can run theese commands:\n \n ```sh\n $ cargo test --no-default-features\n-$ cargo test --features yaml\n+$ cargo test --features \"yaml unstable\"\n ```\n \n-If you have a nightly compiler you can append `--features lints` to both commands\n-to get style warnings and code smells; If you get one from code you think is fine,\n-you can ignore it by prepending `#[cfg_attr(feature=\"lints\", allow(lint_name))]`\n-to the function or impl block.\n+Alternatively, if you have [`just`](https://github.com/casey/just) installed you can run the prebuilt recipies. *Not* using `just` is prfeclty fine as well, it simply bundles commands automatically.\n+\n+For example, to test the code, as above simply run:\n+\n+```sh\n+$ just run-tests`\n+```\n+\n+From here on, I will lis the appropriate `cargo` command as well as the `just` command.\n+\n+Sometimes it's helpful to only run a subset of the tests, which can be done via:\n+\n+```sh\n+$ cargo test --test <test_name>\n+\n+# Or\n+\n+$ just run-test <test_name>\n+```\n \n-If you are debugging (or just trying to understand the code) you can enable the\n-\"debug\" feature which will trace function calls and brances in some parts of the code.\n+### Linting Code\n+\n+During the CI process `clap` runs against many different lints using [`clippy`](https://github.com/Manishearth/rust-clippy). In order to check if these lints pass on your own computer prior to submitting a PR you'll need a nightly compiler.\n+\n+In order to check the code for lints run either:\n+\n+```sh\n+$ rustup override add nightly\n+$ cargo build --features lints\n+$ rustup override remove\n+\n+# Or\n+\n+$ just lint\n+```\n+\n+### Debugging Code\n+\n+Another helpful technique is to see the `clap` debug output while developing features. In order to see the debug output while running the full test suite or individual tests, run:\n+\n+```sh\n+$ cargo test --features debug\n+\n+# Or for individual tests\n+$ cargo test --test <test_name> --features debug\n+\n+# The corresponding just command for individual debugging tests is:\n+$ just debug <test_name>\n+```\n \n ### Goals\n \n", "diff --git a/broker/src/main/java/io/camunda/zeebe/broker/bootstrap/BrokerStartupProcess.java b/broker/src/main/java/io/camunda/zeebe/broker/bootstrap/BrokerStartupProcess.java\nindex 52fa3a9..d81c27a 100644\n--- a/broker/src/main/java/io/camunda/zeebe/broker/bootstrap/BrokerStartupProcess.java\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/bootstrap/BrokerStartupProcess.java\n@@ -50,21 +50,20 @@ public final class BrokerStartupProcess {\n // must be executed before any disk space usage listeners are registered\n result.add(new DiskSpaceUsageMonitorStep());\n }\n-\n result.add(new MonitoringServerStep());\n result.add(new BrokerAdminServiceStep());\n+\n result.add(new ClusterServicesCreationStep());\n+ result.add(new ClusterServicesStep());\n \n result.add(new CommandApiServiceStep());\n result.add(new SubscriptionApiStep());\n-\n- result.add(new ClusterServicesStep());\n+ result.add(new LeaderManagementRequestHandlerStep());\n \n if (config.getGateway().isEnable()) {\n result.add(new EmbeddedGatewayServiceStep());\n }\n \n- result.add(new LeaderManagementRequestHandlerStep());\n result.add(new PartitionManagerStep());\n \n return result;\n"]
2
["eb51316cdfdc7258d287ba13b67ef2f42bd2b8f6", "3e0c4cbf91fe5efc9b93baba93e4df93ef4ab5cd"]
["docs", "refactor"]
cancel in-progress dep update jobs when a new one arrives [skip ci],get ip from forwarded header
["diff --git a/.github/workflows/update-deps.yml b/.github/workflows/update-deps.yml\nindex 3a71e29..25f6f27 100644\n--- a/.github/workflows/update-deps.yml\n+++ b/.github/workflows/update-deps.yml\n@@ -4,6 +4,11 @@ on:\n # run every 24 hours at midnight\n - cron: \"0 */24 * * *\"\n workflow_dispatch:\n+\n+concurrency:\n+ group: ${{ github.repository }}-${{ github.head_ref || github.sha }}-${{ github.workflow }}\n+ cancel-in-progress: true\n+\n jobs:\n generate_updates:\n runs-on: ubuntu-latest\n", "diff --git a/kousa/lib/broth/socket_handler.ex b/kousa/lib/broth/socket_handler.ex\nindex d142135..5828f30 100644\n--- a/kousa/lib/broth/socket_handler.ex\n+++ b/kousa/lib/broth/socket_handler.ex\n@@ -22,7 +22,7 @@ defmodule Broth.SocketHandler do\n ## initialization boilerplate\n \n @impl true\n- def init(request = %{peer: {ip, _reverse_port}}, _state) do\n+ def init(request, _state) do\n props = :cowboy_req.parse_qs(request)\n \n compression =\n@@ -37,10 +37,16 @@ defmodule Broth.SocketHandler do\n _ -> :json\n end\n \n+ ip =\n+ case request.headers do\n+ %{\"x-forwarded-for\" => v} -> v\n+ _ -> nil\n+ end\n+\n state = %__MODULE__{\n awaiting_init: true,\n user_id: nil,\n- ip: IP.to_string(ip),\n+ ip: ip,\n encoding: encoding,\n compression: compression,\n callers: get_callers(request)\ndiff --git a/kousa/test/_support/ws_client.ex b/kousa/test/_support/ws_client.ex\nindex aeca704..125da17 100644\n--- a/kousa/test/_support/ws_client.ex\n+++ b/kousa/test/_support/ws_client.ex\n@@ -19,7 +19,9 @@ defmodule BrothTest.WsClient do\n \n @api_url\n |> Path.join(\"socket\")\n- |> WebSockex.start_link(__MODULE__, nil, extra_headers: [{\"user-agent\", ancestors}])\n+ |> WebSockex.start_link(__MODULE__, nil,\n+ extra_headers: [{\"user-agent\", ancestors}, {\"x-forwarded-for\", \"127.0.0.1\"}]\n+ )\n end\n \n ###########################################################################\n"]
2
["c2300c94c6b7d1599387272b616e1d79e93723c7", "2f5718743a830d40ddf272ad46f253dbb6d08cff"]
["ci", "fix"]
use a closure,create mock img server
["diff --git a/ibis/expr/analysis.py b/ibis/expr/analysis.py\nindex bb17a7a..975c658 100644\n--- a/ibis/expr/analysis.py\n+++ b/ibis/expr/analysis.py\n@@ -39,7 +39,9 @@ def sub_for(expr, substitutions):\n An Ibis expression\n \"\"\"\n \n- def fn(node, mapping={k.op(): v for k, v in substitutions}):\n+ mapping = {k.op(): v for k, v in substitutions}\n+\n+ def fn(node):\n try:\n return mapping[node]\n except KeyError:\n", "diff --git a/scripts/gulp/tasks/test.ts b/scripts/gulp/tasks/test.ts\nindex 8014b12..d10c1aa 100644\n--- a/scripts/gulp/tasks/test.ts\n+++ b/scripts/gulp/tasks/test.ts\n@@ -26,12 +26,18 @@ task('test.imageserver', () => {\n function handleRequest(req, res) {\n const urlParse = url.parse(req.url, true);\n \n+ res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Access-Control-Allow-Methods', 'GET');\n+ res.setHeader('Connection', 'keep-alive');\n+ res.setHeader('Age', '0');\n+ res.setHeader('cache-control', 'no-store');\n+\n if (urlParse.pathname === '/reset') {\n console.log('Image Server Reset');\n console.log('---------------------------');\n requestedUrls.length = 0;\n start = Date.now();\n- res.setHeader('Access-Control-Allow-Origin', '*');\n+ res.setHeader('Content-Type', 'text/plain');\n res.end('reset');\n return;\n }\n@@ -48,9 +54,8 @@ task('test.imageserver', () => {\n \n setTimeout(() => {\n res.setHeader('Content-Type', 'image/svg+xml');\n- res.setHeader('Access-Control-Allow-Origin', '*');\n res.end(`<svg xmlns=\"http://www.w3.org/2000/svg\" xmlns:xlink=\"http://www.w3.org/1999/xlink\"\n- style=\"background-color: ${color}; width: ${width}px; height: ${height}px;\">\n+ viewBox=\"0 0 ${width} ${height}\" style=\"background-color: ${color};\">\n <text x=\"5\" y=\"22\" style=\"font-family: Courier; font-size: 24px\">${id}</text>\n </svg>`);\n }, delay);\n"]
2
["ad52e1d67fd77f0b6a73fbf989b33f9abf395ecc", "32b76173a259ea1993298289b436cf10c1e800bf"]
["refactor", "test"]
rename step,dashboard removed unused code
["diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml\nindex e81d897..5c3ee6b 100644\n--- a/.github/workflows/ci.yml\n+++ b/.github/workflows/ci.yml\n@@ -45,7 +45,7 @@ jobs:\n - name: Install dependencies\n run: pnpm install\n \n- - name: Typecheck\n+ - name: Build (stub)\n run: pnpm build:stub\n \n - name: Typecheck\n", "diff --git a/api/chalicelib/core/dashboard.py b/api/chalicelib/core/dashboard.py\nindex e5bc800..1afa105 100644\n--- a/api/chalicelib/core/dashboard.py\n+++ b/api/chalicelib/core/dashboard.py\n@@ -126,13 +126,6 @@ SESSIONS_META_FIELDS = {\"revId\": \"rev_id\",\n \"browser\": \"user_browser\"}\n \n \n-def __get_domains_errors_neutral(rows):\n- neutral = {l: 0 for l in [i for k in [list(v.keys()) for v in rows] for i in k]}\n- if len(neutral.keys()) == 0:\n- neutral = {\"All\": 0}\n- return neutral\n-\n-\n def get_processed_sessions(project_id, startTimestamp=TimeUTC.now(delta_days=-1),\n endTimestamp=TimeUTC.now(),\n density=7, **args):\n"]
2
["34875bc0e59b43d9041903101c823d25ec194a21", "b7b1d2b315443e1854403c8fe8f871c4632b5d31"]
["ci", "refactor"]
update build,reintroduce timeout for assertion The timeout had been removed by a previous commit. Without the timeout the test might be flaky. Also removed obsolete code
["diff --git a/bootstrap/scripts/publish-patch.sh b/bootstrap/scripts/publish-patch.sh\nindex a1b6f12..0d849a5 100755\n--- a/bootstrap/scripts/publish-patch.sh\n+++ b/bootstrap/scripts/publish-patch.sh\n@@ -5,4 +5,4 @@ lerna version patch\n lerna publish from-package -y\n git push\n \n-./pack_and_install.sh\n\\ No newline at end of file\n+./bootstrap/scripts/pack_and_install.sh\n\\ No newline at end of file\n", "diff --git a/engine/src/test/java/io/camunda/zeebe/engine/processing/streamprocessor/StreamProcessorReplayModeTest.java b/engine/src/test/java/io/camunda/zeebe/engine/processing/streamprocessor/StreamProcessorReplayModeTest.java\nindex d0ee4f3..c2ab83c 100644\n--- a/engine/src/test/java/io/camunda/zeebe/engine/processing/streamprocessor/StreamProcessorReplayModeTest.java\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/processing/streamprocessor/StreamProcessorReplayModeTest.java\n@@ -13,6 +13,7 @@ import static io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent.ACTI\n import static io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent.ELEMENT_ACTIVATING;\n import static java.util.function.Predicate.isEqual;\n import static org.assertj.core.api.Assertions.assertThat;\n+import static org.awaitility.Awaitility.await;\n import static org.mockito.ArgumentMatchers.any;\n import static org.mockito.ArgumentMatchers.anyLong;\n import static org.mockito.ArgumentMatchers.eq;\n@@ -30,7 +31,6 @@ import io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent;\n import io.camunda.zeebe.streamprocessor.StreamProcessor;\n import io.camunda.zeebe.streamprocessor.StreamProcessor.Phase;\n import io.camunda.zeebe.streamprocessor.StreamProcessorMode;\n-import org.awaitility.Awaitility;\n import org.junit.Rule;\n import org.junit.Test;\n import org.mockito.InOrder;\n@@ -71,7 +71,7 @@ public final class StreamProcessorReplayModeTest {\n // when\n startStreamProcessor(replayUntilEnd);\n \n- Awaitility.await()\n+ await()\n .untilAsserted(\n () -> assertThat(getCurrentPhase(replayUntilEnd)).isEqualTo(Phase.PROCESSING));\n \n@@ -163,7 +163,7 @@ public final class StreamProcessorReplayModeTest {\n command().processInstance(ACTIVATE_ELEMENT, RECORD),\n event().processInstance(ELEMENT_ACTIVATING, RECORD).causedBy(0));\n \n- Awaitility.await(\"should have replayed first events\")\n+ await(\"should have replayed first events\")\n .until(replayContinuously::getLastSuccessfulProcessedRecordPosition, (pos) -> pos > 0);\n \n // when\n@@ -210,7 +210,7 @@ public final class StreamProcessorReplayModeTest {\n command().processInstance(ACTIVATE_ELEMENT, RECORD),\n event().processInstance(ELEMENT_ACTIVATING, RECORD).causedBy(0));\n \n- Awaitility.await(\"should have replayed first events\")\n+ await(\"should have replayed first events\")\n .until(replayContinuously::getLastSuccessfulProcessedRecordPosition, (pos) -> pos > 0);\n streamProcessor.pauseProcessing().join();\n replayContinuously.writeBatch(\n@@ -244,7 +244,7 @@ public final class StreamProcessorReplayModeTest {\n // then\n verify(eventApplier, TIMEOUT).applyState(anyLong(), eq(ELEMENT_ACTIVATING), any());\n \n- Awaitility.await()\n+ await()\n .untilAsserted(\n () -> {\n final var lastProcessedPosition = getLastProcessedPosition(replayContinuously);\n@@ -273,8 +273,7 @@ public final class StreamProcessorReplayModeTest {\n \n verify(eventApplier, TIMEOUT).applyState(anyLong(), eq(ELEMENT_ACTIVATING), any());\n \n- Awaitility.await()\n- .until(() -> getLastProcessedPosition(replayContinuously), isEqual(commandPosition));\n+ await().until(() -> getLastProcessedPosition(replayContinuously), isEqual(commandPosition));\n \n // then\n assertThat(replayContinuously.getLastSuccessfulProcessedRecordPosition())\n@@ -285,7 +284,6 @@ public final class StreamProcessorReplayModeTest {\n @Test\n public void shouldNotSetLastProcessedPositionIfLessThanSnapshotPosition() {\n // given\n- final var commandPositionBeforeSnapshot = 1L;\n final var snapshotPosition = 2L;\n \n startStreamProcessor(replayContinuously);\n@@ -298,23 +296,20 @@ public final class StreamProcessorReplayModeTest {\n // when\n startStreamProcessor(replayContinuously);\n \n- Awaitility.await()\n+ await()\n .untilAsserted(\n () -> assertThat(getCurrentPhase(replayContinuously)).isEqualTo(Phase.REPLAY));\n \n- final var eventPosition =\n- replayContinuously.writeEvent(\n- ELEMENT_ACTIVATING,\n- RECORD,\n- writer -> writer.sourceRecordPosition(commandPositionBeforeSnapshot));\n-\n // then\n final var lastProcessedPositionState = replayContinuously.getLastProcessedPositionState();\n \n- assertThat(lastProcessedPositionState.getLastSuccessfulProcessedRecordPosition())\n- .describedAs(\n- \"Expected that the last processed position is not less than the snapshot position\")\n- .isEqualTo(snapshotPosition);\n+ await()\n+ .untilAsserted(\n+ () ->\n+ assertThat(lastProcessedPositionState.getLastSuccessfulProcessedRecordPosition())\n+ .describedAs(\n+ \"Expected that the last processed position is not less than the snapshot position\")\n+ .isEqualTo(snapshotPosition));\n }\n \n private StreamProcessor startStreamProcessor(final StreamProcessorRule streamProcessorRule) {\n"]
2
["3fcfb20b0feb371b357edc42fcb7c87085c9b82a", "0d23f1b3ed22e615b9611bb4eae01d2241e64dff"]
["build", "refactor"]
remove unused branches and ignore envrc file,create dashboards from imported templates
["diff --git a/.github/workflows/ibis-backends-cloud.yml b/.github/workflows/ibis-backends-cloud.yml\nindex 2003e8e..7c7fd26 100644\n--- a/.github/workflows/ibis-backends-cloud.yml\n+++ b/.github/workflows/ibis-backends-cloud.yml\n@@ -5,9 +5,12 @@ on:\n # Skip the backend suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n+ - \"**/*.md\"\n+ - \"**/*.qmd\"\n+ - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n- - quarto\n \n permissions:\n # this allows extractions/setup-just to list releases for `just` at a higher\ndiff --git a/.github/workflows/ibis-backends-skip-helper.yml b/.github/workflows/ibis-backends-skip-helper.yml\nindex 5d5f3f7..0471994 100644\n--- a/.github/workflows/ibis-backends-skip-helper.yml\n+++ b/.github/workflows/ibis-backends-skip-helper.yml\n@@ -9,20 +9,20 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_backends:\ndiff --git a/.github/workflows/ibis-backends.yml b/.github/workflows/ibis-backends.yml\nindex 4a1cae9..30e6c1a 100644\n--- a/.github/workflows/ibis-backends.yml\n+++ b/.github/workflows/ibis-backends.yml\n@@ -8,10 +8,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the backend suite if all changes are docs\n paths-ignore:\n@@ -19,10 +19,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\nindex 1adda11..b528a30 100644\n--- a/.github/workflows/ibis-docs-lint.yml\n+++ b/.github/workflows/ibis-docs-lint.yml\n@@ -5,12 +5,10 @@ on:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-main-skip-helper.yml b/.github/workflows/ibis-main-skip-helper.yml\nindex a5fdc6f..0fb5dea 100644\n--- a/.github/workflows/ibis-main-skip-helper.yml\n+++ b/.github/workflows/ibis-main-skip-helper.yml\n@@ -8,19 +8,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_core:\ndiff --git a/.github/workflows/ibis-main.yml b/.github/workflows/ibis-main.yml\nindex aa31436..0b1536a 100644\n--- a/.github/workflows/ibis-main.yml\n+++ b/.github/workflows/ibis-main.yml\n@@ -7,20 +7,20 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the test suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-tpch-queries-skip-helper.yml b/.github/workflows/ibis-tpch-queries-skip-helper.yml\nindex 1f1c0bc..f10fb8d 100644\n--- a/.github/workflows/ibis-tpch-queries-skip-helper.yml\n+++ b/.github/workflows/ibis-tpch-queries-skip-helper.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-tpch-queries.yml b/.github/workflows/ibis-tpch-queries.yml\nindex b4f8a48..9e65a61 100644\n--- a/.github/workflows/ibis-tpch-queries.yml\n+++ b/.github/workflows/ibis-tpch-queries.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/nix-skip-helper.yml b/.github/workflows/nix-skip-helper.yml\nindex 677b4d7..e0ab8f7 100644\n--- a/.github/workflows/nix-skip-helper.yml\n+++ b/.github/workflows/nix-skip-helper.yml\n@@ -9,19 +9,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n jobs:\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\nindex f2dd3f0..7ea9e26 100644\n--- a/.github/workflows/nix.yml\n+++ b/.github/workflows/nix.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\n", "diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex 590f5ea..bd74e95 100644\n--- a/CHANGELOG.md\n+++ b/CHANGELOG.md\n@@ -1,6 +1,7 @@\n ## v2.0.0-alpha.6 [unreleased]\n \n ### Features\n+1. [12496](https://github.com/influxdata/influxdb/pull/12496): Add ability to import a dashboard\n \n ### Bug Fixes\n \ndiff --git a/ui/src/dashboards/actions/v2/index.ts b/ui/src/dashboards/actions/v2/index.ts\nindex ad0d13d..35babc7 100644\n--- a/ui/src/dashboards/actions/v2/index.ts\n+++ b/ui/src/dashboards/actions/v2/index.ts\n@@ -16,6 +16,7 @@ import {\n removeDashboardLabels as removeDashboardLabelsAJAX,\n updateView as updateViewAJAX,\n } from 'src/dashboards/apis/v2'\n+import {client} from 'src/utils/api'\n \n // Actions\n import {notify} from 'src/shared/actions/notifications'\n@@ -25,6 +26,10 @@ import {\n DeleteTimeRangeAction,\n } from 'src/dashboards/actions/v2/ranges'\n import {setView, SetViewAction} from 'src/dashboards/actions/v2/views'\n+import {\n+ importDashboardSucceeded,\n+ importDashboardFailed,\n+} from 'src/shared/copy/notifications'\n \n // Utils\n import {\n@@ -38,7 +43,7 @@ import * as copy from 'src/shared/copy/notifications'\n // Types\n import {RemoteDataState} from 'src/types'\n import {PublishNotificationAction} from 'src/types/actions/notifications'\n-import {CreateCell} from '@influxdata/influx'\n+import {CreateCell, IDashboardTemplate} from '@influxdata/influx'\n import {Dashboard, NewView, Cell} from 'src/types/v2'\n import {ILabel} from '@influxdata/influx'\n \n@@ -201,6 +206,19 @@ export const getDashboardsAsync = () => async (\n }\n }\n \n+export const createDashboardFromTemplate = (\n+ template: IDashboardTemplate,\n+ orgID: string\n+) => async dispatch => {\n+ try {\n+ await client.dashboards.createFromTemplate(template, orgID)\n+\n+ dispatch(notify(importDashboardSucceeded()))\n+ } catch (error) {\n+ dispatch(notify(importDashboardFailed(error)))\n+ }\n+}\n+\n export const importDashboardAsync = (dashboard: Dashboard) => async (\n dispatch: Dispatch<Action>\n ): Promise<void> => {\ndiff --git a/ui/src/dashboards/components/ImportDashboardOverlay.tsx b/ui/src/dashboards/components/ImportDashboardOverlay.tsx\nindex 37ef80b..84216c3 100644\n--- a/ui/src/dashboards/components/ImportDashboardOverlay.tsx\n+++ b/ui/src/dashboards/components/ImportDashboardOverlay.tsx\n@@ -1,70 +1,80 @@\n+// Libraries\n import React, {PureComponent} from 'react'\n import _ from 'lodash'\n+import {connect} from 'react-redux'\n \n-import Container from 'src/clockface/components/overlays/OverlayContainer'\n-import Heading from 'src/clockface/components/overlays/OverlayHeading'\n-import Body from 'src/clockface/components/overlays/OverlayBody'\n-import DragAndDrop from 'src/shared/components/DragAndDrop'\n+// Constants\n import {dashboardImportFailed} from 'src/shared/copy/notifications'\n \n-import {Dashboard} from 'src/types/v2'\n-import {Notification} from 'src/types/notifications'\n+// Actions\n+import {notify as notifyAction} from 'src/shared/actions/notifications'\n+import {getDashboardsAsync} from 'src/dashboards/actions/v2'\n \n-interface Props {\n+// Types\n+import ImportOverlay from 'src/shared/components/ImportOverlay'\n+import {createDashboardFromTemplate as createDashboardFromTemplateAction} from 'src/dashboards/actions/v2'\n+\n+interface OwnProps {\n onDismissOverlay: () => void\n- onImportDashboard: (dashboard: Dashboard) => void\n- notify: (message: Notification) => void\n+ orgID: string\n+ isVisible: boolean\n+}\n+interface DispatchProps {\n+ notify: typeof notifyAction\n+ createDashboardFromTemplate: typeof createDashboardFromTemplateAction\n+ populateDashboards: typeof getDashboardsAsync\n }\n+\n+type Props = OwnProps & DispatchProps\n+\n class ImportDashboardOverlay extends PureComponent<Props> {\n constructor(props: Props) {\n super(props)\n }\n \n public render() {\n- const {onDismissOverlay} = this.props\n+ const {isVisible, onDismissOverlay} = this.props\n \n return (\n- <Container maxWidth={800}>\n- <Heading title=\"Import Dashboard\" onDismiss={onDismissOverlay} />\n- <Body>\n- <DragAndDrop\n- submitText=\"Upload Dashboard\"\n- fileTypesToAccept={this.validFileExtension}\n- handleSubmit={this.handleUploadDashboard}\n- />\n- </Body>\n- </Container>\n+ <ImportOverlay\n+ isVisible={isVisible}\n+ onDismissOverlay={onDismissOverlay}\n+ resourceName=\"Dashboard\"\n+ onSubmit={this.handleUploadDashboard}\n+ />\n )\n }\n \n- private get validFileExtension(): string {\n- return '.json'\n- }\n-\n- private handleUploadDashboard = (\n- uploadContent: string,\n- fileName: string\n- ): void => {\n- const {notify, onImportDashboard, onDismissOverlay} = this.props\n- const fileExtensionRegex = new RegExp(`${this.validFileExtension}$`)\n- if (!fileName.match(fileExtensionRegex)) {\n- notify(dashboardImportFailed('Please import a JSON file'))\n- return\n- }\n+ private handleUploadDashboard = async (\n+ uploadContent: string\n+ ): Promise<void> => {\n+ const {\n+ notify,\n+ createDashboardFromTemplate,\n+ onDismissOverlay,\n+ populateDashboards,\n+ orgID,\n+ } = this.props\n \n try {\n- const {dashboard} = JSON.parse(uploadContent)\n+ const template = JSON.parse(uploadContent)\n \n- if (!_.isEmpty(dashboard)) {\n- onImportDashboard(dashboard)\n- onDismissOverlay()\n- } else {\n- notify(dashboardImportFailed('No dashboard found in file'))\n- }\n+ await createDashboardFromTemplate(template, orgID)\n+ await populateDashboards()\n+\n+ onDismissOverlay()\n } catch (error) {\n notify(dashboardImportFailed(error))\n }\n }\n }\n+const mdtp: DispatchProps = {\n+ notify: notifyAction,\n+ createDashboardFromTemplate: createDashboardFromTemplateAction,\n+ populateDashboards: getDashboardsAsync,\n+}\n \n-export default ImportDashboardOverlay\n+export default connect<{}, DispatchProps, OwnProps>(\n+ null,\n+ mdtp\n+)(ImportDashboardOverlay)\ndiff --git a/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx b/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\nindex d6b299f..1ff7f47 100644\n--- a/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\n+++ b/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\n@@ -2,15 +2,15 @@\n import React, {PureComponent} from 'react'\n import {InjectedRouter} from 'react-router'\n import {connect} from 'react-redux'\n-import {isEmpty} from 'lodash'\n+import {get} from 'lodash'\n \n // Components\n import DashboardsIndexContents from 'src/dashboards/components/dashboard_index/DashboardsIndexContents'\n import {Page} from 'src/pageLayout'\n import SearchWidget from 'src/shared/components/search_widget/SearchWidget'\n import AddResourceDropdown from 'src/shared/components/AddResourceDropdown'\n-import ImportOverlay from 'src/shared/components/ImportOverlay'\n import ExportOverlay from 'src/shared/components/ExportOverlay'\n+import ImportDashboardOverlay from 'src/dashboards/components/ImportDashboardOverlay'\n \n // APIs\n import {createDashboard, cloneDashboard} from 'src/dashboards/apis/v2/'\n@@ -32,10 +32,7 @@ import {DEFAULT_DASHBOARD_NAME} from 'src/dashboards/constants/index'\n import {\n dashboardSetDefaultFailed,\n dashboardCreateFailed,\n- dashboardImported,\n- dashboardImportFailed,\n } from 'src/shared/copy/notifications'\n-import {cantImportInvalidResource} from 'src/shared/copy/v2/notifications'\n \n // Types\n import {Notification} from 'src/types/notifications'\n@@ -197,24 +194,6 @@ class DashboardIndex extends PureComponent<Props, State> {\n this.props.handleDeleteDashboard(dashboard)\n }\n \n- private handleImportDashboard = async (\n- importString: string\n- ): Promise<void> => {\n- const {notify} = this.props\n- try {\n- const resource = JSON.parse(importString)\n-\n- if (isEmpty(resource)) {\n- notify(cantImportInvalidResource('Dashboard'))\n- return\n- }\n- this.handleToggleImportOverlay()\n- notify(dashboardImported())\n- } catch (error) {\n- notify(dashboardImportFailed(error))\n- }\n- }\n-\n private handleFilterDashboards = (searchTerm: string): void => {\n this.setState({searchTerm})\n }\n@@ -229,13 +208,13 @@ class DashboardIndex extends PureComponent<Props, State> {\n \n private get importOverlay(): JSX.Element {\n const {isImportingDashboard} = this.state\n+ const {orgs} = this.props\n \n return (\n- <ImportOverlay\n- isVisible={isImportingDashboard}\n- resourceName=\"Dashboard\"\n+ <ImportDashboardOverlay\n onDismissOverlay={this.handleToggleImportOverlay}\n- onSubmit={this.handleImportDashboard}\n+ orgID={get(orgs, '0.id', '')}\n+ isVisible={isImportingDashboard}\n />\n )\n }\ndiff --git a/ui/src/organizations/components/Dashboards.tsx b/ui/src/organizations/components/Dashboards.tsx\nindex 08c8402..90c2514 100644\n--- a/ui/src/organizations/components/Dashboards.tsx\n+++ b/ui/src/organizations/components/Dashboards.tsx\n@@ -6,13 +6,10 @@ import _ from 'lodash'\n \n // Components\n import DashboardsIndexContents from 'src/dashboards/components/dashboard_index/DashboardsIndexContents'\n-import {OverlayTechnology, Input, Tabs} from 'src/clockface'\n+import {Input, Tabs} from 'src/clockface'\n import {Button, ComponentColor, IconFont} from '@influxdata/clockface'\n import ImportDashboardOverlay from 'src/dashboards/components/ImportDashboardOverlay'\n \n-// Utils\n-import {getDeep} from 'src/utils/wrappers'\n-\n // APIs\n import {createDashboard, cloneDashboard} from 'src/dashboards/apis/v2/'\n \n@@ -39,7 +36,7 @@ import {DEFAULT_DASHBOARD_NAME} from 'src/dashboards/constants/index'\n \n // Types\n import {Notification} from 'src/types/notifications'\n-import {Links, Cell, Dashboard, AppState, Organization} from 'src/types/v2'\n+import {Links, Dashboard, AppState, Organization} from 'src/types/v2'\n \n // Decorators\n import {ErrorHandling} from 'src/shared/decorators/errors'\n@@ -205,46 +202,20 @@ class Dashboards extends PureComponent<Props, State> {\n this.props.handleDeleteDashboard(dashboard)\n }\n \n- private handleImportDashboard = async (\n- dashboard: Dashboard\n- ): Promise<void> => {\n- const defaultCell = {\n- x: 0,\n- y: 0,\n- w: 4,\n- h: 4,\n- }\n-\n- const name = _.get(dashboard, 'name', DEFAULT_DASHBOARD_NAME)\n- const cellsWithDefaultsApplied = getDeep<Cell[]>(\n- dashboard,\n- 'cells',\n- []\n- ).map(c => ({...defaultCell, ...c}))\n-\n- await this.props.handleImportDashboard({\n- ...dashboard,\n- name,\n- cells: cellsWithDefaultsApplied,\n- })\n- }\n-\n private handleToggleOverlay = (): void => {\n this.setState({isImportingDashboard: !this.state.isImportingDashboard})\n }\n \n private get renderImportOverlay(): JSX.Element {\n- const {notify} = this.props\n const {isImportingDashboard} = this.state\n+ const {orgs} = this.props\n \n return (\n- <OverlayTechnology visible={isImportingDashboard}>\n- <ImportDashboardOverlay\n- onDismissOverlay={this.handleToggleOverlay}\n- onImportDashboard={this.handleImportDashboard}\n- notify={notify}\n- />\n- </OverlayTechnology>\n+ <ImportDashboardOverlay\n+ onDismissOverlay={this.handleToggleOverlay}\n+ orgID={_.get(orgs, '0.id', '')}\n+ isVisible={isImportingDashboard}\n+ />\n )\n }\n }\ndiff --git a/ui/src/shared/components/ImportOverlay.tsx b/ui/src/shared/components/ImportOverlay.tsx\nindex 10f1d50..476fa70 100644\n--- a/ui/src/shared/components/ImportOverlay.tsx\n+++ b/ui/src/shared/components/ImportOverlay.tsx\n@@ -10,6 +10,7 @@ import {\n OverlayHeading,\n OverlayFooter,\n Radio,\n+ ComponentStatus,\n } from 'src/clockface'\n import {Button, ComponentColor} from '@influxdata/clockface'\n \n@@ -93,6 +94,7 @@ export default class ImportOverlay extends PureComponent<Props, State> {\n submitText=\"Upload\"\n handleSubmit={this.handleSetImportContent}\n submitOnDrop={true}\n+ submitOnUpload={true}\n onCancel={this.clearImportContent}\n />\n )\n@@ -110,18 +112,21 @@ export default class ImportOverlay extends PureComponent<Props, State> {\n private get submitButton(): JSX.Element {\n const {resourceName} = this.props\n const {selectedImportOption, importContent} = this.state\n- if (\n+ const isEnabled =\n selectedImportOption === ImportOption.Paste ||\n (selectedImportOption === ImportOption.Upload && importContent)\n- ) {\n- return (\n- <Button\n- text={`Import JSON as ${resourceName}`}\n- onClick={this.submit}\n- color={ComponentColor.Primary}\n- />\n- )\n- }\n+ const status = isEnabled\n+ ? ComponentStatus.Default\n+ : ComponentStatus.Disabled\n+\n+ return (\n+ <Button\n+ text={`Import JSON as ${resourceName}`}\n+ onClick={this.submit}\n+ color={ComponentColor.Primary}\n+ status={status}\n+ />\n+ )\n }\n \n private submit = () => {\ndiff --git a/ui/src/shared/copy/notifications.ts b/ui/src/shared/copy/notifications.ts\nindex 4492078..86700e2 100644\n--- a/ui/src/shared/copy/notifications.ts\n+++ b/ui/src/shared/copy/notifications.ts\n@@ -799,9 +799,18 @@ export const importTaskSucceeded = (): Notification => ({\n })\n \n export const importTaskFailed = (error: string): Notification => ({\n- ...defaultSuccessNotification,\n+ ...defaultErrorNotification,\n message: `Failed to import task: ${error}`,\n })\n+export const importDashboardSucceeded = (): Notification => ({\n+ ...defaultSuccessNotification,\n+ message: `Successfully imported dashboard.`,\n+})\n+\n+export const importDashboardFailed = (error: string): Notification => ({\n+ ...defaultErrorNotification,\n+ message: `Failed to import dashboard: ${error}`,\n+})\n \n // Labels\n export const getLabelsFailed = (): Notification => ({\n"]
2
["d0c6476df61b9c6ab07b87e1724ea7c5318595bb", "9114362b39f5194209cd0b330af7076333f3db77"]
["ci", "feat"]
add clean up test Add another clean up test, which verifies that the state is cleaned up after the timer (non-recurring) is triggered.,add method to extract snapshot name from filename also corrected pattern, where the period was meant to match a period, not any character. related to zeebe-io/zeebe#876
["diff --git a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\nindex d36b4c9..ca5047f 100644\n--- a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n@@ -630,6 +630,40 @@ public final class ProcessExecutionCleanStateTest {\n }\n \n @Test\n+ public void testProcessWithTriggerTimerStartEvent() {\n+ // given\n+ final var deployment =\n+ engineRule\n+ .deployment()\n+ .withXmlResource(\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .timerWithDate(\"=now() + duration(\\\"PT15S\\\")\")\n+ .endEvent()\n+ .done())\n+ .deploy();\n+\n+ final var processDefinitionKey =\n+ deployment.getValue().getProcessesMetadata().get(0).getProcessDefinitionKey();\n+\n+ // when\n+ engineRule.awaitProcessingOf(\n+ RecordingExporter.timerRecords(TimerIntent.CREATED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .getFirst());\n+\n+ engineRule.increaseTime(Duration.ofSeconds(15));\n+\n+ RecordingExporter.processInstanceRecords(ProcessInstanceIntent.ELEMENT_COMPLETED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .withElementType(BpmnElementType.PROCESS)\n+ .await();\n+\n+ // then\n+ assertThatStateIsEmpty();\n+ }\n+\n+ @Test\n public void testProcessWithTimerStartEventRedeployment() {\n // given\n final var deployment =\n", "diff --git a/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java b/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\nindex d8f4d89..e54e85a 100644\n--- a/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\n+++ b/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\n@@ -23,8 +23,9 @@ public class FsSnapshotStorageConfiguration\n {\n protected static final String CHECKSUM_ALGORITHM = \"SHA1\";\n \n- protected static final String SNAPSHOT_FILE_NAME_TEMPLATE = \"%s\" + File.separatorChar + \"%s-%d.snapshot\";\n- protected static final String SNAPSHOT_FILE_NAME_PATTERN = \"%s-(\\\\d+).snapshot\";\n+ protected static final String SNAPSHOT_FILE_NAME_TEMPLATE = \"%s-%d.snapshot\";\n+ protected static final String SNAPSHOT_FILE_PATH_TEMPLATE = \"%s\" + File.separatorChar + SNAPSHOT_FILE_NAME_TEMPLATE;\n+ protected static final String SNAPSHOT_FILE_NAME_PATTERN = \"%s-(\\\\d+)\\\\.snapshot\";\n \n protected static final String CHECKSUM_FILE_NAME_TEMPLATE = \"%s\" + File.separatorChar + \"%s-%d.\" + CHECKSUM_ALGORITHM.toLowerCase();\n \n@@ -50,7 +51,7 @@ public class FsSnapshotStorageConfiguration\n \n public String snapshotFileName(String name, long logPosition)\n {\n- return String.format(SNAPSHOT_FILE_NAME_TEMPLATE, rootPath, name, logPosition);\n+ return String.format(SNAPSHOT_FILE_PATH_TEMPLATE, rootPath, name, logPosition);\n }\n \n public String checksumFileName(String name, long logPosition)\n@@ -86,7 +87,7 @@ public class FsSnapshotStorageConfiguration\n return String.format(CHECKSUM_CONTENT_TEMPLATE, checksum, dataFileName);\n }\n \n- public String extractDigetsFromChecksumContent(String content)\n+ public String extractDigestFromChecksumContent(String content)\n {\n final int indexOfSeparator = content.indexOf(CHECKSUM_CONTENT_SEPARATOR);\n if (indexOfSeparator < 0)\n@@ -108,9 +109,18 @@ public class FsSnapshotStorageConfiguration\n return content.substring(indexOfSeparator + CHECKSUM_CONTENT_SEPARATOR.length());\n }\n \n+ public String getSnapshotNameFromFileName(final String fileName)\n+ {\n+ final String suffixPattern = String.format(SNAPSHOT_FILE_NAME_PATTERN, \"\");\n+ final Pattern pattern = Pattern.compile(suffixPattern);\n+ final String[] parts = pattern.split(fileName);\n+\n+ return parts[0];\n+ }\n+\n public String getSnapshotFileNameTemplate()\n {\n- return SNAPSHOT_FILE_NAME_TEMPLATE;\n+ return SNAPSHOT_FILE_PATH_TEMPLATE;\n }\n \n public String getChecksumFileNameTemplate()\n"]
2
["aa746b764e6c54bbbd631210fce35df842d09b12", "7ab965c55d0e98fdb6179577d0db56599675e400"]
["test", "feat"]
fix cypress on windows,lint source on ci
["diff --git a/packages/cypress/src/builders/cypress/cypress.impl.spec.ts b/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\nindex 22851fb..c9296fe 100644\n--- a/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\n+++ b/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\n@@ -77,7 +77,7 @@ describe('Cypress builder', () => {\n await run.result;\n await run.stop();\n expect(fork).toHaveBeenCalledWith(\n- '/root/node_modules/.bin/tsc',\n+ '/root/node_modules/typescript/bin/tsc',\n ['-p', '/root/apps/my-app-e2e/tsconfig.json'],\n { stdio: [0, 1, 2, 'ipc'] }\n );\ndiff --git a/packages/cypress/src/builders/cypress/cypress.impl.ts b/packages/cypress/src/builders/cypress/cypress.impl.ts\nindex 9d9ded3..d3917c2 100644\n--- a/packages/cypress/src/builders/cypress/cypress.impl.ts\n+++ b/packages/cypress/src/builders/cypress/cypress.impl.ts\n@@ -115,7 +115,7 @@ function compileTypescriptFiles(\n let args = ['-p', path.join(context.workspaceRoot, tsConfigPath)];\n const tscPath = path.join(\n context.workspaceRoot,\n- '/node_modules/.bin/tsc'\n+ '/node_modules/typescript/bin/tsc'\n );\n if (isWatching) {\n args.push('--watch');\n", "diff --git a/.travis.yml b/.travis.yml\nindex d56185e..96510cb 100644\n--- a/.travis.yml\n+++ b/.travis.yml\n@@ -2,5 +2,6 @@ language: node_js\n node_js:\n - 'stable'\n script:\n+ - yarn lint\n - yarn build\n - yarn test\n"]
2
["eebee9ab0bb6d4255ad0402d8422364e96bfef61", "2ac99c0a66a1adc18ee4ef660608f814823dd198"]
["fix", "ci"]
add travis file,updated to lerna v6
["diff --git a/.travis.yml b/.travis.yml\nnew file mode 100644\nindex 0000000..c08cc34\n--- /dev/null\n+++ b/.travis.yml\n@@ -0,0 +1,11 @@\n+sudo: false\n+\n+language: java\n+jdk: oraclejdk8\n+\n+branches:\n+ only:\n+ - master\n+\n+notifications:\n+ email: false\n", "diff --git a/lerna.json b/lerna.json\nindex 29cb6ed..4488fb6 100644\n--- a/lerna.json\n+++ b/lerna.json\n@@ -25,7 +25,6 @@\n \"npmClient\": \"pnpm\",\n \"useWorkspaces\": true,\n \"conventionalCommits\": true,\n- \"useNx\": true,\n \"command\": {\n \"version\": {\n \"message\": \"chore(release): published new version\"\ndiff --git a/nx.json b/nx.json\nindex 00997ba..b3f2627 100644\n--- a/nx.json\n+++ b/nx.json\n@@ -33,32 +33,6 @@\n }\n }\n },\n- \"targetDependencies\": {\n- \"build\": [\n- {\n- \"target\": \"build\",\n- \"projects\": \"dependencies\"\n- }\n- ],\n- \"build:ci\": [\n- {\n- \"target\": \"build:ci\",\n- \"projects\": \"dependencies\"\n- }\n- ],\n- \"prepare\": [\n- {\n- \"target\": \"prepare\",\n- \"projects\": \"dependencies\"\n- }\n- ],\n- \"package\": [\n- {\n- \"target\": \"package\",\n- \"projects\": \"dependencies\"\n- }\n- ]\n- },\n \"affected\": {\n \"defaultBase\": \"main\"\n },\n@@ -66,5 +40,28 @@\n \"@nrwl/js\": {\n \"analyzeSourceFiles\": false\n }\n+ },\n+ \"$schema\": \"./node_modules/nx/schemas/nx-schema.json\",\n+ \"targetDefaults\": {\n+ \"build\": {\n+ \"dependsOn\": [\n+ \"^build\"\n+ ]\n+ },\n+ \"build:ci\": {\n+ \"dependsOn\": [\n+ \"^build:ci\"\n+ ]\n+ },\n+ \"prepare\": {\n+ \"dependsOn\": [\n+ \"^prepare\"\n+ ]\n+ },\n+ \"package\": {\n+ \"dependsOn\": [\n+ \"^package\"\n+ ]\n+ }\n }\n }\n"]
2
["d0814a928601706635287fd3d9d3451d156b821a", "aa30370b796c1f46cc60aa56b2918c3d27c0cf88"]
["ci", "build"]
update build,add gitignore.nix to dep update matrix
["diff --git a/bootstrap/scripts/publish-patch.sh b/bootstrap/scripts/publish-patch.sh\nindex a1b6f12..0d849a5 100755\n--- a/bootstrap/scripts/publish-patch.sh\n+++ b/bootstrap/scripts/publish-patch.sh\n@@ -5,4 +5,4 @@ lerna version patch\n lerna publish from-package -y\n git push\n \n-./pack_and_install.sh\n\\ No newline at end of file\n+./bootstrap/scripts/pack_and_install.sh\n\\ No newline at end of file\n", "diff --git a/.github/workflows/update-deps.yml b/.github/workflows/update-deps.yml\nindex e0ccd62..1236f58 100644\n--- a/.github/workflows/update-deps.yml\n+++ b/.github/workflows/update-deps.yml\n@@ -13,6 +13,7 @@ jobs:\n - nixpkgs\n - poetry2nix\n - pre-commit-hooks\n+ - gitignore.nix\n steps:\n - name: Checkout\n uses: actions/checkout@v2\n"]
2
["3fcfb20b0feb371b357edc42fcb7c87085c9b82a", "c444fdb9e85ce44c5c0c99addc777dd7b6085153"]
["build", "ci"]
add clean up test Add another clean up test, which verifies that the state is cleaned up after the timer (non-recurring) is triggered.,lint README
["diff --git a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\nindex d36b4c9..ca5047f 100644\n--- a/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n+++ b/engine/src/test/java/io/camunda/zeebe/engine/state/ProcessExecutionCleanStateTest.java\n@@ -630,6 +630,40 @@ public final class ProcessExecutionCleanStateTest {\n }\n \n @Test\n+ public void testProcessWithTriggerTimerStartEvent() {\n+ // given\n+ final var deployment =\n+ engineRule\n+ .deployment()\n+ .withXmlResource(\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .timerWithDate(\"=now() + duration(\\\"PT15S\\\")\")\n+ .endEvent()\n+ .done())\n+ .deploy();\n+\n+ final var processDefinitionKey =\n+ deployment.getValue().getProcessesMetadata().get(0).getProcessDefinitionKey();\n+\n+ // when\n+ engineRule.awaitProcessingOf(\n+ RecordingExporter.timerRecords(TimerIntent.CREATED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .getFirst());\n+\n+ engineRule.increaseTime(Duration.ofSeconds(15));\n+\n+ RecordingExporter.processInstanceRecords(ProcessInstanceIntent.ELEMENT_COMPLETED)\n+ .withProcessDefinitionKey(processDefinitionKey)\n+ .withElementType(BpmnElementType.PROCESS)\n+ .await();\n+\n+ // then\n+ assertThatStateIsEmpty();\n+ }\n+\n+ @Test\n public void testProcessWithTimerStartEventRedeployment() {\n // given\n final var deployment =\n", "diff --git a/README.md b/README.md\nindex a163c83..9cd12bc 100644\n--- a/README.md\n+++ b/README.md\n@@ -26,7 +26,7 @@ Ibis has three primary components:\n Ibis aims to be a future-proof solution to interacting with data using Python and can accomplish this goal through its main features:\n \n - **Familiar API**: Ibis\u2019s API design borrows from popular APIs like pandas and dplyr that most users already know and like to use.\n-- **Consistent syntax**: Ibis aims to be universal Python API for tabular data, big or small.\n+- **Consistent syntax**: Ibis aims to be universal Python API for tabular data, big or small.\n - **Deferred execution**: Ibis pushes code execution to the query engine and only moves required data into memory when it has to.\n This leads to more faster, more efficient analytics workflows\n - **Interactive mode**: Ibis also provides an interactive mode, in which users can quickly diagnose problems, do exploratory data analysis, and mock up workflows locally.\n"]
2
["aa746b764e6c54bbbd631210fce35df842d09b12", "cab2cf4d3fffb0ec2b56d455e67ac5fa992b4104"]
["test", "docs"]
remove duplicated variables,add user role enum Signed-off-by: Braks <[email protected]>
["diff --git a/packages/core/src/components/item/item.ios.scss b/packages/core/src/components/item/item.ios.scss\nindex 4de5455..6c4d11a 100644\n--- a/packages/core/src/components/item/item.ios.scss\n+++ b/packages/core/src/components/item/item.ios.scss\n@@ -47,15 +47,6 @@ $item-ios-detail-push-color: $list-ios-border-color !default;\n /// @prop - Icon for the detail arrow\n $item-ios-detail-push-svg: \"<svg xmlns='http://www.w3.org/2000/svg' viewBox='0 0 12 20'><path d='M2,20l-2-2l8-8L0,2l2-2l10,10L2,20z' fill='#{$item-ios-detail-push-color}'/></svg>\" !default;\n \n-/// @prop - Background for the divider\n-$item-ios-divider-background: #f7f7f7 !default;\n-\n-/// @prop - Color for the divider\n-$item-ios-divider-color: #222 !default;\n-\n-/// @prop - Padding for the divider\n-$item-ios-divider-padding: 5px 15px !default;\n-\n \n // iOS Item\n // --------------------------------------------------\ndiff --git a/packages/core/src/components/item/item.md.scss b/packages/core/src/components/item/item.md.scss\nindex 1dd1800..3dadbc0 100644\n--- a/packages/core/src/components/item/item.md.scss\n+++ b/packages/core/src/components/item/item.md.scss\n@@ -35,21 +35,6 @@ $item-md-detail-push-color: $list-md-border-color !default;\n /// @prop - Icon for the detail arrow\n $item-md-detail-push-svg: \"<svg xmlns='http://www.w3.org/2000/svg' viewBox='0 0 12 20'><path d='M2,20l-2-2l8-8L0,2l2-2l10,10L2,20z' fill='#{$item-md-detail-push-color}'/></svg>\" !default;\n \n-/// @prop - Color for the divider\n-$item-md-divider-color: #858585 !default;\n-\n-/// @prop - Background for the divider\n-$item-md-divider-background: #fff !default;\n-\n-/// @prop - Font size for the divider\n-$item-md-divider-font-size: $item-md-body-text-font-size !default;\n-\n-/// @prop - Border bottom for the divider\n-$item-md-divider-border-bottom: 1px solid $list-md-border-color !default;\n-\n-/// @prop - Padding for the divider\n-$item-md-divider-padding: 5px 15px !default;\n-\n \n .item-md {\n @include padding-horizontal($item-md-padding-start, 0);\ndiff --git a/packages/core/src/components/item/item.wp.scss b/packages/core/src/components/item/item.wp.scss\nindex 2c4aae6..07b9266 100644\n--- a/packages/core/src/components/item/item.wp.scss\n+++ b/packages/core/src/components/item/item.wp.scss\n@@ -41,21 +41,6 @@ $item-wp-detail-push-color: $input-wp-border-color !default;\n /// @prop - Icon for the detail arrow\n $item-wp-detail-push-svg: \"<svg xmlns='http://www.w3.org/2000/svg' viewBox='0 0 12 20'><path d='M2,20l-2-2l8-8L0,2l2-2l10,10L2,20z' fill='#{$item-wp-detail-push-color}'/></svg>\" !default;\n \n-/// @prop - Color for the divider\n-$item-wp-divider-color: $list-wp-text-color !default;\n-\n-/// @prop - Background for the divider\n-$item-wp-divider-background: #fff !default;\n-\n-/// @prop - Bodrer bottom for the divider\n-$item-wp-divider-border-bottom: 1px solid $list-wp-border-color !default;\n-\n-/// @prop - Font size for the divider\n-$item-wp-divider-font-size: 2rem !default;\n-\n-/// @prop - Padding for the divider\n-$item-wp-divider-padding: 5px 15px !default;\n-\n \n .item-wp {\n @include padding-horizontal($item-wp-padding-start, 0);\n", "diff --git a/packages/nc-gui-v2/lib/enums.ts b/packages/nc-gui-v2/lib/enums.ts\nindex e87b69a..c6751a3 100644\n--- a/packages/nc-gui-v2/lib/enums.ts\n+++ b/packages/nc-gui-v2/lib/enums.ts\n@@ -1,3 +1,9 @@\n+export enum Role {\n+ Super = 'super',\n+ Admin = 'admin',\n+ User = 'user',\n+}\n+\n export enum Language {\n de = 'Deutsch',\n en = 'English',\ndiff --git a/packages/nc-gui-v2/lib/types.ts b/packages/nc-gui-v2/lib/types.ts\nindex bf152c4..dd8a1ce 100644\n--- a/packages/nc-gui-v2/lib/types.ts\n+++ b/packages/nc-gui-v2/lib/types.ts\n@@ -1,11 +1,12 @@\n import type { ComputedRef, ToRefs } from 'vue'\n+import type { Role } from '~/lib/enums'\n \n export interface User {\n id: string\n email: string\n firstname: string | null\n lastname: string | null\n- roles: string[]\n+ roles: (Role | string)[]\n }\n \n export interface State {\n"]
2
["cd7e8c3d3549ea05115b3f02586eeba894d86906", "176a959eb80d17f9abc5c6b5354e6097be95b42d"]
["refactor", "feat"]
abort parallel stages if one failed,update basic test with colors
["diff --git a/Jenkinsfile b/Jenkinsfile\nindex 168f446..a4da961 100644\n--- a/Jenkinsfile\n+++ b/Jenkinsfile\n@@ -28,6 +28,7 @@ pipeline {\n }\n \n stage('Verify') {\n+ failFast true\n parallel {\n stage('Tests') {\n steps {\n", "diff --git a/core/src/components/label/test/basic/index.html b/core/src/components/label/test/basic/index.html\nindex d0b566c..377e58c 100644\n--- a/core/src/components/label/test/basic/index.html\n+++ b/core/src/components/label/test/basic/index.html\n@@ -19,12 +19,32 @@\n </ion-header>\n \n <ion-content>\n+ <div padding>\n+ <ion-label>Default</ion-label>\n+\n+ <ion-label color=\"secondary\">Secondary</ion-label>\n+\n+ <ion-label color=\"tertiary\">Tertiary</ion-label>\n+\n+ <ion-label color=\"danger\">Danger</ion-label>\n+\n+ <ion-label class=\"custom\">Custom</ion-label>\n+ </div>\n+\n <ion-list>\n <ion-item>\n <ion-label>Default</ion-label>\n <ion-input></ion-input>\n </ion-item>\n <ion-item>\n+ <ion-label color=\"tertiary\">Tertiary</ion-label>\n+ <ion-input></ion-input>\n+ </ion-item>\n+ <ion-item>\n+ <ion-label class=\"custom\">Custom</ion-label>\n+ <ion-input></ion-input>\n+ </ion-item>\n+ <ion-item>\n <ion-label text-wrap>Wrap label this label just goes on and on and on</ion-label>\n <ion-input></ion-input>\n </ion-item>\n@@ -42,6 +62,12 @@\n </ion-item>\n </ion-list>\n </ion-content>\n+\n+ <style>\n+ .custom {\n+ color: lightblue;\n+ }\n+ </style>\n </ion-app>\n </body>\n \n"]
2
["28e623b294816c4e070971782a75c8697a11966f", "c3b5dc77ff3d89d389f6f3a868b17d0a8ca63074"]
["ci", "test"]
support document.html,use lambda to define backend operations
["diff --git a/packages/nuxt3/src/builder/builder.ts b/packages/nuxt3/src/builder/builder.ts\nindex a24bd88..ecc22ef 100644\n--- a/packages/nuxt3/src/builder/builder.ts\n+++ b/packages/nuxt3/src/builder/builder.ts\n@@ -3,6 +3,7 @@ import fsExtra from 'fs-extra'\n import { debounce } from 'lodash'\n import { BundleBuilder } from 'src/webpack'\n import { Nuxt } from '../core'\n+import { DeterminedGlobals, determineGlobals } from '../utils'\n import {\n templateData,\n compileTemplates,\n@@ -15,12 +16,14 @@ import Ignore from './ignore'\n \n export class Builder {\n nuxt: Nuxt\n+ globals: DeterminedGlobals\n ignore: Ignore\n- app: NuxtApp\n templates: NuxtTemplate[]\n+ app: NuxtApp\n \n constructor (nuxt) {\n this.nuxt = nuxt\n+ this.globals = determineGlobals(nuxt.options.globalName, nuxt.options.globals)\n this.ignore = new Ignore({\n rootDir: nuxt.options.srcDir,\n ignoreArray: nuxt.options.ignore.concat(\n@@ -32,6 +35,10 @@ export class Builder {\n build () {\n return build(this)\n }\n+\n+ close () {\n+ // TODO: close watchers\n+ }\n }\n \n // Extends VueRouter\ndiff --git a/packages/nuxt3/src/builder/template.ts b/packages/nuxt3/src/builder/template.ts\nindex 63a9115..fe09f16 100644\n--- a/packages/nuxt3/src/builder/template.ts\n+++ b/packages/nuxt3/src/builder/template.ts\n@@ -11,6 +11,7 @@ export interface NuxtTemplate {\n \n export function templateData (builder) {\n return {\n+ globals: builder.globals,\n app: builder.app\n }\n }\ndiff --git a/packages/nuxt3/src/builder/watch.ts b/packages/nuxt3/src/builder/watch.ts\nindex b4d1415..d148fec 100644\n--- a/packages/nuxt3/src/builder/watch.ts\n+++ b/packages/nuxt3/src/builder/watch.ts\n@@ -38,7 +38,8 @@ export function createWatcher (\n return {\n watchAll,\n watch,\n- debug\n+ debug,\n+ close: () => watcher.close()\n }\n }\n \ndiff --git a/packages/nuxt3/src/config/options.ts b/packages/nuxt3/src/config/options.ts\nindex 5aac8ac..6e7f93c 100644\n--- a/packages/nuxt3/src/config/options.ts\n+++ b/packages/nuxt3/src/config/options.ts\n@@ -12,7 +12,7 @@ import { DefaultConfiguration, defaultNuxtConfigFile, getDefaultNuxtConfig } fro\n import { deleteProp, mergeConfigs, setProp, overrideProp, Optional } from './transformers'\n \n interface InputConfiguration {\n- appTemplatePath?: string\n+ documentPath?: string\n layoutTransition?: string | DefaultConfiguration['layoutTransition']\n loading?: true | false | DefaultConfiguration['loading']\n manifest?: {\n@@ -197,13 +197,16 @@ function normalizeConfig (_options: CliConfiguration) {\n .concat(options.extensions))\n \n // If app.html is defined, set the template path to the user template\n- if (options.appTemplatePath === undefined) {\n- options.appTemplatePath = path.resolve(options.buildDir, 'views/app.template.html')\n- if (fs.existsSync(path.join(options.srcDir, 'app.html'))) {\n- options.appTemplatePath = path.join(options.srcDir, 'app.html')\n+ if (options.documentPath === undefined) {\n+ options.documentPath = path.resolve(options.buildDir, 'views/document.template.html')\n+ const userDocumentPath = path.join(options.srcDir, 'document.html')\n+ if (fs.existsSync(userDocumentPath)) {\n+ options.documentPath = userDocumentPath\n+ } else {\n+ options.watch.push(userDocumentPath)\n }\n } else {\n- options.appTemplatePath = path.resolve(options.srcDir, options.appTemplatePath)\n+ options.documentPath = path.resolve(options.srcDir, options.documentPath)\n }\n \n overrideProp(options.build, 'publicPath', options.build.publicPath.replace(/([^/])$/, '$1/'))\ndiff --git a/packages/nuxt3/src/vue-renderer/renderers/ssr.ts b/packages/nuxt3/src/vue-renderer/renderers/ssr.ts\nindex 3e3ce2d..482bd6b 100644\n--- a/packages/nuxt3/src/vue-renderer/renderers/ssr.ts\n+++ b/packages/nuxt3/src/vue-renderer/renderers/ssr.ts\n@@ -96,6 +96,9 @@ export default class SSRRenderer extends BaseRenderer {\n // Call Vue renderer renderToString\n let APP = await this.vueRenderer.renderToString(renderContext)\n \n+ // Wrap with Nuxt id\n+ APP = `<div id=\"${this.serverContext.globals.id}\">${APP}</div>`\n+\n // Call render:done in app\n await renderContext.nuxt.hooks.callHook('vue-renderer:done')\n \ndiff --git a/packages/nuxt3/src/webpack/configs/client.ts b/packages/nuxt3/src/webpack/configs/client.ts\nindex a257948..4fb35e0 100644\n--- a/packages/nuxt3/src/webpack/configs/client.ts\n+++ b/packages/nuxt3/src/webpack/configs/client.ts\n@@ -94,7 +94,7 @@ function clientHTML (ctx: WebpackConfigContext) {\n config.plugins.push(\n new HTMLPlugin({\n filename: '../server/index.ssr.html',\n- template: options.appTemplatePath,\n+ template: options.documentPath,\n minify: options.build.html.minify as any,\n inject: false // Resources will be injected using bundleRenderer\n })\n@@ -104,7 +104,7 @@ function clientHTML (ctx: WebpackConfigContext) {\n config.plugins.push(\n new HTMLPlugin({\n filename: '../server/index.spa.html',\n- template: options.appTemplatePath,\n+ template: options.documentPath,\n minify: options.build.html.minify as any,\n inject: true\n })\n", "diff --git a/ibis/backends/duckdb/registry.py b/ibis/backends/duckdb/registry.py\nindex 20ffd6f..3f56f2a 100644\n--- a/ibis/backends/duckdb/registry.py\n+++ b/ibis/backends/duckdb/registry.py\n@@ -107,28 +107,13 @@ def _literal(_, op):\n return sa.cast(sa.literal(value), sqla_type)\n \n \n-def _array_column(t, op):\n- (arg,) = op.args\n- sqla_type = to_sqla_type(op.output_dtype)\n- return sa.cast(sa.func.list_value(*map(t.translate, arg)), sqla_type)\n-\n-\n def _neg_idx_to_pos(array, idx):\n if_ = getattr(sa.func, \"if\")\n arg_length = sa.func.array_length(array)\n return if_(idx < 0, arg_length + sa.func.greatest(idx, -arg_length), idx)\n \n \n-def _struct_field(t, op):\n- return sa.func.struct_extract(\n- t.translate(op.arg),\n- sa.text(repr(op.field)),\n- type_=to_sqla_type(op.output_dtype),\n- )\n-\n-\n-def _regex_extract(t, op):\n- string, pattern, index = map(t.translate, op.args)\n+def _regex_extract(string, pattern, index):\n result = sa.case(\n [\n (\n@@ -149,8 +134,7 @@ def _regex_extract(t, op):\n return result\n \n \n-def _json_get_item(t, op):\n- left, path = map(t.translate, op.args)\n+def _json_get_item(left, path):\n # Workaround for https://github.com/duckdb/duckdb/issues/5063\n # In some situations duckdb silently does the wrong thing if\n # the path is parametrized.\n@@ -197,7 +181,12 @@ def _struct_column(t, op):\n \n operation_registry.update(\n {\n- ops.ArrayColumn: _array_column,\n+ ops.ArrayColumn: (\n+ lambda t, op: sa.cast(\n+ sa.func.list_value(*map(t.translate, op.cols)),\n+ to_sqla_type(op.output_dtype),\n+ )\n+ ),\n ops.ArrayConcat: fixed_arity(sa.func.array_concat, 2),\n ops.ArrayRepeat: fixed_arity(\n lambda arg, times: sa.func.flatten(\n@@ -222,7 +211,13 @@ operation_registry.update(\n # TODO: map operations, but DuckDB's maps are multimaps\n ops.Modulus: fixed_arity(operator.mod, 2),\n ops.Round: _round,\n- ops.StructField: _struct_field,\n+ ops.StructField: (\n+ lambda t, op: sa.func.struct_extract(\n+ t.translate(op.arg),\n+ sa.text(repr(op.field)),\n+ type_=to_sqla_type(op.output_dtype),\n+ )\n+ ),\n ops.TableColumn: _table_column,\n ops.TimestampDiff: fixed_arity(sa.func.age, 2),\n ops.TimestampFromUNIX: _timestamp_from_unix,\n@@ -232,7 +227,7 @@ operation_registry.update(\n lambda *_: sa.cast(sa.func.now(), sa.TIMESTAMP),\n 0,\n ),\n- ops.RegexExtract: _regex_extract,\n+ ops.RegexExtract: fixed_arity(_regex_extract, 3),\n ops.RegexReplace: fixed_arity(\n lambda *args: sa.func.regexp_replace(*args, \"g\"), 3\n ),\n@@ -255,7 +250,7 @@ operation_registry.update(\n ops.ArgMin: reduction(sa.func.min_by),\n ops.ArgMax: reduction(sa.func.max_by),\n ops.BitwiseXor: fixed_arity(sa.func.xor, 2),\n- ops.JSONGetItem: _json_get_item,\n+ ops.JSONGetItem: fixed_arity(_json_get_item, 2),\n ops.RowID: lambda *_: sa.literal_column('rowid'),\n ops.StringToTimestamp: fixed_arity(sa.func.strptime, 2),\n }\n"]
2
["09476134eeeb12c025618919ab9a795a680a9b30", "5d14de6722eb34c6604a124f6f11cb711f16bd44"]
["feat", "refactor"]
licensing,coordinator accepts a request transformer instead of a list of operations The request transformer can generate the operations from the current topology. This helps to - ensure that the operations are generated based on the latest topology. When concurrent changes happens, coordinator can detect it. Previously it was unclear because by the time handle apply operations, the cluster topology might have changed. - return the simulated final topology as part of the result
["diff --git a/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java b/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\nindex a4aee6b..bb523fa 100644\n--- a/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\n+++ b/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\n@@ -1,3 +1,18 @@\n+/*\n+ * Copyright \u00a9 2020 camunda services GmbH ([email protected])\n+ *\n+ * Licensed under the Apache License, Version 2.0 (the \"License\");\n+ * you may not use this file except in compliance with the License.\n+ * You may obtain a copy of the License at\n+ *\n+ * http://www.apache.org/licenses/LICENSE-2.0\n+ *\n+ * Unless required by applicable law or agreed to in writing, software\n+ * distributed under the License is distributed on an \"AS IS\" BASIS,\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n+ * See the License for the specific language governing permissions and\n+ * limitations under the License.\n+ */\n package io.atomix.cluster.messaging.impl;\n \n import static org.assertj.core.api.Assertions.assertThat;\n", "diff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\nindex 8bb5c3d..f8f5e24 100644\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\n@@ -10,6 +10,7 @@ package io.camunda.zeebe.topology.changes;\n import io.camunda.zeebe.scheduler.future.ActorFuture;\n import io.camunda.zeebe.topology.state.ClusterTopology;\n import io.camunda.zeebe.topology.state.TopologyChangeOperation;\n+import io.camunda.zeebe.util.Either;\n import java.util.List;\n \n public interface TopologyChangeCoordinator {\n@@ -39,4 +40,16 @@ public interface TopologyChangeCoordinator {\n ActorFuture<Boolean> hasCompletedChanges(final long version);\n \n ActorFuture<ClusterTopology> getCurrentTopology();\n+\n+ ActorFuture<TopologyChangeResult> applyOperations(TopologyChangeRequest request);\n+\n+ record TopologyChangeResult(\n+ ClusterTopology currentTopology,\n+ ClusterTopology finalTopology,\n+ List<TopologyChangeOperation> operations) {}\n+\n+ interface TopologyChangeRequest {\n+ Either<Exception, List<TopologyChangeOperation>> operations(\n+ final ClusterTopology currentTopology);\n+ }\n }\ndiff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\nindex 13ec754..877fc3c 100644\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\n@@ -103,6 +103,62 @@ public class TopologyChangeCoordinatorImpl implements TopologyChangeCoordinator \n return clusterTopologyManager.getClusterTopology();\n }\n \n+ @Override\n+ public ActorFuture<TopologyChangeResult> applyOperations(final TopologyChangeRequest request) {\n+ final ActorFuture<TopologyChangeResult> future = executor.createFuture();\n+ clusterTopologyManager\n+ .getClusterTopology()\n+ .onComplete(\n+ (currentClusterTopology, errorOnGettingTopology) -> {\n+ if (errorOnGettingTopology != null) {\n+ future.completeExceptionally(errorOnGettingTopology);\n+ return;\n+ }\n+\n+ final var operationsEither = request.operations(currentClusterTopology);\n+ if (operationsEither.isLeft()) {\n+ future.completeExceptionally(operationsEither.getLeft());\n+ return;\n+ }\n+ final var operations = operationsEither.get();\n+ if (operations.isEmpty()) {\n+ // No operations to apply\n+ future.complete(\n+ new TopologyChangeResult(\n+ currentClusterTopology, currentClusterTopology, operations));\n+ return;\n+ }\n+\n+ final ActorFuture<ClusterTopology> validation =\n+ validateTopologyChangeRequest(currentClusterTopology, operations);\n+\n+ validation.onComplete(\n+ (simulatedFinalTopology, validationError) -> {\n+ if (validationError != null) {\n+ future.completeExceptionally(validationError);\n+ return;\n+ }\n+\n+ // if the validation was successful, apply the changes\n+ final ActorFuture<ClusterTopology> applyFuture = executor.createFuture();\n+ applyTopologyChange(\n+ operations, currentClusterTopology, simulatedFinalTopology, applyFuture);\n+\n+ applyFuture.onComplete(\n+ (ignore, error) -> {\n+ if (error == null) {\n+ future.complete(\n+ new TopologyChangeResult(\n+ currentClusterTopology, simulatedFinalTopology, operations));\n+ } else {\n+ future.completeExceptionally(error);\n+ }\n+ });\n+ });\n+ });\n+ return future;\n+ }\n+\n private ActorFuture<ClusterTopology> validateTopologyChangeRequest(\n final ClusterTopology currentClusterTopology,\n final List<TopologyChangeOperation> operations) {\n"]
2
["cbe62140ce219da84772e21e7cfb4b5c2a25c1b8", "dec860436916ef216998f80f8b2f9c39d00c064d"]
["docs", "feat"]
do not pin time in tests but only skip ahead related to #573,export order
["diff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\nindex 636cd21..76afff7 100644\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\n@@ -15,7 +15,9 @@\n */\n package io.zeebe.broker.it.startup;\n \n-import static io.zeebe.broker.it.util.TopicEventRecorder.*;\n+import static io.zeebe.broker.it.util.TopicEventRecorder.incidentEvent;\n+import static io.zeebe.broker.it.util.TopicEventRecorder.taskEvent;\n+import static io.zeebe.broker.it.util.TopicEventRecorder.wfInstanceEvent;\n import static io.zeebe.test.util.TestUtil.doRepeatedly;\n import static io.zeebe.test.util.TestUtil.waitUntil;\n import static org.assertj.core.api.Assertions.assertThat;\n@@ -24,11 +26,18 @@ import java.io.File;\n import java.io.InputStream;\n import java.nio.charset.StandardCharsets;\n import java.time.Duration;\n-import java.time.Instant;\n import java.util.Collections;\n import java.util.List;\n import java.util.regex.Pattern;\n \n+import org.assertj.core.util.Files;\n+import org.junit.After;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.rules.ExpectedException;\n+import org.junit.rules.RuleChain;\n+import org.junit.rules.TemporaryFolder;\n+\n import io.zeebe.broker.clustering.ClusterServiceNames;\n import io.zeebe.broker.it.ClientRule;\n import io.zeebe.broker.it.EmbeddedBrokerRule;\n@@ -38,7 +47,9 @@ import io.zeebe.client.ZeebeClient;\n import io.zeebe.client.clustering.impl.TopicLeader;\n import io.zeebe.client.clustering.impl.TopologyResponse;\n import io.zeebe.client.cmd.ClientCommandRejectedException;\n-import io.zeebe.client.event.*;\n+import io.zeebe.client.event.DeploymentEvent;\n+import io.zeebe.client.event.TaskEvent;\n+import io.zeebe.client.event.WorkflowInstanceEvent;\n import io.zeebe.model.bpmn.Bpmn;\n import io.zeebe.model.bpmn.instance.WorkflowDefinition;\n import io.zeebe.raft.Raft;\n@@ -48,9 +59,6 @@ import io.zeebe.test.util.TestFileUtil;\n import io.zeebe.test.util.TestUtil;\n import io.zeebe.transport.SocketAddress;\n import io.zeebe.util.time.ClockUtil;\n-import org.assertj.core.util.Files;\n-import org.junit.*;\n-import org.junit.rules.*;\n \n public class BrokerRecoveryTest\n {\n@@ -360,17 +368,12 @@ public class BrokerRecoveryTest\n waitUntil(() -> !recordingTaskHandler.getHandledTasks().isEmpty());\n \n // when\n- restartBroker(() ->\n- {\n- final Instant now = ClockUtil.getCurrentTime();\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\n- });\n+ restartBroker(() -> ClockUtil.addTime(Duration.ofSeconds(60)));\n \n // wait until stream processor and scheduler process the lock task event which is not re-processed on recovery\n doRepeatedly(() ->\n {\n- final Instant now = ClockUtil.getCurrentTime();\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\n+ ClockUtil.addTime(Duration.ofSeconds(60)); // retriggers lock expiration check in broker\n return null;\n }).until(t -> eventRecorder.hasTaskEvent(taskEvent(\"LOCK_EXPIRED\")));\n \ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\nindex 5ff1301..0ffe98d 100644\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\n@@ -15,7 +15,9 @@\n */\n package io.zeebe.broker.it.startup;\n \n-import static io.zeebe.broker.it.util.TopicEventRecorder.*;\n+import static io.zeebe.broker.it.util.TopicEventRecorder.incidentEvent;\n+import static io.zeebe.broker.it.util.TopicEventRecorder.taskEvent;\n+import static io.zeebe.broker.it.util.TopicEventRecorder.wfInstanceEvent;\n import static io.zeebe.test.util.TestUtil.waitUntil;\n import static org.assertj.core.api.Assertions.assertThat;\n \n@@ -23,11 +25,18 @@ import java.io.File;\n import java.io.InputStream;\n import java.nio.charset.StandardCharsets;\n import java.time.Duration;\n-import java.time.Instant;\n import java.util.Collections;\n import java.util.List;\n import java.util.regex.Pattern;\n \n+import org.junit.After;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.experimental.categories.Category;\n+import org.junit.rules.ExpectedException;\n+import org.junit.rules.RuleChain;\n+import org.junit.rules.TemporaryFolder;\n+\n import io.zeebe.broker.clustering.ClusterServiceNames;\n import io.zeebe.broker.it.ClientRule;\n import io.zeebe.broker.it.EmbeddedBrokerRule;\n@@ -37,7 +46,9 @@ import io.zeebe.client.ZeebeClient;\n import io.zeebe.client.clustering.impl.TopicLeader;\n import io.zeebe.client.clustering.impl.TopologyResponse;\n import io.zeebe.client.cmd.ClientCommandRejectedException;\n-import io.zeebe.client.event.*;\n+import io.zeebe.client.event.DeploymentEvent;\n+import io.zeebe.client.event.TaskEvent;\n+import io.zeebe.client.event.WorkflowInstanceEvent;\n import io.zeebe.model.bpmn.Bpmn;\n import io.zeebe.model.bpmn.instance.WorkflowDefinition;\n import io.zeebe.raft.Raft;\n@@ -47,9 +58,6 @@ import io.zeebe.test.util.TestFileUtil;\n import io.zeebe.test.util.TestUtil;\n import io.zeebe.transport.SocketAddress;\n import io.zeebe.util.time.ClockUtil;\n-import org.junit.*;\n-import org.junit.experimental.categories.Category;\n-import org.junit.rules.*;\n \n public class BrokerRestartTest\n {\n@@ -360,11 +368,7 @@ public class BrokerRestartTest\n waitUntil(() -> !recordingTaskHandler.getHandledTasks().isEmpty());\n \n // when\n- restartBroker(() ->\n- {\n- final Instant now = ClockUtil.getCurrentTime();\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\n- });\n+ restartBroker(() -> ClockUtil.addTime(Duration.ofSeconds(60)));\n \n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\"LOCK_EXPIRED\")));\n recordingTaskHandler.clear();\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\nindex 49b527d..a322fbe 100644\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\n@@ -353,7 +353,7 @@ public class TaskSubscriptionTest\n waitUntil(() -> taskHandler.getHandledTasks().size() == 1);\n \n // when\n- ClockUtil.setCurrentTime(Instant.now().plus(Duration.ofMinutes(5)));\n+ ClockUtil.addTime(Duration.ofMinutes(5));\n \n // then\n waitUntil(() -> taskHandler.getHandledTasks().size() == 2);\n", "diff --git a/src/Object/_api.ts b/src/Object/_api.ts\nindex a1e5c8c..70e9fea 100644\n--- a/src/Object/_api.ts\n+++ b/src/Object/_api.ts\n@@ -38,9 +38,9 @@ export {Optional} from './Optional'\n export {OptionalKeys} from './OptionalKeys'\n export {Overwrite} from './Overwrite'\n export {Partial} from './Partial'\n+export {Path} from './Path'\n export {Paths} from './Paths'\n export {PathValid} from './PathValid'\n-export {Path} from './Path'\n export {Pick} from './Pick'\n export {Readonly} from './Readonly'\n export {ReadonlyKeys} from './ReadonlyKeys'\n"]
2
["7ece3a9a16780dc6c633bbd903d36ce0aefd6a8a", "879edb6ed90f88b9ae6a3c2e8878ae1be48e0c88"]
["test", "refactor"]
build updates,temporary do no run "verify-ffmpeg.py" on Mac CI (#14986)
["diff --git a/demo/vanilla_new/css/404.min.css b/demo/vanilla_new/css/404.min.css\nindex a3485b4..e69de29 100644\n--- a/demo/vanilla_new/css/404.min.css\n+++ b/demo/vanilla_new/css/404.min.css\n@@ -1 +0,0 @@\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\"\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\"Share Tech Mono\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\"Space Mono\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\"Space Mono\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\"\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\"back to hompage\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\"report error\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\ndiff --git a/demo/vanilla_new/css/main.min.css b/demo/vanilla_new/css/main.min.css\nindex 043eb4f..e69de29 100644\n--- a/demo/vanilla_new/css/main.min.css\n+++ b/demo/vanilla_new/css/main.min.css\n@@ -1 +0,0 @@\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}#floating-container{right:20px;top:20px;position:fixed;z-index:4000}\ndiff --git a/demo/vanilla_new/js/404.min.js b/demo/vanilla_new/js/404.min.js\nindex 3642106..e69de29 100644\n--- a/demo/vanilla_new/js/404.min.js\n+++ b/demo/vanilla_new/js/404.min.js\n@@ -1 +0,0 @@\n-tsParticles.loadJSON(\"tsparticles\",\"/configs/404.json\");\ndiff --git a/website/css/404.min.css b/website/css/404.min.css\nindex a3485b4..e69de29 100644\n--- a/website/css/404.min.css\n+++ b/website/css/404.min.css\n@@ -1 +0,0 @@\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\"\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\"Share Tech Mono\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\"Space Mono\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\"Space Mono\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\"\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\"back to hompage\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\"report error\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\ndiff --git a/website/css/main.min.css b/website/css/main.min.css\nindex 818002f..e69de29 100644\n--- a/website/css/main.min.css\n+++ b/website/css/main.min.css\n@@ -1 +0,0 @@\n-@font-face{font-family:Polya;src:url(https://raw.githubusercontent.com/matteobruni/tsparticles/gh-pages/fonts/Polya.otf)}html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}.github{bottom:10px;right:10px;padding:0 12px 6px;position:fixed;border-radius:10px;background:#fff;border:1px solid #000}.github a{color:#000}.github a:active,.github a:hover,.github a:link,.github a:visited{color:#000;text-decoration:none}.github img{height:30px}.github #gh-project{font-size:20px;padding-left:5px;font-weight:700;vertical-align:bottom}.toggle-sidebar{top:50%;left:0;font-size:20px;color:#000;position:absolute;padding:3px;border-top-right-radius:5px;border-bottom-right-radius:5px;background:#e7e7e7;border:1px solid #000;border-left:none}#editor{background:#fff}[hidden]{display:none}#repulse-div{width:200px;height:200px;background-color:rgba(255,255,255,.5);border-radius:100px;position:absolute;top:50%;left:50%;margin-left:-100px;margin-top:-100px;z-index:200}@media (min-width:1600px) and (-webkit-device-pixel-ratio:1){.col-xxl-3{-ms-flex:0 0 25%;flex:0 0 25%;max-width:25%}}.btn-react{color:#fff;background-color:#61dafb;border-color:#fff}.btn-react:hover{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react.focus,.btn-react:focus{color:#fff;background-color:#5aa3c4;border-color:#ccc;box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}.btn-react.disabled,.btn-react:disabled{color:#fff;background-color:#61dafb;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active,.btn-react:not(:disabled):not(.disabled):active,.show>.btn-react.dropdown-toggle{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active:focus,.btn-react:not(:disabled):not(.disabled):active:focus,.show>.btn-react.dropdown-toggle:focus{box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}#stats,.count-particles{-webkit-user-select:none}#stats{overflow:hidden}#stats-graph canvas{border-radius:3px 3px 0 0}.count-particles{border-radius:0 0 3px 3px}\ndiff --git a/website/css/presets.min.css b/website/css/presets.min.css\nindex 6c2ae2c..e69de29 100644\n--- a/website/css/presets.min.css\n+++ b/website/css/presets.min.css\n@@ -1 +0,0 @@\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden}\n", "diff --git a/vsts.yml b/vsts.yml\nindex 6cb0eb3..a058238 100644\n--- a/vsts.yml\n+++ b/vsts.yml\n@@ -86,13 +86,13 @@ jobs:\n killall Electron\n fi\n displayName: Make sure Electron isn't running from previous tests\n-\n- - bash: |\n- cd src\n- python electron/script/verify-ffmpeg.py --source-root \"$PWD\" --build-dir out/Default --ffmpeg-path out/ffmpeg\n- displayName: Verify non proprietary ffmpeg\n- condition: and(succeeded(), eq(variables['RUN_TESTS'], '1'))\n- timeoutInMinutes: 5\n+# FIXME(alexeykuzmin)\n+# - bash: |\n+# cd src\n+# python electron/script/verify-ffmpeg.py --source-root \"$PWD\" --build-dir out/Default --ffmpeg-path out/ffmpeg\n+# displayName: Verify non proprietary ffmpeg\n+# condition: and(succeeded(), eq(variables['RUN_TESTS'], '1'))\n+# timeoutInMinutes: 5\n \n - bash: |\n cd src\n"]
2
["9acf7a062ee9c0538c2cd4661c1f5da61ab06316", "9187415f5ee35d2e88dd834e413fc16bf19c5db1"]
["build", "ci"]
remove sync ts check,refactor to get ride of cloneDeep
["diff --git a/config/webpack.config.prod.js b/config/webpack.config.prod.js\nindex 8b23fba..58a4c17 100644\n--- a/config/webpack.config.prod.js\n+++ b/config/webpack.config.prod.js\n@@ -251,7 +251,7 @@ module.exports = {\n plugins: [\n argv.notypecheck\n ? null\n- : new ForkTsCheckerWebpackPlugin({tslint: true, async: false}),\n+ : new ForkTsCheckerWebpackPlugin({tslint: true}),\n // Makes some environment variables available in index.html.\n // The public URL is available as %PUBLIC_URL% in index.html, e.g.:\n // <link rel=\"shortcut icon\" href=\"%PUBLIC_URL%/favicon.ico\">\n", "diff --git a/config/webpack.config.prod.js b/config/webpack.config.prod.js\nindex 3d2e5a6..e5219bd 100644\n--- a/config/webpack.config.prod.js\n+++ b/config/webpack.config.prod.js\n@@ -56,7 +56,7 @@ const extractTextPluginOptions = shouldUseRelativeAssetPaths\n const entries = fs.readdirSync(paths.appSrc)\n .filter(name => !name.startsWith('_'))\n .map(name => ({name, dirPath: path.join(paths.appSrc, name)}))\n- .filter(({name, dirPath}) => !/^assets|components|manifest|typings$/.test(name) && fs.lstatSync(dirPath).isDirectory())\n+ .filter(({name, dirPath}) => !/^assets|components|manifest|typings|app-config$/.test(name) && fs.lstatSync(dirPath).isDirectory())\n \n // This is the production configuration.\n // It compiles slowly and is focused on producing a fast and minimal bundle.\ndiff --git a/src/app-config/context-menus.ts b/src/app-config/context-menus.ts\nnew file mode 100644\nindex 0000000..a733b01\n--- /dev/null\n+++ b/src/app-config/context-menus.ts\n@@ -0,0 +1,27 @@\n+export function getAllContextMenus () {\n+ const allContextMenus = {\n+ google_page_translate: 'x',\n+ youdao_page_translate: 'x',\n+ google_search: 'https://www.google.com/#newwindow=1&q=%s',\n+ baidu_search: 'https://www.baidu.com/s?ie=utf-8&wd=%s',\n+ bing_search: 'https://www.bing.com/search?q=%s',\n+ google_translate: 'https://translate.google.cn/#auto/zh-CN/%s',\n+ etymonline: 'http://www.etymonline.com/index.php?search=%s',\n+ merriam_webster: 'http://www.merriam-webster.com/dictionary/%s',\n+ oxford: 'http://www.oxforddictionaries.com/us/definition/english/%s',\n+ cambridge: 'http://dictionary.cambridge.org/spellcheck/english-chinese-simplified/?q=%s',\n+ youdao: 'http://dict.youdao.com/w/%s',\n+ dictcn: 'https://dict.eudic.net/dicts/en/%s',\n+ iciba: 'http://www.iciba.com/%s',\n+ liangan: 'https://www.moedict.tw/~%s',\n+ guoyu: 'https://www.moedict.tw/%s',\n+ longman_business: 'http://www.ldoceonline.com/search/?q=%s',\n+ bing_dict: 'https://cn.bing.com/dict/?q=%s'\n+ }\n+\n+ // Just for type check. Keys in allContextMenus are useful so no actual assertion\n+ // tslint:disable-next-line:no-unused-expression\n+ allContextMenus as { [id: string]: string }\n+\n+ return allContextMenus\n+}\ndiff --git a/src/app-config/dicts.ts b/src/app-config/dicts.ts\nnew file mode 100644\nindex 0000000..905d2de\n--- /dev/null\n+++ b/src/app-config/dicts.ts\n@@ -0,0 +1,398 @@\n+import { DeepReadonly } from '@/typings/helpers'\n+\n+export function getALlDicts () {\n+ const allDicts = {\n+ bing: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word.\n+ */\n+ page: 'https://cn.bing.com/dict/search?q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 240,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ tense: true,\n+ phsym: true,\n+ cdef: true,\n+ related: true,\n+ sentence: 4\n+ }\n+ },\n+ business: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.ldoceonline.com/search/?q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ cobuild: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.collinsdictionary.com/dictionary/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 300,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ sentence: 4\n+ }\n+ },\n+ dictcn: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://dict.cn/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 300,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ chart: true,\n+ etym: true\n+ }\n+ },\n+ etymonline: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.etymonline.com/search?q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ resultnum: 2\n+ }\n+ },\n+ google: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://translate.google.com/#auto/zh-CN/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 110,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ guoyu: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.moedict.tw/%z',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ liangan: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.moedict.tw/~%z',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ macmillan: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.macmillandictionary.com/dictionary/british/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 265,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ urban: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.urbandictionary.com/define.php?term=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 180,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ },\n+ /** Optional dict custom options. Can only be boolean or number. */\n+ options: {\n+ resultnum: 4\n+ }\n+ },\n+ vocabulary: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'https://www.vocabulary.com/dictionary/%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 180,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ zdic: {\n+ /**\n+ * Full content page to jump to when user clicks the title.\n+ * %s will be replaced with the current word.\n+ * %z will be replaced with the traditional Chinese version of the current word\n+ */\n+ page: 'http://www.zdic.net/search/?c=1&q=%s',\n+ /**\n+ * If set to true, the dict start searching automatically.\n+ * Otherwise it'll only start seaching when user clicks the unfold button.\n+ * Default MUST be true and let user decide.\n+ */\n+ defaultUnfold: true,\n+ /**\n+ * This is the default height when the dict first renders the result.\n+ * If the content height is greater than the preferred height,\n+ * the preferred height is used and a mask with a view-more button is shown.\n+ * Otherwise the content height is used.\n+ */\n+ preferredHeight: 400,\n+ /**\n+ * Only start searching if the selection contains the language.\n+ * Better set default to true and let user decide.\n+ */\n+ selectionLang: {\n+ eng: true,\n+ chs: true\n+ }\n+ },\n+ }\n+\n+ // Just for type check. Keys in allDicts are useful so no actual assertion\n+ // tslint:disable-next-line:no-unused-expression\n+ allDicts as {\n+ [id: string]: {\n+ page: string\n+ defaultUnfold: boolean\n+ preferredHeight: number\n+ selectionLang: {\n+ eng: boolean\n+ chs: boolean\n+ }\n+ options?: {\n+ [option: string]: number | boolean\n+ }\n+ }\n+ }\n+\n+ return allDicts\n+}\ndiff --git a/src/app-config/index.ts b/src/app-config/index.ts\nindex 350cd8f..879a312 100644\n--- a/src/app-config/index.ts\n+++ b/src/app-config/index.ts\n@@ -1,5 +1,6 @@\n-import cloneDeep from 'lodash/cloneDeep'\n-import { DeepReadonly } from './typings/helpers'\n+import { DeepReadonly } from '@/typings/helpers'\n+import { getALlDicts } from './dicts'\n+import { getAllContextMenus } from './context-menus'\n \n const langUI = (browser.i18n.getUILanguage() || 'en').replace('-', '_')\n const langCode = /^zh_CN|zh_TW|en$/.test(langUI)\n@@ -8,220 +9,11 @@ const langCode = /^zh_CN|zh_TW|en$/.test(langUI)\n : langUI\n : 'en'\n \n-const allDicts = {\n- bing: {\n- page: 'https://cn.bing.com/dict/search?q=%s',\n- defaultUnfold: true,\n- preferredHeight: 240,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- tense: true,\n- phsym: true,\n- cdef: true,\n- related: true,\n- sentence: 4\n- }\n- },\n- business: {\n- page: 'http://www.ldoceonline.com/search/?q=%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- cobuild: {\n- page: 'https://www.collinsdictionary.com/dictionary/%s',\n- defaultUnfold: true,\n- preferredHeight: 300,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- sentence: 4\n- }\n- },\n- dictcn: {\n- page: 'http://dict.cn/%s',\n- defaultUnfold: true,\n- preferredHeight: 300,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- chart: true,\n- etym: true\n- }\n- },\n- etymonline: {\n- page: 'http://www.etymonline.com/search?q=%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- resultnum: 2\n- }\n- },\n- eudic: {\n- page: 'https://dict.eudic.net/dicts/en/%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- google: {\n- page: 'https://translate.google.com/#auto/zh-CN/%s',\n- defaultUnfold: true,\n- preferredHeight: 110,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- guoyu: {\n- page: 'https://www.moedict.tw/%z',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- howjsay: {\n- page: 'http://www.howjsay.com/index.php?word=%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- related: true\n- }\n- },\n- liangan: {\n- page: 'https://www.moedict.tw/~%z',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- macmillan: {\n- page: 'http://www.macmillandictionary.com/dictionary/british/%s',\n- defaultUnfold: true,\n- preferredHeight: 265,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- urban: {\n- page: 'http://www.urbandictionary.com/define.php?term=%s',\n- defaultUnfold: true,\n- preferredHeight: 180,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- resultnum: 4\n- }\n- },\n- vocabulary: {\n- page: 'https://www.vocabulary.com/dictionary/%s',\n- defaultUnfold: true,\n- preferredHeight: 180,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n- wordreference: {\n- page: 'http://www.wordreference.com/definition/%s',\n- defaultUnfold: true,\n- preferredHeight: 180,\n- selectionLang: {\n- eng: true,\n- chs: true\n- },\n- options: {\n- etym: true,\n- idiom: true\n- }\n- },\n- zdic: {\n- page: 'http://www.zdic.net/search/?c=1&q=%s',\n- defaultUnfold: true,\n- preferredHeight: 400,\n- selectionLang: {\n- eng: true,\n- chs: true\n- }\n- },\n-}\n-\n-// Just for type check. Keys in allDicts are useful so no actual assertion\n-// tslint:disable-next-line:no-unused-expression\n-allDicts as {\n- [id: string]: {\n- /** url for the complete result */\n- page: string\n- /** lazy load */\n- defaultUnfold: boolean\n- /** content below the preferrred height will be hidden by default */\n- preferredHeight: number\n- /** only search when the selection contains the language */\n- selectionLang: {\n- eng: boolean\n- chs: boolean\n- }\n- /** other options */\n- options?: {\n- [option: string]: number | boolean\n- }\n- }\n-}\n-\n-export type DictID = keyof typeof allDicts\n-\n-const allContextMenus = {\n- google_page_translate: 'x',\n- youdao_page_translate: 'x',\n- google_search: 'https://www.google.com/#newwindow=1&q=%s',\n- baidu_search: 'https://www.baidu.com/s?ie=utf-8&wd=%s',\n- bing_search: 'https://www.bing.com/search?q=%s',\n- google_translate: 'https://translate.google.cn/#auto/zh-CN/%s',\n- etymonline: 'http://www.etymonline.com/index.php?search=%s',\n- merriam_webster: 'http://www.merriam-webster.com/dictionary/%s',\n- oxford: 'http://www.oxforddictionaries.com/us/definition/english/%s',\n- cambridge: 'http://dictionary.cambridge.org/spellcheck/english-chinese-simplified/?q=%s',\n- youdao: 'http://dict.youdao.com/w/%s',\n- dictcn: 'https://dict.eudic.net/dicts/en/%s',\n- iciba: 'http://www.iciba.com/%s',\n- liangan: 'https://www.moedict.tw/~%s',\n- guoyu: 'https://www.moedict.tw/%s',\n- longman_business: 'http://www.ldoceonline.com/search/?q=%s',\n- bing_dict: 'https://cn.bing.com/dict/?q=%s'\n-}\n-\n-// Just for type check. Keys in allContextMenus are useful so no actual assertion\n-// tslint:disable-next-line:no-unused-expression\n-allContextMenus as { [id: string]: string }\n+export type DictConfigsMutable = ReturnType<typeof getALlDicts>\n+export type DictConfigs = DeepReadonly<DictConfigsMutable>\n+export type DictID = keyof DictConfigsMutable\n \n-export type ContextMenuDictID = keyof typeof allContextMenus\n+export type ContextMenuDictID = keyof ReturnType<typeof getAllContextMenus>\n \n export const enum TCDirection {\n center,\n@@ -238,10 +30,6 @@ export const enum TCDirection {\n /** '' means no preload */\n export type PreloadSource = '' | 'clipboard' | 'selection'\n \n-export type DictConfigs = DeepReadonly<DictConfigsMutable>\n-\n-export type DictConfigsMutable = typeof allDicts\n-\n export type AppConfig = DeepReadonly<AppConfigMutable>\n \n export interface AppConfigMutable {\n@@ -418,7 +206,7 @@ export function appConfigFactory (): AppConfig {\n },\n en: {\n dict: '',\n- list: ['bing', 'dictcn', 'howjsay', 'macmillan', 'eudic', 'urban'],\n+ list: ['bing', 'dictcn', 'macmillan', 'urban'],\n accent: 'uk' as ('us' | 'uk')\n }\n },\n@@ -426,11 +214,11 @@ export function appConfigFactory (): AppConfig {\n dicts: {\n selected: ['bing', 'urban', 'vocabulary', 'dictcn'],\n // settings of each dict will be auto-generated\n- all: cloneDeep(allDicts)\n+ all: getALlDicts()\n },\n contextMenus: {\n selected: ['oxford', 'google_translate', 'merriam_webster', 'cambridge', 'google_search', 'google_page_translate', 'youdao_page_translate'],\n- all: cloneDeep(allContextMenus)\n+ all: getAllContextMenus()\n }\n }\n }\n"]
2
["411be831591b2ea15ca9138eaf8db81f51b5101e", "d986b530775edd8ef1f4e445a5d4b0016f409722"]
["build", "refactor"]
added vue3 readme,fix unit tests
["diff --git a/core/main/README.md b/core/main/README.md\nindex e5e4c93..e9cfda9 100644\n--- a/core/main/README.md\n+++ b/core/main/README.md\n@@ -217,7 +217,7 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\n \n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/svelte/README.md)\n \n-### VueJS\n+### VueJS 2.x\n \n #### `particles.vue`\n \n@@ -225,6 +225,14 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\n \n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue/README.md)\n \n+### VueJS 3.x\n+\n+#### `particles.vue3`\n+\n+[![npm](https://img.shields.io/npm/v/particles.vue3)](https://www.npmjs.com/package/particles.vue3) [![npm](https://img.shields.io/npm/dm/particles.vue3)](https://www.npmjs.com/package/particles.vue3)\n+\n+You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue3/README.md)\n+\n ---\n \n ## **_Demo / Generator_**\ndiff --git a/core/main/tsconfig.json b/core/main/tsconfig.json\nindex 7916bc5..72399c0 100644\n--- a/core/main/tsconfig.json\n+++ b/core/main/tsconfig.json\n@@ -107,10 +107,14 @@\n \"source\": \"../../components/react/README.md\"\n },\n {\n- \"title\": \"Vue\",\n+ \"title\": \"Vue 2.x\",\n \"source\": \"../../components/vue/README.md\"\n },\n {\n+ \"title\": \"Vue 3.x\",\n+ \"source\": \"../../components/vue3/README.md\"\n+ },\n+ {\n \"title\": \"Svelte\",\n \"source\": \"../../components/svelte/README.md\"\n },\n", "diff --git a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\nindex 4d771d6..9f9683c 100644\n--- a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\n+++ b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\n@@ -9,9 +9,6 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -132,6 +129,33 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\n ]\n }\n >\n+ <AnimatedComponent\n+ pointerEvents=\"none\"\n+ style={\n+ Array [\n+ Object {\n+ \"position\": \"absolute\",\n+ },\n+ Object {\n+ \"backgroundColor\": \"rgba(0, 0, 0, 0.12)\",\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -54.666666666666664,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n+ \"transform\": Array [\n+ Object {\n+ \"translateX\": 0,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n+ },\n+ ],\n+ \"width\": 96,\n+ },\n+ ]\n+ }\n+ />\n <TouchableWithoutFeedback\n key=\"key-0\"\n onPress={[Function]}\n@@ -599,9 +623,6 @@ exports[`renders custom icon and label in shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -783,12 +804,15 @@ exports[`renders custom icon and label in shifting bottom navigation 1`] = `\n \"backgroundColor\": undefined,\n \"borderRadius\": 0,\n \"height\": 0,\n- \"left\": 6,\n+ \"left\": -4,\n \"opacity\": 0,\n \"top\": 28,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.008,\n },\n ],\n \"width\": 0,\n@@ -805,17 +829,20 @@ exports[`renders custom icon and label in shifting bottom navigation 1`] = `\n },\n Object {\n \"backgroundColor\": \"rgba(255, 255, 255, 0.12)\",\n- \"borderRadius\": 36,\n- \"height\": 72,\n- \"left\": -30,\n- \"opacity\": 0,\n- \"top\": -8,\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -52,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n },\n ],\n- \"width\": 72,\n+ \"width\": 96,\n },\n ]\n }\n@@ -1374,9 +1401,6 @@ exports[`renders non-shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -1497,6 +1521,33 @@ exports[`renders non-shifting bottom navigation 1`] = `\n ]\n }\n >\n+ <AnimatedComponent\n+ pointerEvents=\"none\"\n+ style={\n+ Array [\n+ Object {\n+ \"position\": \"absolute\",\n+ },\n+ Object {\n+ \"backgroundColor\": \"rgba(0, 0, 0, 0.12)\",\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -54.666666666666664,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n+ \"transform\": Array [\n+ Object {\n+ \"translateX\": 0,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n+ },\n+ ],\n+ \"width\": 96,\n+ },\n+ ]\n+ }\n+ />\n <TouchableWithoutFeedback\n key=\"key-0\"\n onPress={[Function]}\n@@ -2072,9 +2123,6 @@ exports[`renders shifting bottom navigation 1`] = `\n Object {\n \"flex\": 1,\n },\n- Object {\n- \"backgroundColor\": \"#000000\",\n- },\n undefined,\n ]\n }\n@@ -2256,12 +2304,15 @@ exports[`renders shifting bottom navigation 1`] = `\n \"backgroundColor\": undefined,\n \"borderRadius\": 0,\n \"height\": 0,\n- \"left\": 6,\n+ \"left\": -4,\n \"opacity\": 0,\n \"top\": 28,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.008,\n },\n ],\n \"width\": 0,\n@@ -2278,17 +2329,20 @@ exports[`renders shifting bottom navigation 1`] = `\n },\n Object {\n \"backgroundColor\": \"rgba(255, 255, 255, 0.12)\",\n- \"borderRadius\": 36,\n- \"height\": 72,\n- \"left\": -30,\n- \"opacity\": 0,\n- \"top\": -8,\n+ \"borderRadius\": 48,\n+ \"height\": 96,\n+ \"left\": -52,\n+ \"opacity\": 0.002,\n+ \"top\": -20,\n \"transform\": Array [\n Object {\n- \"scale\": 0.002,\n+ \"translateX\": 10,\n+ },\n+ Object {\n+ \"scale\": 0.001,\n },\n ],\n- \"width\": 72,\n+ \"width\": 96,\n },\n ]\n }\n"]
2
["e4c3e2cff769ce46d22d5c8f7dd527510443a8a7", "87427fe39d165bee2acedde8dbaa237cca3fb61e"]
["docs", "test"]
pin version of actionlint used,fix "types" field in dist
["diff --git a/.github/workflows/introspect.yml b/.github/workflows/introspect.yml\nindex b6d9125..82d22a5 100644\n--- a/.github/workflows/introspect.yml\n+++ b/.github/workflows/introspect.yml\n@@ -25,5 +25,5 @@ jobs:\n # From https://github.com/rhysd/actionlint/blob/main/docs/usage.md#use-actionlint-on-github-actions\n - name: Check workflow files\n run: |\n- bash <(curl https://raw.githubusercontent.com/rhysd/actionlint/590d3bd9dde0c91f7a66071d40eb84716526e5a6/scripts/download-actionlint.bash)\n+ bash <(curl https://raw.githubusercontent.com/rhysd/actionlint/590d3bd9dde0c91f7a66071d40eb84716526e5a6/scripts/download-actionlint.bash) 1.6.25\n ./actionlint -color -shellcheck=\"\"\n", "diff --git a/scripts/prepare.js b/scripts/prepare.js\nindex 4bab09b..55f459b 100644\n--- a/scripts/prepare.js\n+++ b/scripts/prepare.js\n@@ -96,7 +96,6 @@ async function prepare() {\n delete json.private\n delete json.scripts\n delete json.devDependencies\n- delete json.types\n \n // Add \"postinstall\" script for donations.\n if (/(native|core)$/.test(name))\n@@ -128,6 +127,7 @@ async function prepare() {\n else {\n json.main = json.main.replace(/^dist\\//, '')\n if (json.main.endsWith('.cjs.js')) {\n+ json.types = json.main.replace('.cjs.js', '.d.ts')\n json.module = json.main.replace('.cjs', '')\n }\n }\n"]
2
["b702adc245f679ae20d84de39f0d63b14aabed5d", "f14ef3809f456aadd73523e47cb16c5d15e9a9df"]
["ci", "build"]
remove unnecessary start argument from `range`,create dashboards from imported templates
["diff --git a/ibis/backends/dask/tests/execution/test_window.py b/ibis/backends/dask/tests/execution/test_window.py\nindex 75a7331..6bfc5e3 100644\n--- a/ibis/backends/dask/tests/execution/test_window.py\n+++ b/ibis/backends/dask/tests/execution/test_window.py\n@@ -489,7 +489,7 @@ def test_project_list_scalar(npartitions):\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\n result = expr.execute()\n \n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\"res\")\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\"res\")\n tm.assert_series_equal(result.res, expected)\n \n \ndiff --git a/ibis/backends/pandas/tests/execution/test_window.py b/ibis/backends/pandas/tests/execution/test_window.py\nindex 8f292b3..effa372 100644\n--- a/ibis/backends/pandas/tests/execution/test_window.py\n+++ b/ibis/backends/pandas/tests/execution/test_window.py\n@@ -436,7 +436,7 @@ def test_project_list_scalar():\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\n result = expr.execute()\n \n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\"res\")\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\"res\")\n tm.assert_series_equal(result.res, expected)\n \n \ndiff --git a/ibis/backends/pyspark/tests/test_basic.py b/ibis/backends/pyspark/tests/test_basic.py\nindex 3850919..14fe677 100644\n--- a/ibis/backends/pyspark/tests/test_basic.py\n+++ b/ibis/backends/pyspark/tests/test_basic.py\n@@ -19,7 +19,7 @@ from ibis.backends.pyspark.compiler import _can_be_replaced_by_column_name # no\n def test_basic(con):\n table = con.table(\"basic_table\")\n result = table.compile().toPandas()\n- expected = pd.DataFrame({\"id\": range(0, 10), \"str_col\": \"value\"})\n+ expected = pd.DataFrame({\"id\": range(10), \"str_col\": \"value\"})\n \n tm.assert_frame_equal(result, expected)\n \n@@ -28,9 +28,7 @@ def test_projection(con):\n table = con.table(\"basic_table\")\n result1 = table.mutate(v=table[\"id\"]).compile().toPandas()\n \n- expected1 = pd.DataFrame(\n- {\"id\": range(0, 10), \"str_col\": \"value\", \"v\": range(0, 10)}\n- )\n+ expected1 = pd.DataFrame({\"id\": range(10), \"str_col\": \"value\", \"v\": range(10)})\n \n result2 = (\n table.mutate(v=table[\"id\"])\n@@ -44,8 +42,8 @@ def test_projection(con):\n {\n \"id\": range(0, 20, 2),\n \"str_col\": \"value\",\n- \"v\": range(0, 10),\n- \"v2\": range(0, 10),\n+ \"v\": range(10),\n+ \"v2\": range(10),\n }\n )\n \n", "diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex 590f5ea..bd74e95 100644\n--- a/CHANGELOG.md\n+++ b/CHANGELOG.md\n@@ -1,6 +1,7 @@\n ## v2.0.0-alpha.6 [unreleased]\n \n ### Features\n+1. [12496](https://github.com/influxdata/influxdb/pull/12496): Add ability to import a dashboard\n \n ### Bug Fixes\n \ndiff --git a/ui/src/dashboards/actions/v2/index.ts b/ui/src/dashboards/actions/v2/index.ts\nindex ad0d13d..35babc7 100644\n--- a/ui/src/dashboards/actions/v2/index.ts\n+++ b/ui/src/dashboards/actions/v2/index.ts\n@@ -16,6 +16,7 @@ import {\n removeDashboardLabels as removeDashboardLabelsAJAX,\n updateView as updateViewAJAX,\n } from 'src/dashboards/apis/v2'\n+import {client} from 'src/utils/api'\n \n // Actions\n import {notify} from 'src/shared/actions/notifications'\n@@ -25,6 +26,10 @@ import {\n DeleteTimeRangeAction,\n } from 'src/dashboards/actions/v2/ranges'\n import {setView, SetViewAction} from 'src/dashboards/actions/v2/views'\n+import {\n+ importDashboardSucceeded,\n+ importDashboardFailed,\n+} from 'src/shared/copy/notifications'\n \n // Utils\n import {\n@@ -38,7 +43,7 @@ import * as copy from 'src/shared/copy/notifications'\n // Types\n import {RemoteDataState} from 'src/types'\n import {PublishNotificationAction} from 'src/types/actions/notifications'\n-import {CreateCell} from '@influxdata/influx'\n+import {CreateCell, IDashboardTemplate} from '@influxdata/influx'\n import {Dashboard, NewView, Cell} from 'src/types/v2'\n import {ILabel} from '@influxdata/influx'\n \n@@ -201,6 +206,19 @@ export const getDashboardsAsync = () => async (\n }\n }\n \n+export const createDashboardFromTemplate = (\n+ template: IDashboardTemplate,\n+ orgID: string\n+) => async dispatch => {\n+ try {\n+ await client.dashboards.createFromTemplate(template, orgID)\n+\n+ dispatch(notify(importDashboardSucceeded()))\n+ } catch (error) {\n+ dispatch(notify(importDashboardFailed(error)))\n+ }\n+}\n+\n export const importDashboardAsync = (dashboard: Dashboard) => async (\n dispatch: Dispatch<Action>\n ): Promise<void> => {\ndiff --git a/ui/src/dashboards/components/ImportDashboardOverlay.tsx b/ui/src/dashboards/components/ImportDashboardOverlay.tsx\nindex 37ef80b..84216c3 100644\n--- a/ui/src/dashboards/components/ImportDashboardOverlay.tsx\n+++ b/ui/src/dashboards/components/ImportDashboardOverlay.tsx\n@@ -1,70 +1,80 @@\n+// Libraries\n import React, {PureComponent} from 'react'\n import _ from 'lodash'\n+import {connect} from 'react-redux'\n \n-import Container from 'src/clockface/components/overlays/OverlayContainer'\n-import Heading from 'src/clockface/components/overlays/OverlayHeading'\n-import Body from 'src/clockface/components/overlays/OverlayBody'\n-import DragAndDrop from 'src/shared/components/DragAndDrop'\n+// Constants\n import {dashboardImportFailed} from 'src/shared/copy/notifications'\n \n-import {Dashboard} from 'src/types/v2'\n-import {Notification} from 'src/types/notifications'\n+// Actions\n+import {notify as notifyAction} from 'src/shared/actions/notifications'\n+import {getDashboardsAsync} from 'src/dashboards/actions/v2'\n \n-interface Props {\n+// Types\n+import ImportOverlay from 'src/shared/components/ImportOverlay'\n+import {createDashboardFromTemplate as createDashboardFromTemplateAction} from 'src/dashboards/actions/v2'\n+\n+interface OwnProps {\n onDismissOverlay: () => void\n- onImportDashboard: (dashboard: Dashboard) => void\n- notify: (message: Notification) => void\n+ orgID: string\n+ isVisible: boolean\n+}\n+interface DispatchProps {\n+ notify: typeof notifyAction\n+ createDashboardFromTemplate: typeof createDashboardFromTemplateAction\n+ populateDashboards: typeof getDashboardsAsync\n }\n+\n+type Props = OwnProps & DispatchProps\n+\n class ImportDashboardOverlay extends PureComponent<Props> {\n constructor(props: Props) {\n super(props)\n }\n \n public render() {\n- const {onDismissOverlay} = this.props\n+ const {isVisible, onDismissOverlay} = this.props\n \n return (\n- <Container maxWidth={800}>\n- <Heading title=\"Import Dashboard\" onDismiss={onDismissOverlay} />\n- <Body>\n- <DragAndDrop\n- submitText=\"Upload Dashboard\"\n- fileTypesToAccept={this.validFileExtension}\n- handleSubmit={this.handleUploadDashboard}\n- />\n- </Body>\n- </Container>\n+ <ImportOverlay\n+ isVisible={isVisible}\n+ onDismissOverlay={onDismissOverlay}\n+ resourceName=\"Dashboard\"\n+ onSubmit={this.handleUploadDashboard}\n+ />\n )\n }\n \n- private get validFileExtension(): string {\n- return '.json'\n- }\n-\n- private handleUploadDashboard = (\n- uploadContent: string,\n- fileName: string\n- ): void => {\n- const {notify, onImportDashboard, onDismissOverlay} = this.props\n- const fileExtensionRegex = new RegExp(`${this.validFileExtension}$`)\n- if (!fileName.match(fileExtensionRegex)) {\n- notify(dashboardImportFailed('Please import a JSON file'))\n- return\n- }\n+ private handleUploadDashboard = async (\n+ uploadContent: string\n+ ): Promise<void> => {\n+ const {\n+ notify,\n+ createDashboardFromTemplate,\n+ onDismissOverlay,\n+ populateDashboards,\n+ orgID,\n+ } = this.props\n \n try {\n- const {dashboard} = JSON.parse(uploadContent)\n+ const template = JSON.parse(uploadContent)\n \n- if (!_.isEmpty(dashboard)) {\n- onImportDashboard(dashboard)\n- onDismissOverlay()\n- } else {\n- notify(dashboardImportFailed('No dashboard found in file'))\n- }\n+ await createDashboardFromTemplate(template, orgID)\n+ await populateDashboards()\n+\n+ onDismissOverlay()\n } catch (error) {\n notify(dashboardImportFailed(error))\n }\n }\n }\n+const mdtp: DispatchProps = {\n+ notify: notifyAction,\n+ createDashboardFromTemplate: createDashboardFromTemplateAction,\n+ populateDashboards: getDashboardsAsync,\n+}\n \n-export default ImportDashboardOverlay\n+export default connect<{}, DispatchProps, OwnProps>(\n+ null,\n+ mdtp\n+)(ImportDashboardOverlay)\ndiff --git a/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx b/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\nindex d6b299f..1ff7f47 100644\n--- a/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\n+++ b/ui/src/dashboards/components/dashboard_index/DashboardsIndex.tsx\n@@ -2,15 +2,15 @@\n import React, {PureComponent} from 'react'\n import {InjectedRouter} from 'react-router'\n import {connect} from 'react-redux'\n-import {isEmpty} from 'lodash'\n+import {get} from 'lodash'\n \n // Components\n import DashboardsIndexContents from 'src/dashboards/components/dashboard_index/DashboardsIndexContents'\n import {Page} from 'src/pageLayout'\n import SearchWidget from 'src/shared/components/search_widget/SearchWidget'\n import AddResourceDropdown from 'src/shared/components/AddResourceDropdown'\n-import ImportOverlay from 'src/shared/components/ImportOverlay'\n import ExportOverlay from 'src/shared/components/ExportOverlay'\n+import ImportDashboardOverlay from 'src/dashboards/components/ImportDashboardOverlay'\n \n // APIs\n import {createDashboard, cloneDashboard} from 'src/dashboards/apis/v2/'\n@@ -32,10 +32,7 @@ import {DEFAULT_DASHBOARD_NAME} from 'src/dashboards/constants/index'\n import {\n dashboardSetDefaultFailed,\n dashboardCreateFailed,\n- dashboardImported,\n- dashboardImportFailed,\n } from 'src/shared/copy/notifications'\n-import {cantImportInvalidResource} from 'src/shared/copy/v2/notifications'\n \n // Types\n import {Notification} from 'src/types/notifications'\n@@ -197,24 +194,6 @@ class DashboardIndex extends PureComponent<Props, State> {\n this.props.handleDeleteDashboard(dashboard)\n }\n \n- private handleImportDashboard = async (\n- importString: string\n- ): Promise<void> => {\n- const {notify} = this.props\n- try {\n- const resource = JSON.parse(importString)\n-\n- if (isEmpty(resource)) {\n- notify(cantImportInvalidResource('Dashboard'))\n- return\n- }\n- this.handleToggleImportOverlay()\n- notify(dashboardImported())\n- } catch (error) {\n- notify(dashboardImportFailed(error))\n- }\n- }\n-\n private handleFilterDashboards = (searchTerm: string): void => {\n this.setState({searchTerm})\n }\n@@ -229,13 +208,13 @@ class DashboardIndex extends PureComponent<Props, State> {\n \n private get importOverlay(): JSX.Element {\n const {isImportingDashboard} = this.state\n+ const {orgs} = this.props\n \n return (\n- <ImportOverlay\n- isVisible={isImportingDashboard}\n- resourceName=\"Dashboard\"\n+ <ImportDashboardOverlay\n onDismissOverlay={this.handleToggleImportOverlay}\n- onSubmit={this.handleImportDashboard}\n+ orgID={get(orgs, '0.id', '')}\n+ isVisible={isImportingDashboard}\n />\n )\n }\ndiff --git a/ui/src/organizations/components/Dashboards.tsx b/ui/src/organizations/components/Dashboards.tsx\nindex 08c8402..90c2514 100644\n--- a/ui/src/organizations/components/Dashboards.tsx\n+++ b/ui/src/organizations/components/Dashboards.tsx\n@@ -6,13 +6,10 @@ import _ from 'lodash'\n \n // Components\n import DashboardsIndexContents from 'src/dashboards/components/dashboard_index/DashboardsIndexContents'\n-import {OverlayTechnology, Input, Tabs} from 'src/clockface'\n+import {Input, Tabs} from 'src/clockface'\n import {Button, ComponentColor, IconFont} from '@influxdata/clockface'\n import ImportDashboardOverlay from 'src/dashboards/components/ImportDashboardOverlay'\n \n-// Utils\n-import {getDeep} from 'src/utils/wrappers'\n-\n // APIs\n import {createDashboard, cloneDashboard} from 'src/dashboards/apis/v2/'\n \n@@ -39,7 +36,7 @@ import {DEFAULT_DASHBOARD_NAME} from 'src/dashboards/constants/index'\n \n // Types\n import {Notification} from 'src/types/notifications'\n-import {Links, Cell, Dashboard, AppState, Organization} from 'src/types/v2'\n+import {Links, Dashboard, AppState, Organization} from 'src/types/v2'\n \n // Decorators\n import {ErrorHandling} from 'src/shared/decorators/errors'\n@@ -205,46 +202,20 @@ class Dashboards extends PureComponent<Props, State> {\n this.props.handleDeleteDashboard(dashboard)\n }\n \n- private handleImportDashboard = async (\n- dashboard: Dashboard\n- ): Promise<void> => {\n- const defaultCell = {\n- x: 0,\n- y: 0,\n- w: 4,\n- h: 4,\n- }\n-\n- const name = _.get(dashboard, 'name', DEFAULT_DASHBOARD_NAME)\n- const cellsWithDefaultsApplied = getDeep<Cell[]>(\n- dashboard,\n- 'cells',\n- []\n- ).map(c => ({...defaultCell, ...c}))\n-\n- await this.props.handleImportDashboard({\n- ...dashboard,\n- name,\n- cells: cellsWithDefaultsApplied,\n- })\n- }\n-\n private handleToggleOverlay = (): void => {\n this.setState({isImportingDashboard: !this.state.isImportingDashboard})\n }\n \n private get renderImportOverlay(): JSX.Element {\n- const {notify} = this.props\n const {isImportingDashboard} = this.state\n+ const {orgs} = this.props\n \n return (\n- <OverlayTechnology visible={isImportingDashboard}>\n- <ImportDashboardOverlay\n- onDismissOverlay={this.handleToggleOverlay}\n- onImportDashboard={this.handleImportDashboard}\n- notify={notify}\n- />\n- </OverlayTechnology>\n+ <ImportDashboardOverlay\n+ onDismissOverlay={this.handleToggleOverlay}\n+ orgID={_.get(orgs, '0.id', '')}\n+ isVisible={isImportingDashboard}\n+ />\n )\n }\n }\ndiff --git a/ui/src/shared/components/ImportOverlay.tsx b/ui/src/shared/components/ImportOverlay.tsx\nindex 10f1d50..476fa70 100644\n--- a/ui/src/shared/components/ImportOverlay.tsx\n+++ b/ui/src/shared/components/ImportOverlay.tsx\n@@ -10,6 +10,7 @@ import {\n OverlayHeading,\n OverlayFooter,\n Radio,\n+ ComponentStatus,\n } from 'src/clockface'\n import {Button, ComponentColor} from '@influxdata/clockface'\n \n@@ -93,6 +94,7 @@ export default class ImportOverlay extends PureComponent<Props, State> {\n submitText=\"Upload\"\n handleSubmit={this.handleSetImportContent}\n submitOnDrop={true}\n+ submitOnUpload={true}\n onCancel={this.clearImportContent}\n />\n )\n@@ -110,18 +112,21 @@ export default class ImportOverlay extends PureComponent<Props, State> {\n private get submitButton(): JSX.Element {\n const {resourceName} = this.props\n const {selectedImportOption, importContent} = this.state\n- if (\n+ const isEnabled =\n selectedImportOption === ImportOption.Paste ||\n (selectedImportOption === ImportOption.Upload && importContent)\n- ) {\n- return (\n- <Button\n- text={`Import JSON as ${resourceName}`}\n- onClick={this.submit}\n- color={ComponentColor.Primary}\n- />\n- )\n- }\n+ const status = isEnabled\n+ ? ComponentStatus.Default\n+ : ComponentStatus.Disabled\n+\n+ return (\n+ <Button\n+ text={`Import JSON as ${resourceName}`}\n+ onClick={this.submit}\n+ color={ComponentColor.Primary}\n+ status={status}\n+ />\n+ )\n }\n \n private submit = () => {\ndiff --git a/ui/src/shared/copy/notifications.ts b/ui/src/shared/copy/notifications.ts\nindex 4492078..86700e2 100644\n--- a/ui/src/shared/copy/notifications.ts\n+++ b/ui/src/shared/copy/notifications.ts\n@@ -799,9 +799,18 @@ export const importTaskSucceeded = (): Notification => ({\n })\n \n export const importTaskFailed = (error: string): Notification => ({\n- ...defaultSuccessNotification,\n+ ...defaultErrorNotification,\n message: `Failed to import task: ${error}`,\n })\n+export const importDashboardSucceeded = (): Notification => ({\n+ ...defaultSuccessNotification,\n+ message: `Successfully imported dashboard.`,\n+})\n+\n+export const importDashboardFailed = (error: string): Notification => ({\n+ ...defaultErrorNotification,\n+ message: `Failed to import dashboard: ${error}`,\n+})\n \n // Labels\n export const getLabelsFailed = (): Notification => ({\n"]
2
["15f8d95754a0b6865ea475ca9e515272a07bf6ba", "9114362b39f5194209cd0b330af7076333f3db77"]
["refactor", "feat"]
release for ppc64 closes #3703 Signed-off-by: Carlos A Becker <[email protected]>,return Animated nodes passed to "getAnimated" ...instead of undefined. Also, stop using instanceof in "isAnimated" for perf.
["diff --git a/.goreleaser.yaml b/.goreleaser.yaml\nindex 46901cb..7d4d355 100644\n--- a/.goreleaser.yaml\n+++ b/.goreleaser.yaml\n@@ -25,6 +25,7 @@ builds:\n - amd64\n - arm\n - arm64\n+ - ppc64\n goarm:\n - \"7\"\n mod_timestamp: '{{ .CommitTimestamp }}'\n", "diff --git a/packages/animated/src/Animated.ts b/packages/animated/src/Animated.ts\nindex 00daa96..05ff7f9 100644\n--- a/packages/animated/src/Animated.ts\n+++ b/packages/animated/src/Animated.ts\n@@ -4,7 +4,7 @@ import { AnimatedValue } from './AnimatedValue'\n const $node: any = Symbol.for('Animated:node')\n \n export const isAnimated = (value: any): value is Animated =>\n- value instanceof Animated\n+ !!value && value[$node] === value\n \n /** Get the owner's `Animated` node. */\n export const getAnimated = (owner: any): Animated | undefined =>\n@@ -23,6 +23,10 @@ export abstract class Animated<T = any> {\n protected payload?: Payload\n \n /** Returns every value of the node. Pass true for only the animated values. */\n+ constructor() {\n+ setAnimated(this, this)\n+ }\n+\n abstract getValue(animated?: boolean): T\n \n abstract setValue(value: T): void\n"]
2
["e27e3a6478d59eb0f93af0a51a9c474bad6f8350", "eb513f7eeea7865f15e5bd561a471d1f4381ea70"]
["build", "fix"]
README,remove unused branches and ignore envrc file
["diff --git a/README.md b/README.md\nindex 11a24b3..56e8d2a 100644\n--- a/README.md\n+++ b/README.md\n@@ -16,7 +16,9 @@ oclif: create your own CLI\n \n # Description\n \n-This is a framework for building CLIs in Node.js. This framework was built out of the [Heroku CLI](https://cli.heroku.com) but generalized to build any custom CLI. It's designed both for simple CLIs that can be just a single file with a few flag options, or for very complex CLIs that have many commands (like git or heroku). Most CLI tools in Node are simple flag parsers but oclif is much more than that\u2014though without the overhead of making simple CLIs quick to write with minimal boilerplate.\n+This is a framework for building CLIs in Node.js. This framework was built out of the [Heroku CLI](https://cli.heroku.com) but generalized to build any custom CLI. It's designed both for simple CLIs that can be just a single file with a few flag options, or for very complex CLIs that have many commands (like git or heroku).\n+\n+Most CLI tools for Node are simple flag parsers but oclif is much more than that\u2014though without the overhead of making simple CLIs quick to write with minimal boilerplate.\n \n # Features\n \n", "diff --git a/.github/workflows/ibis-backends-cloud.yml b/.github/workflows/ibis-backends-cloud.yml\nindex 2003e8e..7c7fd26 100644\n--- a/.github/workflows/ibis-backends-cloud.yml\n+++ b/.github/workflows/ibis-backends-cloud.yml\n@@ -5,9 +5,12 @@ on:\n # Skip the backend suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n+ - \"**/*.md\"\n+ - \"**/*.qmd\"\n+ - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n- - quarto\n \n permissions:\n # this allows extractions/setup-just to list releases for `just` at a higher\ndiff --git a/.github/workflows/ibis-backends-skip-helper.yml b/.github/workflows/ibis-backends-skip-helper.yml\nindex 5d5f3f7..0471994 100644\n--- a/.github/workflows/ibis-backends-skip-helper.yml\n+++ b/.github/workflows/ibis-backends-skip-helper.yml\n@@ -9,20 +9,20 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_backends:\ndiff --git a/.github/workflows/ibis-backends.yml b/.github/workflows/ibis-backends.yml\nindex 4a1cae9..30e6c1a 100644\n--- a/.github/workflows/ibis-backends.yml\n+++ b/.github/workflows/ibis-backends.yml\n@@ -8,10 +8,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the backend suite if all changes are docs\n paths-ignore:\n@@ -19,10 +19,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\nindex 1adda11..b528a30 100644\n--- a/.github/workflows/ibis-docs-lint.yml\n+++ b/.github/workflows/ibis-docs-lint.yml\n@@ -5,12 +5,10 @@ on:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-main-skip-helper.yml b/.github/workflows/ibis-main-skip-helper.yml\nindex a5fdc6f..0fb5dea 100644\n--- a/.github/workflows/ibis-main-skip-helper.yml\n+++ b/.github/workflows/ibis-main-skip-helper.yml\n@@ -8,19 +8,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_core:\ndiff --git a/.github/workflows/ibis-main.yml b/.github/workflows/ibis-main.yml\nindex aa31436..0b1536a 100644\n--- a/.github/workflows/ibis-main.yml\n+++ b/.github/workflows/ibis-main.yml\n@@ -7,20 +7,20 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the test suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-tpch-queries-skip-helper.yml b/.github/workflows/ibis-tpch-queries-skip-helper.yml\nindex 1f1c0bc..f10fb8d 100644\n--- a/.github/workflows/ibis-tpch-queries-skip-helper.yml\n+++ b/.github/workflows/ibis-tpch-queries-skip-helper.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-tpch-queries.yml b/.github/workflows/ibis-tpch-queries.yml\nindex b4f8a48..9e65a61 100644\n--- a/.github/workflows/ibis-tpch-queries.yml\n+++ b/.github/workflows/ibis-tpch-queries.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/nix-skip-helper.yml b/.github/workflows/nix-skip-helper.yml\nindex 677b4d7..e0ab8f7 100644\n--- a/.github/workflows/nix-skip-helper.yml\n+++ b/.github/workflows/nix-skip-helper.yml\n@@ -9,19 +9,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n jobs:\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\nindex f2dd3f0..7ea9e26 100644\n--- a/.github/workflows/nix.yml\n+++ b/.github/workflows/nix.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\n"]
2
["363f84c7da411468b4103da8e0b39ca48cfd8327", "d0c6476df61b9c6ab07b87e1724ea7c5318595bb"]
["docs", "ci"]
reuse backup id parameter definition,update README.md about the NPM package
["diff --git a/dist/src/main/resources/api/backup-management-api.yaml b/dist/src/main/resources/api/backup-management-api.yaml\nindex d09fce8..499b0b6 100644\n--- a/dist/src/main/resources/api/backup-management-api.yaml\n+++ b/dist/src/main/resources/api/backup-management-api.yaml\n@@ -65,13 +65,7 @@ paths:\n summary: Get information of a backup\n description: A detailed information of the backup with the give backup id.\n parameters:\n- - name: backupId\n- in : path\n- description: Backup ID\n- required: true\n- schema:\n- type: integer\n- format: int64\n+ - $ref: '#/components/parameters/BackupId'\n responses:\n '200':\n description: OK\n@@ -92,13 +86,7 @@ paths:\n summary: Delete a backup\n description: Delete a backup with the given id\n parameters:\n- - name: backupId\n- in: path\n- description: Backup ID\n- required: true\n- schema:\n- type: integer\n- format: int64\n+ - $ref: '#/components/parameters/BackupId'\n responses:\n '204':\n description: Backup is deleted\n@@ -110,6 +98,15 @@ paths:\n $ref: '#/components/responses/Error'\n \n components:\n+ parameters:\n+ BackupId:\n+ name: backupId\n+ required: true\n+ in: path\n+ description: Id of the backup\n+ schema:\n+ $ref: '#/components/schemas/BackupId'\n+\n responses:\n Error:\n description: Generic error response\n", "diff --git a/README.md b/README.md\nindex 9faf168..bbb5b5c 100644\n--- a/README.md\n+++ b/README.md\n@@ -126,23 +126,24 @@ pacman -S git-cliff\n \n ### From NPM\n \n-[git-cliff](https://www.npmjs.com/package/git-cliff) can be installed from NPM:\n+\n+You can install and run [git-cliff](https://www.npmjs.com/package/git-cliff) with a single command:\n \n ```sh\n-yarn add -D git-cliff\n+npx git-cliff@latest\n ```\n \n-or:\n+Also, if you want to add `git-cliff` to your project:\n \n ```sh\n+# with yarn\n+yarn add -D git-cliff\n+\n+# with npm\n npm install git-cliff --save-dev\n ```\n \n-You can also use `git-cliff` directly with `npx`:\n-\n-```sh\n-npx git-cliff\n-```\n+Afterwards, you can run `git-cliff` via `npm exec git-cliff` or `npx git-cliff@latest`.\n \n ### From MacPorts\n \n"]
2
["f7ee0c0983c49a39a6780a751c112904c3b234ca", "e0177c25e13812306aab0b0991562d58b6d14767"]
["refactor", "docs"]
init environ cache,post installers compatiblity with Windows #2520
["diff --git a/src/environment.go b/src/environment.go\nindex ae5e26a..0c961c5 100644\n--- a/src/environment.go\n+++ b/src/environment.go\n@@ -229,6 +229,7 @@ func (env *environment) environ() map[string]string {\n \tif env.environCache != nil {\n \t\treturn env.environCache\n \t}\n+\tenv.environCache = make(map[string]string)\n \tconst separator = \"=\"\n \tvalues := os.Environ()\n \tfor value := range values {\n", "diff --git a/packages/cubejs-databricks-jdbc-driver/package.json b/packages/cubejs-databricks-jdbc-driver/package.json\nindex cc164f0..fd7ad45 100644\n--- a/packages/cubejs-databricks-jdbc-driver/package.json\n+++ b/packages/cubejs-databricks-jdbc-driver/package.json\n@@ -14,13 +14,16 @@\n },\n \"main\": \"dist/src/index.js\",\n \"typings\": \"dist/src/index.d.ts\",\n+ \"bin\": {\n+ \"databricks-jdbc-installer\": \"bin/post-install\"\n+ },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n \"tsc\": \"tsc\",\n \"watch\": \"tsc -w\",\n \"lint\": \"eslint src/* --ext .ts\",\n \"lint:fix\": \"eslint --fix src/* --ext .ts\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"databricks-jdbc-installer\"\n },\n \"files\": [\n \"README.md\",\ndiff --git a/rust/package.json b/rust/package.json\nindex b139279..5bf6446 100644\n--- a/rust/package.json\n+++ b/rust/package.json\n@@ -8,7 +8,8 @@\n \"node\": \">=10.8.0\"\n },\n \"bin\": {\n- \"cubestore-dev\": \"bin/cubestore-dev\"\n+ \"cubestore-dev\": \"bin/cubestore-dev\",\n+ \"cubestore-installer\": \"bin/post-install\"\n },\n \"scripts\": {\n \"build\": \"rm -rf dist && npm run tsc\",\n@@ -18,7 +19,7 @@\n \"lint:fix\": \"eslint --fix js-wrapper/* --ext .ts,js\",\n \"unit\": \"jest\",\n \"unit:debug\": \"jest --runInBand\",\n- \"postinstall\": \"bin/post-install\"\n+ \"postinstall\": \"cubestore-installer\"\n },\n \"files\": [\n \"dist\",\ndiff --git a/yarn.lock b/yarn.lock\nindex d2a4038..b59bb77 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -4036,9 +4036,9 @@\n integrity sha512-7btbphLrKvo5yl/5CC2OCxUSMx1wV1wvGT1qDXkSt7yi00/YW7E8k6qzXqJHsp+WU0eoG7r6MTQQXI9lIvd0qA==\n \n \"@types/fs-extra@^9.0.1\", \"@types/fs-extra@^9.0.2\", \"@types/fs-extra@^9.0.8\":\n- version \"9.0.10\"\n- resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.10.tgz#8023a72e3d06cf54929ea47ec7634e47f33f4046\"\n- integrity sha512-O9T2LLkRDiTlalOBdjEkcnT0MRdT2+wglCl7pJUJ3mkWkR8hX4K+5bg2raQNJcLv4V8zGuTXe7Ud3wSqkTyuyQ==\n+ version \"9.0.11\"\n+ resolved \"https://registry.yarnpkg.com/@types/fs-extra/-/fs-extra-9.0.11.tgz#8cc99e103499eab9f347dbc6ca4e99fb8d2c2b87\"\n+ integrity sha512-mZsifGG4QeQ7hlkhO56u7zt/ycBgGxSVsFI/6lGTU34VtwkiqrrSDgw0+ygs8kFGWcXnFQWMrzF2h7TtDFNixA==\n dependencies:\n \"@types/node\" \"*\"\n \n@@ -5306,9 +5306,9 @@ acorn@^7.0.0, acorn@^7.1.0, acorn@^7.1.1, acorn@^7.4.0:\n integrity sha512-nQyp0o1/mNdbTO1PO6kHkwSrmgZ0MT/jCCpNiwbUjGoRN4dlBhqJtoQuCnEOKzgTVwg0ZWiCoQy6SxMebQVh8A==\n \n acorn@^8.1.0:\n- version \"8.1.0\"\n- resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.0.tgz#52311fd7037ae119cbb134309e901aa46295b3fe\"\n- integrity sha512-LWCF/Wn0nfHOmJ9rzQApGnxnvgfROzGilS8936rqN/lfcYkY9MYZzdMqN+2NJ4SlTc+m5HiSa+kNfDtI64dwUA==\n+ version \"8.1.1\"\n+ resolved \"https://registry.yarnpkg.com/acorn/-/acorn-8.1.1.tgz#fb0026885b9ac9f48bac1e185e4af472971149ff\"\n+ integrity sha512-xYiIVjNuqtKXMxlRMDc6mZUhXehod4a3gbZ1qRlM7icK4EbxUFNLhWoPblCvFtB2Y9CIqHP3CF/rdxLItaQv8g==\n \n adal-node@^0.1.28:\n version \"0.1.28\"\n@@ -5441,9 +5441,9 @@ ajv@^6.1.0, ajv@^6.10.0, ajv@^6.10.2, ajv@^6.12.2, ajv@^6.12.3, ajv@^6.12.4, ajv\n uri-js \"^4.2.2\"\n \n ajv@^8.0.1:\n- version \"8.0.5\"\n- resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.0.5.tgz#f07d6fdeffcdbb80485570ce3f1bc845fcc812b9\"\n- integrity sha512-RkiLa/AeJx7+9OvniQ/qeWu0w74A8DiPPBclQ6ji3ZQkv5KamO+QGpqmi7O4JIw3rHGUXZ6CoP9tsAkn3gyazg==\n+ version \"8.1.0\"\n+ resolved \"https://registry.yarnpkg.com/ajv/-/ajv-8.1.0.tgz#45d5d3d36c7cdd808930cc3e603cf6200dbeb736\"\n+ integrity sha512-B/Sk2Ix7A36fs/ZkuGLIR86EdjbgR6fsAcbx9lOP/QBSXujDNbVmIS/U4Itz5k8fPFDeVZl/zQ/gJW4Jrq6XjQ==\n dependencies:\n fast-deep-equal \"^3.1.1\"\n json-schema-traverse \"^1.0.0\"\n@@ -6828,15 +6828,15 @@ [email protected]:\n node-releases \"^1.1.61\"\n \n browserslist@^4.0.0, browserslist@^4.11.1, browserslist@^4.12.0, browserslist@^4.14.5, browserslist@^4.16.3, browserslist@^4.3.4, browserslist@^4.6.2, browserslist@^4.6.4, browserslist@^4.7.0, browserslist@^4.9.1:\n- version \"4.16.3\"\n- resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.3.tgz#340aa46940d7db878748567c5dea24a48ddf3717\"\n- integrity sha512-vIyhWmIkULaq04Gt93txdh+j02yX/JzlyhLYbV3YQCn/zvES3JnY7TifHHvvr1w5hTDluNKMkV05cs4vy8Q7sw==\n+ version \"4.16.4\"\n+ resolved \"https://registry.yarnpkg.com/browserslist/-/browserslist-4.16.4.tgz#7ebf913487f40caf4637b892b268069951c35d58\"\n+ integrity sha512-d7rCxYV8I9kj41RH8UKYnvDYCRENUlHRgyXy/Rhr/1BaeLGfiCptEdFE8MIrvGfWbBFNjVYx76SQWvNX1j+/cQ==\n dependencies:\n- caniuse-lite \"^1.0.30001181\"\n- colorette \"^1.2.1\"\n- electron-to-chromium \"^1.3.649\"\n+ caniuse-lite \"^1.0.30001208\"\n+ colorette \"^1.2.2\"\n+ electron-to-chromium \"^1.3.712\"\n escalade \"^3.1.1\"\n- node-releases \"^1.1.70\"\n+ node-releases \"^1.1.71\"\n \n [email protected]:\n version \"0.2.6\"\n@@ -7217,7 +7217,7 @@ caniuse-api@^3.0.0:\n lodash.memoize \"^4.1.2\"\n lodash.uniq \"^4.5.0\"\n \n-caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001181:\n+caniuse-lite@^1.0.0, caniuse-lite@^1.0.30000981, caniuse-lite@^1.0.30001032, caniuse-lite@^1.0.30001061, caniuse-lite@^1.0.30001109, caniuse-lite@^1.0.30001125, caniuse-lite@^1.0.30001208:\n version \"1.0.30001208\"\n resolved \"https://registry.yarnpkg.com/caniuse-lite/-/caniuse-lite-1.0.30001208.tgz#a999014a35cebd4f98c405930a057a0d75352eb9\"\n integrity sha512-OE5UE4+nBOro8Dyvv0lfx+SRtfVIOM9uhKqFmJeUbGriqhhStgp1A0OyBpgy3OUF8AhYCT+PVwPC1gMl2ZcQMA==\n@@ -9549,10 +9549,10 @@ ejs@^2.6.1:\n resolved \"https://registry.yarnpkg.com/ejs/-/ejs-2.7.4.tgz#48661287573dcc53e366c7a1ae52c3a120eec9ba\"\n integrity sha512-7vmuyh5+kuUyJKePhQfRQBhXV5Ce+RnaeeQArKu1EAMpL3WbgMt5WG6uQZpEVvYSSsxMXRKOewtDk9RaTKXRlA==\n \n-electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.649:\n- version \"1.3.711\"\n- resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.711.tgz#92c3caf7ffed5e18bf63f66b4b57b4db2409c450\"\n- integrity sha512-XbklBVCDiUeho0PZQCjC25Ha6uBwqqJeyDhPLwLwfWRAo4x+FZFsmu1pPPkXT+B4MQMQoQULfyaMltDopfeiHQ==\n+electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.712:\n+ version \"1.3.712\"\n+ resolved \"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.712.tgz#ae467ffe5f95961c6d41ceefe858fc36eb53b38f\"\n+ integrity sha512-3kRVibBeCM4vsgoHHGKHmPocLqtFAGTrebXxxtgKs87hNUzXrX2NuS3jnBys7IozCnw7viQlozxKkmty2KNfrw==\n \n elegant-spinner@^1.0.1:\n version \"1.0.1\"\n@@ -9945,9 +9945,9 @@ eslint-plugin-import@^2.16.0, eslint-plugin-import@^2.18.2, eslint-plugin-import\n tsconfig-paths \"^3.9.0\"\n \n eslint-plugin-jest@^24.1.0:\n- version \"24.3.4\"\n- resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.4.tgz#6d90c3554de0302e879603dd6405474c98849f19\"\n- integrity sha512-3n5oY1+fictanuFkTWPwSlehugBTAgwLnYLFsCllzE3Pl1BwywHl5fL0HFxmMjoQY8xhUDk8uAWc3S4JOHGh3A==\n+ version \"24.3.5\"\n+ resolved \"https://registry.yarnpkg.com/eslint-plugin-jest/-/eslint-plugin-jest-24.3.5.tgz#71f0b580f87915695c286c3f0eb88cf23664d044\"\n+ integrity sha512-XG4rtxYDuJykuqhsOqokYIR84/C8pRihRtEpVskYLbIIKGwPNW2ySxdctuVzETZE+MbF/e7wmsnbNVpzM0rDug==\n dependencies:\n \"@typescript-eslint/experimental-utils\" \"^4.0.1\"\n \n@@ -12140,12 +12140,11 @@ [email protected]:\n micromatch \"^3.1.10\"\n \n http-proxy-middleware@^1.0.0:\n- version \"1.1.0\"\n- resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.0.tgz#b896b2cc6836019af4a4f2d5f7b21b99c77ea13f\"\n- integrity sha512-OnjU5vyVgcZVe2AjLJyMrk8YLNOC2lspCHirB5ldM+B/dwEfZ5bgVTrFyzE9R7xRWAP/i/FXtvIqKjTNEZBhBg==\n+ version \"1.1.1\"\n+ resolved \"https://registry.yarnpkg.com/http-proxy-middleware/-/http-proxy-middleware-1.1.1.tgz#48900a68cd9d388c735d1dd97302c919b7e94a13\"\n+ integrity sha512-FIDg9zPvOwMhQ3XKB2+vdxK6WWbVAH7s5QpqQCif7a1TNL76GNAATWA1sy6q2gSfss8UJ/Nwza3N6QnFkKclpA==\n dependencies:\n \"@types/http-proxy\" \"^1.17.5\"\n- camelcase \"^6.2.0\"\n http-proxy \"^1.18.1\"\n is-glob \"^4.0.1\"\n is-plain-obj \"^3.0.0\"\n@@ -14341,9 +14340,9 @@ jsdom@^15.2.1:\n xml-name-validator \"^3.0.0\"\n \n jsdom@^16.4.0:\n- version \"16.5.2\"\n- resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.2.tgz#583fac89a0aea31dbf6237e7e4bedccd9beab472\"\n- integrity sha512-JxNtPt9C1ut85boCbJmffaQ06NBnzkQY/MWO3YxPW8IWS38A26z+B1oBvA9LwKrytewdfymnhi4UNH3/RAgZrg==\n+ version \"16.5.3\"\n+ resolved \"https://registry.yarnpkg.com/jsdom/-/jsdom-16.5.3.tgz#13a755b3950eb938b4482c407238ddf16f0d2136\"\n+ integrity sha512-Qj1H+PEvUsOtdPJ056ewXM4UJPCi4hhLA8wpiz9F2YvsRBhuFsXxtrIFAgGBDynQA9isAMGE91PfUYbdMPXuTA==\n dependencies:\n abab \"^2.0.5\"\n acorn \"^8.1.0\"\n@@ -15590,12 +15589,12 @@ micromatch@^3.1.10, micromatch@^3.1.4:\n to-regex \"^3.0.2\"\n \n micromatch@^4.0.2:\n- version \"4.0.3\"\n- resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.3.tgz#fdad8352bf0cbeb89b391b5d244bc22ff3dd4ec8\"\n- integrity sha512-ueuSaP4i67F/FAUac9zzZ0Dz/5KeKDkITYIS/k4fps+9qeh1SkeH6gbljcqz97mNBOsaWZ+iv2UobMKK/yD+aw==\n+ version \"4.0.4\"\n+ resolved \"https://registry.yarnpkg.com/micromatch/-/micromatch-4.0.4.tgz#896d519dfe9db25fce94ceb7a500919bf881ebf9\"\n+ integrity sha512-pRmzw/XUcwXGpD9aI9q/0XOwLNygjETJ8y0ao0wdqprrzDa4YnxLcz7fQRZr8voh8V10kGhABbNcHVk5wHgWwg==\n dependencies:\n braces \"^3.0.1\"\n- picomatch \"^2.2.1\"\n+ picomatch \"^2.2.3\"\n \n miller-rabin@^4.0.0:\n version \"4.0.1\"\n@@ -16356,7 +16355,7 @@ node-pre-gyp@^0.11.0:\n semver \"^5.3.0\"\n tar \"^4\"\n \n-node-releases@^1.1.61, node-releases@^1.1.70:\n+node-releases@^1.1.61, node-releases@^1.1.71:\n version \"1.1.71\"\n resolved \"https://registry.yarnpkg.com/node-releases/-/node-releases-1.1.71.tgz#cb1334b179896b1c89ecfdd4b725fb7bbdfc7dbb\"\n integrity sha512-zR6HoT6LrLCRBwukmrVbHv0EpEQjksO6GmFcZQQuCAy139BEsoVKPYnf3jongYW83fAa1torLGYwxxky/p28sg==\n@@ -17571,10 +17570,10 @@ [email protected]:\n dependencies:\n split2 \"^3.1.1\"\n \n-picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2:\n- version \"2.2.2\"\n- resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.2.tgz#21f333e9b6b8eaff02468f5146ea406d345f4dad\"\n- integrity sha512-q0M/9eZHzmr0AulXyPwNfZjtwZ/RBZlbN3K3CErVrk50T2ASYI7Bye0EvekFY3IP1Nt2DHu0re+V2ZHIpMkuWg==\n+picomatch@^2.0.4, picomatch@^2.2.1, picomatch@^2.2.2, picomatch@^2.2.3:\n+ version \"2.2.3\"\n+ resolved \"https://registry.yarnpkg.com/picomatch/-/picomatch-2.2.3.tgz#465547f359ccc206d3c48e46a1bcb89bf7ee619d\"\n+ integrity sha512-KpELjfwcCDUb9PeigTs2mBJzXUPzAuP2oPcA989He8Rte0+YUAjw1JVedDhuTKPkHjSYzMN3npC9luThGYEKdg==\n \n pify@^2.0.0, pify@^2.2.0, pify@^2.3.0:\n version \"2.3.0\"\n@@ -18446,9 +18445,9 @@ postcss@^7, postcss@^7.0.0, postcss@^7.0.1, postcss@^7.0.14, postcss@^7.0.17, po\n supports-color \"^6.1.0\"\n \n postcss@^8.1.0, postcss@^8.2.8:\n- version \"8.2.9\"\n- resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.9.tgz#fd95ff37b5cee55c409b3fdd237296ab4096fba3\"\n- integrity sha512-b+TmuIL4jGtCHtoLi+G/PisuIl9avxs8IZMSmlABRwNz5RLUUACrC+ws81dcomz1nRezm5YPdXiMEzBEKgYn+Q==\n+ version \"8.2.10\"\n+ resolved \"https://registry.yarnpkg.com/postcss/-/postcss-8.2.10.tgz#ca7a042aa8aff494b334d0ff3e9e77079f6f702b\"\n+ integrity sha512-b/h7CPV7QEdrqIxtAf2j31U5ef05uBDuvoXv6L51Q4rcS1jdlXAVKJv+atCFdUXYl9dyTHGyoMzIepwowRJjFw==\n dependencies:\n colorette \"^1.2.2\"\n nanoid \"^3.1.22\"\n@@ -19318,9 +19317,9 @@ rc-tree@^4.0.0, rc-tree@~4.1.0:\n rc-virtual-list \"^3.0.1\"\n \n rc-trigger@^5.0.0, rc-trigger@^5.0.4, rc-trigger@^5.1.2, rc-trigger@^5.2.1:\n- version \"5.2.3\"\n- resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.3.tgz#8c55046ab432d7b52d51c69afb57ebb5bbe37e17\"\n- integrity sha512-6Fokao07HUbqKIDkDRFEM0AGZvsvK0Fbp8A/KFgl1ngaqfO1nY037cISCG1Jm5fxImVsXp9awdkP7Vu5cxjjog==\n+ version \"5.2.4\"\n+ resolved \"https://registry.yarnpkg.com/rc-trigger/-/rc-trigger-5.2.4.tgz#f1cca4a6c1f378a5d6fadec010292250772069d3\"\n+ integrity sha512-nLZa4XYo3hOAVauQr7HsGrBtE8/pyoIWhHZnpr7x/H/dd6pPeRzH0//+1TzaBAXylbFgsY6hogKAMeJwaKeDFw==\n dependencies:\n \"@babel/runtime\" \"^7.11.2\"\n classnames \"^2.2.6\"\n@@ -20516,9 +20515,9 @@ rollup@^1.31.1:\n acorn \"^7.1.0\"\n \n rollup@^2.40.0, rollup@^2.8.0:\n- version \"2.45.0\"\n- resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.0.tgz#bfcce2347c96f15f5c78ac860bc38e3349ba27c9\"\n- integrity sha512-JJznbtGIsHZfKH0Sa9RpCAy5JarH8SWvBzRAGuRkgzAafb8e8D7VSMJ0O1Bsix1nn91koN/Ecvl2+ZWhljcuTw==\n+ version \"2.45.1\"\n+ resolved \"https://registry.yarnpkg.com/rollup/-/rollup-2.45.1.tgz#eae2b94dc2088b4e0a3b7197a5a1ee0bdd589d5c\"\n+ integrity sha512-vPD+JoDj3CY8k6m1bLcAFttXMe78P4CMxoau0iLVS60+S9kLsv2379xaGy4NgYWu+h2WTlucpoLPAoUoixFBag==\n optionalDependencies:\n fsevents \"~2.3.1\"\n \n@@ -22971,9 +22970,9 @@ typescript@~4.1.5:\n integrity sha512-6OSu9PTIzmn9TCDiovULTnET6BgXtDYL4Gg4szY+cGsc3JP1dQL8qvE8kShTRx1NIw4Q9IBHlwODjkjWEtMUyA==\n \n ua-parser-js@^0.7.18:\n- version \"0.7.27\"\n- resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.27.tgz#b54f8ce9eb6c7abf3584edeaf9a3d8b3bd92edba\"\n- integrity sha512-eXMaRYK2skomGocoX0x9sBXzx5A1ZVQgXfrW4mTc8dT0zS7olEcyfudAzRC5tIIRgLxQ69B6jut3DI+n5hslPA==\n+ version \"0.7.28\"\n+ resolved \"https://registry.yarnpkg.com/ua-parser-js/-/ua-parser-js-0.7.28.tgz#8ba04e653f35ce210239c64661685bf9121dec31\"\n+ integrity sha512-6Gurc1n//gjp9eQNXjD9O3M/sMwVtN5S8Lv9bvOYBfKfDNiIIhqiyi01vMBO45u4zkDE420w/e0se7Vs+sIg+g==\n \n [email protected]:\n version \"3.4.10\"\n@@ -22984,9 +22983,9 @@ [email protected]:\n source-map \"~0.6.1\"\n \n uglify-js@^3.1.4, uglify-js@^3.4.9:\n- version \"3.13.3\"\n- resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.3.tgz#ce72a1ad154348ea2af61f50933c76cc8802276e\"\n- integrity sha512-otIc7O9LyxpUcQoXzj2hL4LPWKklO6LJWoJUzNa8A17Xgi4fOeDC8FBDOLHnC/Slo1CQgsZMcM6as0M76BZaig==\n+ version \"3.13.4\"\n+ resolved \"https://registry.yarnpkg.com/uglify-js/-/uglify-js-3.13.4.tgz#592588bb9f47ae03b24916e2471218d914955574\"\n+ integrity sha512-kv7fCkIXyQIilD5/yQy8O+uagsYIOt5cZvs890W40/e/rvjMSzJw81o9Bg0tkURxzZBROtDQhW2LFjOGoK3RZw==\n \n [email protected]:\n version \"0.0.6\"\n"]
2
["dc50bd35462a49058c91a939fc8830ae7a9eb692", "7e9bd7c86df1032d53e752654fe4a446951480bb"]
["fix", "build"]
process CommandDistribution ACKNOWLEDGED event Adds an EventApplier for the CommandDistribution ACKNOWLEDGED event. This applier will be responsible to remove a pending distribution from the state. This will be used to mark the distribution to a specific partition as completed.,add react ecosystem
["diff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java\nnew file mode 100644\nindex 0000000..4abf2e3\n--- /dev/null\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java\n@@ -0,0 +1,28 @@\n+/*\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\n+ * one or more contributor license agreements. See the NOTICE file distributed\n+ * with this work for additional information regarding copyright ownership.\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\n+ * except in compliance with the Zeebe Community License 1.1.\n+ */\n+package io.camunda.zeebe.engine.state.appliers;\n+\n+import io.camunda.zeebe.engine.state.TypedEventApplier;\n+import io.camunda.zeebe.engine.state.mutable.MutableDistributionState;\n+import io.camunda.zeebe.protocol.impl.record.value.distribution.CommandDistributionRecord;\n+import io.camunda.zeebe.protocol.record.intent.CommandDistributionIntent;\n+\n+public final class CommandDistributionAcknowledgedApplier\n+ implements TypedEventApplier<CommandDistributionIntent, CommandDistributionRecord> {\n+\n+ private final MutableDistributionState distributionState;\n+\n+ public CommandDistributionAcknowledgedApplier(final MutableDistributionState distributionState) {\n+ this.distributionState = distributionState;\n+ }\n+\n+ @Override\n+ public void applyState(final long key, final CommandDistributionRecord value) {\n+ distributionState.removePendingDistribution(key, value.getPartitionId());\n+ }\n+}\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\nindex a72309b..4793315 100644\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\n@@ -284,6 +284,9 @@ public final class EventAppliers implements EventApplier {\n CommandDistributionIntent.DISTRIBUTING,\n new CommandDistributionDistributingApplier(distributionState));\n register(\n+ CommandDistributionIntent.ACKNOWLEDGED,\n+ new CommandDistributionAcknowledgedApplier(distributionState));\n+ register(\n CommandDistributionIntent.FINISHED,\n new CommandDistributionFinishedApplier(distributionState));\n }\n", "diff --git a/package.json b/package.json\nindex 1ba8c4f..d1de9a0 100644\n--- a/package.json\n+++ b/package.json\n@@ -36,14 +36,19 @@\n \"@types/node\": \"^9.3.0\",\n \"@types/react\": \"^16.0.34\",\n \"@types/react-dom\": \"^16.0.3\",\n+ \"@types/react-motion\": \"^0.0.25\",\n \"bootstrap-sass\": \"^3.3.7\",\n \"highcharts\": \"^6.0.4\",\n \"html2canvas\": \"^1.0.0-alpha.9\",\n+ \"immer\": \"^1.2.1\",\n \"lodash\": \"^4.17.4\",\n \"moment\": \"^2.20.1\",\n \"normalize.css\": \"^8.0.0\",\n- \"react\": \"^16.2.0\",\n- \"react-dom\": \"^16.2.0\",\n+ \"react\": \"^16.3.1\",\n+ \"react-dom\": \"^16.3.1\",\n+ \"react-motion\": \"^0.5.2\",\n+ \"react-redux\": \"^5.0.7\",\n+ \"redux\": \"^3.7.2\",\n \"rxjs\": \"^5.5.6\",\n \"vue\": \"^2.5.13\",\n \"vue-plugin-webextension-i18n\": \"^0.1.0\",\ndiff --git a/yarn.lock b/yarn.lock\nindex c8898d8..5d0fc9f 100644\n--- a/yarn.lock\n+++ b/yarn.lock\n@@ -187,6 +187,12 @@\n \"@types/node\" \"*\"\n \"@types/react\" \"*\"\n \n+\"@types/react-motion@^0.0.25\":\n+ version \"0.0.25\"\n+ resolved \"https://registry.npmjs.org/@types/react-motion/-/react-motion-0.0.25.tgz#2445745ee8e8e6149faa47a36ff6b0d4c21dbf94\"\n+ dependencies:\n+ \"@types/react\" \"*\"\n+\n \"@types/react@*\", \"@types/react@^16.0.34\":\n version \"16.0.40\"\n resolved \"https://registry.npmjs.org/@types/react/-/react-16.0.40.tgz#caabc2296886f40b67f6fc80f0f3464476461df9\"\n@@ -3837,6 +3843,10 @@ [email protected]:\n version \"4.2.1\"\n resolved \"https://registry.npmjs.org/hoek/-/hoek-4.2.1.tgz#9634502aa12c445dd5a7c5734b572bb8738aacbb\"\n \n+hoist-non-react-statics@^2.5.0:\n+ version \"2.5.0\"\n+ resolved \"https://registry.npmjs.org/hoist-non-react-statics/-/hoist-non-react-statics-2.5.0.tgz#d2ca2dfc19c5a91c5a6615ce8e564ef0347e2a40\"\n+\n home-or-tmp@^2.0.0:\n version \"2.0.0\"\n resolved \"https://registry.npmjs.org/home-or-tmp/-/home-or-tmp-2.0.0.tgz#e36c3f2d2cae7d746a857e38d18d5f32a7882db8\"\n@@ -4004,6 +4014,10 @@ ignore@^3.3.5:\n version \"3.3.7\"\n resolved \"https://registry.npmjs.org/ignore/-/ignore-3.3.7.tgz#612289bfb3c220e186a58118618d5be8c1bab021\"\n \n+immer@^1.2.1:\n+ version \"1.2.1\"\n+ resolved \"https://registry.npmjs.org/immer/-/immer-1.2.1.tgz#96e2ae29cdfc428f28120b832701931b92fa597c\"\n+\n import-local@^1.0.0:\n version \"1.0.0\"\n resolved \"https://registry.npmjs.org/import-local/-/import-local-1.0.0.tgz#5e4ffdc03f4fe6c009c6729beb29631c2f8227bc\"\n@@ -4104,7 +4118,7 @@ interpret@^1.0.0:\n version \"1.1.0\"\n resolved \"https://registry.npmjs.org/interpret/-/interpret-1.1.0.tgz#7ed1b1410c6a0e0f78cf95d3b8440c63f78b8614\"\n \n-invariant@^2.2.2:\n+invariant@^2.0.0, invariant@^2.2.2:\n version \"2.2.4\"\n resolved \"https://registry.npmjs.org/invariant/-/invariant-2.2.4.tgz#610f3c92c9359ce1db616e538008d23ff35158e6\"\n dependencies:\n@@ -5040,6 +5054,10 @@ locate-path@^2.0.0:\n p-locate \"^2.0.0\"\n path-exists \"^3.0.0\"\n \n+lodash-es@^4.17.5, lodash-es@^4.2.1:\n+ version \"4.17.8\"\n+ resolved \"https://registry.npmjs.org/lodash-es/-/lodash-es-4.17.8.tgz#6fa8c8c5d337481df0bdf1c0d899d42473121e45\"\n+\n lodash._reinterpolate@~3.0.0:\n version \"3.0.0\"\n resolved \"https://registry.npmjs.org/lodash._reinterpolate/-/lodash._reinterpolate-3.0.0.tgz#0ccf2d89166af03b3663c796538b75ac6e114d9d\"\n@@ -5149,7 +5167,7 @@ [email protected]:\n version \"4.17.2\"\n resolved \"https://registry.npmjs.org/lodash/-/lodash-4.17.2.tgz#34a3055babe04ce42467b607d700072c7ff6bf42\"\n \[email protected], lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\[email protected], lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.17.5, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\n version \"4.17.5\"\n resolved \"https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz#99a92d65c0272debe8c96b6057bc8fbfa3bed511\"\n \n@@ -6467,7 +6485,7 @@ promise@^7.1.1:\n dependencies:\n asap \"~2.0.3\"\n \n-prop-types@^15.6.0:\n+prop-types@^15.5.8, prop-types@^15.6.0:\n version \"15.6.1\"\n resolved \"https://registry.npmjs.org/prop-types/-/prop-types-15.6.1.tgz#36644453564255ddda391191fb3a125cbdf654ca\"\n dependencies:\n@@ -6574,7 +6592,7 @@ quick-lru@^1.0.0:\n version \"1.1.0\"\n resolved \"https://registry.npmjs.org/quick-lru/-/quick-lru-1.1.0.tgz#4360b17c61136ad38078397ff11416e186dcfbb8\"\n \[email protected]:\[email protected], raf@^3.1.0:\n version \"3.4.0\"\n resolved \"https://registry.npmjs.org/raf/-/raf-3.4.0.tgz#a28876881b4bc2ca9117d4138163ddb80f781575\"\n dependencies:\n@@ -6645,9 +6663,9 @@ react-dev-utils@^5.0.0:\n strip-ansi \"3.0.1\"\n text-table \"0.2.0\"\n \n-react-dom@^16.2.0:\n- version \"16.2.0\"\n- resolved \"https://registry.npmjs.org/react-dom/-/react-dom-16.2.0.tgz#69003178601c0ca19b709b33a83369fe6124c044\"\n+react-dom@^16.3.1:\n+ version \"16.3.1\"\n+ resolved \"https://registry.npmjs.org/react-dom/-/react-dom-16.3.1.tgz#6a3c90a4fb62f915bdbcf6204422d93a7d4ca573\"\n dependencies:\n fbjs \"^0.8.16\"\n loose-envify \"^1.1.0\"\n@@ -6658,9 +6676,28 @@ react-error-overlay@^4.0.0:\n version \"4.0.0\"\n resolved \"https://registry.npmjs.org/react-error-overlay/-/react-error-overlay-4.0.0.tgz#d198408a85b4070937a98667f500c832f86bd5d4\"\n \n-react@^16.2.0:\n- version \"16.2.0\"\n- resolved \"https://registry.npmjs.org/react/-/react-16.2.0.tgz#a31bd2dab89bff65d42134fa187f24d054c273ba\"\n+react-motion@^0.5.2:\n+ version \"0.5.2\"\n+ resolved \"https://registry.npmjs.org/react-motion/-/react-motion-0.5.2.tgz#0dd3a69e411316567927917c6626551ba0607316\"\n+ dependencies:\n+ performance-now \"^0.2.0\"\n+ prop-types \"^15.5.8\"\n+ raf \"^3.1.0\"\n+\n+react-redux@^5.0.7:\n+ version \"5.0.7\"\n+ resolved \"https://registry.npmjs.org/react-redux/-/react-redux-5.0.7.tgz#0dc1076d9afb4670f993ffaef44b8f8c1155a4c8\"\n+ dependencies:\n+ hoist-non-react-statics \"^2.5.0\"\n+ invariant \"^2.0.0\"\n+ lodash \"^4.17.5\"\n+ lodash-es \"^4.17.5\"\n+ loose-envify \"^1.1.0\"\n+ prop-types \"^15.6.0\"\n+\n+react@^16.3.1:\n+ version \"16.3.1\"\n+ resolved \"https://registry.npmjs.org/react/-/react-16.3.1.tgz#4a2da433d471251c69b6033ada30e2ed1202cfd8\"\n dependencies:\n fbjs \"^0.8.16\"\n loose-envify \"^1.1.0\"\n@@ -6788,6 +6825,15 @@ reduce-function-call@^1.0.1:\n dependencies:\n balanced-match \"^0.4.2\"\n \n+redux@^3.7.2:\n+ version \"3.7.2\"\n+ resolved \"https://registry.npmjs.org/redux/-/redux-3.7.2.tgz#06b73123215901d25d065be342eb026bc1c8537b\"\n+ dependencies:\n+ lodash \"^4.2.1\"\n+ lodash-es \"^4.2.1\"\n+ loose-envify \"^1.1.0\"\n+ symbol-observable \"^1.0.3\"\n+\n regenerate@^1.2.1:\n version \"1.3.3\"\n resolved \"https://registry.npmjs.org/regenerate/-/regenerate-1.3.3.tgz#0c336d3980553d755c39b586ae3b20aa49c82b7f\"\n@@ -7811,6 +7857,10 @@ [email protected]:\n version \"1.0.1\"\n resolved \"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.0.1.tgz#8340fc4702c3122df5d22288f88283f513d3fdd4\"\n \n+symbol-observable@^1.0.3:\n+ version \"1.2.0\"\n+ resolved \"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.2.0.tgz#c22688aed4eab3cdc2dfeacbb561660560a00804\"\n+\n symbol-tree@^3.2.2:\n version \"3.2.2\"\n resolved \"https://registry.npmjs.org/symbol-tree/-/symbol-tree-3.2.2.tgz#ae27db38f660a7ae2e1c3b7d1bc290819b8519e6\"\n"]
2
["6f4c06076abff94f8bb5c634beaba55483a78b72", "7e04a5e829d7416e312ac342a00a11787745753b"]
["feat", "build"]
Fix Cube Store build on Windows,deploy dmn using java client This test is an acceptance test that verifies that the java client can deploy a dmn decision model using the newDeployCommand client method. It verifies that the model was resource was parsed and deployed, resulting in a response that contains metadata of the deployed decision requirements graph and the decisions it contains.
["diff --git a/.github/workflows/publish.yml b/.github/workflows/publish.yml\nindex a986501..db33097 100644\n--- a/.github/workflows/publish.yml\n+++ b/.github/workflows/publish.yml\n@@ -409,7 +409,7 @@ jobs:\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\n - name: Instal LLVM for Windows\n if: ${{ startsWith(matrix.os, 'windows') }}\n- run: choco install -y llvm --version 9.0.1\n+ run: choco install -y --force llvm --version 9.0.1\n - name: Set Env Variables for Windows\n uses: allenevans/[email protected]\n if: ${{ startsWith(matrix.os, 'windows') }}\ndiff --git a/.github/workflows/rust.yml b/.github/workflows/rust.yml\nindex d45cdf9..8d87ed6 100644\n--- a/.github/workflows/rust.yml\n+++ b/.github/workflows/rust.yml\n@@ -158,7 +158,7 @@ jobs:\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\n - name: Instal LLVM for Windows\n if: ${{ startsWith(matrix.os, 'windows') }}\n- run: choco install -y llvm --version 9.0.1\n+ run: choco install -y --force llvm --version 9.0.1\n - name: Set Env Variables for Windows\n uses: allenevans/[email protected]\n if: ${{ startsWith(matrix.os, 'windows') }}\n", "diff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java\nindex f36465b..6b6ab48 100644\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/client/command/CreateDeploymentTest.java\n@@ -67,6 +67,49 @@ public final class CreateDeploymentTest {\n }\n \n @Test\n+ public void shouldDeployDecisionModel() {\n+ // given\n+ final String resourceName = \"dmn/drg-force-user.dmn\";\n+\n+ // when\n+ final DeploymentEvent result =\n+ CLIENT_RULE\n+ .getClient()\n+ .newDeployCommand()\n+ .addResourceFromClasspath(resourceName)\n+ .send()\n+ .join();\n+\n+ // then\n+ assertThat(result.getKey()).isPositive();\n+ assertThat(result.getDecisionRequirements()).hasSize(1);\n+ assertThat(result.getDecisions()).hasSize(2);\n+\n+ final var decisionRequirements = result.getDecisionRequirements().get(0);\n+ assertThat(decisionRequirements.getDmnDecisionRequirementsId()).isEqualTo(\"force_users\");\n+ assertThat(decisionRequirements.getDmnDecisionRequirementsName()).isEqualTo(\"Force Users\");\n+ assertThat(decisionRequirements.getVersion()).isEqualTo(1);\n+ assertThat(decisionRequirements.getDecisionRequirementsKey()).isPositive();\n+ assertThat(decisionRequirements.getResourceName()).isEqualTo(resourceName);\n+\n+ final var decision1 = result.getDecisions().get(0);\n+ assertThat(decision1.getDmnDecisionId()).isEqualTo(\"jedi_or_sith\");\n+ assertThat(decision1.getDmnDecisionName()).isEqualTo(\"Jedi or Sith\");\n+ assertThat(decision1.getVersion()).isEqualTo(1);\n+ assertThat(decision1.getDecisionKey()).isPositive();\n+ assertThat(decision1.getDmnDecisionRequirementsId()).isEqualTo(\"force_users\");\n+ assertThat(decision1.getDecisionRequirementsKey()).isPositive();\n+\n+ final var decision2 = result.getDecisions().get(1);\n+ assertThat(decision2.getDmnDecisionId()).isEqualTo(\"force_user\");\n+ assertThat(decision2.getDmnDecisionName()).isEqualTo(\"Which force user?\");\n+ assertThat(decision2.getVersion()).isEqualTo(1);\n+ assertThat(decision2.getDecisionKey()).isPositive();\n+ assertThat(decision2.getDmnDecisionRequirementsId()).isEqualTo(\"force_users\");\n+ assertThat(decision2.getDecisionRequirementsKey()).isPositive();\n+ }\n+\n+ @Test\n public void shouldRejectDeployIfProcessIsInvalid() {\n // given\n final BpmnModelInstance process =\ndiff --git a/qa/integration-tests/src/test/resources/dmn/drg-force-user.dmn b/qa/integration-tests/src/test/resources/dmn/drg-force-user.dmn\nnew file mode 100644\nindex 0000000..8d55c55\n--- /dev/null\n+++ b/qa/integration-tests/src/test/resources/dmn/drg-force-user.dmn\n@@ -0,0 +1,144 @@\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n+<definitions xmlns=\"https://www.omg.org/spec/DMN/20191111/MODEL/\" xmlns:dmndi=\"https://www.omg.org/spec/DMN/20191111/DMNDI/\" xmlns:dc=\"http://www.omg.org/spec/DMN/20180521/DC/\" xmlns:biodi=\"http://bpmn.io/schema/dmn/biodi/2.0\" xmlns:di=\"http://www.omg.org/spec/DMN/20180521/DI/\" id=\"force_users\" name=\"Force Users\" namespace=\"http://camunda.org/schema/1.0/dmn\" exporter=\"Camunda Modeler\" exporterVersion=\"4.12.0\">\n+ <decision id=\"jedi_or_sith\" name=\"Jedi or Sith\">\n+ <decisionTable id=\"DecisionTable_14n3bxx\">\n+ <input id=\"Input_1\" label=\"Lightsaber color\" biodi:width=\"192\">\n+ <inputExpression id=\"InputExpression_1\" typeRef=\"string\">\n+ <text>lightsaberColor</text>\n+ </inputExpression>\n+ </input>\n+ <output id=\"Output_1\" label=\"Jedi or Sith\" name=\"jedi_or_sith\" typeRef=\"string\" biodi:width=\"192\">\n+ <outputValues id=\"UnaryTests_0hj346a\">\n+ <text>\"Jedi\",\"Sith\"</text>\n+ </outputValues>\n+ </output>\n+ <rule id=\"DecisionRule_0zumznl\">\n+ <inputEntry id=\"UnaryTests_0leuxqi\">\n+ <text>\"blue\"</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0c9vpz8\">\n+ <text>\"Jedi\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_1utwb1e\">\n+ <inputEntry id=\"UnaryTests_1v3sd4m\">\n+ <text>\"green\"</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0tgh8k1\">\n+ <text>\"Jedi\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_1bwgcym\">\n+ <inputEntry id=\"UnaryTests_0n1ewm3\">\n+ <text>\"red\"</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_19xnlkw\">\n+ <text>\"Sith\"</text>\n+ </outputEntry>\n+ </rule>\n+ </decisionTable>\n+ </decision>\n+ <decision id=\"force_user\" name=\"Which force user?\">\n+ <informationRequirement id=\"InformationRequirement_1o8esai\">\n+ <requiredDecision href=\"#jedi_or_sith\" />\n+ </informationRequirement>\n+ <decisionTable id=\"DecisionTable_07g94t1\" hitPolicy=\"FIRST\">\n+ <input id=\"InputClause_0qnqj25\" label=\"Jedi or Sith\">\n+ <inputExpression id=\"LiteralExpression_00lcyt5\" typeRef=\"string\">\n+ <text>jedi_or_sith</text>\n+ </inputExpression>\n+ <inputValues id=\"UnaryTests_1xjidd8\">\n+ <text>\"Jedi\",\"Sith\"</text>\n+ </inputValues>\n+ </input>\n+ <input id=\"InputClause_0k64hys\" label=\"Body height\">\n+ <inputExpression id=\"LiteralExpression_0ib6fnk\" typeRef=\"number\">\n+ <text>height</text>\n+ </inputExpression>\n+ </input>\n+ <output id=\"OutputClause_0hhe1yo\" label=\"Force user\" name=\"force_user\" typeRef=\"string\" />\n+ <rule id=\"DecisionRule_13zidc5\">\n+ <inputEntry id=\"UnaryTests_056skcq\">\n+ <text>\"Jedi\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_0l4xksq\">\n+ <text>&gt; 190</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0hclhw3\">\n+ <text>\"Mace Windu\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_0uin2hk\">\n+ <description></description>\n+ <inputEntry id=\"UnaryTests_16maepk\">\n+ <text>\"Jedi\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_0rv0nwf\">\n+ <text>&gt; 180</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0t82c11\">\n+ <text>\"Obi-Wan Kenobi\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_0mpio0p\">\n+ <inputEntry id=\"UnaryTests_09eicyc\">\n+ <text>\"Jedi\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_1bekl8k\">\n+ <text>&lt; 70</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_0brx3vt\">\n+ <text>\"Yoda\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_06paffx\">\n+ <inputEntry id=\"UnaryTests_1baiid4\">\n+ <text>\"Sith\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_0fcdq0i\">\n+ <text>&gt; 200</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_02oibi4\">\n+ <text>\"Darth Vader\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_1ua4pcl\">\n+ <inputEntry id=\"UnaryTests_1s1h3nm\">\n+ <text>\"Sith\"</text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_1pnvw8p\">\n+ <text>&gt; 170</text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_1w1n2rc\">\n+ <text>\"Darth Sidius\"</text>\n+ </outputEntry>\n+ </rule>\n+ <rule id=\"DecisionRule_00ew25e\">\n+ <inputEntry id=\"UnaryTests_07uxyug\">\n+ <text></text>\n+ </inputEntry>\n+ <inputEntry id=\"UnaryTests_1he6fym\">\n+ <text></text>\n+ </inputEntry>\n+ <outputEntry id=\"LiteralExpression_07i3sc8\">\n+ <text>\"unknown\"</text>\n+ </outputEntry>\n+ </rule>\n+ </decisionTable>\n+ </decision>\n+ <dmndi:DMNDI>\n+ <dmndi:DMNDiagram>\n+ <dmndi:DMNShape dmnElementRef=\"jedi_or_sith\">\n+ <dc:Bounds height=\"80\" width=\"180\" x=\"160\" y=\"280\" />\n+ </dmndi:DMNShape>\n+ <dmndi:DMNShape id=\"DMNShape_1sb3tre\" dmnElementRef=\"force_user\">\n+ <dc:Bounds height=\"80\" width=\"180\" x=\"280\" y=\"80\" />\n+ </dmndi:DMNShape>\n+ <dmndi:DMNEdge id=\"DMNEdge_0gt1p1u\" dmnElementRef=\"InformationRequirement_1o8esai\">\n+ <di:waypoint x=\"250\" y=\"280\" />\n+ <di:waypoint x=\"370\" y=\"180\" />\n+ <di:waypoint x=\"370\" y=\"160\" />\n+ </dmndi:DMNEdge>\n+ </dmndi:DMNDiagram>\n+ </dmndi:DMNDI>\n+</definitions>\n"]
2
["e34bb04baac7574e67bc566d13dea72092e0cfa3", "73eac947689e3fc6b53bf626a6b4604056166d6e"]
["ci", "test"]
Publish crates,make sure root is being watched and setRoot called when it changes
["diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex 7b98b44..f17ad6f 100644\n--- a/CHANGELOG.md\n+++ b/CHANGELOG.md\n@@ -7,6 +7,9 @@\n \n - **(css/parser)** Fix parsing of at rules (#3328) ([506a310](https://github.com/swc-project/swc/commit/506a31078aaebf50129658f096bbd5929995205f))\n \n+\n+- **(es/compat)** Fix regression of `destructuring` (#3326) ([6d1ad36](https://github.com/swc-project/swc/commit/6d1ad368aca53ee64a63ae565cd015909f2f4458))\n+\n ### Performance\n \n \ndiff --git a/Cargo.lock b/Cargo.lock\nindex 3c6598b..4baa252 100644\n--- a/Cargo.lock\n+++ b/Cargo.lock\n@@ -2652,7 +2652,7 @@ dependencies = [\n \n [[package]]\n name = \"swc\"\n-version = \"0.116.15\"\n+version = \"0.116.16\"\n dependencies = [\n \"ahash\",\n \"anyhow\",\n@@ -3097,7 +3097,7 @@ dependencies = [\n \n [[package]]\n name = \"swc_ecma_transforms\"\n-version = \"0.113.3\"\n+version = \"0.113.4\"\n dependencies = [\n \"pretty_assertions 0.7.2\",\n \"sourcemap\",\n@@ -3157,7 +3157,7 @@ dependencies = [\n \n [[package]]\n name = \"swc_ecma_transforms_compat\"\n-version = \"0.68.2\"\n+version = \"0.68.3\"\n dependencies = [\n \"ahash\",\n \"arrayvec 0.7.2\",\n@@ -3366,7 +3366,7 @@ dependencies = [\n \n [[package]]\n name = \"swc_ecmascript\"\n-version = \"0.110.14\"\n+version = \"0.110.15\"\n dependencies = [\n \"swc_ecma_ast\",\n \"swc_ecma_codegen\",\ndiff --git a/crates/swc/Cargo.toml b/crates/swc/Cargo.toml\nindex 756cfc8..2f02d22 100644\n--- a/crates/swc/Cargo.toml\n+++ b/crates/swc/Cargo.toml\n@@ -9,7 +9,7 @@ include = [\"Cargo.toml\", \"src/**/*.rs\"]\n license = \"Apache-2.0\"\n name = \"swc\"\n repository = \"https://github.com/swc-project/swc.git\"\n-version = \"0.116.15\"\n+version = \"0.116.16\"\n \n [lib]\n name = \"swc\"\n@@ -55,7 +55,7 @@ swc_ecma_loader = {version = \"0.27.0\", path = \"../swc_ecma_loader\", features = [\n swc_ecma_minifier = {version = \"0.70.9\", path = \"../swc_ecma_minifier\"}\n swc_ecma_parser = {version = \"0.87.0\", path = \"../swc_ecma_parser\"}\n swc_ecma_preset_env = {version = \"0.86.1\", path = \"../swc_ecma_preset_env\"}\n-swc_ecma_transforms = {version = \"0.113.3\", path = \"../swc_ecma_transforms\", features = [\n+swc_ecma_transforms = {version = \"0.113.4\", path = \"../swc_ecma_transforms\", features = [\n \"compat\",\n \"module\",\n \"optimization\",\n@@ -64,11 +64,11 @@ swc_ecma_transforms = {version = \"0.113.3\", path = \"../swc_ecma_transforms\", fea\n \"typescript\",\n ]}\n swc_ecma_transforms_base = {version = \"0.57.1\", path = \"../swc_ecma_transforms_base\"}\n-swc_ecma_transforms_compat = {version = \"0.68.2\", path = \"../swc_ecma_transforms_compat\"}\n+swc_ecma_transforms_compat = {version = \"0.68.3\", path = \"../swc_ecma_transforms_compat\"}\n swc_ecma_transforms_optimization = {version = \"0.83.0\", path = \"../swc_ecma_transforms_optimization\"}\n swc_ecma_utils = {version = \"0.64.0\", path = \"../swc_ecma_utils\"}\n swc_ecma_visit = {version = \"0.51.1\", path = \"../swc_ecma_visit\"}\n-swc_ecmascript = {version = \"0.110.14\", path = \"../swc_ecmascript\"}\n+swc_ecmascript = {version = \"0.110.15\", path = \"../swc_ecmascript\"}\n swc_node_comments = {version = \"0.4.0\", path = \"../swc_node_comments\"}\n swc_plugin_runner = {version = \"0.30.0\", path = \"../swc_plugin_runner\", optional = true}\n swc_visit = {version = \"0.3.0\", path = \"../swc_visit\"}\ndiff --git a/crates/swc_ecma_transforms/Cargo.toml b/crates/swc_ecma_transforms/Cargo.toml\nindex 1604f4e..a0aafae 100644\n--- a/crates/swc_ecma_transforms/Cargo.toml\n+++ b/crates/swc_ecma_transforms/Cargo.toml\n@@ -6,7 +6,7 @@ edition = \"2021\"\n license = \"Apache-2.0\"\n name = \"swc_ecma_transforms\"\n repository = \"https://github.com/swc-project/swc.git\"\n-version = \"0.113.3\"\n+version = \"0.113.4\"\n \n [package.metadata.docs.rs]\n all-features = true\n@@ -28,7 +28,7 @@ swc_common = {version = \"0.17.0\", path = \"../swc_common\"}\n swc_ecma_ast = {version = \"0.65.0\", path = \"../swc_ecma_ast\"}\n swc_ecma_parser = {version = \"0.87.0\", path = \"../swc_ecma_parser\"}\n swc_ecma_transforms_base = {version = \"0.57.1\", path = \"../swc_ecma_transforms_base\"}\n-swc_ecma_transforms_compat = {version = \"0.68.2\", path = \"../swc_ecma_transforms_compat\", optional = true}\n+swc_ecma_transforms_compat = {version = \"0.68.3\", path = \"../swc_ecma_transforms_compat\", optional = true}\n swc_ecma_transforms_module = {version = \"0.74.0\", path = \"../swc_ecma_transforms_module\", optional = true}\n swc_ecma_transforms_optimization = {version = \"0.83.0\", path = \"../swc_ecma_transforms_optimization\", optional = true}\n swc_ecma_transforms_proposal = {version = \"0.74.0\", path = \"../swc_ecma_transforms_proposal\", optional = true}\ndiff --git a/crates/swc_ecma_transforms_compat/Cargo.toml b/crates/swc_ecma_transforms_compat/Cargo.toml\nindex 0ea6609..58374e3 100644\n--- a/crates/swc_ecma_transforms_compat/Cargo.toml\n+++ b/crates/swc_ecma_transforms_compat/Cargo.toml\n@@ -6,7 +6,7 @@ edition = \"2021\"\n license = \"Apache-2.0\"\n name = \"swc_ecma_transforms_compat\"\n repository = \"https://github.com/swc-project/swc.git\"\n-version = \"0.68.2\"\n+version = \"0.68.3\"\n # See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html\n \n [features]\ndiff --git a/crates/swc_ecmascript/Cargo.toml b/crates/swc_ecmascript/Cargo.toml\nindex 63680a0..775208a 100644\n--- a/crates/swc_ecmascript/Cargo.toml\n+++ b/crates/swc_ecmascript/Cargo.toml\n@@ -6,7 +6,7 @@ edition = \"2021\"\n license = \"Apache-2.0\"\n name = \"swc_ecmascript\"\n repository = \"https://github.com/swc-project/swc.git\"\n-version = \"0.110.14\"\n+version = \"0.110.15\"\n \n [package.metadata.docs.rs]\n all-features = true\n@@ -39,7 +39,7 @@ swc_ecma_dep_graph = {version = \"0.58.0\", path = \"../swc_ecma_dep_graph\", option\n swc_ecma_minifier = {version = \"0.70.9\", path = \"../swc_ecma_minifier\", optional = true}\n swc_ecma_parser = {version = \"0.87.0\", path = \"../swc_ecma_parser\", optional = true, default-features = false}\n swc_ecma_preset_env = {version = \"0.86.1\", path = \"../swc_ecma_preset_env\", optional = true}\n-swc_ecma_transforms = {version = \"0.113.3\", path = \"../swc_ecma_transforms\", optional = true}\n+swc_ecma_transforms = {version = \"0.113.4\", path = \"../swc_ecma_transforms\", optional = true}\n swc_ecma_utils = {version = \"0.64.0\", path = \"../swc_ecma_utils\", optional = true}\n swc_ecma_visit = {version = \"0.51.1\", path = \"../swc_ecma_visit\", optional = true}\n \n", "diff --git a/packages/core/src/components/nav/nav.tsx b/packages/core/src/components/nav/nav.tsx\nindex 5aaacb6..27241ee 100644\n--- a/packages/core/src/components/nav/nav.tsx\n+++ b/packages/core/src/components/nav/nav.tsx\n@@ -1,4 +1,4 @@\n-import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\n+import { Component, Element, Event, EventEmitter, Listen, Method, Prop, Watch } from '@stencil/core';\n import {\n Animation,\n AnimationController,\n@@ -103,10 +103,19 @@ export class Nav implements PublicNav, NavContainer {\n }\n this.init = true;\n if (!this.useRouter) {\n+ console.log('componentDidLoadImpl: ', this.root);\n componentDidLoadImpl(this);\n }\n }\n \n+ @Watch('root')\n+ updateRootComponent(): any {\n+ console.log('updateRootComponent: ', this.root);\n+ if (this.init) {\n+ return this.setRoot(this.root);\n+ }\n+ }\n+\n getViews(): PublicViewController[] {\n return getViews(this);\n }\ndiff --git a/packages/core/src/components/nav/test/set-root/index.html b/packages/core/src/components/nav/test/set-root/index.html\nnew file mode 100644\nindex 0000000..823c9ed\n--- /dev/null\n+++ b/packages/core/src/components/nav/test/set-root/index.html\n@@ -0,0 +1,110 @@\n+<!DOCTYPE html>\n+<html dir=\"ltr\">\n+<head>\n+ <meta charset=\"UTF-8\">\n+ <title>Nav</title>\n+ <meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0, minimum-scale=1.0, maximum-scale=1.0, user-scalable=no\">\n+ <script src=\"/dist/ionic.js\"></script>\n+</head>\n+<body onload=\"initiaize()\">\n+ <ion-app>\n+ <ion-nav root=\"page-one\"></ion-nav>\n+ </ion-app>\n+</body>\n+\n+<script>\n+\n+ class PageOne extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page One</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page One</h1>\n+ <ion-button class=\"next\">Go to Page Two</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const button = this.querySelector('ion-button');\n+ button.addEventListener('click', async () => {\n+ this.closest('ion-nav').push('page-two');\n+ });\n+ }\n+ }\n+\n+ class PageTwo extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page Two</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page Two</h1>\n+ <ion-button class=\"next\">Go to Page Three</ion-button>\n+ <ion-button class=\"previous\">Go Back</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const previousButton = this.querySelector('ion-button.previous');\n+ previousButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').pop();\n+ });\n+\n+ const nextButton = this.querySelector('ion-button.next');\n+ nextButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').push('page-three');\n+ });\n+ }\n+ }\n+\n+ class PageThree extends HTMLElement {\n+ async connectedCallback() {\n+ this.innerHTML = `\n+ <ion-page>\n+ <ion-header>\n+ <ion-toolbar>\n+ <ion-title>Page Three</ion-title>\n+ </ion-toolbar>\n+ </ion-header>\n+ <ion-content padding>\n+ <h1>Page Three</h1>\n+ <ion-button class=\"previous\">Go Back</ion-button>\n+ </ion-content>\n+ </ion-page>`;\n+\n+ const previousButton = this.querySelector('ion-button.previous');\n+ previousButton.addEventListener('click', async () => {\n+ await this.closest('ion-nav').pop();\n+ });\n+ }\n+ }\n+\n+ customElements.define('page-one', PageOne);\n+ customElements.define('page-two', PageTwo);\n+ customElements.define('page-three', PageThree);\n+\n+ async function initiaize() {\n+ const nav = document.querySelector('ion-nav');\n+ await nav.componentOnReady();\n+ nav.root = 'page-one';\n+\n+ setInterval(() => {\n+ if (nav.root === 'page-one') {\n+ nav.root = 'page-two';\n+ } else if ( nav.root === 'page-two') {\n+ nav.root = 'page-three';\n+ } else {\n+ nav.root = 'page-one';\n+ }\n+ }, 1000);\n+ }\n+\n+</script>\n+</html>\n"]
2
["af53b9487f74ff28438928903fb1f2db93fe4fa8", "4be836f5655fb5356fde5ddd7437125f8574705d"]
["build", "refactor"]
support react@17 in peer deps resolves #1478,updated test to use rows for action items references #279
["diff --git a/packages/animated/package.json b/packages/animated/package.json\nindex 2249a2f..e35a1fd 100644\n--- a/packages/animated/package.json\n+++ b/packages/animated/package.json\n@@ -33,6 +33,6 @@\n \"react-layout-effect\": \"^1.0.1\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n }\n }\ndiff --git a/packages/core/package.json b/packages/core/package.json\nindex 584bbc2..c934253 100644\n--- a/packages/core/package.json\n+++ b/packages/core/package.json\n@@ -36,7 +36,7 @@\n \"react-layout-effect\": \"^1.0.1\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n },\n \"devDependencies\": {\n \"rafz\": \"^0.1.13\"\ndiff --git a/packages/parallax/package.json b/packages/parallax/package.json\nindex 49f8391..5a181fe 100644\n--- a/packages/parallax/package.json\n+++ b/packages/parallax/package.json\n@@ -31,6 +31,6 @@\n \"@react-spring/web\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n }\n }\ndiff --git a/packages/shared/package.json b/packages/shared/package.json\nindex 67d286c..12f7db3 100644\n--- a/packages/shared/package.json\n+++ b/packages/shared/package.json\n@@ -33,6 +33,6 @@\n \"rafz\": \"^0.1.13\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\"\n+ \"react\": \"^16.8.0 || ^17.0.0\"\n }\n }\ndiff --git a/targets/konva/package.json b/targets/konva/package.json\nindex 17675ac..271d58c 100644\n--- a/targets/konva/package.json\n+++ b/targets/konva/package.json\n@@ -34,7 +34,7 @@\n },\n \"peerDependencies\": {\n \"konva\": \">=2.6\",\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-konva\": \">=16.8\"\n },\n \"devDependencies\": {\ndiff --git a/targets/native/package.json b/targets/native/package.json\nindex e97aa97..802a66c 100644\n--- a/targets/native/package.json\n+++ b/targets/native/package.json\n@@ -33,7 +33,7 @@\n \"@react-spring/types\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-native\": \">=0.58\"\n },\n \"devDependencies\": {\ndiff --git a/targets/web/package.json b/targets/web/package.json\nindex d74c25c..f7ac000 100644\n--- a/targets/web/package.json\n+++ b/targets/web/package.json\n@@ -33,7 +33,7 @@\n \"@react-spring/types\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-dom\": \">=16.8\"\n }\n }\ndiff --git a/targets/zdog/package.json b/targets/zdog/package.json\nindex aa57890..f65945a 100644\n--- a/targets/zdog/package.json\n+++ b/targets/zdog/package.json\n@@ -33,7 +33,7 @@\n \"@react-spring/types\": \"~9.2.0-beta.0\"\n },\n \"peerDependencies\": {\n- \"react\": \">=16.8\",\n+ \"react\": \"^16.8.0 || ^17.0.0\",\n \"react-dom\": \">=16.8\",\n \"react-zdog\": \">=1.0\",\n \"zdog\": \">=1.0\"\n", "diff --git a/ionic/components/card/test/advanced/main.html b/ionic/components/card/test/advanced/main.html\nindex 7c56a7d..c19ea12 100644\n--- a/ionic/components/card/test/advanced/main.html\n+++ b/ionic/components/card/test/advanced/main.html\n@@ -19,16 +19,20 @@\n </p>\n </ion-card-content>\n \n- <ion-item>\n- <button clear item-left>\n- <icon star></icon>\n- Star\n- </button>\n- <button clear item-right class=\"activated\">\n- <icon share></icon>\n- Share.activated\n- </button>\n- </ion-item>\n+ <ion-row no-padding>\n+ <ion-col>\n+ <button clear small>\n+ <icon star></icon>\n+ Star\n+ </button>\n+ </ion-col>\n+ <ion-col text-right>\n+ <button clear small class=\"activated\">\n+ <icon share></icon>\n+ Share.activated\n+ </button>\n+ </ion-col>\n+ </ion-row>\n \n </ion-card>\n \n@@ -51,19 +55,24 @@\n <p>Hello. I am a paragraph.</p>\n </ion-card-content>\n \n- <ion-item>\n- <button clear item-left danger class=\"activated\">\n- <icon star></icon>\n- Favorite.activated\n- </button>\n- <button clear item-left danger>\n- <icon musical-notes></icon>\n- Listen\n- </button>\n- <ion-note item-right>\n- Right Note\n- </ion-note>\n- </ion-item>\n+ <ion-row center no-padding>\n+ <ion-col width-75>\n+ <button clear small danger class=\"activated\">\n+ <icon star></icon>\n+ Favorite.activated\n+ </button>\n+ <button clear small danger>\n+ <icon musical-notes></icon>\n+ Listen\n+ </button>\n+ </ion-col>\n+ <ion-col text-right>\n+ <button clear small>\n+ <icon share></icon>\n+ Share\n+ </button>\n+ </ion-col>\n+ </ion-row>\n </ion-card>\n \n <ion-card>\n@@ -76,20 +85,27 @@\n This card was breaking the border radius.\n </ion-card-content>\n \n- <ion-item>\n- <button clear item-left dark>\n- <icon star></icon>\n- Favorite\n- </button>\n- <button clear item-right dark>\n- <icon musical-notes></icon>\n- Listen\n- </button>\n- <button clear item-right dark>\n- <icon share-alt></icon>\n- Share\n- </button>\n- </ion-item>\n+ <ion-row text-center no-padding>\n+ <ion-col>\n+ <button clear small dark>\n+ <icon star></icon>\n+ Favorite\n+ </button>\n+ </ion-col>\n+\n+ <ion-col>\n+ <button clear small dark>\n+ <icon musical-notes></icon>\n+ Listen\n+ </button>\n+ </ion-col>\n+ <ion-col>\n+ <button clear small dark>\n+ <icon share-alt></icon>\n+ Share\n+ </button>\n+ </ion-col>\n+ </ion-row>\n \n </ion-card>\n \n"]
2
["27169897c0e58bc4fbca724f290ad54fa39abec7", "19feaea1885eb015759b5c7a5d785521f2b8a212"]
["build", "test"]
add flag to wait for workflow instance result - with the flag withResult the create instance command will wait for the workflow to complete - optional a list of variable names can be specified to limit the fetched variables,Handle different events.
["diff --git a/clients/zbctl/cmd/createInstance.go b/clients/zbctl/cmd/createInstance.go\nindex 016f115..85ac0be 100644\n--- a/clients/zbctl/cmd/createInstance.go\n+++ b/clients/zbctl/cmd/createInstance.go\n@@ -15,13 +15,15 @@ package cmd\n \n import (\n \t\"github.com/zeebe-io/zeebe/clients/go/commands\"\n+\t\"strings\"\n \n \t\"github.com/spf13/cobra\"\n )\n \n var (\n-\tcreateInstanceVersionFlag int32\n-\tcreateInstanceVariablesFlag string\n+\tcreateInstanceVersionFlag int32\n+\tcreateInstanceVariablesFlag string\n+\tcreateInstanceWithResultFlag []string\n )\n \n var createInstanceCmd = &cobra.Command{\n@@ -39,12 +41,29 @@ var createInstanceCmd = &cobra.Command{\n \t\t\treturn err\n \t\t}\n \n-\t\tresponse, err := zbCmd.Send()\n-\t\tif err != nil {\n-\t\t\treturn err\n-\t\t}\n+\t\tif createInstanceWithResultFlag == nil {\n+\t\t\tresponse, err := zbCmd.Send()\n+\t\t\tif err != nil {\n+\t\t\t\treturn err\n+\t\t\t}\n+\n+\t\t\treturn printJson(response)\n+\t\t} else {\n+\t\t\tvariableNames := []string{}\n+\t\t\tfor _, variableName := range createInstanceWithResultFlag {\n+\t\t\t\ttrimedVariableName := strings.TrimSpace(variableName)\n+\t\t\t\tif trimedVariableName != \"\" {\n+\t\t\t\t\tvariableNames = append(variableNames, trimedVariableName)\n+\t\t\t\t}\n+\t\t\t}\n+\t\t\tresponse, err := zbCmd.WithResult().FetchVariables(variableNames...).Send()\n+\t\t\tif err != nil {\n+\t\t\t\treturn err\n+\t\t\t}\n+\n+\t\t\treturn printJson(response)\n \n-\t\treturn printJson(response)\n+\t\t}\n \t},\n }\n \n@@ -58,4 +77,11 @@ func init() {\n \tcreateInstanceCmd.\n \t\tFlags().\n \t\tInt32Var(&createInstanceVersionFlag, \"version\", commands.LatestVersion, \"Specify version of workflow which should be executed.\")\n+\n+\tcreateInstanceCmd.\n+\t\tFlags().\n+\t\tStringSliceVar(&createInstanceWithResultFlag, \"withResult\", nil, \"Specify to await result of workflow, optional a list of variable names can be provided to limit the returned variables\")\n+\n+\t// hack to use --withResult without values\n+\tcreateInstanceCmd.Flag(\"withResult\").NoOptDefVal = \" \"\n }\n", "diff --git a/src/notebook/epics/kernel-launch.js b/src/notebook/epics/kernel-launch.js\nindex 9075d7c..9f16e67 100644\n--- a/src/notebook/epics/kernel-launch.js\n+++ b/src/notebook/epics/kernel-launch.js\n@@ -113,6 +113,12 @@ export function newKernelObservable(kernelSpec: KernelInfo, cwd: string) {\n observer.error({ type: 'ERROR', payload: error, err: true });\n observer.complete();\n });\n+ spawn.on('exit', () => {\n+ observer.complete();\n+ });\n+ spawn.on('disconnect', () => {\n+ observer.complete();\n+ });\n });\n });\n }\n"]
2
["f3107f1a8eb124b55e775d23416540f49204a19e", "a280a52c8309465276c3509848ddcddbe19732b6"]
["feat", "fix"]
Introduce timediff fn (stub),remove unused branches and ignore envrc file
["diff --git a/rust/Cargo.lock b/rust/Cargo.lock\nindex b42616f..4795eb6 100644\n--- a/rust/Cargo.lock\n+++ b/rust/Cargo.lock\n@@ -1287,7 +1287,7 @@ dependencies = [\n [[package]]\n name = \"datafusion\"\n version = \"5.1.0\"\n-source = \"git+https://github.com/cube-js/arrow-datafusion.git?rev=8df4132b83d896a0d3db5c82a4eaaa3eaa285d15#8df4132b83d896a0d3db5c82a4eaaa3eaa285d15\"\n+source = \"git+https://github.com/cube-js/arrow-datafusion.git?rev=868f3c4de13d13cda84cee33475b9782b94fa60c#868f3c4de13d13cda84cee33475b9782b94fa60c\"\n dependencies = [\n \"ahash 0.7.4\",\n \"arrow 6.0.0\",\ndiff --git a/rust/cubesql/Cargo.toml b/rust/cubesql/Cargo.toml\nindex 3cb386a..9aef494 100644\n--- a/rust/cubesql/Cargo.toml\n+++ b/rust/cubesql/Cargo.toml\n@@ -9,7 +9,7 @@ documentation = \"https://cube.dev/docs\"\n homepage = \"https://cube.dev\"\n \n [dependencies]\n-datafusion = { git = 'https://github.com/cube-js/arrow-datafusion.git', rev = \"8df4132b83d896a0d3db5c82a4eaaa3eaa285d15\", default-features = false, features = [\"unicode_expressions\"] }\n+datafusion = { git = 'https://github.com/cube-js/arrow-datafusion.git', rev = \"868f3c4de13d13cda84cee33475b9782b94fa60c\", default-features = false, features = [\"unicode_expressions\"] }\n anyhow = \"1.0\"\n thiserror = \"1.0\"\n cubeclient = { path = \"../cubeclient\" }\ndiff --git a/rust/cubesql/src/compile/engine/df/intervals.rs b/rust/cubesql/src/compile/engine/df/intervals.rs\nnew file mode 100644\nindex 0000000..9e6cb7e\n--- /dev/null\n+++ b/rust/cubesql/src/compile/engine/df/intervals.rs\n@@ -0,0 +1,51 @@\n+#[macro_export]\n+macro_rules! make_string_interval_year_month {\n+ ($array: ident, $row: ident) => {{\n+ let s = if $array.is_null($row) {\n+ \"NULL\".to_string()\n+ } else {\n+ let interval = $array.value($row) as f64;\n+ let years = (interval / 12_f64).floor();\n+ let month = interval - (years * 12_f64);\n+\n+ format!(\n+ \"{} years {} mons 0 days 0 hours 0 mins 0.00 secs\",\n+ years, month,\n+ )\n+ };\n+\n+ s\n+ }};\n+}\n+\n+#[macro_export]\n+macro_rules! make_string_interval_day_time {\n+ ($array: ident, $row: ident) => {{\n+ let s = if $array.is_null($row) {\n+ \"NULL\".to_string()\n+ } else {\n+ let value: u64 = $array.value($row) as u64;\n+\n+ let days_parts: i32 = ((value & 0xFFFFFFFF00000000) >> 32) as i32;\n+ let milliseconds_part: i32 = (value & 0xFFFFFFFF) as i32;\n+\n+ let secs = milliseconds_part / 1000;\n+ let mins = secs / 60;\n+ let hours = mins / 60;\n+\n+ let secs = secs - (mins * 60);\n+ let mins = mins - (hours * 60);\n+\n+ format!(\n+ \"0 years 0 mons {} days {} hours {} mins {}.{:02} secs\",\n+ days_parts,\n+ hours,\n+ mins,\n+ secs,\n+ (milliseconds_part % 1000),\n+ )\n+ };\n+\n+ s\n+ }};\n+}\ndiff --git a/rust/cubesql/src/compile/engine/df/mod.rs b/rust/cubesql/src/compile/engine/df/mod.rs\nindex a19a970..3097523 100644\n--- a/rust/cubesql/src/compile/engine/df/mod.rs\n+++ b/rust/cubesql/src/compile/engine/df/mod.rs\n@@ -1 +1,2 @@\n pub mod coerce;\n+pub mod intervals;\ndiff --git a/rust/cubesql/src/compile/engine/udf.rs b/rust/cubesql/src/compile/engine/udf.rs\nindex 55b8bc1..0e160b3 100644\n--- a/rust/cubesql/src/compile/engine/udf.rs\n+++ b/rust/cubesql/src/compile/engine/udf.rs\n@@ -1,14 +1,19 @@\n use std::any::type_name;\n use std::sync::Arc;\n \n+\n use datafusion::{\n arrow::{\n array::{\n ArrayRef, BooleanArray, BooleanBuilder, GenericStringArray, Int32Builder,\n- PrimitiveArray, StringBuilder, UInt32Builder,\n+ IntervalDayTimeBuilder, PrimitiveArray, StringBuilder,\n+ UInt32Builder,\n },\n compute::cast,\n- datatypes::{DataType, Int64Type},\n+ datatypes::{\n+ DataType, Int64Type, IntervalUnit, TimeUnit,\n+ TimestampNanosecondType,\n+ },\n },\n error::DataFusionError,\n logical_plan::create_udf,\n@@ -399,3 +404,63 @@ pub fn create_convert_tz_udf() -> ScalarUDF {\n &fun,\n )\n }\n+\n+pub fn create_timediff_udf() -> ScalarUDF {\n+ let fun = make_scalar_function(move |args: &[ArrayRef]| {\n+ assert!(args.len() == 2);\n+\n+ let left_dt = &args[0];\n+ let right_dt = &args[1];\n+\n+ let left_date = match left_dt.data_type() {\n+ DataType::Timestamp(TimeUnit::Nanosecond, _) => {\n+ let arr = downcast_primitive_arg!(left_dt, \"left_dt\", TimestampNanosecondType);\n+ let ts = arr.value(0);\n+\n+ // NaiveDateTime::from_timestamp(ts, 0)\n+ ts\n+ }\n+ _ => {\n+ return Err(DataFusionError::Execution(format!(\n+ \"left_dt argument must be a Timestamp, actual: {}\",\n+ left_dt.data_type()\n+ )));\n+ }\n+ };\n+\n+ let right_date = match right_dt.data_type() {\n+ DataType::Timestamp(TimeUnit::Nanosecond, _) => {\n+ let arr = downcast_primitive_arg!(right_dt, \"right_dt\", TimestampNanosecondType);\n+ arr.value(0)\n+ }\n+ _ => {\n+ return Err(DataFusionError::Execution(format!(\n+ \"right_dt argument must be a Timestamp, actual: {}\",\n+ right_dt.data_type()\n+ )));\n+ }\n+ };\n+\n+ let diff = right_date - left_date;\n+ if diff != 0 {\n+ return Err(DataFusionError::NotImplemented(format!(\n+ \"timediff is not implemented, it's stub\"\n+ )));\n+ }\n+\n+ let mut interal_arr = IntervalDayTimeBuilder::new(1);\n+ interal_arr.append_value(diff)?;\n+\n+ Ok(Arc::new(interal_arr.finish()) as ArrayRef)\n+ });\n+\n+ let return_type: ReturnTypeFunction =\n+ Arc::new(move |_| Ok(Arc::new(DataType::Interval(IntervalUnit::DayTime))));\n+\n+ ScalarUDF::new(\n+ \"timediff\",\n+ &Signature::any(2, Volatility::Immutable),\n+ &return_type,\n+ &fun,\n+ )\n+}\ndiff --git a/rust/cubesql/src/compile/mod.rs b/rust/cubesql/src/compile/mod.rs\nindex a88da57..6121aa0 100644\n--- a/rust/cubesql/src/compile/mod.rs\n+++ b/rust/cubesql/src/compile/mod.rs\n@@ -32,8 +32,8 @@ use self::engine::context::SystemVar;\n use self::engine::provider::CubeContext;\n use self::engine::udf::{\n create_connection_id_udf, create_convert_tz_udf, create_current_user_udf, create_db_udf,\n- create_if_udf, create_instr_udf, create_isnull_udf, create_least_udf, create_user_udf,\n- create_version_udf,\n+ create_if_udf, create_instr_udf, create_isnull_udf, create_least_udf, create_timediff_udf,\n+ create_user_udf, create_version_udf,\n };\n use self::parser::parse_sql_to_statement;\n \n@@ -1450,6 +1450,7 @@ impl QueryPlanner {\n ctx.register_udf(create_if_udf());\n ctx.register_udf(create_least_udf());\n ctx.register_udf(create_convert_tz_udf());\n+ ctx.register_udf(create_timediff_udf());\n \n let state = ctx.state.lock().unwrap().clone();\n let cube_ctx = CubeContext::new(&state, &self.context.cubes);\n@@ -3226,6 +3227,25 @@ mod tests {\n }\n \n #[tokio::test]\n+ async fn test_timediff() -> Result<(), CubeError> {\n+ assert_eq!(\n+ execute_df_query(\n+ \"select \\\n+ timediff('1994-11-26T13:25:00.000Z'::timestamp, '1994-11-26T13:25:00.000Z'::timestamp) as r1\n+ \".to_string()\n+ )\n+ .await?,\n+ \"+------------------------------------------------+\\n\\\n+ | r1 |\\n\\\n+ +------------------------------------------------+\\n\\\n+ | 0 years 0 mons 0 days 0 hours 0 mins 0.00 secs |\\n\\\n+ +------------------------------------------------+\"\n+ );\n+\n+ Ok(())\n+ }\n+\n+ #[tokio::test]\n async fn test_metabase() -> Result<(), CubeError> {\n assert_eq!(\n execute_df_query(\ndiff --git a/rust/cubesql/src/mysql/dataframe.rs b/rust/cubesql/src/mysql/dataframe.rs\nindex fa246aa..2443458 100644\n--- a/rust/cubesql/src/mysql/dataframe.rs\n+++ b/rust/cubesql/src/mysql/dataframe.rs\n@@ -3,9 +3,10 @@ use std::fmt::{self, Debug, Formatter};\n use chrono::{SecondsFormat, TimeZone, Utc};\n use comfy_table::{Cell, Table};\n use datafusion::arrow::array::{\n- Array, Float64Array, Int32Array, Int64Array, StringArray, TimestampMicrosecondArray,\n- UInt32Array,\n+ Array, Float64Array, Int32Array, Int64Array, IntervalDayTimeArray, IntervalYearMonthArray,\n+ StringArray, TimestampMicrosecondArray, UInt32Array,\n };\n+use datafusion::arrow::datatypes::IntervalUnit;\n use datafusion::arrow::{\n array::{BooleanArray, TimestampNanosecondArray, UInt64Array},\n datatypes::{DataType, TimeUnit},\n@@ -15,6 +16,7 @@ use log::{error, warn};\n use msql_srv::{ColumnFlags, ColumnType};\n \n use crate::{compile::builder::CompiledQueryFieldMeta, CubeError};\n+use crate::{make_string_interval_day_time, make_string_interval_year_month};\n \n #[derive(Clone, Debug)]\n pub struct Column {\n@@ -309,6 +311,7 @@ pub fn arrow_to_column_type(arrow_type: DataType) -> Result<ColumnType, CubeErro\n DataType::Binary => Ok(ColumnType::MYSQL_TYPE_BLOB),\n DataType::Utf8 | DataType::LargeUtf8 => Ok(ColumnType::MYSQL_TYPE_STRING),\n DataType::Timestamp(_, _) => Ok(ColumnType::MYSQL_TYPE_STRING),\n+ DataType::Interval(_) => Ok(ColumnType::MYSQL_TYPE_STRING),\n DataType::Float16 | DataType::Float64 => Ok(ColumnType::MYSQL_TYPE_DOUBLE),\n DataType::Boolean => Ok(ColumnType::MYSQL_TYPE_TINY),\n DataType::Int8\n@@ -402,6 +405,24 @@ pub fn batch_to_dataframe(batches: &Vec<RecordBatch>) -> Result<DataFrame, CubeE\n });\n }\n }\n+ DataType::Interval(IntervalUnit::DayTime) => {\n+ let a = array\n+ .as_any()\n+ .downcast_ref::<IntervalDayTimeArray>()\n+ .unwrap();\n+ for i in 0..num_rows {\n+ rows[i].push(TableValue::String(make_string_interval_day_time!(a, i)));\n+ }\n+ }\n+ DataType::Interval(IntervalUnit::YearMonth) => {\n+ let a = array\n+ .as_any()\n+ .downcast_ref::<IntervalYearMonthArray>()\n+ .unwrap();\n+ for i in 0..num_rows {\n+ rows[i].push(TableValue::String(make_string_interval_year_month!(a, i)));\n+ }\n+ }\n DataType::Boolean => {\n let a = array.as_any().downcast_ref::<BooleanArray>().unwrap();\n for i in 0..num_rows {\n", "diff --git a/.github/workflows/ibis-backends-cloud.yml b/.github/workflows/ibis-backends-cloud.yml\nindex 2003e8e..7c7fd26 100644\n--- a/.github/workflows/ibis-backends-cloud.yml\n+++ b/.github/workflows/ibis-backends-cloud.yml\n@@ -5,9 +5,12 @@ on:\n # Skip the backend suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n+ - \"**/*.md\"\n+ - \"**/*.qmd\"\n+ - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n- - quarto\n \n permissions:\n # this allows extractions/setup-just to list releases for `just` at a higher\ndiff --git a/.github/workflows/ibis-backends-skip-helper.yml b/.github/workflows/ibis-backends-skip-helper.yml\nindex 5d5f3f7..0471994 100644\n--- a/.github/workflows/ibis-backends-skip-helper.yml\n+++ b/.github/workflows/ibis-backends-skip-helper.yml\n@@ -9,20 +9,20 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_backends:\ndiff --git a/.github/workflows/ibis-backends.yml b/.github/workflows/ibis-backends.yml\nindex 4a1cae9..30e6c1a 100644\n--- a/.github/workflows/ibis-backends.yml\n+++ b/.github/workflows/ibis-backends.yml\n@@ -8,10 +8,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the backend suite if all changes are docs\n paths-ignore:\n@@ -19,10 +19,10 @@ on:\n - \"**/*.md\"\n - \"**/*.qmd\"\n - \"codecov.yml\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\nindex 1adda11..b528a30 100644\n--- a/.github/workflows/ibis-docs-lint.yml\n+++ b/.github/workflows/ibis-docs-lint.yml\n@@ -5,12 +5,10 @@ on:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-main-skip-helper.yml b/.github/workflows/ibis-main-skip-helper.yml\nindex a5fdc6f..0fb5dea 100644\n--- a/.github/workflows/ibis-main-skip-helper.yml\n+++ b/.github/workflows/ibis-main-skip-helper.yml\n@@ -8,19 +8,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n jobs:\n test_core:\ndiff --git a/.github/workflows/ibis-main.yml b/.github/workflows/ibis-main.yml\nindex aa31436..0b1536a 100644\n--- a/.github/workflows/ibis-main.yml\n+++ b/.github/workflows/ibis-main.yml\n@@ -7,20 +7,20 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n # Skip the test suite if all changes are in the docs directory\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n permissions:\ndiff --git a/.github/workflows/ibis-tpch-queries-skip-helper.yml b/.github/workflows/ibis-tpch-queries-skip-helper.yml\nindex 1f1c0bc..f10fb8d 100644\n--- a/.github/workflows/ibis-tpch-queries-skip-helper.yml\n+++ b/.github/workflows/ibis-tpch-queries-skip-helper.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/ibis-tpch-queries.yml b/.github/workflows/ibis-tpch-queries.yml\nindex b4f8a48..9e65a61 100644\n--- a/.github/workflows/ibis-tpch-queries.yml\n+++ b/.github/workflows/ibis-tpch-queries.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\ndiff --git a/.github/workflows/nix-skip-helper.yml b/.github/workflows/nix-skip-helper.yml\nindex 677b4d7..e0ab8f7 100644\n--- a/.github/workflows/nix-skip-helper.yml\n+++ b/.github/workflows/nix-skip-helper.yml\n@@ -9,19 +9,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n jobs:\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\nindex f2dd3f0..7ea9e26 100644\n--- a/.github/workflows/nix.yml\n+++ b/.github/workflows/nix.yml\n@@ -6,19 +6,19 @@ on:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n pull_request:\n paths-ignore:\n - \"docs/**\"\n - \"**/*.md\"\n - \"**/*.qmd\"\n+ - \".envrc\"\n branches:\n - master\n - \"*.x.x\"\n- - quarto\n merge_group:\n \n concurrency:\n"]
2
["29dfb9716298c5a579c0ffba6742e13a29325670", "d0c6476df61b9c6ab07b87e1724ea7c5318595bb"]
["feat", "ci"]
use lambda to define backend operations,tests
["diff --git a/ibis/backends/duckdb/registry.py b/ibis/backends/duckdb/registry.py\nindex 20ffd6f..3f56f2a 100644\n--- a/ibis/backends/duckdb/registry.py\n+++ b/ibis/backends/duckdb/registry.py\n@@ -107,28 +107,13 @@ def _literal(_, op):\n return sa.cast(sa.literal(value), sqla_type)\n \n \n-def _array_column(t, op):\n- (arg,) = op.args\n- sqla_type = to_sqla_type(op.output_dtype)\n- return sa.cast(sa.func.list_value(*map(t.translate, arg)), sqla_type)\n-\n-\n def _neg_idx_to_pos(array, idx):\n if_ = getattr(sa.func, \"if\")\n arg_length = sa.func.array_length(array)\n return if_(idx < 0, arg_length + sa.func.greatest(idx, -arg_length), idx)\n \n \n-def _struct_field(t, op):\n- return sa.func.struct_extract(\n- t.translate(op.arg),\n- sa.text(repr(op.field)),\n- type_=to_sqla_type(op.output_dtype),\n- )\n-\n-\n-def _regex_extract(t, op):\n- string, pattern, index = map(t.translate, op.args)\n+def _regex_extract(string, pattern, index):\n result = sa.case(\n [\n (\n@@ -149,8 +134,7 @@ def _regex_extract(t, op):\n return result\n \n \n-def _json_get_item(t, op):\n- left, path = map(t.translate, op.args)\n+def _json_get_item(left, path):\n # Workaround for https://github.com/duckdb/duckdb/issues/5063\n # In some situations duckdb silently does the wrong thing if\n # the path is parametrized.\n@@ -197,7 +181,12 @@ def _struct_column(t, op):\n \n operation_registry.update(\n {\n- ops.ArrayColumn: _array_column,\n+ ops.ArrayColumn: (\n+ lambda t, op: sa.cast(\n+ sa.func.list_value(*map(t.translate, op.cols)),\n+ to_sqla_type(op.output_dtype),\n+ )\n+ ),\n ops.ArrayConcat: fixed_arity(sa.func.array_concat, 2),\n ops.ArrayRepeat: fixed_arity(\n lambda arg, times: sa.func.flatten(\n@@ -222,7 +211,13 @@ operation_registry.update(\n # TODO: map operations, but DuckDB's maps are multimaps\n ops.Modulus: fixed_arity(operator.mod, 2),\n ops.Round: _round,\n- ops.StructField: _struct_field,\n+ ops.StructField: (\n+ lambda t, op: sa.func.struct_extract(\n+ t.translate(op.arg),\n+ sa.text(repr(op.field)),\n+ type_=to_sqla_type(op.output_dtype),\n+ )\n+ ),\n ops.TableColumn: _table_column,\n ops.TimestampDiff: fixed_arity(sa.func.age, 2),\n ops.TimestampFromUNIX: _timestamp_from_unix,\n@@ -232,7 +227,7 @@ operation_registry.update(\n lambda *_: sa.cast(sa.func.now(), sa.TIMESTAMP),\n 0,\n ),\n- ops.RegexExtract: _regex_extract,\n+ ops.RegexExtract: fixed_arity(_regex_extract, 3),\n ops.RegexReplace: fixed_arity(\n lambda *args: sa.func.regexp_replace(*args, \"g\"), 3\n ),\n@@ -255,7 +250,7 @@ operation_registry.update(\n ops.ArgMin: reduction(sa.func.min_by),\n ops.ArgMax: reduction(sa.func.max_by),\n ops.BitwiseXor: fixed_arity(sa.func.xor, 2),\n- ops.JSONGetItem: _json_get_item,\n+ ops.JSONGetItem: fixed_arity(_json_get_item, 2),\n ops.RowID: lambda *_: sa.literal_column('rowid'),\n ops.StringToTimestamp: fixed_arity(sa.func.strptime, 2),\n }\n", "diff --git a/client/src/components/Profile/__test__/EducationCard.test.tsx b/client/src/components/Profile/__test__/EducationCard.test.tsx\nindex 44b6e00..14539dd 100644\n--- a/client/src/components/Profile/__test__/EducationCard.test.tsx\n+++ b/client/src/components/Profile/__test__/EducationCard.test.tsx\n@@ -53,7 +53,7 @@ describe('EducationCard', () => {\n });\n \n describe('filterPermissions', () => {\n- it('should left only contacts in \"permissionsSettings\" object', () => {\n+ it('should left only \"isEducationVisible\" in \"permissionsSettings\" object', () => {\n const permissionsSettings = {\n isProfileVisible: { all: true },\n isAboutVisible: { all: true, mentor: true, student: true },\ndiff --git a/client/src/components/Profile/__test__/MainCard.test.tsx b/client/src/components/Profile/__test__/MainCard.test.tsx\nindex 8fb2840..552804b 100644\n--- a/client/src/components/Profile/__test__/MainCard.test.tsx\n+++ b/client/src/components/Profile/__test__/MainCard.test.tsx\n@@ -3,6 +3,8 @@ import { shallow } from 'enzyme';\n import { shallowToJson } from 'enzyme-to-json';\n import MainCard from '../MainCard';\n \n+// TODO: Known Issue: https://stackoverflow.com/questions/59942808/how-can-i-use-jest-coverage-in-next-js-styled-jsx\n+\n describe('MainCard', () => {\n describe('Should render correctly', () => {\n it('if is editing mode disabled', () => {\n@@ -21,49 +23,89 @@ describe('MainCard', () => {\n );\n expect(shallowToJson(output)).toMatchSnapshot();\n });\n+ it('if is editing mode enabled', () => {\n+ const output = shallow(\n+ <MainCard\n+ data={{\n+ name: 'Petr Pervyi',\n+ githubId: 'piter',\n+ locationName: 'SPB',\n+ locationId: '1',\n+ }}\n+ isEditingModeEnabled={true}\n+ onPermissionsSettingsChange={() => {}}\n+ onProfileSettingsChange={() => {}}\n+ />,\n+ );\n+ expect(shallowToJson(output)).toMatchSnapshot();\n+ });\n });\n \n- // const wrapper = shallow(\n- // <MainCard\n- // data={{\n- // name: 'Petr Pervyi',\n- // githubId: 'piter',\n- // locationName: 'SPB',\n- // locationId: '1',\n- // }}\n- // isEditingModeEnabled={false}\n- // onPermissionsSettingsChange={() => {}}\n- // onProfileSettingsChange={() => {}}\n- // />);\n- // const instance = wrapper.instance();\n- // describe('showVisibilitySettings', () => {\n- // it('should set \"state.isVisibilitySettingsVisible\" as \"true\"', () => {\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n- // instance.showVisibilitySettings();\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n- // });\n- // });\n- // describe('hideVisibilitySettings', () => {\n- // it('should set \"state.isVisibilitySettingsVisible\" as \"false\"', () => {\n- // instance.state.isVisibilitySettingsVisible = true;\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n- // instance.hideVisibilitySettings();\n- // expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n- // });\n- // });\n- // describe('showProfileSettings', () => {\n- // it('should set \"state.isProfileSettingsVisible\" as \"true\"', () => {\n- // expect(instance.state.isProfileSettingsVisible).toBe(false);\n- // instance.showProfileSettings();\n- // expect(instance.state.isProfileSettingsVisible).toBe(true);\n- // });\n- // });\n- // describe('hideProfileSettings', () => {\n- // it('should set \"state.isProfileSettingsVisible\" as \"false\"', () => {\n- // instance.state.isProfileSettingsVisible = true;\n- // expect(instance.state.isProfileSettingsVisible).toBe(true);\n- // instance.hideProfileSettings();\n- // expect(instance.state.isProfileSettingsVisible).toBe(false);\n- // });\n- // });\n+ const wrapper = shallow(\n+ <MainCard\n+ data={{\n+ name: 'Petr Pervyi',\n+ githubId: 'piter',\n+ locationName: 'SPB',\n+ locationId: '1',\n+ }}\n+ isEditingModeEnabled={false}\n+ onPermissionsSettingsChange={() => {}}\n+ onProfileSettingsChange={() => {}}\n+ />);\n+ const instance = wrapper.instance();\n+ describe('showVisibilitySettings', () => {\n+ it('should set \"state.isVisibilitySettingsVisible\" as \"true\"', () => {\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n+ instance.showVisibilitySettings();\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n+ });\n+ });\n+ describe('hideVisibilitySettings', () => {\n+ it('should set \"state.isVisibilitySettingsVisible\" as \"false\"', () => {\n+ instance.state.isVisibilitySettingsVisible = true;\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(true);\n+ instance.hideVisibilitySettings();\n+ expect(instance.state.isVisibilitySettingsVisible).toBe(false);\n+ });\n+ });\n+ describe('showProfileSettings', () => {\n+ it('should set \"state.isProfileSettingsVisible\" as \"true\"', () => {\n+ expect(instance.state.isProfileSettingsVisible).toBe(false);\n+ instance.showProfileSettings();\n+ expect(instance.state.isProfileSettingsVisible).toBe(true);\n+ });\n+ });\n+ describe('hideProfileSettings', () => {\n+ it('should set \"state.isProfileSettingsVisible\" as \"false\"', () => {\n+ instance.state.isProfileSettingsVisible = true;\n+ expect(instance.state.isProfileSettingsVisible).toBe(true);\n+ instance.hideProfileSettings();\n+ expect(instance.state.isProfileSettingsVisible).toBe(false);\n+ });\n+ });\n+ describe('filterPermissions', () => {\n+ it('should left only \"isProfileVisible\" in \"permissionsSettings\" object', () => {\n+ const permissionsSettings = {\n+ isProfileVisible: { all: true },\n+ isAboutVisible: { all: true, mentor: true, student: true },\n+ isEducationVisible: { all: true, mentor: true, student: true },\n+ isEnglishVisible: { all: false, student: false },\n+ isEmailVisible: { all: true, student: true },\n+ isTelegramVisible: { all: false, student: false },\n+ isSkypeVisible: { all: true, student: true },\n+ isPhoneVisible: { all: false, student: false },\n+ isContactsNotesVisible: { all: true, student: true },\n+ isLinkedInVisible: { all: false, mentor: false, student: false },\n+ isPublicFeedbackVisible: { all: true, mentor: true, student: true },\n+ isMentorStatsVisible: { all: true, mentor: true, student: true },\n+ isStudentStatsVisible: { all: true, student: true },\n+ };\n+ const instance = wrapper.instance();\n+ const result = instance.filterPermissions(permissionsSettings);\n+ expect(result).toEqual({\n+ isProfileVisible: { all: true },\n+ });\n+ });\n+ });\n });\ndiff --git a/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap b/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap\nindex 40331eb..fef20dd 100644\n--- a/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap\n+++ b/client/src/components/Profile/__test__/__snapshots__/MainCard.test.tsx.snap\n@@ -71,3 +71,158 @@ exports[`MainCard Should render correctly if is editing mode disabled 1`] = `\n </Card>\n </Fragment>\n `;\n+\n+exports[`MainCard Should render correctly if is editing mode enabled 1`] = `\n+<Fragment>\n+ <Card\n+ actions={\n+ Array [\n+ <ForwardRef(EditOutlined)\n+ onClick={[Function]}\n+ />,\n+ <ForwardRef(SettingOutlined)\n+ onClick={[Function]}\n+ />,\n+ ]\n+ }\n+ >\n+ <GithubAvatar\n+ githubId=\"piter\"\n+ size={96}\n+ style={\n+ Object {\n+ \"display\": \"block\",\n+ \"margin\": \"0 auto 10px\",\n+ }\n+ }\n+ />\n+ <Title\n+ level={1}\n+ style={\n+ Object {\n+ \"fontSize\": 24,\n+ \"margin\": 0,\n+ \"textAlign\": \"center\",\n+ }\n+ }\n+ >\n+ Petr Pervyi\n+ </Title>\n+ <Paragraph\n+ style={\n+ Object {\n+ \"marginBottom\": 20,\n+ \"textAlign\": \"center\",\n+ }\n+ }\n+ >\n+ <a\n+ href=\"https://github.com/piter\"\n+ style={\n+ Object {\n+ \"fontSize\": 16,\n+ \"marginLeft\": \"-14px\",\n+ }\n+ }\n+ target=\"_blank\"\n+ >\n+ <ForwardRef(GithubFilled) />\n+ \n+ piter\n+ </a>\n+ </Paragraph>\n+ <Paragraph\n+ style={\n+ Object {\n+ \"margin\": 0,\n+ \"textAlign\": \"center\",\n+ }\n+ }\n+ >\n+ <span\n+ style={\n+ Object {\n+ \"marginLeft\": \"-14px\",\n+ }\n+ }\n+ >\n+ <ForwardRef(EnvironmentFilled) />\n+ \n+ SPB\n+ </span>\n+ </Paragraph>\n+ <PermissionsSettingsDrawer\n+ hideSettings={[Function]}\n+ isSettingsVisible={false}\n+ onPermissionsSettingsChange={[Function]}\n+ />\n+ <ProfileSettingsDrawer\n+ content={\n+ <div>\n+ <p\n+ style={\n+ Object {\n+ \"fontSize\": 18,\n+ \"marginBottom\": 5,\n+ }\n+ }\n+ >\n+ <Text\n+ strong={true}\n+ >\n+ Name:\n+ </Text>\n+ </p>\n+ <p\n+ style={\n+ Object {\n+ \"marginBottom\": 20,\n+ }\n+ }\n+ >\n+ <Input\n+ onChange={[Function]}\n+ placeholder=\"Firstname Lastname\"\n+ type=\"text\"\n+ value=\"Petr Pervyi\"\n+ />\n+ </p>\n+ <p\n+ style={\n+ Object {\n+ \"fontSize\": 18,\n+ \"marginBottom\": 5,\n+ }\n+ }\n+ >\n+ <Text\n+ strong={true}\n+ >\n+ Location:\n+ </Text>\n+ </p>\n+ <div\n+ style={\n+ Object {\n+ \"marginBottom\": 5,\n+ }\n+ }\n+ >\n+ <LocationSelect\n+ defaultValue=\"1\"\n+ onChange={[Function]}\n+ style={\n+ Object {\n+ \"width\": \"100%\",\n+ }\n+ }\n+ />\n+ </div>\n+ </div>\n+ }\n+ hideSettings={[Function]}\n+ isSettingsVisible={false}\n+ />\n+ </Card>\n+</Fragment>\n+`;\ndiff --git a/client/src/jest.config.js b/client/src/jest.config.js\nindex df39788..654f9f3 100644\n--- a/client/src/jest.config.js\n+++ b/client/src/jest.config.js\n@@ -7,4 +7,5 @@ module.exports = {\n '^services(.*)$': '<rootDir>/services/$1',\n '^utils(.*)$': '<rootDir>/utils/$1',\n },\n+ verbose: true,\n };\n"]
2
["5d14de6722eb34c6604a124f6f11cb711f16bd44", "f87659953e9af59bc7cb314a22dd076d988ef607"]
["refactor", "test"]
implement array flatten support,fix deploy
["diff --git a/ibis/backends/snowflake/registry.py b/ibis/backends/snowflake/registry.py\nindex 2373dd7..4ce03b0 100644\n--- a/ibis/backends/snowflake/registry.py\n+++ b/ibis/backends/snowflake/registry.py\n@@ -422,6 +422,7 @@ operation_registry.update(\n ops.ArrayZip: _array_zip,\n ops.ArraySort: unary(sa.func.array_sort),\n ops.ArrayRepeat: fixed_arity(sa.func.ibis_udfs.public.array_repeat, 2),\n+ ops.ArrayFlatten: fixed_arity(sa.func.array_flatten, 1),\n ops.StringSplit: fixed_arity(sa.func.split, 2),\n # snowflake typeof only accepts VARIANT, so we cast\n ops.TypeOf: unary(lambda arg: sa.func.typeof(sa.func.to_variant(arg))),\n", "diff --git a/.github/workflows/deploy.yaml b/.github/workflows/deploy.yaml\nindex 3830f4c..3b14ee5 100644\n--- a/.github/workflows/deploy.yaml\n+++ b/.github/workflows/deploy.yaml\n@@ -67,7 +67,7 @@ jobs:\n run: aws s3 cp .next/static s3://cdn.rs.school/_next/static/ --recursive --cache-control \"public,max-age=15552000,immutable\"\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n@@ -117,7 +117,7 @@ jobs:\n run: npm run build\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n@@ -167,7 +167,7 @@ jobs:\n run: npm run build\n \n - name: Build container\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\n \n - name: Login to GitHub Container Registry\n uses: docker/login-action@v1\n"]
2
["d3c754f09502be979e5dcc79f968b15052590bd0", "7785be09053049b30cf41b420c59f051cd0129fc"]
["feat", "ci"]
don't consider cases where there are no txids,update sandbox-option.md (#18275) Co-Authored-By: Mark Lee <[email protected]>
["diff --git a/src/main.rs b/src/main.rs\nindex 25d9580..9ba4e38 100644\n--- a/src/main.rs\n+++ b/src/main.rs\n@@ -441,6 +441,9 @@ fn main() {\n let mut delta_tx_fees = vec![];\n let empty_txids = vec![];\n let txids = tx_mined_deltas.get(&delta).unwrap_or(&empty_txids);\n+ if txids.len() == 0 {\n+ continue;\n+ }\n for txid in txids.iter() {\n delta_tx_fees.push(*tx_fees.get(txid).unwrap_or(&0));\n }\n", "diff --git a/docs/api/sandbox-option.md b/docs/api/sandbox-option.md\nindex 7d24bee..e293d34 100644\n--- a/docs/api/sandbox-option.md\n+++ b/docs/api/sandbox-option.md\n@@ -113,8 +113,8 @@ window.open = customWindowOpen\n Important things to notice in the preload script:\n \n - Even though the sandboxed renderer doesn't have Node.js running, it still has\n- access to a limited node-like environment: `Buffer`, `process`, `setImmediate`\n- and `require` are available.\n+ access to a limited node-like environment: `Buffer`, `process`, `setImmediate`,\n+ `clearImmediate` and `require` are available.\n - The preload script can indirectly access all APIs from the main process through the\n `remote` and `ipcRenderer` modules.\n - The preload script must be contained in a single script, but it is possible to have\n@@ -162,16 +162,17 @@ feature. We are still not aware of the security implications of exposing some\n Electron renderer APIs to the preload script, but here are some things to\n consider before rendering untrusted content:\n \n-- A preload script can accidentally leak privileged APIs to untrusted code.\n+- A preload script can accidentally leak privileged APIs to untrusted code,\n+ unless [`contextIsolation`](../tutorial/security.md#3-enable-context-isolation-for-remote-content)\n+ is also enabled.\n - Some bug in V8 engine may allow malicious code to access the renderer preload\n APIs, effectively granting full access to the system through the `remote`\n- module.\n+ module. Therefore, it is highly recommended to\n+ [disable the `remote` module](../tutorial/security.md#15-disable-the-remote-module).\n+ If disabling is not feasible, you should selectively\n+ [filter the `remote` module](../tutorial/security.md#16-filter-the-remote-module).\n \n Since rendering untrusted content in Electron is still uncharted territory,\n the APIs exposed to the sandbox preload script should be considered more\n unstable than the rest of Electron APIs, and may have breaking changes to fix\n security issues.\n-\n-One planned enhancement that should greatly increase security is to block IPC\n-messages from sandboxed renderers by default, allowing the main process to\n-explicitly define a set of messages the renderer is allowed to send.\n"]
2
["37a1b5bbb5270befcee5d9b9621af196c787a61f", "dbb8617214aaa8b56b827deef1265d9ee38765bd"]
["fix", "docs"]
fix cypress on windows,treeview width fix Signed-off-by: Raju Udava <[email protected]>
["diff --git a/packages/cypress/src/builders/cypress/cypress.impl.spec.ts b/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\nindex 22851fb..c9296fe 100644\n--- a/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\n+++ b/packages/cypress/src/builders/cypress/cypress.impl.spec.ts\n@@ -77,7 +77,7 @@ describe('Cypress builder', () => {\n await run.result;\n await run.stop();\n expect(fork).toHaveBeenCalledWith(\n- '/root/node_modules/.bin/tsc',\n+ '/root/node_modules/typescript/bin/tsc',\n ['-p', '/root/apps/my-app-e2e/tsconfig.json'],\n { stdio: [0, 1, 2, 'ipc'] }\n );\ndiff --git a/packages/cypress/src/builders/cypress/cypress.impl.ts b/packages/cypress/src/builders/cypress/cypress.impl.ts\nindex 9d9ded3..d3917c2 100644\n--- a/packages/cypress/src/builders/cypress/cypress.impl.ts\n+++ b/packages/cypress/src/builders/cypress/cypress.impl.ts\n@@ -115,7 +115,7 @@ function compileTypescriptFiles(\n let args = ['-p', path.join(context.workspaceRoot, tsConfigPath)];\n const tscPath = path.join(\n context.workspaceRoot,\n- '/node_modules/.bin/tsc'\n+ '/node_modules/typescript/bin/tsc'\n );\n if (isWatching) {\n args.push('--watch');\n", "diff --git a/tests/playwright/pages/Dashboard/TreeView.ts b/tests/playwright/pages/Dashboard/TreeView.ts\nindex 9cc622b..75c02c0 100644\n--- a/tests/playwright/pages/Dashboard/TreeView.ts\n+++ b/tests/playwright/pages/Dashboard/TreeView.ts\n@@ -23,10 +23,24 @@ export class TreeViewPage extends BasePage {\n }\n \n async verifyVisibility({ isVisible }: { isVisible: boolean }) {\n- if (isVisible) {\n- await expect(this.get()).toBeVisible();\n+ await this.rootPage.waitForTimeout(1000);\n+\n+ const domElement = await this.get();\n+ // get width of treeview dom element\n+ const width = (await domElement.boundingBox()).width;\n+\n+ // if (isVisible) {\n+ // await expect(this.get()).toBeVisible();\n+ // } else {\n+ // await expect(this.get()).not.toBeVisible();\n+ // }\n+\n+ // border for treeview is 1px\n+ // if not-visible, width should be < 5;\n+ if (!isVisible) {\n+ expect(width).toBeLessThan(5);\n } else {\n- await expect(this.get()).not.toBeVisible();\n+ expect(width).toBeGreaterThan(5);\n }\n }\n \n"]
2
["eebee9ab0bb6d4255ad0402d8422364e96bfef61", "efeb30f26252ef4791ef2a02d83827b7f0c45462"]
["fix", "test"]
add test case with multiple partitions for message,extract lambdas
["diff --git a/broker-core/src/main/java/io/zeebe/broker/subscription/command/SubscriptionCommandSender.java b/broker-core/src/main/java/io/zeebe/broker/subscription/command/SubscriptionCommandSender.java\nindex 693d1da..e3552d4 100644\n--- a/broker-core/src/main/java/io/zeebe/broker/subscription/command/SubscriptionCommandSender.java\n+++ b/broker-core/src/main/java/io/zeebe/broker/subscription/command/SubscriptionCommandSender.java\n@@ -74,7 +74,7 @@ public class SubscriptionCommandSender {\n new CloseWorkflowInstanceSubscriptionCommand();\n \n private final ClientTransport subscriptionClient;\n- private final IntArrayList partitionIds;\n+ private final IntArrayList partitionIds = new IntArrayList();\n \n private int partitionId;\n private TopologyPartitionListenerImpl partitionListener;\n@@ -82,7 +82,6 @@ public class SubscriptionCommandSender {\n public SubscriptionCommandSender(\n final ClusterCfg clusterCfg, final ClientTransport subscriptionClient) {\n this.subscriptionClient = subscriptionClient;\n- partitionIds = new IntArrayList();\n partitionIds.addAll(clusterCfg.getPartitionIds());\n }\n \n@@ -100,7 +99,8 @@ public class SubscriptionCommandSender {\n final DirectBuffer messageName,\n final DirectBuffer correlationKey) {\n \n- final int subscriptionPartitionId = getSubscriptionPartitionId(correlationKey);\n+ final int subscriptionPartitionId =\n+ SubscriptionUtil.getSubscriptionPartitionId(correlationKey, partitionIds.size());\n \n openMessageSubscriptionCommand.setSubscriptionPartitionId(subscriptionPartitionId);\n openMessageSubscriptionCommand.setWorkflowInstanceKey(workflowInstanceKey);\n@@ -111,14 +111,6 @@ public class SubscriptionCommandSender {\n return sendSubscriptionCommand(subscriptionPartitionId, openMessageSubscriptionCommand);\n }\n \n- private int getSubscriptionPartitionId(final DirectBuffer correlationKey) {\n- if (partitionIds == null) {\n- throw new IllegalStateException(\"no partition ids available\");\n- }\n-\n- return SubscriptionUtil.getSubscriptionPartitionId(correlationKey, partitionIds.size());\n- }\n-\n public boolean openWorkflowInstanceSubscription(\n final long workflowInstanceKey,\n final long elementInstanceKey,\ndiff --git a/broker-core/src/test/java/io/zeebe/broker/workflow/message/MessageCatchElementTest.java b/broker-core/src/test/java/io/zeebe/broker/workflow/message/MessageCatchElementTest.java\nindex 4baed4f..838c9ca 100644\n--- a/broker-core/src/test/java/io/zeebe/broker/workflow/message/MessageCatchElementTest.java\n+++ b/broker-core/src/test/java/io/zeebe/broker/workflow/message/MessageCatchElementTest.java\n@@ -36,7 +36,6 @@ import io.zeebe.model.bpmn.Bpmn;\n import io.zeebe.model.bpmn.BpmnModelInstance;\n import io.zeebe.protocol.clientapi.RecordType;\n import io.zeebe.protocol.clientapi.ValueType;\n-import io.zeebe.protocol.impl.SubscriptionUtil;\n import io.zeebe.protocol.intent.DeploymentIntent;\n import io.zeebe.protocol.intent.MessageSubscriptionIntent;\n import io.zeebe.protocol.intent.WorkflowInstanceIntent;\n@@ -44,7 +43,6 @@ import io.zeebe.protocol.intent.WorkflowInstanceSubscriptionIntent;\n import io.zeebe.test.broker.protocol.clientapi.ClientApiRule;\n import io.zeebe.test.broker.protocol.clientapi.PartitionTestClient;\n import io.zeebe.test.util.record.RecordingExporter;\n-import io.zeebe.util.buffer.BufferUtil;\n import java.util.List;\n import java.util.stream.Collectors;\n import org.agrona.DirectBuffer;\n@@ -171,39 +169,6 @@ public class MessageCatchElementTest {\n }\n \n @Test\n- public void shouldOpenMessageSubscriptionsOnSamePartition() {\n- // given\n- final List<Integer> partitionIds = apiRule.getPartitionIds();\n-\n- final String correlationKey = \"order-123\";\n-\n- final PartitionTestClient workflowPartition = apiRule.partitionClient(partitionIds.get(0));\n- final PartitionTestClient subscriptionPartition =\n- apiRule.partitionClient(getPartitionId(correlationKey));\n-\n- testClient.deploy(CATCH_EVENT_WORKFLOW);\n-\n- // when\n- final long workflowInstanceKey1 =\n- workflowPartition.createWorkflowInstance(PROCESS_ID, asMsgPack(\"orderId\", correlationKey));\n-\n- final long workflowInstanceKey2 =\n- workflowPartition.createWorkflowInstance(PROCESS_ID, asMsgPack(\"orderId\", correlationKey));\n-\n- // then\n- final List<Record<MessageSubscriptionRecordValue>> subscriptions =\n- subscriptionPartition\n- .receiveMessageSubscriptions()\n- .withIntent(MessageSubscriptionIntent.OPENED)\n- .limit(2)\n- .collect(Collectors.toList());\n-\n- assertThat(subscriptions)\n- .extracting(s -> s.getValue().getWorkflowInstanceKey())\n- .contains(workflowInstanceKey1, workflowInstanceKey2);\n- }\n-\n- @Test\n public void shouldOpenWorkflowInstanceSubscription() {\n final long workflowInstanceKey =\n testClient.createWorkflowInstance(PROCESS_ID, asMsgPack(\"orderId\", \"order-123\"));\n@@ -352,10 +317,4 @@ public class MessageCatchElementTest {\n .exists())\n .isTrue();\n }\n-\n- private int getPartitionId(final String correlationKey) {\n- final List<Integer> partitionIds = apiRule.getPartitionIds();\n- return SubscriptionUtil.getSubscriptionPartitionId(\n- BufferUtil.wrapString(correlationKey), partitionIds.size());\n- }\n }\ndiff --git a/broker-core/src/test/java/io/zeebe/broker/workflow/message/MessageCorrelationMultiplePartitionsTest.java b/broker-core/src/test/java/io/zeebe/broker/workflow/message/MessageCorrelationMultiplePartitionsTest.java\nnew file mode 100644\nindex 0000000..cf8261a\n--- /dev/null\n+++ b/broker-core/src/test/java/io/zeebe/broker/workflow/message/MessageCorrelationMultiplePartitionsTest.java\n@@ -0,0 +1,134 @@\n+/*\n+ * Zeebe Broker Core\n+ * Copyright \u00a9 2017 camunda services GmbH ([email protected])\n+ *\n+ * This program is free software: you can redistribute it and/or modify\n+ * it under the terms of the GNU Affero General Public License as published by\n+ * the Free Software Foundation, either version 3 of the License, or\n+ * (at your option) any later version.\n+ *\n+ * This program is distributed in the hope that it will be useful,\n+ * but WITHOUT ANY WARRANTY; without even the implied warranty of\n+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n+ * GNU Affero General Public License for more details.\n+ *\n+ * You should have received a copy of the GNU Affero General Public License\n+ * along with this program. If not, see <http://www.gnu.org/licenses/>.\n+ */\n+package io.zeebe.broker.workflow.message;\n+\n+import static io.zeebe.broker.test.EmbeddedBrokerConfigurator.setPartitionCount;\n+import static io.zeebe.test.util.MsgPackUtil.asMsgPack;\n+import static org.assertj.core.api.Assertions.assertThat;\n+import static org.assertj.core.api.Assertions.tuple;\n+\n+import io.zeebe.broker.test.EmbeddedBrokerRule;\n+import io.zeebe.model.bpmn.Bpmn;\n+import io.zeebe.model.bpmn.BpmnModelInstance;\n+import io.zeebe.protocol.impl.SubscriptionUtil;\n+import io.zeebe.protocol.intent.MessageSubscriptionIntent;\n+import io.zeebe.protocol.intent.WorkflowInstanceIntent;\n+import io.zeebe.test.broker.protocol.clientapi.ClientApiRule;\n+import io.zeebe.test.broker.protocol.clientapi.PartitionTestClient;\n+import io.zeebe.test.util.record.RecordingExporter;\n+import io.zeebe.util.buffer.BufferUtil;\n+import java.util.List;\n+import java.util.stream.IntStream;\n+import org.junit.Before;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.rules.RuleChain;\n+\n+public class MessageCorrelationMultiplePartitionsTest {\n+\n+ private static final String CORRELATION_KEY_PARTITION_0 = \"item-2\";\n+ private static final String CORRELATION_KEY_PARTITION_1 = \"item-1\";\n+ private static final String CORRELATION_KEY_PARTITION_2 = \"item-0\";\n+\n+ private static final String PROCESS_ID = \"process\";\n+\n+ private static final BpmnModelInstance WORKFLOW =\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateCatchEvent(\"receive-message\")\n+ .message(m -> m.name(\"message\").zeebeCorrelationKey(\"$.key\"))\n+ .endEvent(\"end\")\n+ .done();\n+\n+ public EmbeddedBrokerRule brokerRule = new EmbeddedBrokerRule(setPartitionCount(3));\n+\n+ public ClientApiRule apiRule = new ClientApiRule(brokerRule::getClientAddress);\n+\n+ @Rule public RuleChain ruleChain = RuleChain.outerRule(brokerRule).around(apiRule);\n+\n+ private PartitionTestClient testClient;\n+\n+ @Before\n+ public void init() {\n+ assertThat(getPartitionId(CORRELATION_KEY_PARTITION_0)).isEqualTo(0);\n+ assertThat(getPartitionId(CORRELATION_KEY_PARTITION_1)).isEqualTo(1);\n+ assertThat(getPartitionId(CORRELATION_KEY_PARTITION_2)).isEqualTo(2);\n+\n+ testClient = apiRule.partitionClient();\n+\n+ testClient.deploy(WORKFLOW);\n+ }\n+\n+ @Test\n+ public void shouldOpenMessageSubscriptionsOnDifferentPartitions() {\n+ // when\n+ IntStream.range(0, 10)\n+ .forEach(\n+ i -> {\n+ testClient.createWorkflowInstance(\n+ PROCESS_ID, asMsgPack(\"key\", CORRELATION_KEY_PARTITION_0));\n+ testClient.createWorkflowInstance(\n+ PROCESS_ID, asMsgPack(\"key\", CORRELATION_KEY_PARTITION_1));\n+ testClient.createWorkflowInstance(\n+ PROCESS_ID, asMsgPack(\"key\", CORRELATION_KEY_PARTITION_2));\n+ });\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.messageSubscriptionRecords(MessageSubscriptionIntent.OPENED)\n+ .limit(30))\n+ .extracting(r -> tuple(r.getMetadata().getPartitionId(), r.getValue().getCorrelationKey()))\n+ .containsOnly(\n+ tuple(0, CORRELATION_KEY_PARTITION_0),\n+ tuple(1, CORRELATION_KEY_PARTITION_1),\n+ tuple(2, CORRELATION_KEY_PARTITION_2));\n+ }\n+\n+ @Test\n+ public void shouldCorrelateMessageOnDifferentPartitions() {\n+ // given\n+ apiRule\n+ .partitionClient(0)\n+ .publishMessage(\"message\", CORRELATION_KEY_PARTITION_0, asMsgPack(\"p\", \"p0\"));\n+ apiRule\n+ .partitionClient(1)\n+ .publishMessage(\"message\", CORRELATION_KEY_PARTITION_1, asMsgPack(\"p\", \"p1\"));\n+ apiRule\n+ .partitionClient(2)\n+ .publishMessage(\"message\", CORRELATION_KEY_PARTITION_2, asMsgPack(\"p\", \"p2\"));\n+\n+ // when\n+ testClient.createWorkflowInstance(PROCESS_ID, asMsgPack(\"key\", CORRELATION_KEY_PARTITION_0));\n+ testClient.createWorkflowInstance(PROCESS_ID, asMsgPack(\"key\", CORRELATION_KEY_PARTITION_1));\n+ testClient.createWorkflowInstance(PROCESS_ID, asMsgPack(\"key\", CORRELATION_KEY_PARTITION_2));\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.workflowInstanceRecords(WorkflowInstanceIntent.END_EVENT_OCCURRED)\n+ .withElementId(\"end\")\n+ .limit(3))\n+ .extracting(r -> r.getValue().getPayloadAsMap().get(\"p\"))\n+ .contains(\"p0\", \"p1\", \"p2\");\n+ }\n+\n+ private int getPartitionId(final String correlationKey) {\n+ final List<Integer> partitionIds = apiRule.getPartitionIds();\n+ return SubscriptionUtil.getSubscriptionPartitionId(\n+ BufferUtil.wrapString(correlationKey), partitionIds.size());\n+ }\n+}\ndiff --git a/protocol-test-util/src/main/java/io/zeebe/test/broker/protocol/clientapi/PartitionTestClient.java b/protocol-test-util/src/main/java/io/zeebe/test/broker/protocol/clientapi/PartitionTestClient.java\nindex dac11a2..e2b8397 100644\n--- a/protocol-test-util/src/main/java/io/zeebe/test/broker/protocol/clientapi/PartitionTestClient.java\n+++ b/protocol-test-util/src/main/java/io/zeebe/test/broker/protocol/clientapi/PartitionTestClient.java\n@@ -329,6 +329,7 @@ public class PartitionTestClient {\n final String messageName, final String correlationKey, final byte[] payload, final long ttl) {\n return apiRule\n .createCmdRequest()\n+ .partitionId(partitionId)\n .type(ValueType.MESSAGE, MessageIntent.PUBLISH)\n .command()\n .put(\"name\", messageName)\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerReprocessingTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerReprocessingTest.java\nindex 9a122d9..b7db67e 100644\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerReprocessingTest.java\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerReprocessingTest.java\n@@ -619,14 +619,9 @@ public class BrokerReprocessingTest {\n }\n \n @Test\n- public void shouldCorrelateMessageAfterRestartIfEnteredBeforeA() throws Exception {\n+ public void shouldCorrelateMessageAfterRestartIfEnteredBefore() throws Exception {\n // given\n- clientRule\n- .getWorkflowClient()\n- .newDeployCommand()\n- .addWorkflowModel(WORKFLOW_MESSAGE, \"message.bpmn\")\n- .send()\n- .join();\n+ deploy(WORKFLOW_MESSAGE, \"message.bpmn\");\n \n final long workflowInstanceKey =\n startWorkflowInstance(PROCESS_ID, singletonMap(\"orderId\", \"order-123\"))\n@@ -658,12 +653,7 @@ public class BrokerReprocessingTest {\n @Test\n public void shouldCorrelateMessageAfterRestartIfPublishedBefore() throws Exception {\n // given\n- clientRule\n- .getWorkflowClient()\n- .newDeployCommand()\n- .addWorkflowModel(WORKFLOW_MESSAGE, \"message.bpmn\")\n- .send()\n- .join();\n+ deploy(WORKFLOW_MESSAGE, \"message.bpmn\");\n \n publishMessage(\"order canceled\", \"order-123\", singletonMap(\"foo\", \"bar\"));\n reprocessingTrigger.accept(this);\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/MessageCorrelationTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/MessageCorrelationTest.java\ndeleted file mode 100644\nindex c6a05fb..0000000\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/MessageCorrelationTest.java\n+++ /dev/null\n@@ -1,176 +0,0 @@\n-/*\n- * Copyright \u00a9 2017 camunda services GmbH ([email protected])\n- *\n- * Licensed under the Apache License, Version 2.0 (the \"License\");\n- * you may not use this file except in compliance with the License.\n- * You may obtain a copy of the License at\n- *\n- * http://www.apache.org/licenses/LICENSE-2.0\n- *\n- * Unless required by applicable law or agreed to in writing, software\n- * distributed under the License is distributed on an \"AS IS\" BASIS,\n- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n- * See the License for the specific language governing permissions and\n- * limitations under the License.\n- */\n-package io.zeebe.broker.it.workflow;\n-\n-import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertElementActivated;\n-import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertElementCompleted;\n-import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertWorkflowInstanceCompleted;\n-import static org.assertj.core.api.Assertions.assertThat;\n-import static org.assertj.core.api.Assertions.entry;\n-\n-import io.zeebe.broker.it.GrpcClientRule;\n-import io.zeebe.broker.test.EmbeddedBrokerRule;\n-import io.zeebe.client.api.events.DeploymentEvent;\n-import io.zeebe.model.bpmn.Bpmn;\n-import io.zeebe.model.bpmn.BpmnModelInstance;\n-import java.util.Collections;\n-import org.junit.Before;\n-import org.junit.Rule;\n-import org.junit.Test;\n-import org.junit.rules.RuleChain;\n-import org.junit.runner.RunWith;\n-import org.junit.runners.Parameterized;\n-import org.junit.runners.Parameterized.Parameter;\n-import org.junit.runners.Parameterized.Parameters;\n-\n-@RunWith(Parameterized.class)\n-public class MessageCorrelationTest {\n-\n- public EmbeddedBrokerRule brokerRule = new EmbeddedBrokerRule();\n- public GrpcClientRule clientRule = new GrpcClientRule(brokerRule);\n-\n- @Rule public RuleChain ruleChain = RuleChain.outerRule(brokerRule).around(clientRule);\n-\n- private static final BpmnModelInstance CATCH_EVENT_WORKFLOW =\n- Bpmn.createExecutableProcess(\"wf\")\n- .startEvent()\n- .intermediateCatchEvent(\"receive-message\")\n- .message(m -> m.name(\"order canceled\").zeebeCorrelationKey(\"$.orderId\"))\n- .sequenceFlowId(\"to-end\")\n- .endEvent()\n- .done();\n-\n- private static final BpmnModelInstance RECEIVE_TASK_WORKFLOW =\n- Bpmn.createExecutableProcess(\"wf\")\n- .startEvent()\n- .receiveTask(\"receive-message\")\n- .message(m -> m.name(\"order canceled\").zeebeCorrelationKey(\"$.orderId\"))\n- .sequenceFlowId(\"to-end\")\n- .endEvent()\n- .done();\n-\n- @Parameter(0)\n- public String elementType;\n-\n- @Parameter(1)\n- public BpmnModelInstance workflow;\n-\n- @Parameters(name = \"{0}\")\n- public static final Object[][] parameters() {\n- return new Object[][] {\n- {\"intermediate message catch event\", CATCH_EVENT_WORKFLOW},\n- {\"receive task\", RECEIVE_TASK_WORKFLOW}\n- };\n- }\n-\n- @Before\n- public void init() {\n- final DeploymentEvent deploymentEvent =\n- clientRule\n- .getWorkflowClient()\n- .newDeployCommand()\n- .addWorkflowModel(workflow, \"wf.bpmn\")\n- .send()\n- .join();\n-\n- clientRule.waitUntilDeploymentIsDone(deploymentEvent.getKey());\n- }\n-\n- @Test\n- public void shouldCorrelateMessageIfEnteredBefore() {\n- // given\n- clientRule\n- .getWorkflowClient()\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- assertElementActivated(\"receive-message\");\n-\n- // when\n- clientRule\n- .getWorkflowClient()\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .send()\n- .join();\n-\n- // then\n- assertWorkflowInstanceCompleted(\"wf\");\n- }\n-\n- @Test\n- public void shouldCorrelateMessageIfPublishedBefore() {\n- // given\n- clientRule\n- .getWorkflowClient()\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .send()\n- .join();\n-\n- // when\n- clientRule\n- .getWorkflowClient()\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- // then\n- assertWorkflowInstanceCompleted(\"wf\");\n- }\n-\n- @Test\n- public void shouldCorrelateMessageAndMergePayload() {\n- // given\n- clientRule\n- .getWorkflowClient()\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- // when\n- clientRule\n- .getWorkflowClient()\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .payload(Collections.singletonMap(\"foo\", \"bar\"))\n- .send()\n- .join();\n-\n- // then\n- assertWorkflowInstanceCompleted(\"wf\");\n-\n- assertElementCompleted(\n- \"wf\",\n- \"receive-message\",\n- (catchEventOccurredEvent) ->\n- assertThat(catchEventOccurredEvent.getPayloadAsMap())\n- .containsExactly(entry(\"orderId\", \"order-123\"), entry(\"foo\", \"bar\")));\n- }\n-}\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/PublishMessageTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/PublishMessageTest.java\ndeleted file mode 100644\nindex 7845eec..0000000\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/PublishMessageTest.java\n+++ /dev/null\n@@ -1,234 +0,0 @@\n-/*\n- * Copyright \u00a9 2017 camunda services GmbH ([email protected])\n- *\n- * Licensed under the Apache License, Version 2.0 (the \"License\");\n- * you may not use this file except in compliance with the License.\n- * You may obtain a copy of the License at\n- *\n- * http://www.apache.org/licenses/LICENSE-2.0\n- *\n- * Unless required by applicable law or agreed to in writing, software\n- * distributed under the License is distributed on an \"AS IS\" BASIS,\n- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n- * See the License for the specific language governing permissions and\n- * limitations under the License.\n- */\n-package io.zeebe.broker.it.workflow;\n-\n-import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertElementActivated;\n-import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertElementCompleted;\n-import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertWorkflowInstanceCompleted;\n-import static io.zeebe.broker.test.EmbeddedBrokerConfigurator.setPartitionCount;\n-import static org.assertj.core.api.Assertions.assertThat;\n-import static org.assertj.core.api.Assertions.assertThatThrownBy;\n-import static org.assertj.core.api.Assertions.entry;\n-\n-import io.zeebe.broker.it.GrpcClientRule;\n-import io.zeebe.broker.test.EmbeddedBrokerRule;\n-import io.zeebe.client.api.ZeebeFuture;\n-import io.zeebe.client.api.clients.WorkflowClient;\n-import io.zeebe.client.api.events.DeploymentEvent;\n-import io.zeebe.client.api.events.WorkflowInstanceEvent;\n-import io.zeebe.client.cmd.ClientException;\n-import io.zeebe.model.bpmn.Bpmn;\n-import io.zeebe.model.bpmn.BpmnModelInstance;\n-import java.time.Duration;\n-import java.util.Collections;\n-import org.junit.Before;\n-import org.junit.Rule;\n-import org.junit.Test;\n-import org.junit.rules.RuleChain;\n-\n-public class PublishMessageTest {\n-\n- private static final BpmnModelInstance WORKFLOW =\n- Bpmn.createExecutableProcess(\"wf\")\n- .startEvent()\n- .intermediateCatchEvent(\"catch-event\")\n- .message(c -> c.name(\"order canceled\").zeebeCorrelationKey(\"$.orderId\"))\n- .endEvent()\n- .done();\n- public EmbeddedBrokerRule brokerRule = new EmbeddedBrokerRule(setPartitionCount(3));\n- public GrpcClientRule clientRule = new GrpcClientRule(brokerRule);\n-\n- @Rule public RuleChain ruleChain = RuleChain.outerRule(brokerRule).around(clientRule);\n-\n- private WorkflowClient workflowClient;\n-\n- @Before\n- public void init() {\n-\n- workflowClient = clientRule.getClient().workflowClient();\n-\n- final DeploymentEvent deploymentEvent =\n- workflowClient.newDeployCommand().addWorkflowModel(WORKFLOW, \"wf.bpmn\").send().join();\n-\n- clientRule.waitUntilDeploymentIsDone(deploymentEvent.getKey());\n- }\n-\n- @Test\n- public void shouldCorrelateMessageToAllSubscriptions() {\n- // given\n- final WorkflowInstanceEvent wf =\n- workflowClient\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- final WorkflowInstanceEvent wf2 =\n- workflowClient\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- // when\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .send()\n- .join();\n-\n- // then\n- assertWorkflowInstanceCompleted(\"wf\", wf.getWorkflowInstanceKey());\n- assertWorkflowInstanceCompleted(\"wf\", wf2.getWorkflowInstanceKey());\n- }\n-\n- @Test\n- public void shouldCorrelateMessageWithZeroTTL() {\n- // given\n- workflowClient\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- assertElementActivated(\"catch-event\");\n-\n- // when\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .timeToLive(Duration.ZERO)\n- .send()\n- .join();\n-\n- // then\n- assertElementCompleted(\"wf\", \"catch-event\");\n- }\n-\n- @Test\n- public void shouldNotCorrelateMessageAfterTTL() {\n- // given\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .timeToLive(Duration.ZERO)\n- .payload(Collections.singletonMap(\"msg\", \"failure\"))\n- .send()\n- .join();\n-\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .timeToLive(Duration.ofMinutes(1))\n- .payload(Collections.singletonMap(\"msg\", \"expected\"))\n- .send()\n- .join();\n-\n- // when\n- workflowClient\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- // then\n-\n- assertElementCompleted(\n- \"wf\",\n- \"catch-event\",\n- (catchEventOccurred) ->\n- assertThat(catchEventOccurred.getPayloadAsMap()).contains(entry(\"msg\", \"expected\")));\n- }\n-\n- @Test\n- public void shouldCorrelateMessageOnDifferentPartitions() {\n- // given\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .send()\n- .join();\n-\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-124\")\n- .send()\n- .join();\n-\n- // when\n- final WorkflowInstanceEvent wf =\n- workflowClient\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-123\\\"}\")\n- .send()\n- .join();\n-\n- final WorkflowInstanceEvent wf2 =\n- workflowClient\n- .newCreateInstanceCommand()\n- .bpmnProcessId(\"wf\")\n- .latestVersion()\n- .payload(\"{\\\"orderId\\\":\\\"order-124\\\"}\")\n- .send()\n- .join();\n-\n- // then\n- assertWorkflowInstanceCompleted(\"wf\", wf.getWorkflowInstanceKey());\n- assertWorkflowInstanceCompleted(\"wf\", wf2.getWorkflowInstanceKey());\n- }\n-\n- @Test\n- public void shouldRejectMessageWithSameId() {\n- // given\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .messageId(\"foo\")\n- .send()\n- .join();\n-\n- // when\n- final ZeebeFuture<Void> future =\n- workflowClient\n- .newPublishMessageCommand()\n- .messageName(\"order canceled\")\n- .correlationKey(\"order-123\")\n- .messageId(\"foo\")\n- .send();\n-\n- // then\n- assertThatThrownBy(future::join)\n- .isInstanceOf(ClientException.class)\n- .hasMessageContaining(\"message with id 'foo' is already published\");\n- }\n-}\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/message/MessageCorrelationMultiplePartitionsTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/message/MessageCorrelationMultiplePartitionsTest.java\nnew file mode 100644\nindex 0000000..0e37c95\n--- /dev/null\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/message/MessageCorrelationMultiplePartitionsTest.java\n@@ -0,0 +1,196 @@\n+/*\n+ * Copyright \u00a9 2017 camunda services GmbH ([email protected])\n+ *\n+ * Licensed under the Apache License, Version 2.0 (the \"License\");\n+ * you may not use this file except in compliance with the License.\n+ * You may obtain a copy of the License at\n+ *\n+ * http://www.apache.org/licenses/LICENSE-2.0\n+ *\n+ * Unless required by applicable law or agreed to in writing, software\n+ * distributed under the License is distributed on an \"AS IS\" BASIS,\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n+ * See the License for the specific language governing permissions and\n+ * limitations under the License.\n+ */\n+package io.zeebe.broker.it.workflow.message;\n+\n+import static io.zeebe.broker.test.EmbeddedBrokerConfigurator.setPartitionCount;\n+import static org.assertj.core.api.Assertions.assertThat;\n+import static org.assertj.core.api.Assertions.tuple;\n+\n+import io.zeebe.broker.it.GrpcClientRule;\n+import io.zeebe.broker.test.EmbeddedBrokerRule;\n+import io.zeebe.client.api.events.DeploymentEvent;\n+import io.zeebe.model.bpmn.Bpmn;\n+import io.zeebe.model.bpmn.BpmnModelInstance;\n+import io.zeebe.protocol.intent.MessageIntent;\n+import io.zeebe.protocol.intent.MessageSubscriptionIntent;\n+import io.zeebe.protocol.intent.WorkflowInstanceIntent;\n+import io.zeebe.test.util.record.RecordingExporter;\n+import java.util.Collections;\n+import java.util.stream.IntStream;\n+import org.junit.Before;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.rules.RuleChain;\n+\n+public class MessageCorrelationMultiplePartitionsTest {\n+\n+ private static final String CORRELATION_KEY_PARTITION_0 = \"item-2\";\n+ private static final String CORRELATION_KEY_PARTITION_1 = \"item-1\";\n+ private static final String CORRELATION_KEY_PARTITION_2 = \"item-0\";\n+\n+ private static final String PROCESS_ID = \"process\";\n+\n+ public EmbeddedBrokerRule brokerRule = new EmbeddedBrokerRule(setPartitionCount(3));\n+ public GrpcClientRule clientRule = new GrpcClientRule(brokerRule);\n+\n+ @Rule public RuleChain ruleChain = RuleChain.outerRule(brokerRule).around(clientRule);\n+\n+ private static final BpmnModelInstance WORKFLOW =\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateCatchEvent()\n+ .message(m -> m.name(\"message\").zeebeCorrelationKey(\"$.key\"))\n+ .endEvent(\"end\")\n+ .done();\n+\n+ @Before\n+ public void init() {\n+ final DeploymentEvent deploymentEvent =\n+ clientRule\n+ .getWorkflowClient()\n+ .newDeployCommand()\n+ .addWorkflowModel(WORKFLOW, \"wf.bpmn\")\n+ .send()\n+ .join();\n+\n+ clientRule.waitUntilDeploymentIsDone(deploymentEvent.getKey());\n+ }\n+\n+ @Test\n+ public void shouldOpenMessageSubscriptionsOnDifferentPartitions() {\n+ // when\n+ IntStream.range(0, 10)\n+ .forEach(\n+ i -> {\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_0));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_1));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_2));\n+ });\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.messageSubscriptionRecords(MessageSubscriptionIntent.OPENED)\n+ .limit(30))\n+ .extracting(r -> tuple(r.getMetadata().getPartitionId(), r.getValue().getCorrelationKey()))\n+ .containsOnly(\n+ tuple(0, CORRELATION_KEY_PARTITION_0),\n+ tuple(1, CORRELATION_KEY_PARTITION_1),\n+ tuple(2, CORRELATION_KEY_PARTITION_2));\n+ }\n+\n+ @Test\n+ public void shouldPublishMessageOnDifferentPartitions() {\n+ // when\n+ IntStream.range(0, 10)\n+ .forEach(\n+ i -> {\n+ publishMessage(CORRELATION_KEY_PARTITION_0, Collections.singletonMap(\"p\", \"p0\"));\n+ publishMessage(CORRELATION_KEY_PARTITION_1, Collections.singletonMap(\"p\", \"p1\"));\n+ publishMessage(CORRELATION_KEY_PARTITION_2, Collections.singletonMap(\"p\", \"p2\"));\n+ });\n+\n+ // then\n+ assertThat(RecordingExporter.messageRecords(MessageIntent.PUBLISHED).limit(30))\n+ .extracting(r -> tuple(r.getMetadata().getPartitionId(), r.getValue().getCorrelationKey()))\n+ .containsOnly(\n+ tuple(0, CORRELATION_KEY_PARTITION_0),\n+ tuple(1, CORRELATION_KEY_PARTITION_1),\n+ tuple(2, CORRELATION_KEY_PARTITION_2));\n+ }\n+\n+ @Test\n+ public void shouldCorrelateMessageOnDifferentPartitions() {\n+ // given\n+ publishMessage(CORRELATION_KEY_PARTITION_0, Collections.singletonMap(\"p\", \"p0\"));\n+ publishMessage(CORRELATION_KEY_PARTITION_1, Collections.singletonMap(\"p\", \"p1\"));\n+ publishMessage(CORRELATION_KEY_PARTITION_2, Collections.singletonMap(\"p\", \"p2\"));\n+\n+ // when\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_0));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_1));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_2));\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.workflowInstanceRecords(WorkflowInstanceIntent.END_EVENT_OCCURRED)\n+ .withElementId(\"end\")\n+ .limit(3))\n+ .extracting(r -> r.getValue().getPayloadAsMap().get(\"p\"))\n+ .contains(\"p0\", \"p1\", \"p2\");\n+ }\n+\n+ @Test\n+ public void shouldOpenMessageSubscriptionsOnSamePartitionsAfterRestart() {\n+ // given\n+ IntStream.range(0, 5)\n+ .forEach(\n+ i -> {\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_0));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_1));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_2));\n+ });\n+\n+ assertThat(\n+ RecordingExporter.messageSubscriptionRecords(MessageSubscriptionIntent.OPENED)\n+ .limit(15)\n+ .exists())\n+ .isTrue();\n+\n+ // when\n+ brokerRule.stopBroker();\n+ brokerRule.startBroker();\n+\n+ IntStream.range(0, 5)\n+ .forEach(\n+ i -> {\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_0));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_1));\n+ createWorkflowInstance(Collections.singletonMap(\"key\", CORRELATION_KEY_PARTITION_2));\n+ });\n+\n+ // then\n+ assertThat(\n+ RecordingExporter.messageSubscriptionRecords(MessageSubscriptionIntent.OPENED)\n+ .limit(30))\n+ .extracting(r -> tuple(r.getMetadata().getPartitionId(), r.getValue().getCorrelationKey()))\n+ .containsOnly(\n+ tuple(0, CORRELATION_KEY_PARTITION_0),\n+ tuple(1, CORRELATION_KEY_PARTITION_1),\n+ tuple(2, CORRELATION_KEY_PARTITION_2));\n+ }\n+\n+ private void createWorkflowInstance(Object payload) {\n+ clientRule\n+ .getWorkflowClient()\n+ .newCreateInstanceCommand()\n+ .bpmnProcessId(PROCESS_ID)\n+ .latestVersion()\n+ .payload(payload)\n+ .send()\n+ .join();\n+ }\n+\n+ private void publishMessage(String correlationKey, Object payload) {\n+ clientRule\n+ .getWorkflowClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"message\")\n+ .correlationKey(correlationKey)\n+ .payload(payload)\n+ .send()\n+ .join();\n+ }\n+}\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/message/MessageCorrelationTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/message/MessageCorrelationTest.java\nnew file mode 100644\nindex 0000000..3b08572\n--- /dev/null\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/workflow/message/MessageCorrelationTest.java\n@@ -0,0 +1,198 @@\n+/*\n+ * Copyright \u00a9 2017 camunda services GmbH ([email protected])\n+ *\n+ * Licensed under the Apache License, Version 2.0 (the \"License\");\n+ * you may not use this file except in compliance with the License.\n+ * You may obtain a copy of the License at\n+ *\n+ * http://www.apache.org/licenses/LICENSE-2.0\n+ *\n+ * Unless required by applicable law or agreed to in writing, software\n+ * distributed under the License is distributed on an \"AS IS\" BASIS,\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n+ * See the License for the specific language governing permissions and\n+ * limitations under the License.\n+ */\n+package io.zeebe.broker.it.workflow.message;\n+\n+import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertElementActivated;\n+import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertElementCompleted;\n+import static io.zeebe.broker.it.util.ZeebeAssertHelper.assertWorkflowInstanceCompleted;\n+import static org.assertj.core.api.Assertions.assertThat;\n+import static org.assertj.core.api.Assertions.assertThatThrownBy;\n+import static org.assertj.core.api.Assertions.entry;\n+\n+import io.zeebe.broker.it.GrpcClientRule;\n+import io.zeebe.broker.test.EmbeddedBrokerRule;\n+import io.zeebe.client.api.ZeebeFuture;\n+import io.zeebe.client.api.events.DeploymentEvent;\n+import io.zeebe.client.cmd.ClientException;\n+import io.zeebe.model.bpmn.Bpmn;\n+import io.zeebe.model.bpmn.BpmnModelInstance;\n+import java.time.Duration;\n+import java.util.Collections;\n+import org.junit.Before;\n+import org.junit.Rule;\n+import org.junit.Test;\n+import org.junit.rules.RuleChain;\n+\n+public class MessageCorrelationTest {\n+\n+ private static final String PROCESS_ID = \"process\";\n+\n+ public EmbeddedBrokerRule brokerRule = new EmbeddedBrokerRule();\n+ public GrpcClientRule clientRule = new GrpcClientRule(brokerRule);\n+\n+ @Rule public RuleChain ruleChain = RuleChain.outerRule(brokerRule).around(clientRule);\n+\n+ private static final BpmnModelInstance WORKFLOW =\n+ Bpmn.createExecutableProcess(PROCESS_ID)\n+ .startEvent()\n+ .intermediateCatchEvent(\"catch-event\")\n+ .message(c -> c.name(\"order canceled\").zeebeCorrelationKey(\"$.orderId\"))\n+ .endEvent()\n+ .done();\n+\n+ @Before\n+ public void init() {\n+ final DeploymentEvent deploymentEvent =\n+ clientRule\n+ .getWorkflowClient()\n+ .newDeployCommand()\n+ .addWorkflowModel(WORKFLOW, \"wf.bpmn\")\n+ .send()\n+ .join();\n+\n+ clientRule.waitUntilDeploymentIsDone(deploymentEvent.getKey());\n+ }\n+\n+ @Test\n+ public void shouldCorrelateMessage() {\n+ // given\n+ clientRule\n+ .getWorkflowClient()\n+ .newCreateInstanceCommand()\n+ .bpmnProcessId(PROCESS_ID)\n+ .latestVersion()\n+ .payload(Collections.singletonMap(\"orderId\", \"order-123\"))\n+ .send()\n+ .join();\n+\n+ // when\n+ clientRule\n+ .getWorkflowClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"order canceled\")\n+ .correlationKey(\"order-123\")\n+ .payload(Collections.singletonMap(\"foo\", \"bar\"))\n+ .send()\n+ .join();\n+\n+ // then\n+ assertWorkflowInstanceCompleted(PROCESS_ID);\n+\n+ assertElementCompleted(\n+ PROCESS_ID,\n+ \"catch-event\",\n+ (catchEventOccurredEvent) ->\n+ assertThat(catchEventOccurredEvent.getPayloadAsMap())\n+ .containsExactly(entry(\"orderId\", \"order-123\"), entry(\"foo\", \"bar\")));\n+ }\n+\n+ @Test\n+ public void shouldCorrelateMessageWithZeroTTL() {\n+ // given\n+ clientRule\n+ .getWorkflowClient()\n+ .newCreateInstanceCommand()\n+ .bpmnProcessId(PROCESS_ID)\n+ .latestVersion()\n+ .payload(Collections.singletonMap(\"orderId\", \"order-123\"))\n+ .send()\n+ .join();\n+\n+ assertElementActivated(\"catch-event\");\n+\n+ // when\n+ clientRule\n+ .getWorkflowClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"order canceled\")\n+ .correlationKey(\"order-123\")\n+ .timeToLive(Duration.ZERO)\n+ .send()\n+ .join();\n+\n+ // then\n+ assertElementCompleted(PROCESS_ID, \"catch-event\");\n+ }\n+\n+ @Test\n+ public void shouldNotCorrelateMessageAfterTTL() {\n+ // given\n+ clientRule\n+ .getWorkflowClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"order canceled\")\n+ .correlationKey(\"order-123\")\n+ .timeToLive(Duration.ZERO)\n+ .payload(Collections.singletonMap(\"msg\", \"failure\"))\n+ .send()\n+ .join();\n+\n+ clientRule\n+ .getWorkflowClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"order canceled\")\n+ .correlationKey(\"order-123\")\n+ .timeToLive(Duration.ofMinutes(1))\n+ .payload(Collections.singletonMap(\"msg\", \"expected\"))\n+ .send()\n+ .join();\n+\n+ // when\n+ clientRule\n+ .getWorkflowClient()\n+ .newCreateInstanceCommand()\n+ .bpmnProcessId(PROCESS_ID)\n+ .latestVersion()\n+ .payload(Collections.singletonMap(\"orderId\", \"order-123\"))\n+ .send()\n+ .join();\n+\n+ // then\n+ assertElementCompleted(\n+ PROCESS_ID,\n+ \"catch-event\",\n+ (catchEventOccurred) ->\n+ assertThat(catchEventOccurred.getPayloadAsMap()).contains(entry(\"msg\", \"expected\")));\n+ }\n+\n+ @Test\n+ public void shouldRejectMessageWithSameId() {\n+ // given\n+ clientRule\n+ .getWorkflowClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"order canceled\")\n+ .correlationKey(\"order-123\")\n+ .messageId(\"foo\")\n+ .send()\n+ .join();\n+\n+ // when\n+ final ZeebeFuture<Void> future =\n+ clientRule\n+ .getWorkflowClient()\n+ .newPublishMessageCommand()\n+ .messageName(\"order canceled\")\n+ .correlationKey(\"order-123\")\n+ .messageId(\"foo\")\n+ .send();\n+\n+ // then\n+ assertThatThrownBy(future::join)\n+ .isInstanceOf(ClientException.class)\n+ .hasMessageContaining(\"message with id 'foo' is already published\");\n+ }\n+}\n", "diff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\nindex 6ee5797..bcfcc72 100644\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\n@@ -224,7 +224,6 @@ public final class AsyncSnapshotDirector extends Actor\n private void takeSnapshot() {\n final var transientSnapshotFuture =\n stateController.takeTransientSnapshot(lowerBoundSnapshotPosition);\n-\n transientSnapshotFuture.onComplete(\n (optionalTransientSnapshot, snapshotTakenError) -> {\n if (snapshotTakenError != null) {\n@@ -237,27 +236,31 @@ public final class AsyncSnapshotDirector extends Actor\n takingSnapshot = false;\n return;\n }\n- pendingSnapshot = optionalTransientSnapshot.get();\n- onRecovered();\n-\n- final ActorFuture<Long> lastWrittenPosition =\n- streamProcessor.getLastWrittenPositionAsync();\n- actor.runOnCompletion(\n- lastWrittenPosition,\n- (endPosition, error) -> {\n- if (error == null) {\n- LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\n- lastWrittenEventPosition = endPosition;\n- persistingSnapshot = false;\n- persistSnapshotIfLastWrittenPositionCommitted();\n- } else {\n- resetStateOnFailure();\n- LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\n- }\n- });\n+ onTransientSnapshotTaken(optionalTransientSnapshot.get());\n });\n }\n \n+ private void onTransientSnapshotTaken(final TransientSnapshot transientSnapshot) {\n+\n+ pendingSnapshot = transientSnapshot;\n+ onRecovered();\n+\n+ final ActorFuture<Long> lastWrittenPosition = streamProcessor.getLastWrittenPositionAsync();\n+ actor.runOnCompletion(lastWrittenPosition, this::onLastWrittenPositionReceived);\n+ }\n+\n+ private void onLastWrittenPositionReceived(final Long endPosition, final Throwable error) {\n+ if (error == null) {\n+ LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\n+ lastWrittenEventPosition = endPosition;\n+ persistingSnapshot = false;\n+ persistSnapshotIfLastWrittenPositionCommitted();\n+ } else {\n+ resetStateOnFailure();\n+ LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\n+ }\n+ }\n+\n private void onRecovered() {\n if (healthStatus != HealthStatus.HEALTHY) {\n healthStatus = HealthStatus.HEALTHY;\n"]
2
["2d416be63eeec9e7fdb90a62c40c8ad8f0672efa", "14abf5c31523a551134aebe9e8f3505ef26ed421"]
["test", "refactor"]
rebuild when environment variables change (#11471),Template using kube api version Signed-off-by: rjshrjndrn <[email protected]>
["diff --git a/cli/build.rs b/cli/build.rs\nindex 548fbb5..d7bed21 100644\n--- a/cli/build.rs\n+++ b/cli/build.rs\n@@ -269,8 +269,17 @@ fn main() {\n // To debug snapshot issues uncomment:\n // op_fetch_asset::trace_serializer();\n \n- println!(\"cargo:rustc-env=TS_VERSION={}\", ts_version());\n+ if let Ok(c) = env::var(\"DENO_CANARY\") {\n+ println!(\"cargo:rustc-env=DENO_CANARY={}\", c);\n+ }\n+ println!(\"cargo:rerun-if-env-changed=DENO_CANARY\");\n+\n println!(\"cargo:rustc-env=GIT_COMMIT_HASH={}\", git_commit_hash());\n+ println!(\"cargo:rerun-if-env-changed=GIT_COMMIT_HASH\");\n+\n+ println!(\"cargo:rustc-env=TS_VERSION={}\", ts_version());\n+ println!(\"cargo:rerun-if-env-changed=TS_VERSION\");\n+\n println!(\n \"cargo:rustc-env=DENO_CONSOLE_LIB_PATH={}\",\n deno_console::get_declaration().display()\n@@ -322,9 +331,6 @@ fn main() {\n \n println!(\"cargo:rustc-env=TARGET={}\", env::var(\"TARGET\").unwrap());\n println!(\"cargo:rustc-env=PROFILE={}\", env::var(\"PROFILE\").unwrap());\n- if let Ok(c) = env::var(\"DENO_CANARY\") {\n- println!(\"cargo:rustc-env=DENO_CANARY={}\", c);\n- }\n \n let c = PathBuf::from(env::var_os(\"CARGO_MANIFEST_DIR\").unwrap());\n let o = PathBuf::from(env::var_os(\"OUT_DIR\").unwrap());\n", "diff --git a/.github/workflows/api-ee.yaml b/.github/workflows/api-ee.yaml\nindex c014f34..2a12e0d 100644\n--- a/.github/workflows/api-ee.yaml\n+++ b/.github/workflows/api-ee.yaml\n@@ -8,7 +8,7 @@ on:\n default: 'false'\n push:\n branches:\n- - dev\n+ - test_ci\n paths:\n - ee/api/**\n - api/**\n@@ -112,7 +112,8 @@ jobs:\n # Deploy command\n kubectl config set-context --namespace=app --current\n kubectl config get-contexts\n- helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks | kubectl apply -f -\n+ k_version=$(kubectl version --short 2>/dev/null | awk '/Server/{print $NF}')\n+ helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks --kube-version=$k_version | kubectl apply -f -\n env:\n DOCKER_REPO: ${{ secrets.EE_REGISTRY_URL }}\n # We're not passing -ee flag, because helm will add that.\n"]
2
["63546c15bfb1284ac6d956eee274e6d7cf263a8f", "c3531347fe5a4cc82d426db195026a5bdad15e7a"]
["build", "ci"]
rebuild when environment variables change (#11471),note about archive without license and readme closes #602
["diff --git a/cli/build.rs b/cli/build.rs\nindex 548fbb5..d7bed21 100644\n--- a/cli/build.rs\n+++ b/cli/build.rs\n@@ -269,8 +269,17 @@ fn main() {\n // To debug snapshot issues uncomment:\n // op_fetch_asset::trace_serializer();\n \n- println!(\"cargo:rustc-env=TS_VERSION={}\", ts_version());\n+ if let Ok(c) = env::var(\"DENO_CANARY\") {\n+ println!(\"cargo:rustc-env=DENO_CANARY={}\", c);\n+ }\n+ println!(\"cargo:rerun-if-env-changed=DENO_CANARY\");\n+\n println!(\"cargo:rustc-env=GIT_COMMIT_HASH={}\", git_commit_hash());\n+ println!(\"cargo:rerun-if-env-changed=GIT_COMMIT_HASH\");\n+\n+ println!(\"cargo:rustc-env=TS_VERSION={}\", ts_version());\n+ println!(\"cargo:rerun-if-env-changed=TS_VERSION\");\n+\n println!(\n \"cargo:rustc-env=DENO_CONSOLE_LIB_PATH={}\",\n deno_console::get_declaration().display()\n@@ -322,9 +331,6 @@ fn main() {\n \n println!(\"cargo:rustc-env=TARGET={}\", env::var(\"TARGET\").unwrap());\n println!(\"cargo:rustc-env=PROFILE={}\", env::var(\"PROFILE\").unwrap());\n- if let Ok(c) = env::var(\"DENO_CANARY\") {\n- println!(\"cargo:rustc-env=DENO_CANARY={}\", c);\n- }\n \n let c = PathBuf::from(env::var_os(\"CARGO_MANIFEST_DIR\").unwrap());\n let o = PathBuf::from(env::var_os(\"OUT_DIR\").unwrap());\n", "diff --git a/docs/060-archive.md b/docs/060-archive.md\nindex e5d7cfb..bf0d569 100644\n--- a/docs/060-archive.md\n+++ b/docs/060-archive.md\n@@ -85,3 +85,24 @@ Then you can run:\n ```console\n GOVERSION_NR=$(go version | awk '{print $3;}') goreleaser\n ```\n+\n+## Packaging only the binaries\n+\n+Since GoReleaser will always add the `README` and `LICENSE` files to the\n+archive if the file list is empty, you'll need to provide a filled `files`\n+on the archive section.\n+\n+A working hack is to use something like this:\n+\n+```yaml\n+# goreleaser.yml\n+archive:\n+ files:\n+ - none*\n+```\n+\n+This would add all files matching the glob `none*`, provide that you don't\n+have any files matching that glob, only the binary will be added to the\n+archive.\n+\n+For more information, check [#602](https://github.com/goreleaser/goreleaser/issues/602)\n"]
2
["63546c15bfb1284ac6d956eee274e6d7cf263a8f", "7648760108613f771a5e6e40bb87a8f1fcee21ad"]
["build", "docs"]